From b42baa75820664aca91123e2c3528f81052f17c9 Mon Sep 17 00:00:00 2001
From: fjosw
Date: Mon, 16 Jan 2023 14:58:20 +0000
Subject: [PATCH] Documentation updated
---
docs/pyerrors/dirac.html | 254 +-
docs/pyerrors/fits.html | 2032 ++++++++--------
docs/pyerrors/input/bdio.html | 2557 ++++++++++----------
docs/pyerrors/input/dobs.html | 2898 ++++++++++++-----------
docs/pyerrors/input/hadrons.html | 1477 ++++++------
docs/pyerrors/input/json.html | 2058 +++++++++--------
docs/pyerrors/input/misc.html | 427 ++--
docs/pyerrors/input/openQCD.html | 3720 +++++++++++++++---------------
docs/pyerrors/input/pandas.html | 454 ++--
docs/pyerrors/input/sfcf.html | 1243 +++++-----
docs/pyerrors/input/utils.html | 67 +-
docs/pyerrors/misc.html | 421 ++--
docs/pyerrors/mpm.html | 157 +-
docs/pyerrors/roots.html | 7 +-
docs/search.js | 2 +-
15 files changed, 9322 insertions(+), 8452 deletions(-)
diff --git a/docs/pyerrors/dirac.html b/docs/pyerrors/dirac.html
index 206770ff..9ebb0312 100644
--- a/docs/pyerrors/dirac.html
+++ b/docs/pyerrors/dirac.html
@@ -110,63 +110,74 @@
26 """Rank-3 epsilon tensor
27
28 Based on https://codegolf.stackexchange.com/a/160375
-29 """
-30 test_set = set((i, j, k))
-31 if not (test_set <= set((1, 2, 3)) or test_set <= set((0, 1, 2))):
-32 raise Exception("Unexpected input", i, j, k)
-33
-34 return (i - j) * (j - k) * (k - i) / 2
-35
-36
-37def epsilon_tensor_rank4(i, j, k, o):
-38 """Rank-4 epsilon tensor
-39
-40 Extension of https://codegolf.stackexchange.com/a/160375
-41 """
-42 test_set = set((i, j, k, o))
-43 if not (test_set <= set((1, 2, 3, 4)) or test_set <= set((0, 1, 2, 3))):
-44 raise Exception("Unexpected input", i, j, k, o)
-45
-46 return (i - j) * (j - k) * (k - i) * (i - o) * (j - o) * (o - k) / 12
+29
+30 Returns
+31 -------
+32 elem : int
+33 Element (i,j,k) of the epsilon tensor of rank 3
+34 """
+35 test_set = set((i, j, k))
+36 if not (test_set <= set((1, 2, 3)) or test_set <= set((0, 1, 2))):
+37 raise Exception("Unexpected input", i, j, k)
+38
+39 return (i - j) * (j - k) * (k - i) / 2
+40
+41
+42def epsilon_tensor_rank4(i, j, k, o):
+43 """Rank-4 epsilon tensor
+44
+45 Extension of https://codegolf.stackexchange.com/a/160375
+46
47
-48
-49def Grid_gamma(gamma_tag):
-50 """Returns gamma matrix in Grid labeling."""
-51 if gamma_tag == 'Identity':
-52 g = identity
-53 elif gamma_tag == 'Gamma5':
-54 g = gamma5
-55 elif gamma_tag == 'GammaX':
-56 g = gamma[0]
-57 elif gamma_tag == 'GammaY':
-58 g = gamma[1]
-59 elif gamma_tag == 'GammaZ':
-60 g = gamma[2]
-61 elif gamma_tag == 'GammaT':
-62 g = gamma[3]
-63 elif gamma_tag == 'GammaXGamma5':
-64 g = gamma[0] @ gamma5
-65 elif gamma_tag == 'GammaYGamma5':
-66 g = gamma[1] @ gamma5
-67 elif gamma_tag == 'GammaZGamma5':
-68 g = gamma[2] @ gamma5
-69 elif gamma_tag == 'GammaTGamma5':
-70 g = gamma[3] @ gamma5
-71 elif gamma_tag == 'SigmaXT':
-72 g = 0.5 * (gamma[0] @ gamma[3] - gamma[3] @ gamma[0])
-73 elif gamma_tag == 'SigmaXY':
-74 g = 0.5 * (gamma[0] @ gamma[1] - gamma[1] @ gamma[0])
-75 elif gamma_tag == 'SigmaXZ':
-76 g = 0.5 * (gamma[0] @ gamma[2] - gamma[2] @ gamma[0])
-77 elif gamma_tag == 'SigmaYT':
-78 g = 0.5 * (gamma[1] @ gamma[3] - gamma[3] @ gamma[1])
-79 elif gamma_tag == 'SigmaYZ':
-80 g = 0.5 * (gamma[1] @ gamma[2] - gamma[2] @ gamma[1])
-81 elif gamma_tag == 'SigmaZT':
-82 g = 0.5 * (gamma[2] @ gamma[3] - gamma[3] @ gamma[2])
-83 else:
-84 raise Exception('Unkown gamma structure', gamma_tag)
-85 return g
+48 Returns
+49 -------
+50 elem : int
+51 Element (i,j,k,o) of the epsilon tensor of rank 4
+52 """
+53 test_set = set((i, j, k, o))
+54 if not (test_set <= set((1, 2, 3, 4)) or test_set <= set((0, 1, 2, 3))):
+55 raise Exception("Unexpected input", i, j, k, o)
+56
+57 return (i - j) * (j - k) * (k - i) * (i - o) * (j - o) * (o - k) / 12
+58
+59
+60def Grid_gamma(gamma_tag):
+61 """Returns gamma matrix in Grid labeling."""
+62 if gamma_tag == 'Identity':
+63 g = identity
+64 elif gamma_tag == 'Gamma5':
+65 g = gamma5
+66 elif gamma_tag == 'GammaX':
+67 g = gamma[0]
+68 elif gamma_tag == 'GammaY':
+69 g = gamma[1]
+70 elif gamma_tag == 'GammaZ':
+71 g = gamma[2]
+72 elif gamma_tag == 'GammaT':
+73 g = gamma[3]
+74 elif gamma_tag == 'GammaXGamma5':
+75 g = gamma[0] @ gamma5
+76 elif gamma_tag == 'GammaYGamma5':
+77 g = gamma[1] @ gamma5
+78 elif gamma_tag == 'GammaZGamma5':
+79 g = gamma[2] @ gamma5
+80 elif gamma_tag == 'GammaTGamma5':
+81 g = gamma[3] @ gamma5
+82 elif gamma_tag == 'SigmaXT':
+83 g = 0.5 * (gamma[0] @ gamma[3] - gamma[3] @ gamma[0])
+84 elif gamma_tag == 'SigmaXY':
+85 g = 0.5 * (gamma[0] @ gamma[1] - gamma[1] @ gamma[0])
+86 elif gamma_tag == 'SigmaXZ':
+87 g = 0.5 * (gamma[0] @ gamma[2] - gamma[2] @ gamma[0])
+88 elif gamma_tag == 'SigmaYT':
+89 g = 0.5 * (gamma[1] @ gamma[3] - gamma[3] @ gamma[1])
+90 elif gamma_tag == 'SigmaYZ':
+91 g = 0.5 * (gamma[1] @ gamma[2] - gamma[2] @ gamma[1])
+92 elif gamma_tag == 'SigmaZT':
+93 g = 0.5 * (gamma[2] @ gamma[3] - gamma[3] @ gamma[2])
+94 else:
+95 raise Exception('Unkown gamma structure', gamma_tag)
+96 return g
@@ -186,18 +197,30 @@
27 """Rank-3 epsilon tensor
28
29 Based on https://codegolf.stackexchange.com/a/160375
-30 """
-31 test_set = set((i, j, k))
-32 if not (test_set <= set((1, 2, 3)) or test_set <= set((0, 1, 2))):
-33 raise Exception("Unexpected input", i, j, k)
-34
-35 return (i - j) * (j - k) * (k - i) / 2
+30
+31 Returns
+32 -------
+33 elem : int
+34 Element (i,j,k) of the epsilon tensor of rank 3
+35 """
+36 test_set = set((i, j, k))
+37 if not (test_set <= set((1, 2, 3)) or test_set <= set((0, 1, 2))):
+38 raise Exception("Unexpected input", i, j, k)
+39
+40 return (i - j) * (j - k) * (k - i) / 2
@@ -213,22 +236,35 @@
- 38def epsilon_tensor_rank4(i, j, k, o):
-39 """Rank-4 epsilon tensor
-40
-41 Extension of https://codegolf.stackexchange.com/a/160375
-42 """
-43 test_set = set((i, j, k, o))
-44 if not (test_set <= set((1, 2, 3, 4)) or test_set <= set((0, 1, 2, 3))):
-45 raise Exception("Unexpected input", i, j, k, o)
-46
-47 return (i - j) * (j - k) * (k - i) * (i - o) * (j - o) * (o - k) / 12
+ 43def epsilon_tensor_rank4(i, j, k, o):
+44 """Rank-4 epsilon tensor
+45
+46 Extension of https://codegolf.stackexchange.com/a/160375
+47
+48
+49 Returns
+50 -------
+51 elem : int
+52 Element (i,j,k,o) of the epsilon tensor of rank 4
+53 """
+54 test_set = set((i, j, k, o))
+55 if not (test_set <= set((1, 2, 3, 4)) or test_set <= set((0, 1, 2, 3))):
+56 raise Exception("Unexpected input", i, j, k, o)
+57
+58 return (i - j) * (j - k) * (k - i) * (i - o) * (j - o) * (o - k) / 12
@@ -244,43 +280,43 @@
- 50def Grid_gamma(gamma_tag):
-51 """Returns gamma matrix in Grid labeling."""
-52 if gamma_tag == 'Identity':
-53 g = identity
-54 elif gamma_tag == 'Gamma5':
-55 g = gamma5
-56 elif gamma_tag == 'GammaX':
-57 g = gamma[0]
-58 elif gamma_tag == 'GammaY':
-59 g = gamma[1]
-60 elif gamma_tag == 'GammaZ':
-61 g = gamma[2]
-62 elif gamma_tag == 'GammaT':
-63 g = gamma[3]
-64 elif gamma_tag == 'GammaXGamma5':
-65 g = gamma[0] @ gamma5
-66 elif gamma_tag == 'GammaYGamma5':
-67 g = gamma[1] @ gamma5
-68 elif gamma_tag == 'GammaZGamma5':
-69 g = gamma[2] @ gamma5
-70 elif gamma_tag == 'GammaTGamma5':
-71 g = gamma[3] @ gamma5
-72 elif gamma_tag == 'SigmaXT':
-73 g = 0.5 * (gamma[0] @ gamma[3] - gamma[3] @ gamma[0])
-74 elif gamma_tag == 'SigmaXY':
-75 g = 0.5 * (gamma[0] @ gamma[1] - gamma[1] @ gamma[0])
-76 elif gamma_tag == 'SigmaXZ':
-77 g = 0.5 * (gamma[0] @ gamma[2] - gamma[2] @ gamma[0])
-78 elif gamma_tag == 'SigmaYT':
-79 g = 0.5 * (gamma[1] @ gamma[3] - gamma[3] @ gamma[1])
-80 elif gamma_tag == 'SigmaYZ':
-81 g = 0.5 * (gamma[1] @ gamma[2] - gamma[2] @ gamma[1])
-82 elif gamma_tag == 'SigmaZT':
-83 g = 0.5 * (gamma[2] @ gamma[3] - gamma[3] @ gamma[2])
-84 else:
-85 raise Exception('Unkown gamma structure', gamma_tag)
-86 return g
+ 61def Grid_gamma(gamma_tag):
+62 """Returns gamma matrix in Grid labeling."""
+63 if gamma_tag == 'Identity':
+64 g = identity
+65 elif gamma_tag == 'Gamma5':
+66 g = gamma5
+67 elif gamma_tag == 'GammaX':
+68 g = gamma[0]
+69 elif gamma_tag == 'GammaY':
+70 g = gamma[1]
+71 elif gamma_tag == 'GammaZ':
+72 g = gamma[2]
+73 elif gamma_tag == 'GammaT':
+74 g = gamma[3]
+75 elif gamma_tag == 'GammaXGamma5':
+76 g = gamma[0] @ gamma5
+77 elif gamma_tag == 'GammaYGamma5':
+78 g = gamma[1] @ gamma5
+79 elif gamma_tag == 'GammaZGamma5':
+80 g = gamma[2] @ gamma5
+81 elif gamma_tag == 'GammaTGamma5':
+82 g = gamma[3] @ gamma5
+83 elif gamma_tag == 'SigmaXT':
+84 g = 0.5 * (gamma[0] @ gamma[3] - gamma[3] @ gamma[0])
+85 elif gamma_tag == 'SigmaXY':
+86 g = 0.5 * (gamma[0] @ gamma[1] - gamma[1] @ gamma[0])
+87 elif gamma_tag == 'SigmaXZ':
+88 g = 0.5 * (gamma[0] @ gamma[2] - gamma[2] @ gamma[0])
+89 elif gamma_tag == 'SigmaYT':
+90 g = 0.5 * (gamma[1] @ gamma[3] - gamma[3] @ gamma[1])
+91 elif gamma_tag == 'SigmaYZ':
+92 g = 0.5 * (gamma[1] @ gamma[2] - gamma[2] @ gamma[1])
+93 elif gamma_tag == 'SigmaZT':
+94 g = 0.5 * (gamma[2] @ gamma[3] - gamma[3] @ gamma[2])
+95 else:
+96 raise Exception('Unkown gamma structure', gamma_tag)
+97 return g
diff --git a/docs/pyerrors/fits.html b/docs/pyerrors/fits.html
index 0be3cb04..a8802145 100644
--- a/docs/pyerrors/fits.html
+++ b/docs/pyerrors/fits.html
@@ -240,673 +240,707 @@
129 If True, a quantile-quantile plot of the fit result is generated (default False).
130 num_grad : bool
131 Use numerical differentation instead of automatic differentiation to perform the error propagation (default False).
-132 '''
-133 if priors is not None:
-134 return _prior_fit(x, y, func, priors, silent=silent, **kwargs)
-135 else:
-136 return _standard_fit(x, y, func, silent=silent, **kwargs)
-137
-138
-139def total_least_squares(x, y, func, silent=False, **kwargs):
-140 r'''Performs a non-linear fit to y = func(x) and returns a list of Obs corresponding to the fit parameters.
-141
-142 Parameters
-143 ----------
-144 x : list
-145 list of Obs, or a tuple of lists of Obs
-146 y : list
-147 list of Obs. The dvalues of the Obs are used as x- and yerror for the fit.
-148 func : object
-149 func has to be of the form
-150
-151 ```python
-152 import autograd.numpy as anp
-153
-154 def func(a, x):
-155 return a[0] + a[1] * x + a[2] * anp.sinh(x)
-156 ```
-157
-158 For multiple x values func can be of the form
-159
-160 ```python
-161 def func(a, x):
-162 (x1, x2) = x
-163 return a[0] * x1 ** 2 + a[1] * x2
-164 ```
-165
-166 It is important that all numpy functions refer to autograd.numpy, otherwise the differentiation
-167 will not work.
-168 silent : bool, optional
-169 If true all output to the console is omitted (default False).
-170 initial_guess : list
-171 can provide an initial guess for the input parameters. Relevant for non-linear
-172 fits with many parameters.
-173 expected_chisquare : bool
-174 If true prints the expected chisquare which is
-175 corrected by effects caused by correlated input data.
-176 This can take a while as the full correlation matrix
-177 has to be calculated (default False).
-178 num_grad : bool
-179 Use numerical differentation instead of automatic differentiation to perform the error propagation (default False).
-180
-181 Notes
-182 -----
-183 Based on the orthogonal distance regression module of scipy
-184 '''
+132
+133 Returns
+134 -------
+135 output : Fit_result
+136 Parameters and information on the fitted result.
+137 '''
+138 if priors is not None:
+139 return _prior_fit(x, y, func, priors, silent=silent, **kwargs)
+140 else:
+141 return _standard_fit(x, y, func, silent=silent, **kwargs)
+142
+143
+144def total_least_squares(x, y, func, silent=False, **kwargs):
+145 r'''Performs a non-linear fit to y = func(x) and returns a list of Obs corresponding to the fit parameters.
+146
+147 Parameters
+148 ----------
+149 x : list
+150 list of Obs, or a tuple of lists of Obs
+151 y : list
+152 list of Obs. The dvalues of the Obs are used as x- and yerror for the fit.
+153 func : object
+154 func has to be of the form
+155
+156 ```python
+157 import autograd.numpy as anp
+158
+159 def func(a, x):
+160 return a[0] + a[1] * x + a[2] * anp.sinh(x)
+161 ```
+162
+163 For multiple x values func can be of the form
+164
+165 ```python
+166 def func(a, x):
+167 (x1, x2) = x
+168 return a[0] * x1 ** 2 + a[1] * x2
+169 ```
+170
+171 It is important that all numpy functions refer to autograd.numpy, otherwise the differentiation
+172 will not work.
+173 silent : bool, optional
+174 If true all output to the console is omitted (default False).
+175 initial_guess : list
+176 can provide an initial guess for the input parameters. Relevant for non-linear
+177 fits with many parameters.
+178 expected_chisquare : bool
+179 If true prints the expected chisquare which is
+180 corrected by effects caused by correlated input data.
+181 This can take a while as the full correlation matrix
+182 has to be calculated (default False).
+183 num_grad : bool
+184 Use numerical differentation instead of automatic differentiation to perform the error propagation (default False).
185
-186 output = Fit_result()
-187
-188 output.fit_function = func
+186 Notes
+187 -----
+188 Based on the orthogonal distance regression module of scipy.
189
-190 x = np.array(x)
-191
-192 x_shape = x.shape
-193
-194 if kwargs.get('num_grad') is True:
-195 jacobian = num_jacobian
-196 hessian = num_hessian
-197 else:
-198 jacobian = auto_jacobian
-199 hessian = auto_hessian
-200
-201 if not callable(func):
-202 raise TypeError('func has to be a function.')
+190 Returns
+191 -------
+192 output : Fit_result
+193 Parameters and information on the fitted result.
+194 '''
+195
+196 output = Fit_result()
+197
+198 output.fit_function = func
+199
+200 x = np.array(x)
+201
+202 x_shape = x.shape
203
-204 for i in range(42):
-205 try:
-206 func(np.arange(i), x.T[0])
-207 except TypeError:
-208 continue
-209 except IndexError:
-210 continue
-211 else:
-212 break
-213 else:
-214 raise RuntimeError("Fit function is not valid.")
-215
-216 n_parms = i
-217 if not silent:
-218 print('Fit with', n_parms, 'parameter' + 's' * (n_parms > 1))
-219
-220 x_f = np.vectorize(lambda o: o.value)(x)
-221 dx_f = np.vectorize(lambda o: o.dvalue)(x)
-222 y_f = np.array([o.value for o in y])
-223 dy_f = np.array([o.dvalue for o in y])
-224
-225 if np.any(np.asarray(dx_f) <= 0.0):
-226 raise Exception('No x errors available, run the gamma method first.')
-227
-228 if np.any(np.asarray(dy_f) <= 0.0):
-229 raise Exception('No y errors available, run the gamma method first.')
-230
-231 if 'initial_guess' in kwargs:
-232 x0 = kwargs.get('initial_guess')
-233 if len(x0) != n_parms:
-234 raise Exception('Initial guess does not have the correct length: %d vs. %d' % (len(x0), n_parms))
-235 else:
-236 x0 = [1] * n_parms
+204 if kwargs.get('num_grad') is True:
+205 jacobian = num_jacobian
+206 hessian = num_hessian
+207 else:
+208 jacobian = auto_jacobian
+209 hessian = auto_hessian
+210
+211 if not callable(func):
+212 raise TypeError('func has to be a function.')
+213
+214 for i in range(42):
+215 try:
+216 func(np.arange(i), x.T[0])
+217 except TypeError:
+218 continue
+219 except IndexError:
+220 continue
+221 else:
+222 break
+223 else:
+224 raise RuntimeError("Fit function is not valid.")
+225
+226 n_parms = i
+227 if not silent:
+228 print('Fit with', n_parms, 'parameter' + 's' * (n_parms > 1))
+229
+230 x_f = np.vectorize(lambda o: o.value)(x)
+231 dx_f = np.vectorize(lambda o: o.dvalue)(x)
+232 y_f = np.array([o.value for o in y])
+233 dy_f = np.array([o.dvalue for o in y])
+234
+235 if np.any(np.asarray(dx_f) <= 0.0):
+236 raise Exception('No x errors available, run the gamma method first.')
237
-238 data = RealData(x_f, y_f, sx=dx_f, sy=dy_f)
-239 model = Model(func)
-240 odr = ODR(data, model, x0, partol=np.finfo(np.float64).eps)
-241 odr.set_job(fit_type=0, deriv=1)
-242 out = odr.run()
-243
-244 output.residual_variance = out.res_var
-245
-246 output.method = 'ODR'
+238 if np.any(np.asarray(dy_f) <= 0.0):
+239 raise Exception('No y errors available, run the gamma method first.')
+240
+241 if 'initial_guess' in kwargs:
+242 x0 = kwargs.get('initial_guess')
+243 if len(x0) != n_parms:
+244 raise Exception('Initial guess does not have the correct length: %d vs. %d' % (len(x0), n_parms))
+245 else:
+246 x0 = [1] * n_parms
247
-248 output.message = out.stopreason
-249
-250 output.xplus = out.xplus
-251
-252 if not silent:
-253 print('Method: ODR')
-254 print(*out.stopreason)
-255 print('Residual variance:', output.residual_variance)
-256
-257 if out.info > 3:
-258 raise Exception('The minimization procedure did not converge.')
+248 data = RealData(x_f, y_f, sx=dx_f, sy=dy_f)
+249 model = Model(func)
+250 odr = ODR(data, model, x0, partol=np.finfo(np.float64).eps)
+251 odr.set_job(fit_type=0, deriv=1)
+252 out = odr.run()
+253
+254 output.residual_variance = out.res_var
+255
+256 output.method = 'ODR'
+257
+258 output.message = out.stopreason
259
-260 m = x_f.size
+260 output.xplus = out.xplus
261
-262 def odr_chisquare(p):
-263 model = func(p[:n_parms], p[n_parms:].reshape(x_shape))
-264 chisq = anp.sum(((y_f - model) / dy_f) ** 2) + anp.sum(((x_f - p[n_parms:].reshape(x_shape)) / dx_f) ** 2)
-265 return chisq
+262 if not silent:
+263 print('Method: ODR')
+264 print(*out.stopreason)
+265 print('Residual variance:', output.residual_variance)
266
-267 if kwargs.get('expected_chisquare') is True:
-268 W = np.diag(1 / np.asarray(np.concatenate((dy_f.ravel(), dx_f.ravel()))))
+267 if out.info > 3:
+268 raise Exception('The minimization procedure did not converge.')
269
-270 if kwargs.get('covariance') is not None:
-271 cov = kwargs.get('covariance')
-272 else:
-273 cov = covariance(np.concatenate((y, x.ravel())))
-274
-275 number_of_x_parameters = int(m / x_f.shape[-1])
+270 m = x_f.size
+271
+272 def odr_chisquare(p):
+273 model = func(p[:n_parms], p[n_parms:].reshape(x_shape))
+274 chisq = anp.sum(((y_f - model) / dy_f) ** 2) + anp.sum(((x_f - p[n_parms:].reshape(x_shape)) / dx_f) ** 2)
+275 return chisq
276
-277 old_jac = jacobian(func)(out.beta, out.xplus)
-278 fused_row1 = np.concatenate((old_jac, np.concatenate((number_of_x_parameters * [np.zeros(old_jac.shape)]), axis=0)))
-279 fused_row2 = np.concatenate((jacobian(lambda x, y: func(y, x))(out.xplus, out.beta).reshape(x_f.shape[-1], x_f.shape[-1] * number_of_x_parameters), np.identity(number_of_x_parameters * old_jac.shape[0])))
-280 new_jac = np.concatenate((fused_row1, fused_row2), axis=1)
-281
-282 A = W @ new_jac
-283 P_phi = A @ np.linalg.pinv(A.T @ A) @ A.T
-284 expected_chisquare = np.trace((np.identity(P_phi.shape[0]) - P_phi) @ W @ cov @ W)
-285 if expected_chisquare <= 0.0:
-286 warnings.warn("Negative expected_chisquare.", RuntimeWarning)
-287 expected_chisquare = np.abs(expected_chisquare)
-288 output.chisquare_by_expected_chisquare = odr_chisquare(np.concatenate((out.beta, out.xplus.ravel()))) / expected_chisquare
-289 if not silent:
-290 print('chisquare/expected_chisquare:',
-291 output.chisquare_by_expected_chisquare)
-292
-293 fitp = out.beta
-294 try:
-295 hess = hessian(odr_chisquare)(np.concatenate((fitp, out.xplus.ravel())))
-296 except TypeError:
-297 raise Exception("It is required to use autograd.numpy instead of numpy within fit functions, see the documentation for details.") from None
-298
-299 def odr_chisquare_compact_x(d):
-300 model = func(d[:n_parms], d[n_parms:n_parms + m].reshape(x_shape))
-301 chisq = anp.sum(((y_f - model) / dy_f) ** 2) + anp.sum(((d[n_parms + m:].reshape(x_shape) - d[n_parms:n_parms + m].reshape(x_shape)) / dx_f) ** 2)
-302 return chisq
-303
-304 jac_jac_x = hessian(odr_chisquare_compact_x)(np.concatenate((fitp, out.xplus.ravel(), x_f.ravel())))
-305
-306 # Compute hess^{-1} @ jac_jac_x[:n_parms + m, n_parms + m:] using LAPACK dgesv
-307 try:
-308 deriv_x = -scipy.linalg.solve(hess, jac_jac_x[:n_parms + m, n_parms + m:])
-309 except np.linalg.LinAlgError:
-310 raise Exception("Cannot invert hessian matrix.")
-311
-312 def odr_chisquare_compact_y(d):
-313 model = func(d[:n_parms], d[n_parms:n_parms + m].reshape(x_shape))
-314 chisq = anp.sum(((d[n_parms + m:] - model) / dy_f) ** 2) + anp.sum(((x_f - d[n_parms:n_parms + m].reshape(x_shape)) / dx_f) ** 2)
-315 return chisq
-316
-317 jac_jac_y = hessian(odr_chisquare_compact_y)(np.concatenate((fitp, out.xplus.ravel(), y_f)))
-318
-319 # Compute hess^{-1} @ jac_jac_y[:n_parms + m, n_parms + m:] using LAPACK dgesv
-320 try:
-321 deriv_y = -scipy.linalg.solve(hess, jac_jac_y[:n_parms + m, n_parms + m:])
-322 except np.linalg.LinAlgError:
-323 raise Exception("Cannot invert hessian matrix.")
-324
-325 result = []
-326 for i in range(n_parms):
-327 result.append(derived_observable(lambda my_var, **kwargs: (my_var[0] + np.finfo(np.float64).eps) / (x.ravel()[0].value + np.finfo(np.float64).eps) * out.beta[i], list(x.ravel()) + list(y), man_grad=list(deriv_x[i]) + list(deriv_y[i])))
+277 if kwargs.get('expected_chisquare') is True:
+278 W = np.diag(1 / np.asarray(np.concatenate((dy_f.ravel(), dx_f.ravel()))))
+279
+280 if kwargs.get('covariance') is not None:
+281 cov = kwargs.get('covariance')
+282 else:
+283 cov = covariance(np.concatenate((y, x.ravel())))
+284
+285 number_of_x_parameters = int(m / x_f.shape[-1])
+286
+287 old_jac = jacobian(func)(out.beta, out.xplus)
+288 fused_row1 = np.concatenate((old_jac, np.concatenate((number_of_x_parameters * [np.zeros(old_jac.shape)]), axis=0)))
+289 fused_row2 = np.concatenate((jacobian(lambda x, y: func(y, x))(out.xplus, out.beta).reshape(x_f.shape[-1], x_f.shape[-1] * number_of_x_parameters), np.identity(number_of_x_parameters * old_jac.shape[0])))
+290 new_jac = np.concatenate((fused_row1, fused_row2), axis=1)
+291
+292 A = W @ new_jac
+293 P_phi = A @ np.linalg.pinv(A.T @ A) @ A.T
+294 expected_chisquare = np.trace((np.identity(P_phi.shape[0]) - P_phi) @ W @ cov @ W)
+295 if expected_chisquare <= 0.0:
+296 warnings.warn("Negative expected_chisquare.", RuntimeWarning)
+297 expected_chisquare = np.abs(expected_chisquare)
+298 output.chisquare_by_expected_chisquare = odr_chisquare(np.concatenate((out.beta, out.xplus.ravel()))) / expected_chisquare
+299 if not silent:
+300 print('chisquare/expected_chisquare:',
+301 output.chisquare_by_expected_chisquare)
+302
+303 fitp = out.beta
+304 try:
+305 hess = hessian(odr_chisquare)(np.concatenate((fitp, out.xplus.ravel())))
+306 except TypeError:
+307 raise Exception("It is required to use autograd.numpy instead of numpy within fit functions, see the documentation for details.") from None
+308
+309 def odr_chisquare_compact_x(d):
+310 model = func(d[:n_parms], d[n_parms:n_parms + m].reshape(x_shape))
+311 chisq = anp.sum(((y_f - model) / dy_f) ** 2) + anp.sum(((d[n_parms + m:].reshape(x_shape) - d[n_parms:n_parms + m].reshape(x_shape)) / dx_f) ** 2)
+312 return chisq
+313
+314 jac_jac_x = hessian(odr_chisquare_compact_x)(np.concatenate((fitp, out.xplus.ravel(), x_f.ravel())))
+315
+316 # Compute hess^{-1} @ jac_jac_x[:n_parms + m, n_parms + m:] using LAPACK dgesv
+317 try:
+318 deriv_x = -scipy.linalg.solve(hess, jac_jac_x[:n_parms + m, n_parms + m:])
+319 except np.linalg.LinAlgError:
+320 raise Exception("Cannot invert hessian matrix.")
+321
+322 def odr_chisquare_compact_y(d):
+323 model = func(d[:n_parms], d[n_parms:n_parms + m].reshape(x_shape))
+324 chisq = anp.sum(((d[n_parms + m:] - model) / dy_f) ** 2) + anp.sum(((x_f - d[n_parms:n_parms + m].reshape(x_shape)) / dx_f) ** 2)
+325 return chisq
+326
+327 jac_jac_y = hessian(odr_chisquare_compact_y)(np.concatenate((fitp, out.xplus.ravel(), y_f)))
328
-329 output.fit_parameters = result
-330
-331 output.odr_chisquare = odr_chisquare(np.concatenate((out.beta, out.xplus.ravel())))
-332 output.dof = x.shape[-1] - n_parms
-333 output.p_value = 1 - scipy.stats.chi2.cdf(output.odr_chisquare, output.dof)
+329 # Compute hess^{-1} @ jac_jac_y[:n_parms + m, n_parms + m:] using LAPACK dgesv
+330 try:
+331 deriv_y = -scipy.linalg.solve(hess, jac_jac_y[:n_parms + m, n_parms + m:])
+332 except np.linalg.LinAlgError:
+333 raise Exception("Cannot invert hessian matrix.")
334
-335 return output
-336
-337
-338def _prior_fit(x, y, func, priors, silent=False, **kwargs):
-339 output = Fit_result()
+335 result = []
+336 for i in range(n_parms):
+337 result.append(derived_observable(lambda my_var, **kwargs: (my_var[0] + np.finfo(np.float64).eps) / (x.ravel()[0].value + np.finfo(np.float64).eps) * out.beta[i], list(x.ravel()) + list(y), man_grad=list(deriv_x[i]) + list(deriv_y[i])))
+338
+339 output.fit_parameters = result
340
-341 output.fit_function = func
-342
-343 x = np.asarray(x)
+341 output.odr_chisquare = odr_chisquare(np.concatenate((out.beta, out.xplus.ravel())))
+342 output.dof = x.shape[-1] - n_parms
+343 output.p_value = 1 - scipy.stats.chi2.cdf(output.odr_chisquare, output.dof)
344
-345 if kwargs.get('num_grad') is True:
-346 hessian = num_hessian
-347 else:
-348 hessian = auto_hessian
-349
-350 if not callable(func):
-351 raise TypeError('func has to be a function.')
+345 return output
+346
+347
+348def _prior_fit(x, y, func, priors, silent=False, **kwargs):
+349 output = Fit_result()
+350
+351 output.fit_function = func
352
-353 for i in range(100):
-354 try:
-355 func(np.arange(i), 0)
-356 except TypeError:
-357 continue
-358 except IndexError:
-359 continue
-360 else:
-361 break
-362 else:
-363 raise RuntimeError("Fit function is not valid.")
-364
-365 n_parms = i
-366
-367 if n_parms != len(priors):
-368 raise Exception('Priors does not have the correct length.')
-369
-370 def extract_val_and_dval(string):
-371 split_string = string.split('(')
-372 if '.' in split_string[0] and '.' not in split_string[1][:-1]:
-373 factor = 10 ** -len(split_string[0].partition('.')[2])
-374 else:
-375 factor = 1
-376 return float(split_string[0]), float(split_string[1][:-1]) * factor
-377
-378 loc_priors = []
-379 for i_n, i_prior in enumerate(priors):
-380 if isinstance(i_prior, Obs):
-381 loc_priors.append(i_prior)
-382 else:
-383 loc_val, loc_dval = extract_val_and_dval(i_prior)
-384 loc_priors.append(cov_Obs(loc_val, loc_dval ** 2, '#prior' + str(i_n) + f"_{np.random.randint(2147483647):010d}"))
-385
-386 output.priors = loc_priors
+353 x = np.asarray(x)
+354
+355 if kwargs.get('num_grad') is True:
+356 hessian = num_hessian
+357 else:
+358 hessian = auto_hessian
+359
+360 if not callable(func):
+361 raise TypeError('func has to be a function.')
+362
+363 for i in range(100):
+364 try:
+365 func(np.arange(i), 0)
+366 except TypeError:
+367 continue
+368 except IndexError:
+369 continue
+370 else:
+371 break
+372 else:
+373 raise RuntimeError("Fit function is not valid.")
+374
+375 n_parms = i
+376
+377 if n_parms != len(priors):
+378 raise Exception('Priors does not have the correct length.')
+379
+380 def extract_val_and_dval(string):
+381 split_string = string.split('(')
+382 if '.' in split_string[0] and '.' not in split_string[1][:-1]:
+383 factor = 10 ** -len(split_string[0].partition('.')[2])
+384 else:
+385 factor = 1
+386 return float(split_string[0]), float(split_string[1][:-1]) * factor
387
-388 if not silent:
-389 print('Fit with', n_parms, 'parameter' + 's' * (n_parms > 1))
-390
-391 y_f = [o.value for o in y]
-392 dy_f = [o.dvalue for o in y]
-393
-394 if np.any(np.asarray(dy_f) <= 0.0):
-395 raise Exception('No y errors available, run the gamma method first.')
-396
-397 p_f = [o.value for o in loc_priors]
-398 dp_f = [o.dvalue for o in loc_priors]
-399
-400 if np.any(np.asarray(dp_f) <= 0.0):
-401 raise Exception('No prior errors available, run the gamma method first.')
-402
-403 if 'initial_guess' in kwargs:
-404 x0 = kwargs.get('initial_guess')
-405 if len(x0) != n_parms:
-406 raise Exception('Initial guess does not have the correct length.')
-407 else:
-408 x0 = p_f
+388 loc_priors = []
+389 for i_n, i_prior in enumerate(priors):
+390 if isinstance(i_prior, Obs):
+391 loc_priors.append(i_prior)
+392 else:
+393 loc_val, loc_dval = extract_val_and_dval(i_prior)
+394 loc_priors.append(cov_Obs(loc_val, loc_dval ** 2, '#prior' + str(i_n) + f"_{np.random.randint(2147483647):010d}"))
+395
+396 output.priors = loc_priors
+397
+398 if not silent:
+399 print('Fit with', n_parms, 'parameter' + 's' * (n_parms > 1))
+400
+401 y_f = [o.value for o in y]
+402 dy_f = [o.dvalue for o in y]
+403
+404 if np.any(np.asarray(dy_f) <= 0.0):
+405 raise Exception('No y errors available, run the gamma method first.')
+406
+407 p_f = [o.value for o in loc_priors]
+408 dp_f = [o.dvalue for o in loc_priors]
409
-410 def chisqfunc(p):
-411 model = func(p, x)
-412 chisq = anp.sum(((y_f - model) / dy_f) ** 2) + anp.sum(((p_f - p) / dp_f) ** 2)
-413 return chisq
-414
-415 if not silent:
-416 print('Method: migrad')
-417
-418 m = iminuit.Minuit(chisqfunc, x0)
-419 m.errordef = 1
-420 m.print_level = 0
-421 if 'tol' in kwargs:
-422 m.tol = kwargs.get('tol')
-423 else:
-424 m.tol = 1e-4
-425 m.migrad()
-426 params = np.asarray(m.values)
+410 if np.any(np.asarray(dp_f) <= 0.0):
+411 raise Exception('No prior errors available, run the gamma method first.')
+412
+413 if 'initial_guess' in kwargs:
+414 x0 = kwargs.get('initial_guess')
+415 if len(x0) != n_parms:
+416 raise Exception('Initial guess does not have the correct length.')
+417 else:
+418 x0 = p_f
+419
+420 def chisqfunc(p):
+421 model = func(p, x)
+422 chisq = anp.sum(((y_f - model) / dy_f) ** 2) + anp.sum(((p_f - p) / dp_f) ** 2)
+423 return chisq
+424
+425 if not silent:
+426 print('Method: migrad')
427
-428 output.chisquare_by_dof = m.fval / len(x)
-429
-430 output.method = 'migrad'
-431
-432 if not silent:
-433 print('chisquare/d.o.f.:', output.chisquare_by_dof)
-434
-435 if not m.fmin.is_valid:
-436 raise Exception('The minimization procedure did not converge.')
+428 m = iminuit.Minuit(chisqfunc, x0)
+429 m.errordef = 1
+430 m.print_level = 0
+431 if 'tol' in kwargs:
+432 m.tol = kwargs.get('tol')
+433 else:
+434 m.tol = 1e-4
+435 m.migrad()
+436 params = np.asarray(m.values)
437
-438 hess = hessian(chisqfunc)(params)
-439 hess_inv = np.linalg.pinv(hess)
-440
-441 def chisqfunc_compact(d):
-442 model = func(d[:n_parms], x)
-443 chisq = anp.sum(((d[n_parms: n_parms + len(x)] - model) / dy_f) ** 2) + anp.sum(((d[n_parms + len(x):] - d[:n_parms]) / dp_f) ** 2)
-444 return chisq
-445
-446 jac_jac = hessian(chisqfunc_compact)(np.concatenate((params, y_f, p_f)))
+438 output.chisquare_by_dof = m.fval / len(x)
+439
+440 output.method = 'migrad'
+441
+442 if not silent:
+443 print('chisquare/d.o.f.:', output.chisquare_by_dof)
+444
+445 if not m.fmin.is_valid:
+446 raise Exception('The minimization procedure did not converge.')
447
-448 deriv = -hess_inv @ jac_jac[:n_parms, n_parms:]
-449
-450 result = []
-451 for i in range(n_parms):
-452 result.append(derived_observable(lambda x, **kwargs: (x[0] + np.finfo(np.float64).eps) / (y[0].value + np.finfo(np.float64).eps) * params[i], list(y) + list(loc_priors), man_grad=list(deriv[i])))
-453
-454 output.fit_parameters = result
-455 output.chisquare = chisqfunc(np.asarray(params))
-456
-457 if kwargs.get('resplot') is True:
-458 residual_plot(x, y, func, result)
+448 hess = hessian(chisqfunc)(params)
+449 hess_inv = np.linalg.pinv(hess)
+450
+451 def chisqfunc_compact(d):
+452 model = func(d[:n_parms], x)
+453 chisq = anp.sum(((d[n_parms: n_parms + len(x)] - model) / dy_f) ** 2) + anp.sum(((d[n_parms + len(x):] - d[:n_parms]) / dp_f) ** 2)
+454 return chisq
+455
+456 jac_jac = hessian(chisqfunc_compact)(np.concatenate((params, y_f, p_f)))
+457
+458 deriv = -hess_inv @ jac_jac[:n_parms, n_parms:]
459
-460 if kwargs.get('qqplot') is True:
-461 qqplot(x, y, func, result)
-462
-463 return output
-464
-465
-466def _standard_fit(x, y, func, silent=False, **kwargs):
-467
-468 output = Fit_result()
+460 result = []
+461 for i in range(n_parms):
+462 result.append(derived_observable(lambda x, **kwargs: (x[0] + np.finfo(np.float64).eps) / (y[0].value + np.finfo(np.float64).eps) * params[i], list(y) + list(loc_priors), man_grad=list(deriv[i])))
+463
+464 output.fit_parameters = result
+465 output.chisquare = chisqfunc(np.asarray(params))
+466
+467 if kwargs.get('resplot') is True:
+468 residual_plot(x, y, func, result)
469
-470 output.fit_function = func
-471
-472 x = np.asarray(x)
-473
-474 if kwargs.get('num_grad') is True:
-475 jacobian = num_jacobian
-476 hessian = num_hessian
-477 else:
-478 jacobian = auto_jacobian
-479 hessian = auto_hessian
-480
-481 if x.shape[-1] != len(y):
-482 raise Exception('x and y input have to have the same length')
+470 if kwargs.get('qqplot') is True:
+471 qqplot(x, y, func, result)
+472
+473 return output
+474
+475
+476def _standard_fit(x, y, func, silent=False, **kwargs):
+477
+478 output = Fit_result()
+479
+480 output.fit_function = func
+481
+482 x = np.asarray(x)
483
-484 if len(x.shape) > 2:
-485 raise Exception('Unknown format for x values')
-486
-487 if not callable(func):
-488 raise TypeError('func has to be a function.')
-489
-490 for i in range(42):
-491 try:
-492 func(np.arange(i), x.T[0])
-493 except TypeError:
-494 continue
-495 except IndexError:
-496 continue
-497 else:
-498 break
-499 else:
-500 raise RuntimeError("Fit function is not valid.")
-501
-502 n_parms = i
-503
-504 if not silent:
-505 print('Fit with', n_parms, 'parameter' + 's' * (n_parms > 1))
-506
-507 y_f = [o.value for o in y]
-508 dy_f = [o.dvalue for o in y]
-509
-510 if np.any(np.asarray(dy_f) <= 0.0):
-511 raise Exception('No y errors available, run the gamma method first.')
-512
-513 if 'initial_guess' in kwargs:
-514 x0 = kwargs.get('initial_guess')
-515 if len(x0) != n_parms:
-516 raise Exception('Initial guess does not have the correct length: %d vs. %d' % (len(x0), n_parms))
-517 else:
-518 x0 = [0.1] * n_parms
+484 if kwargs.get('num_grad') is True:
+485 jacobian = num_jacobian
+486 hessian = num_hessian
+487 else:
+488 jacobian = auto_jacobian
+489 hessian = auto_hessian
+490
+491 if x.shape[-1] != len(y):
+492 raise Exception('x and y input have to have the same length')
+493
+494 if len(x.shape) > 2:
+495 raise Exception('Unknown format for x values')
+496
+497 if not callable(func):
+498 raise TypeError('func has to be a function.')
+499
+500 for i in range(42):
+501 try:
+502 func(np.arange(i), x.T[0])
+503 except TypeError:
+504 continue
+505 except IndexError:
+506 continue
+507 else:
+508 break
+509 else:
+510 raise RuntimeError("Fit function is not valid.")
+511
+512 n_parms = i
+513
+514 if not silent:
+515 print('Fit with', n_parms, 'parameter' + 's' * (n_parms > 1))
+516
+517 y_f = [o.value for o in y]
+518 dy_f = [o.dvalue for o in y]
519
-520 if kwargs.get('correlated_fit') is True:
-521 corr = covariance(y, correlation=True, **kwargs)
-522 covdiag = np.diag(1 / np.asarray(dy_f))
-523 condn = np.linalg.cond(corr)
-524 if condn > 0.1 / np.finfo(float).eps:
-525 raise Exception(f"Cannot invert correlation matrix as its condition number exceeds machine precision ({condn:1.2e})")
-526 if condn > 1e13:
-527 warnings.warn("Correlation matrix may be ill-conditioned, condition number: {%1.2e}" % (condn), RuntimeWarning)
-528 chol = np.linalg.cholesky(corr)
-529 chol_inv = scipy.linalg.solve_triangular(chol, covdiag, lower=True)
-530
-531 def chisqfunc_corr(p):
-532 model = func(p, x)
-533 chisq = anp.sum(anp.dot(chol_inv, (y_f - model)) ** 2)
-534 return chisq
-535
-536 def chisqfunc(p):
-537 model = func(p, x)
-538 chisq = anp.sum(((y_f - model) / dy_f) ** 2)
-539 return chisq
+520 if np.any(np.asarray(dy_f) <= 0.0):
+521 raise Exception('No y errors available, run the gamma method first.')
+522
+523 if 'initial_guess' in kwargs:
+524 x0 = kwargs.get('initial_guess')
+525 if len(x0) != n_parms:
+526 raise Exception('Initial guess does not have the correct length: %d vs. %d' % (len(x0), n_parms))
+527 else:
+528 x0 = [0.1] * n_parms
+529
+530 if kwargs.get('correlated_fit') is True:
+531 corr = covariance(y, correlation=True, **kwargs)
+532 covdiag = np.diag(1 / np.asarray(dy_f))
+533 condn = np.linalg.cond(corr)
+534 if condn > 0.1 / np.finfo(float).eps:
+535 raise Exception(f"Cannot invert correlation matrix as its condition number exceeds machine precision ({condn:1.2e})")
+536 if condn > 1e13:
+537 warnings.warn("Correlation matrix may be ill-conditioned, condition number: {%1.2e}" % (condn), RuntimeWarning)
+538 chol = np.linalg.cholesky(corr)
+539 chol_inv = scipy.linalg.solve_triangular(chol, covdiag, lower=True)
540
-541 output.method = kwargs.get('method', 'Levenberg-Marquardt')
-542 if not silent:
-543 print('Method:', output.method)
-544
-545 if output.method != 'Levenberg-Marquardt':
-546 if output.method == 'migrad':
-547 fit_result = iminuit.minimize(chisqfunc, x0, tol=1e-4) # Stopping criterion 0.002 * tol * errordef
-548 if kwargs.get('correlated_fit') is True:
-549 fit_result = iminuit.minimize(chisqfunc_corr, fit_result.x, tol=1e-4) # Stopping criterion 0.002 * tol * errordef
-550 output.iterations = fit_result.nfev
-551 else:
-552 fit_result = scipy.optimize.minimize(chisqfunc, x0, method=kwargs.get('method'), tol=1e-12)
-553 if kwargs.get('correlated_fit') is True:
-554 fit_result = scipy.optimize.minimize(chisqfunc_corr, fit_result.x, method=kwargs.get('method'), tol=1e-12)
-555 output.iterations = fit_result.nit
-556
-557 chisquare = fit_result.fun
-558
-559 else:
-560 if kwargs.get('correlated_fit') is True:
-561 def chisqfunc_residuals_corr(p):
-562 model = func(p, x)
-563 chisq = anp.dot(chol_inv, (y_f - model))
-564 return chisq
-565
-566 def chisqfunc_residuals(p):
-567 model = func(p, x)
-568 chisq = ((y_f - model) / dy_f)
-569 return chisq
-570
-571 fit_result = scipy.optimize.least_squares(chisqfunc_residuals, x0, method='lm', ftol=1e-15, gtol=1e-15, xtol=1e-15)
-572 if kwargs.get('correlated_fit') is True:
-573 fit_result = scipy.optimize.least_squares(chisqfunc_residuals_corr, fit_result.x, method='lm', ftol=1e-15, gtol=1e-15, xtol=1e-15)
-574
-575 chisquare = np.sum(fit_result.fun ** 2)
-576 if kwargs.get('correlated_fit') is True:
-577 assert np.isclose(chisquare, chisqfunc_corr(fit_result.x), atol=1e-14)
-578 else:
-579 assert np.isclose(chisquare, chisqfunc(fit_result.x), atol=1e-14)
+541 def chisqfunc_corr(p):
+542 model = func(p, x)
+543 chisq = anp.sum(anp.dot(chol_inv, (y_f - model)) ** 2)
+544 return chisq
+545
+546 def chisqfunc(p):
+547 model = func(p, x)
+548 chisq = anp.sum(((y_f - model) / dy_f) ** 2)
+549 return chisq
+550
+551 output.method = kwargs.get('method', 'Levenberg-Marquardt')
+552 if not silent:
+553 print('Method:', output.method)
+554
+555 if output.method != 'Levenberg-Marquardt':
+556 if output.method == 'migrad':
+557 fit_result = iminuit.minimize(chisqfunc, x0, tol=1e-4) # Stopping criterion 0.002 * tol * errordef
+558 if kwargs.get('correlated_fit') is True:
+559 fit_result = iminuit.minimize(chisqfunc_corr, fit_result.x, tol=1e-4) # Stopping criterion 0.002 * tol * errordef
+560 output.iterations = fit_result.nfev
+561 else:
+562 fit_result = scipy.optimize.minimize(chisqfunc, x0, method=kwargs.get('method'), tol=1e-12)
+563 if kwargs.get('correlated_fit') is True:
+564 fit_result = scipy.optimize.minimize(chisqfunc_corr, fit_result.x, method=kwargs.get('method'), tol=1e-12)
+565 output.iterations = fit_result.nit
+566
+567 chisquare = fit_result.fun
+568
+569 else:
+570 if kwargs.get('correlated_fit') is True:
+571 def chisqfunc_residuals_corr(p):
+572 model = func(p, x)
+573 chisq = anp.dot(chol_inv, (y_f - model))
+574 return chisq
+575
+576 def chisqfunc_residuals(p):
+577 model = func(p, x)
+578 chisq = ((y_f - model) / dy_f)
+579 return chisq
580
-581 output.iterations = fit_result.nfev
-582
-583 if not fit_result.success:
-584 raise Exception('The minimization procedure did not converge.')
-585
-586 if x.shape[-1] - n_parms > 0:
-587 output.chisquare_by_dof = chisquare / (x.shape[-1] - n_parms)
-588 else:
-589 output.chisquare_by_dof = float('nan')
+581 fit_result = scipy.optimize.least_squares(chisqfunc_residuals, x0, method='lm', ftol=1e-15, gtol=1e-15, xtol=1e-15)
+582 if kwargs.get('correlated_fit') is True:
+583 fit_result = scipy.optimize.least_squares(chisqfunc_residuals_corr, fit_result.x, method='lm', ftol=1e-15, gtol=1e-15, xtol=1e-15)
+584
+585 chisquare = np.sum(fit_result.fun ** 2)
+586 if kwargs.get('correlated_fit') is True:
+587 assert np.isclose(chisquare, chisqfunc_corr(fit_result.x), atol=1e-14)
+588 else:
+589 assert np.isclose(chisquare, chisqfunc(fit_result.x), atol=1e-14)
590
-591 output.message = fit_result.message
-592 if not silent:
-593 print(fit_result.message)
-594 print('chisquare/d.o.f.:', output.chisquare_by_dof)
+591 output.iterations = fit_result.nfev
+592
+593 if not fit_result.success:
+594 raise Exception('The minimization procedure did not converge.')
595
-596 if kwargs.get('expected_chisquare') is True:
-597 if kwargs.get('correlated_fit') is not True:
-598 W = np.diag(1 / np.asarray(dy_f))
-599 cov = covariance(y)
-600 A = W @ jacobian(func)(fit_result.x, x)
-601 P_phi = A @ np.linalg.pinv(A.T @ A) @ A.T
-602 expected_chisquare = np.trace((np.identity(x.shape[-1]) - P_phi) @ W @ cov @ W)
-603 output.chisquare_by_expected_chisquare = chisquare / expected_chisquare
-604 if not silent:
-605 print('chisquare/expected_chisquare:',
-606 output.chisquare_by_expected_chisquare)
-607
-608 fitp = fit_result.x
-609 try:
-610 if kwargs.get('correlated_fit') is True:
-611 hess = hessian(chisqfunc_corr)(fitp)
-612 else:
-613 hess = hessian(chisqfunc)(fitp)
-614 except TypeError:
-615 raise Exception("It is required to use autograd.numpy instead of numpy within fit functions, see the documentation for details.") from None
-616
-617 if kwargs.get('correlated_fit') is True:
-618 def chisqfunc_compact(d):
-619 model = func(d[:n_parms], x)
-620 chisq = anp.sum(anp.dot(chol_inv, (d[n_parms:] - model)) ** 2)
-621 return chisq
-622
-623 else:
-624 def chisqfunc_compact(d):
-625 model = func(d[:n_parms], x)
-626 chisq = anp.sum(((d[n_parms:] - model) / dy_f) ** 2)
-627 return chisq
-628
-629 jac_jac = hessian(chisqfunc_compact)(np.concatenate((fitp, y_f)))
-630
-631 # Compute hess^{-1} @ jac_jac[:n_parms, n_parms:] using LAPACK dgesv
-632 try:
-633 deriv = -scipy.linalg.solve(hess, jac_jac[:n_parms, n_parms:])
-634 except np.linalg.LinAlgError:
-635 raise Exception("Cannot invert hessian matrix.")
-636
-637 result = []
-638 for i in range(n_parms):
-639 result.append(derived_observable(lambda x, **kwargs: (x[0] + np.finfo(np.float64).eps) / (y[0].value + np.finfo(np.float64).eps) * fit_result.x[i], list(y), man_grad=list(deriv[i])))
+596 if x.shape[-1] - n_parms > 0:
+597 output.chisquare_by_dof = chisquare / (x.shape[-1] - n_parms)
+598 else:
+599 output.chisquare_by_dof = float('nan')
+600
+601 output.message = fit_result.message
+602 if not silent:
+603 print(fit_result.message)
+604 print('chisquare/d.o.f.:', output.chisquare_by_dof)
+605
+606 if kwargs.get('expected_chisquare') is True:
+607 if kwargs.get('correlated_fit') is not True:
+608 W = np.diag(1 / np.asarray(dy_f))
+609 cov = covariance(y)
+610 A = W @ jacobian(func)(fit_result.x, x)
+611 P_phi = A @ np.linalg.pinv(A.T @ A) @ A.T
+612 expected_chisquare = np.trace((np.identity(x.shape[-1]) - P_phi) @ W @ cov @ W)
+613 output.chisquare_by_expected_chisquare = chisquare / expected_chisquare
+614 if not silent:
+615 print('chisquare/expected_chisquare:',
+616 output.chisquare_by_expected_chisquare)
+617
+618 fitp = fit_result.x
+619 try:
+620 if kwargs.get('correlated_fit') is True:
+621 hess = hessian(chisqfunc_corr)(fitp)
+622 else:
+623 hess = hessian(chisqfunc)(fitp)
+624 except TypeError:
+625 raise Exception("It is required to use autograd.numpy instead of numpy within fit functions, see the documentation for details.") from None
+626
+627 if kwargs.get('correlated_fit') is True:
+628 def chisqfunc_compact(d):
+629 model = func(d[:n_parms], x)
+630 chisq = anp.sum(anp.dot(chol_inv, (d[n_parms:] - model)) ** 2)
+631 return chisq
+632
+633 else:
+634 def chisqfunc_compact(d):
+635 model = func(d[:n_parms], x)
+636 chisq = anp.sum(((d[n_parms:] - model) / dy_f) ** 2)
+637 return chisq
+638
+639 jac_jac = hessian(chisqfunc_compact)(np.concatenate((fitp, y_f)))
640
-641 output.fit_parameters = result
-642
-643 output.chisquare = chisquare
-644 output.dof = x.shape[-1] - n_parms
-645 output.p_value = 1 - scipy.stats.chi2.cdf(output.chisquare, output.dof)
-646 # Hotelling t-squared p-value for correlated fits.
-647 if kwargs.get('correlated_fit') is True:
-648 n_cov = np.min(np.vectorize(lambda x: x.N)(y))
-649 output.t2_p_value = 1 - scipy.stats.f.cdf((n_cov - output.dof) / (output.dof * (n_cov - 1)) * output.chisquare,
-650 output.dof, n_cov - output.dof)
-651
-652 if kwargs.get('resplot') is True:
-653 residual_plot(x, y, func, result)
-654
-655 if kwargs.get('qqplot') is True:
-656 qqplot(x, y, func, result)
-657
-658 return output
-659
-660
-661def fit_lin(x, y, **kwargs):
-662 """Performs a linear fit to y = n + m * x and returns two Obs n, m.
-663
-664 Parameters
-665 ----------
-666 x : list
-667 Can either be a list of floats in which case no xerror is assumed, or
-668 a list of Obs, where the dvalues of the Obs are used as xerror for the fit.
-669 y : list
-670 List of Obs, the dvalues of the Obs are used as yerror for the fit.
-671 """
-672
-673 def f(a, x):
-674 y = a[0] + a[1] * x
-675 return y
-676
-677 if all(isinstance(n, Obs) for n in x):
-678 out = total_least_squares(x, y, f, **kwargs)
-679 return out.fit_parameters
-680 elif all(isinstance(n, float) or isinstance(n, int) for n in x) or isinstance(x, np.ndarray):
-681 out = least_squares(x, y, f, **kwargs)
-682 return out.fit_parameters
-683 else:
-684 raise Exception('Unsupported types for x')
-685
-686
-687def qqplot(x, o_y, func, p):
-688 """Generates a quantile-quantile plot of the fit result which can be used to
-689 check if the residuals of the fit are gaussian distributed.
-690 """
+641 # Compute hess^{-1} @ jac_jac[:n_parms, n_parms:] using LAPACK dgesv
+642 try:
+643 deriv = -scipy.linalg.solve(hess, jac_jac[:n_parms, n_parms:])
+644 except np.linalg.LinAlgError:
+645 raise Exception("Cannot invert hessian matrix.")
+646
+647 result = []
+648 for i in range(n_parms):
+649 result.append(derived_observable(lambda x, **kwargs: (x[0] + np.finfo(np.float64).eps) / (y[0].value + np.finfo(np.float64).eps) * fit_result.x[i], list(y), man_grad=list(deriv[i])))
+650
+651 output.fit_parameters = result
+652
+653 output.chisquare = chisquare
+654 output.dof = x.shape[-1] - n_parms
+655 output.p_value = 1 - scipy.stats.chi2.cdf(output.chisquare, output.dof)
+656 # Hotelling t-squared p-value for correlated fits.
+657 if kwargs.get('correlated_fit') is True:
+658 n_cov = np.min(np.vectorize(lambda x: x.N)(y))
+659 output.t2_p_value = 1 - scipy.stats.f.cdf((n_cov - output.dof) / (output.dof * (n_cov - 1)) * output.chisquare,
+660 output.dof, n_cov - output.dof)
+661
+662 if kwargs.get('resplot') is True:
+663 residual_plot(x, y, func, result)
+664
+665 if kwargs.get('qqplot') is True:
+666 qqplot(x, y, func, result)
+667
+668 return output
+669
+670
+671def fit_lin(x, y, **kwargs):
+672 """Performs a linear fit to y = n + m * x and returns two Obs n, m.
+673
+674 Parameters
+675 ----------
+676 x : list
+677 Can either be a list of floats in which case no xerror is assumed, or
+678 a list of Obs, where the dvalues of the Obs are used as xerror for the fit.
+679 y : list
+680 List of Obs, the dvalues of the Obs are used as yerror for the fit.
+681
+682 Returns
+683 -------
+684 fit_parameters : list[Obs]
+685 LIist of fitted observables.
+686 """
+687
+688 def f(a, x):
+689 y = a[0] + a[1] * x
+690 return y
691
-692 residuals = []
-693 for i_x, i_y in zip(x, o_y):
-694 residuals.append((i_y - func(p, i_x)) / i_y.dvalue)
-695 residuals = sorted(residuals)
-696 my_y = [o.value for o in residuals]
-697 probplot = scipy.stats.probplot(my_y)
-698 my_x = probplot[0][0]
-699 plt.figure(figsize=(8, 8 / 1.618))
-700 plt.errorbar(my_x, my_y, fmt='o')
-701 fit_start = my_x[0]
-702 fit_stop = my_x[-1]
-703 samples = np.arange(fit_start, fit_stop, 0.01)
-704 plt.plot(samples, samples, 'k--', zorder=11, label='Standard normal distribution')
-705 plt.plot(samples, probplot[1][0] * samples + probplot[1][1], zorder=10, label='Least squares fit, r=' + str(np.around(probplot[1][2], 3)), marker='', ls='-')
-706
-707 plt.xlabel('Theoretical quantiles')
-708 plt.ylabel('Ordered Values')
-709 plt.legend()
-710 plt.draw()
-711
-712
-713def residual_plot(x, y, func, fit_res):
-714 """ Generates a plot which compares the fit to the data and displays the corresponding residuals"""
-715 sorted_x = sorted(x)
-716 xstart = sorted_x[0] - 0.5 * (sorted_x[1] - sorted_x[0])
-717 xstop = sorted_x[-1] + 0.5 * (sorted_x[-1] - sorted_x[-2])
-718 x_samples = np.arange(xstart, xstop + 0.01, 0.01)
-719
-720 plt.figure(figsize=(8, 8 / 1.618))
-721 gs = gridspec.GridSpec(2, 1, height_ratios=[3, 1], wspace=0.0, hspace=0.0)
-722 ax0 = plt.subplot(gs[0])
-723 ax0.errorbar(x, [o.value for o in y], yerr=[o.dvalue for o in y], ls='none', fmt='o', capsize=3, markersize=5, label='Data')
-724 ax0.plot(x_samples, func([o.value for o in fit_res], x_samples), label='Fit', zorder=10, ls='-', ms=0)
-725 ax0.set_xticklabels([])
-726 ax0.set_xlim([xstart, xstop])
-727 ax0.set_xticklabels([])
-728 ax0.legend()
-729
-730 residuals = (np.asarray([o.value for o in y]) - func([o.value for o in fit_res], x)) / np.asarray([o.dvalue for o in y])
-731 ax1 = plt.subplot(gs[1])
-732 ax1.plot(x, residuals, 'ko', ls='none', markersize=5)
-733 ax1.tick_params(direction='out')
-734 ax1.tick_params(axis="x", bottom=True, top=True, labelbottom=True)
-735 ax1.axhline(y=0.0, ls='--', color='k', marker=" ")
-736 ax1.fill_between(x_samples, -1.0, 1.0, alpha=0.1, facecolor='k')
-737 ax1.set_xlim([xstart, xstop])
-738 ax1.set_ylabel('Residuals')
-739 plt.subplots_adjust(wspace=None, hspace=None)
-740 plt.draw()
-741
-742
-743def error_band(x, func, beta):
-744 """Returns the error band for an array of sample values x, for given fit function func with optimized parameters beta."""
-745 cov = covariance(beta)
-746 if np.any(np.abs(cov - cov.T) > 1000 * np.finfo(np.float64).eps):
-747 warnings.warn("Covariance matrix is not symmetric within floating point precision", RuntimeWarning)
-748
-749 deriv = []
-750 for i, item in enumerate(x):
-751 deriv.append(np.array(egrad(func)([o.value for o in beta], item)))
-752
-753 err = []
-754 for i, item in enumerate(x):
-755 err.append(np.sqrt(deriv[i] @ cov @ deriv[i]))
-756 err = np.array(err)
-757
-758 return err
-759
-760
-761def ks_test(objects=None):
-762 """Performs a Kolmogorov–Smirnov test for the p-values of all fit object.
-763
-764 Parameters
-765 ----------
-766 objects : list
-767 List of fit results to include in the analysis (optional).
-768 """
+692 if all(isinstance(n, Obs) for n in x):
+693 out = total_least_squares(x, y, f, **kwargs)
+694 return out.fit_parameters
+695 elif all(isinstance(n, float) or isinstance(n, int) for n in x) or isinstance(x, np.ndarray):
+696 out = least_squares(x, y, f, **kwargs)
+697 return out.fit_parameters
+698 else:
+699 raise Exception('Unsupported types for x')
+700
+701
+702def qqplot(x, o_y, func, p):
+703 """Generates a quantile-quantile plot of the fit result which can be used to
+704 check if the residuals of the fit are gaussian distributed.
+705
+706 Returns
+707 -------
+708 None
+709 """
+710
+711 residuals = []
+712 for i_x, i_y in zip(x, o_y):
+713 residuals.append((i_y - func(p, i_x)) / i_y.dvalue)
+714 residuals = sorted(residuals)
+715 my_y = [o.value for o in residuals]
+716 probplot = scipy.stats.probplot(my_y)
+717 my_x = probplot[0][0]
+718 plt.figure(figsize=(8, 8 / 1.618))
+719 plt.errorbar(my_x, my_y, fmt='o')
+720 fit_start = my_x[0]
+721 fit_stop = my_x[-1]
+722 samples = np.arange(fit_start, fit_stop, 0.01)
+723 plt.plot(samples, samples, 'k--', zorder=11, label='Standard normal distribution')
+724 plt.plot(samples, probplot[1][0] * samples + probplot[1][1], zorder=10, label='Least squares fit, r=' + str(np.around(probplot[1][2], 3)), marker='', ls='-')
+725
+726 plt.xlabel('Theoretical quantiles')
+727 plt.ylabel('Ordered Values')
+728 plt.legend()
+729 plt.draw()
+730
+731
+732def residual_plot(x, y, func, fit_res):
+733 """Generates a plot which compares the fit to the data and displays the corresponding residuals
+734
+735 Returns
+736 -------
+737 None
+738 """
+739 sorted_x = sorted(x)
+740 xstart = sorted_x[0] - 0.5 * (sorted_x[1] - sorted_x[0])
+741 xstop = sorted_x[-1] + 0.5 * (sorted_x[-1] - sorted_x[-2])
+742 x_samples = np.arange(xstart, xstop + 0.01, 0.01)
+743
+744 plt.figure(figsize=(8, 8 / 1.618))
+745 gs = gridspec.GridSpec(2, 1, height_ratios=[3, 1], wspace=0.0, hspace=0.0)
+746 ax0 = plt.subplot(gs[0])
+747 ax0.errorbar(x, [o.value for o in y], yerr=[o.dvalue for o in y], ls='none', fmt='o', capsize=3, markersize=5, label='Data')
+748 ax0.plot(x_samples, func([o.value for o in fit_res], x_samples), label='Fit', zorder=10, ls='-', ms=0)
+749 ax0.set_xticklabels([])
+750 ax0.set_xlim([xstart, xstop])
+751 ax0.set_xticklabels([])
+752 ax0.legend()
+753
+754 residuals = (np.asarray([o.value for o in y]) - func([o.value for o in fit_res], x)) / np.asarray([o.dvalue for o in y])
+755 ax1 = plt.subplot(gs[1])
+756 ax1.plot(x, residuals, 'ko', ls='none', markersize=5)
+757 ax1.tick_params(direction='out')
+758 ax1.tick_params(axis="x", bottom=True, top=True, labelbottom=True)
+759 ax1.axhline(y=0.0, ls='--', color='k', marker=" ")
+760 ax1.fill_between(x_samples, -1.0, 1.0, alpha=0.1, facecolor='k')
+761 ax1.set_xlim([xstart, xstop])
+762 ax1.set_ylabel('Residuals')
+763 plt.subplots_adjust(wspace=None, hspace=None)
+764 plt.draw()
+765
+766
+767def error_band(x, func, beta):
+768 """Calculate the error band for an array of sample values x, for given fit function func with optimized parameters beta.
769
-770 if objects is None:
-771 obs_list = []
-772 for obj in gc.get_objects():
-773 if isinstance(obj, Fit_result):
-774 obs_list.append(obj)
-775 else:
-776 obs_list = objects
-777
-778 p_values = [o.p_value for o in obs_list]
-779
-780 bins = len(p_values)
-781 x = np.arange(0, 1.001, 0.001)
-782 plt.plot(x, x, 'k', zorder=1)
-783 plt.xlim(0, 1)
-784 plt.ylim(0, 1)
-785 plt.xlabel('p-value')
-786 plt.ylabel('Cumulative probability')
-787 plt.title(str(bins) + ' p-values')
-788
-789 n = np.arange(1, bins + 1) / np.float64(bins)
-790 Xs = np.sort(p_values)
-791 plt.step(Xs, n)
-792 diffs = n - Xs
-793 loc_max_diff = np.argmax(np.abs(diffs))
-794 loc = Xs[loc_max_diff]
-795 plt.annotate('', xy=(loc, loc), xytext=(loc, loc + diffs[loc_max_diff]), arrowprops=dict(arrowstyle='<->', shrinkA=0, shrinkB=0))
-796 plt.draw()
-797
-798 print(scipy.stats.kstest(p_values, 'uniform'))
+770 Returns
+771 -------
+772 err : np.array(Obs)
+773 Error band for an array of sample values x
+774 """
+775 cov = covariance(beta)
+776 if np.any(np.abs(cov - cov.T) > 1000 * np.finfo(np.float64).eps):
+777 warnings.warn("Covariance matrix is not symmetric within floating point precision", RuntimeWarning)
+778
+779 deriv = []
+780 for i, item in enumerate(x):
+781 deriv.append(np.array(egrad(func)([o.value for o in beta], item)))
+782
+783 err = []
+784 for i, item in enumerate(x):
+785 err.append(np.sqrt(deriv[i] @ cov @ deriv[i]))
+786 err = np.array(err)
+787
+788 return err
+789
+790
+791def ks_test(objects=None):
+792 """Performs a Kolmogorov–Smirnov test for the p-values of all fit object.
+793
+794 Parameters
+795 ----------
+796 objects : list
+797 List of fit results to include in the analysis (optional).
+798
+799 Returns
+800 -------
+801 None
+802 """
+803
+804 if objects is None:
+805 obs_list = []
+806 for obj in gc.get_objects():
+807 if isinstance(obj, Fit_result):
+808 obs_list.append(obj)
+809 else:
+810 obs_list = objects
+811
+812 p_values = [o.p_value for o in obs_list]
+813
+814 bins = len(p_values)
+815 x = np.arange(0, 1.001, 0.001)
+816 plt.plot(x, x, 'k', zorder=1)
+817 plt.xlim(0, 1)
+818 plt.ylim(0, 1)
+819 plt.xlabel('p-value')
+820 plt.ylabel('Cumulative probability')
+821 plt.title(str(bins) + ' p-values')
+822
+823 n = np.arange(1, bins + 1) / np.float64(bins)
+824 Xs = np.sort(p_values)
+825 plt.step(Xs, n)
+826 diffs = n - Xs
+827 loc_max_diff = np.argmax(np.abs(diffs))
+828 loc = Xs[loc_max_diff]
+829 plt.annotate('', xy=(loc, loc), xytext=(loc, loc + diffs[loc_max_diff]), arrowprops=dict(arrowstyle='<->', shrinkA=0, shrinkB=0))
+830 plt.draw()
+831
+832 print(scipy.stats.kstest(p_values, 'uniform'))
@@ -1137,11 +1171,16 @@ Hotelling t-squared p-value for correlated fits.
130 If True, a quantile-quantile plot of the fit result is generated (default False).
131 num_grad : bool
132 Use numerical differentation instead of automatic differentiation to perform the error propagation (default False).
-133 '''
-134 if priors is not None:
-135 return _prior_fit(x, y, func, priors, silent=silent, **kwargs)
-136 else:
-137 return _standard_fit(x, y, func, silent=silent, **kwargs)
+133
+134 Returns
+135 -------
+136 output : Fit_result
+137 Parameters and information on the fitted result.
+138 '''
+139 if priors is not None:
+140 return _prior_fit(x, y, func, priors, silent=silent, **kwargs)
+141 else:
+142 return _standard_fit(x, y, func, silent=silent, **kwargs)
@@ -1207,6 +1246,13 @@ If True, a quantile-quantile plot of the fit result is generated (default False)
num_grad (bool):
Use numerical differentation instead of automatic differentiation to perform the error propagation (default False).
+
+Returns
+
+
+- output (Fit_result):
+Parameters and information on the fitted result.
+
@@ -1222,203 +1268,208 @@ Use numerical differentation instead of automatic differentiation to perform the
- 140def total_least_squares(x, y, func, silent=False, **kwargs):
-141 r'''Performs a non-linear fit to y = func(x) and returns a list of Obs corresponding to the fit parameters.
-142
-143 Parameters
-144 ----------
-145 x : list
-146 list of Obs, or a tuple of lists of Obs
-147 y : list
-148 list of Obs. The dvalues of the Obs are used as x- and yerror for the fit.
-149 func : object
-150 func has to be of the form
-151
-152 ```python
-153 import autograd.numpy as anp
-154
-155 def func(a, x):
-156 return a[0] + a[1] * x + a[2] * anp.sinh(x)
-157 ```
-158
-159 For multiple x values func can be of the form
-160
-161 ```python
-162 def func(a, x):
-163 (x1, x2) = x
-164 return a[0] * x1 ** 2 + a[1] * x2
-165 ```
-166
-167 It is important that all numpy functions refer to autograd.numpy, otherwise the differentiation
-168 will not work.
-169 silent : bool, optional
-170 If true all output to the console is omitted (default False).
-171 initial_guess : list
-172 can provide an initial guess for the input parameters. Relevant for non-linear
-173 fits with many parameters.
-174 expected_chisquare : bool
-175 If true prints the expected chisquare which is
-176 corrected by effects caused by correlated input data.
-177 This can take a while as the full correlation matrix
-178 has to be calculated (default False).
-179 num_grad : bool
-180 Use numerical differentation instead of automatic differentiation to perform the error propagation (default False).
-181
-182 Notes
-183 -----
-184 Based on the orthogonal distance regression module of scipy
-185 '''
+ 145def total_least_squares(x, y, func, silent=False, **kwargs):
+146 r'''Performs a non-linear fit to y = func(x) and returns a list of Obs corresponding to the fit parameters.
+147
+148 Parameters
+149 ----------
+150 x : list
+151 list of Obs, or a tuple of lists of Obs
+152 y : list
+153 list of Obs. The dvalues of the Obs are used as x- and yerror for the fit.
+154 func : object
+155 func has to be of the form
+156
+157 ```python
+158 import autograd.numpy as anp
+159
+160 def func(a, x):
+161 return a[0] + a[1] * x + a[2] * anp.sinh(x)
+162 ```
+163
+164 For multiple x values func can be of the form
+165
+166 ```python
+167 def func(a, x):
+168 (x1, x2) = x
+169 return a[0] * x1 ** 2 + a[1] * x2
+170 ```
+171
+172 It is important that all numpy functions refer to autograd.numpy, otherwise the differentiation
+173 will not work.
+174 silent : bool, optional
+175 If true all output to the console is omitted (default False).
+176 initial_guess : list
+177 can provide an initial guess for the input parameters. Relevant for non-linear
+178 fits with many parameters.
+179 expected_chisquare : bool
+180 If true prints the expected chisquare which is
+181 corrected by effects caused by correlated input data.
+182 This can take a while as the full correlation matrix
+183 has to be calculated (default False).
+184 num_grad : bool
+185 Use numerical differentation instead of automatic differentiation to perform the error propagation (default False).
186
-187 output = Fit_result()
-188
-189 output.fit_function = func
+187 Notes
+188 -----
+189 Based on the orthogonal distance regression module of scipy.
190
-191 x = np.array(x)
-192
-193 x_shape = x.shape
-194
-195 if kwargs.get('num_grad') is True:
-196 jacobian = num_jacobian
-197 hessian = num_hessian
-198 else:
-199 jacobian = auto_jacobian
-200 hessian = auto_hessian
-201
-202 if not callable(func):
-203 raise TypeError('func has to be a function.')
+191 Returns
+192 -------
+193 output : Fit_result
+194 Parameters and information on the fitted result.
+195 '''
+196
+197 output = Fit_result()
+198
+199 output.fit_function = func
+200
+201 x = np.array(x)
+202
+203 x_shape = x.shape
204
-205 for i in range(42):
-206 try:
-207 func(np.arange(i), x.T[0])
-208 except TypeError:
-209 continue
-210 except IndexError:
-211 continue
-212 else:
-213 break
-214 else:
-215 raise RuntimeError("Fit function is not valid.")
-216
-217 n_parms = i
-218 if not silent:
-219 print('Fit with', n_parms, 'parameter' + 's' * (n_parms > 1))
-220
-221 x_f = np.vectorize(lambda o: o.value)(x)
-222 dx_f = np.vectorize(lambda o: o.dvalue)(x)
-223 y_f = np.array([o.value for o in y])
-224 dy_f = np.array([o.dvalue for o in y])
-225
-226 if np.any(np.asarray(dx_f) <= 0.0):
-227 raise Exception('No x errors available, run the gamma method first.')
-228
-229 if np.any(np.asarray(dy_f) <= 0.0):
-230 raise Exception('No y errors available, run the gamma method first.')
-231
-232 if 'initial_guess' in kwargs:
-233 x0 = kwargs.get('initial_guess')
-234 if len(x0) != n_parms:
-235 raise Exception('Initial guess does not have the correct length: %d vs. %d' % (len(x0), n_parms))
-236 else:
-237 x0 = [1] * n_parms
+205 if kwargs.get('num_grad') is True:
+206 jacobian = num_jacobian
+207 hessian = num_hessian
+208 else:
+209 jacobian = auto_jacobian
+210 hessian = auto_hessian
+211
+212 if not callable(func):
+213 raise TypeError('func has to be a function.')
+214
+215 for i in range(42):
+216 try:
+217 func(np.arange(i), x.T[0])
+218 except TypeError:
+219 continue
+220 except IndexError:
+221 continue
+222 else:
+223 break
+224 else:
+225 raise RuntimeError("Fit function is not valid.")
+226
+227 n_parms = i
+228 if not silent:
+229 print('Fit with', n_parms, 'parameter' + 's' * (n_parms > 1))
+230
+231 x_f = np.vectorize(lambda o: o.value)(x)
+232 dx_f = np.vectorize(lambda o: o.dvalue)(x)
+233 y_f = np.array([o.value for o in y])
+234 dy_f = np.array([o.dvalue for o in y])
+235
+236 if np.any(np.asarray(dx_f) <= 0.0):
+237 raise Exception('No x errors available, run the gamma method first.')
238
-239 data = RealData(x_f, y_f, sx=dx_f, sy=dy_f)
-240 model = Model(func)
-241 odr = ODR(data, model, x0, partol=np.finfo(np.float64).eps)
-242 odr.set_job(fit_type=0, deriv=1)
-243 out = odr.run()
-244
-245 output.residual_variance = out.res_var
-246
-247 output.method = 'ODR'
+239 if np.any(np.asarray(dy_f) <= 0.0):
+240 raise Exception('No y errors available, run the gamma method first.')
+241
+242 if 'initial_guess' in kwargs:
+243 x0 = kwargs.get('initial_guess')
+244 if len(x0) != n_parms:
+245 raise Exception('Initial guess does not have the correct length: %d vs. %d' % (len(x0), n_parms))
+246 else:
+247 x0 = [1] * n_parms
248
-249 output.message = out.stopreason
-250
-251 output.xplus = out.xplus
-252
-253 if not silent:
-254 print('Method: ODR')
-255 print(*out.stopreason)
-256 print('Residual variance:', output.residual_variance)
-257
-258 if out.info > 3:
-259 raise Exception('The minimization procedure did not converge.')
+249 data = RealData(x_f, y_f, sx=dx_f, sy=dy_f)
+250 model = Model(func)
+251 odr = ODR(data, model, x0, partol=np.finfo(np.float64).eps)
+252 odr.set_job(fit_type=0, deriv=1)
+253 out = odr.run()
+254
+255 output.residual_variance = out.res_var
+256
+257 output.method = 'ODR'
+258
+259 output.message = out.stopreason
260
-261 m = x_f.size
+261 output.xplus = out.xplus
262
-263 def odr_chisquare(p):
-264 model = func(p[:n_parms], p[n_parms:].reshape(x_shape))
-265 chisq = anp.sum(((y_f - model) / dy_f) ** 2) + anp.sum(((x_f - p[n_parms:].reshape(x_shape)) / dx_f) ** 2)
-266 return chisq
+263 if not silent:
+264 print('Method: ODR')
+265 print(*out.stopreason)
+266 print('Residual variance:', output.residual_variance)
267
-268 if kwargs.get('expected_chisquare') is True:
-269 W = np.diag(1 / np.asarray(np.concatenate((dy_f.ravel(), dx_f.ravel()))))
+268 if out.info > 3:
+269 raise Exception('The minimization procedure did not converge.')
270
-271 if kwargs.get('covariance') is not None:
-272 cov = kwargs.get('covariance')
-273 else:
-274 cov = covariance(np.concatenate((y, x.ravel())))
-275
-276 number_of_x_parameters = int(m / x_f.shape[-1])
+271 m = x_f.size
+272
+273 def odr_chisquare(p):
+274 model = func(p[:n_parms], p[n_parms:].reshape(x_shape))
+275 chisq = anp.sum(((y_f - model) / dy_f) ** 2) + anp.sum(((x_f - p[n_parms:].reshape(x_shape)) / dx_f) ** 2)
+276 return chisq
277
-278 old_jac = jacobian(func)(out.beta, out.xplus)
-279 fused_row1 = np.concatenate((old_jac, np.concatenate((number_of_x_parameters * [np.zeros(old_jac.shape)]), axis=0)))
-280 fused_row2 = np.concatenate((jacobian(lambda x, y: func(y, x))(out.xplus, out.beta).reshape(x_f.shape[-1], x_f.shape[-1] * number_of_x_parameters), np.identity(number_of_x_parameters * old_jac.shape[0])))
-281 new_jac = np.concatenate((fused_row1, fused_row2), axis=1)
-282
-283 A = W @ new_jac
-284 P_phi = A @ np.linalg.pinv(A.T @ A) @ A.T
-285 expected_chisquare = np.trace((np.identity(P_phi.shape[0]) - P_phi) @ W @ cov @ W)
-286 if expected_chisquare <= 0.0:
-287 warnings.warn("Negative expected_chisquare.", RuntimeWarning)
-288 expected_chisquare = np.abs(expected_chisquare)
-289 output.chisquare_by_expected_chisquare = odr_chisquare(np.concatenate((out.beta, out.xplus.ravel()))) / expected_chisquare
-290 if not silent:
-291 print('chisquare/expected_chisquare:',
-292 output.chisquare_by_expected_chisquare)
-293
-294 fitp = out.beta
-295 try:
-296 hess = hessian(odr_chisquare)(np.concatenate((fitp, out.xplus.ravel())))
-297 except TypeError:
-298 raise Exception("It is required to use autograd.numpy instead of numpy within fit functions, see the documentation for details.") from None
-299
-300 def odr_chisquare_compact_x(d):
-301 model = func(d[:n_parms], d[n_parms:n_parms + m].reshape(x_shape))
-302 chisq = anp.sum(((y_f - model) / dy_f) ** 2) + anp.sum(((d[n_parms + m:].reshape(x_shape) - d[n_parms:n_parms + m].reshape(x_shape)) / dx_f) ** 2)
-303 return chisq
-304
-305 jac_jac_x = hessian(odr_chisquare_compact_x)(np.concatenate((fitp, out.xplus.ravel(), x_f.ravel())))
-306
-307 # Compute hess^{-1} @ jac_jac_x[:n_parms + m, n_parms + m:] using LAPACK dgesv
-308 try:
-309 deriv_x = -scipy.linalg.solve(hess, jac_jac_x[:n_parms + m, n_parms + m:])
-310 except np.linalg.LinAlgError:
-311 raise Exception("Cannot invert hessian matrix.")
-312
-313 def odr_chisquare_compact_y(d):
-314 model = func(d[:n_parms], d[n_parms:n_parms + m].reshape(x_shape))
-315 chisq = anp.sum(((d[n_parms + m:] - model) / dy_f) ** 2) + anp.sum(((x_f - d[n_parms:n_parms + m].reshape(x_shape)) / dx_f) ** 2)
-316 return chisq
-317
-318 jac_jac_y = hessian(odr_chisquare_compact_y)(np.concatenate((fitp, out.xplus.ravel(), y_f)))
-319
-320 # Compute hess^{-1} @ jac_jac_y[:n_parms + m, n_parms + m:] using LAPACK dgesv
-321 try:
-322 deriv_y = -scipy.linalg.solve(hess, jac_jac_y[:n_parms + m, n_parms + m:])
-323 except np.linalg.LinAlgError:
-324 raise Exception("Cannot invert hessian matrix.")
-325
-326 result = []
-327 for i in range(n_parms):
-328 result.append(derived_observable(lambda my_var, **kwargs: (my_var[0] + np.finfo(np.float64).eps) / (x.ravel()[0].value + np.finfo(np.float64).eps) * out.beta[i], list(x.ravel()) + list(y), man_grad=list(deriv_x[i]) + list(deriv_y[i])))
+278 if kwargs.get('expected_chisquare') is True:
+279 W = np.diag(1 / np.asarray(np.concatenate((dy_f.ravel(), dx_f.ravel()))))
+280
+281 if kwargs.get('covariance') is not None:
+282 cov = kwargs.get('covariance')
+283 else:
+284 cov = covariance(np.concatenate((y, x.ravel())))
+285
+286 number_of_x_parameters = int(m / x_f.shape[-1])
+287
+288 old_jac = jacobian(func)(out.beta, out.xplus)
+289 fused_row1 = np.concatenate((old_jac, np.concatenate((number_of_x_parameters * [np.zeros(old_jac.shape)]), axis=0)))
+290 fused_row2 = np.concatenate((jacobian(lambda x, y: func(y, x))(out.xplus, out.beta).reshape(x_f.shape[-1], x_f.shape[-1] * number_of_x_parameters), np.identity(number_of_x_parameters * old_jac.shape[0])))
+291 new_jac = np.concatenate((fused_row1, fused_row2), axis=1)
+292
+293 A = W @ new_jac
+294 P_phi = A @ np.linalg.pinv(A.T @ A) @ A.T
+295 expected_chisquare = np.trace((np.identity(P_phi.shape[0]) - P_phi) @ W @ cov @ W)
+296 if expected_chisquare <= 0.0:
+297 warnings.warn("Negative expected_chisquare.", RuntimeWarning)
+298 expected_chisquare = np.abs(expected_chisquare)
+299 output.chisquare_by_expected_chisquare = odr_chisquare(np.concatenate((out.beta, out.xplus.ravel()))) / expected_chisquare
+300 if not silent:
+301 print('chisquare/expected_chisquare:',
+302 output.chisquare_by_expected_chisquare)
+303
+304 fitp = out.beta
+305 try:
+306 hess = hessian(odr_chisquare)(np.concatenate((fitp, out.xplus.ravel())))
+307 except TypeError:
+308 raise Exception("It is required to use autograd.numpy instead of numpy within fit functions, see the documentation for details.") from None
+309
+310 def odr_chisquare_compact_x(d):
+311 model = func(d[:n_parms], d[n_parms:n_parms + m].reshape(x_shape))
+312 chisq = anp.sum(((y_f - model) / dy_f) ** 2) + anp.sum(((d[n_parms + m:].reshape(x_shape) - d[n_parms:n_parms + m].reshape(x_shape)) / dx_f) ** 2)
+313 return chisq
+314
+315 jac_jac_x = hessian(odr_chisquare_compact_x)(np.concatenate((fitp, out.xplus.ravel(), x_f.ravel())))
+316
+317 # Compute hess^{-1} @ jac_jac_x[:n_parms + m, n_parms + m:] using LAPACK dgesv
+318 try:
+319 deriv_x = -scipy.linalg.solve(hess, jac_jac_x[:n_parms + m, n_parms + m:])
+320 except np.linalg.LinAlgError:
+321 raise Exception("Cannot invert hessian matrix.")
+322
+323 def odr_chisquare_compact_y(d):
+324 model = func(d[:n_parms], d[n_parms:n_parms + m].reshape(x_shape))
+325 chisq = anp.sum(((d[n_parms + m:] - model) / dy_f) ** 2) + anp.sum(((x_f - d[n_parms:n_parms + m].reshape(x_shape)) / dx_f) ** 2)
+326 return chisq
+327
+328 jac_jac_y = hessian(odr_chisquare_compact_y)(np.concatenate((fitp, out.xplus.ravel(), y_f)))
329
-330 output.fit_parameters = result
-331
-332 output.odr_chisquare = odr_chisquare(np.concatenate((out.beta, out.xplus.ravel())))
-333 output.dof = x.shape[-1] - n_parms
-334 output.p_value = 1 - scipy.stats.chi2.cdf(output.odr_chisquare, output.dof)
+330 # Compute hess^{-1} @ jac_jac_y[:n_parms + m, n_parms + m:] using LAPACK dgesv
+331 try:
+332 deriv_y = -scipy.linalg.solve(hess, jac_jac_y[:n_parms + m, n_parms + m:])
+333 except np.linalg.LinAlgError:
+334 raise Exception("Cannot invert hessian matrix.")
335
-336 return output
+336 result = []
+337 for i in range(n_parms):
+338 result.append(derived_observable(lambda my_var, **kwargs: (my_var[0] + np.finfo(np.float64).eps) / (x.ravel()[0].value + np.finfo(np.float64).eps) * out.beta[i], list(x.ravel()) + list(y), man_grad=list(deriv_x[i]) + list(deriv_y[i])))
+339
+340 output.fit_parameters = result
+341
+342 output.odr_chisquare = odr_chisquare(np.concatenate((out.beta, out.xplus.ravel())))
+343 output.dof = x.shape[-1] - n_parms
+344 output.p_value = 1 - scipy.stats.chi2.cdf(output.odr_chisquare, output.dof)
+345
+346 return output
@@ -1469,7 +1520,14 @@ Use numerical differentation instead of automatic differentiation to perform the
Notes
-Based on the orthogonal distance regression module of scipy
+Based on the orthogonal distance regression module of scipy.
+
+Returns
+
+
+- output (Fit_result):
+Parameters and information on the fitted result.
+
@@ -1485,30 +1543,35 @@ Use numerical differentation instead of automatic differentiation to perform the
- 662def fit_lin(x, y, **kwargs):
-663 """Performs a linear fit to y = n + m * x and returns two Obs n, m.
-664
-665 Parameters
-666 ----------
-667 x : list
-668 Can either be a list of floats in which case no xerror is assumed, or
-669 a list of Obs, where the dvalues of the Obs are used as xerror for the fit.
-670 y : list
-671 List of Obs, the dvalues of the Obs are used as yerror for the fit.
-672 """
-673
-674 def f(a, x):
-675 y = a[0] + a[1] * x
-676 return y
-677
-678 if all(isinstance(n, Obs) for n in x):
-679 out = total_least_squares(x, y, f, **kwargs)
-680 return out.fit_parameters
-681 elif all(isinstance(n, float) or isinstance(n, int) for n in x) or isinstance(x, np.ndarray):
-682 out = least_squares(x, y, f, **kwargs)
-683 return out.fit_parameters
-684 else:
-685 raise Exception('Unsupported types for x')
+ 672def fit_lin(x, y, **kwargs):
+673 """Performs a linear fit to y = n + m * x and returns two Obs n, m.
+674
+675 Parameters
+676 ----------
+677 x : list
+678 Can either be a list of floats in which case no xerror is assumed, or
+679 a list of Obs, where the dvalues of the Obs are used as xerror for the fit.
+680 y : list
+681 List of Obs, the dvalues of the Obs are used as yerror for the fit.
+682
+683 Returns
+684 -------
+685 fit_parameters : list[Obs]
+686 LIist of fitted observables.
+687 """
+688
+689 def f(a, x):
+690 y = a[0] + a[1] * x
+691 return y
+692
+693 if all(isinstance(n, Obs) for n in x):
+694 out = total_least_squares(x, y, f, **kwargs)
+695 return out.fit_parameters
+696 elif all(isinstance(n, float) or isinstance(n, int) for n in x) or isinstance(x, np.ndarray):
+697 out = least_squares(x, y, f, **kwargs)
+698 return out.fit_parameters
+699 else:
+700 raise Exception('Unsupported types for x')
@@ -1523,6 +1586,13 @@ a list of Obs, where the dvalues of the Obs are used as xerror for the fit.
y (list):
List of Obs, the dvalues of the Obs are used as yerror for the fit.
+
+Returns
+
+
+- fit_parameters (list[Obs]):
+LIist of fitted observables.
+
@@ -1538,35 +1608,45 @@ List of Obs, the dvalues of the Obs are used as yerror for the fit.
- 688def qqplot(x, o_y, func, p):
-689 """Generates a quantile-quantile plot of the fit result which can be used to
-690 check if the residuals of the fit are gaussian distributed.
-691 """
-692
-693 residuals = []
-694 for i_x, i_y in zip(x, o_y):
-695 residuals.append((i_y - func(p, i_x)) / i_y.dvalue)
-696 residuals = sorted(residuals)
-697 my_y = [o.value for o in residuals]
-698 probplot = scipy.stats.probplot(my_y)
-699 my_x = probplot[0][0]
-700 plt.figure(figsize=(8, 8 / 1.618))
-701 plt.errorbar(my_x, my_y, fmt='o')
-702 fit_start = my_x[0]
-703 fit_stop = my_x[-1]
-704 samples = np.arange(fit_start, fit_stop, 0.01)
-705 plt.plot(samples, samples, 'k--', zorder=11, label='Standard normal distribution')
-706 plt.plot(samples, probplot[1][0] * samples + probplot[1][1], zorder=10, label='Least squares fit, r=' + str(np.around(probplot[1][2], 3)), marker='', ls='-')
-707
-708 plt.xlabel('Theoretical quantiles')
-709 plt.ylabel('Ordered Values')
-710 plt.legend()
-711 plt.draw()
+ 703def qqplot(x, o_y, func, p):
+704 """Generates a quantile-quantile plot of the fit result which can be used to
+705 check if the residuals of the fit are gaussian distributed.
+706
+707 Returns
+708 -------
+709 None
+710 """
+711
+712 residuals = []
+713 for i_x, i_y in zip(x, o_y):
+714 residuals.append((i_y - func(p, i_x)) / i_y.dvalue)
+715 residuals = sorted(residuals)
+716 my_y = [o.value for o in residuals]
+717 probplot = scipy.stats.probplot(my_y)
+718 my_x = probplot[0][0]
+719 plt.figure(figsize=(8, 8 / 1.618))
+720 plt.errorbar(my_x, my_y, fmt='o')
+721 fit_start = my_x[0]
+722 fit_stop = my_x[-1]
+723 samples = np.arange(fit_start, fit_stop, 0.01)
+724 plt.plot(samples, samples, 'k--', zorder=11, label='Standard normal distribution')
+725 plt.plot(samples, probplot[1][0] * samples + probplot[1][1], zorder=10, label='Least squares fit, r=' + str(np.around(probplot[1][2], 3)), marker='', ls='-')
+726
+727 plt.xlabel('Theoretical quantiles')
+728 plt.ylabel('Ordered Values')
+729 plt.legend()
+730 plt.draw()
Generates a quantile-quantile plot of the fit result which can be used to
-check if the residuals of the fit are gaussian distributed.
+ check if the residuals of the fit are gaussian distributed.
+
+
Returns
+
+
@@ -1582,38 +1662,49 @@ check if the residuals of the fit are gaussian distributed.
- 714def residual_plot(x, y, func, fit_res):
-715 """ Generates a plot which compares the fit to the data and displays the corresponding residuals"""
-716 sorted_x = sorted(x)
-717 xstart = sorted_x[0] - 0.5 * (sorted_x[1] - sorted_x[0])
-718 xstop = sorted_x[-1] + 0.5 * (sorted_x[-1] - sorted_x[-2])
-719 x_samples = np.arange(xstart, xstop + 0.01, 0.01)
-720
-721 plt.figure(figsize=(8, 8 / 1.618))
-722 gs = gridspec.GridSpec(2, 1, height_ratios=[3, 1], wspace=0.0, hspace=0.0)
-723 ax0 = plt.subplot(gs[0])
-724 ax0.errorbar(x, [o.value for o in y], yerr=[o.dvalue for o in y], ls='none', fmt='o', capsize=3, markersize=5, label='Data')
-725 ax0.plot(x_samples, func([o.value for o in fit_res], x_samples), label='Fit', zorder=10, ls='-', ms=0)
-726 ax0.set_xticklabels([])
-727 ax0.set_xlim([xstart, xstop])
-728 ax0.set_xticklabels([])
-729 ax0.legend()
-730
-731 residuals = (np.asarray([o.value for o in y]) - func([o.value for o in fit_res], x)) / np.asarray([o.dvalue for o in y])
-732 ax1 = plt.subplot(gs[1])
-733 ax1.plot(x, residuals, 'ko', ls='none', markersize=5)
-734 ax1.tick_params(direction='out')
-735 ax1.tick_params(axis="x", bottom=True, top=True, labelbottom=True)
-736 ax1.axhline(y=0.0, ls='--', color='k', marker=" ")
-737 ax1.fill_between(x_samples, -1.0, 1.0, alpha=0.1, facecolor='k')
-738 ax1.set_xlim([xstart, xstop])
-739 ax1.set_ylabel('Residuals')
-740 plt.subplots_adjust(wspace=None, hspace=None)
-741 plt.draw()
+ 733def residual_plot(x, y, func, fit_res):
+734 """Generates a plot which compares the fit to the data and displays the corresponding residuals
+735
+736 Returns
+737 -------
+738 None
+739 """
+740 sorted_x = sorted(x)
+741 xstart = sorted_x[0] - 0.5 * (sorted_x[1] - sorted_x[0])
+742 xstop = sorted_x[-1] + 0.5 * (sorted_x[-1] - sorted_x[-2])
+743 x_samples = np.arange(xstart, xstop + 0.01, 0.01)
+744
+745 plt.figure(figsize=(8, 8 / 1.618))
+746 gs = gridspec.GridSpec(2, 1, height_ratios=[3, 1], wspace=0.0, hspace=0.0)
+747 ax0 = plt.subplot(gs[0])
+748 ax0.errorbar(x, [o.value for o in y], yerr=[o.dvalue for o in y], ls='none', fmt='o', capsize=3, markersize=5, label='Data')
+749 ax0.plot(x_samples, func([o.value for o in fit_res], x_samples), label='Fit', zorder=10, ls='-', ms=0)
+750 ax0.set_xticklabels([])
+751 ax0.set_xlim([xstart, xstop])
+752 ax0.set_xticklabels([])
+753 ax0.legend()
+754
+755 residuals = (np.asarray([o.value for o in y]) - func([o.value for o in fit_res], x)) / np.asarray([o.dvalue for o in y])
+756 ax1 = plt.subplot(gs[1])
+757 ax1.plot(x, residuals, 'ko', ls='none', markersize=5)
+758 ax1.tick_params(direction='out')
+759 ax1.tick_params(axis="x", bottom=True, top=True, labelbottom=True)
+760 ax1.axhline(y=0.0, ls='--', color='k', marker=" ")
+761 ax1.fill_between(x_samples, -1.0, 1.0, alpha=0.1, facecolor='k')
+762 ax1.set_xlim([xstart, xstop])
+763 ax1.set_ylabel('Residuals')
+764 plt.subplots_adjust(wspace=None, hspace=None)
+765 plt.draw()
Generates a plot which compares the fit to the data and displays the corresponding residuals
+
+
Returns
+
+
@@ -1629,26 +1720,39 @@ check if the residuals of the fit are gaussian distributed.
- 744def error_band(x, func, beta):
-745 """Returns the error band for an array of sample values x, for given fit function func with optimized parameters beta."""
-746 cov = covariance(beta)
-747 if np.any(np.abs(cov - cov.T) > 1000 * np.finfo(np.float64).eps):
-748 warnings.warn("Covariance matrix is not symmetric within floating point precision", RuntimeWarning)
-749
-750 deriv = []
-751 for i, item in enumerate(x):
-752 deriv.append(np.array(egrad(func)([o.value for o in beta], item)))
-753
-754 err = []
-755 for i, item in enumerate(x):
-756 err.append(np.sqrt(deriv[i] @ cov @ deriv[i]))
-757 err = np.array(err)
-758
-759 return err
+ 768def error_band(x, func, beta):
+769 """Calculate the error band for an array of sample values x, for given fit function func with optimized parameters beta.
+770
+771 Returns
+772 -------
+773 err : np.array(Obs)
+774 Error band for an array of sample values x
+775 """
+776 cov = covariance(beta)
+777 if np.any(np.abs(cov - cov.T) > 1000 * np.finfo(np.float64).eps):
+778 warnings.warn("Covariance matrix is not symmetric within floating point precision", RuntimeWarning)
+779
+780 deriv = []
+781 for i, item in enumerate(x):
+782 deriv.append(np.array(egrad(func)([o.value for o in beta], item)))
+783
+784 err = []
+785 for i, item in enumerate(x):
+786 err.append(np.sqrt(deriv[i] @ cov @ deriv[i]))
+787 err = np.array(err)
+788
+789 return err
- Returns the error band for an array of sample values x, for given fit function func with optimized parameters beta.
+
Calculate the error band for an array of sample values x, for given fit function func with optimized parameters beta.
+
+
Returns
+
+
+- err (np.array(Obs)):
+Error band for an array of sample values x
+
@@ -1664,44 +1768,48 @@ check if the residuals of the fit are gaussian distributed.
- 762def ks_test(objects=None):
-763 """Performs a Kolmogorov–Smirnov test for the p-values of all fit object.
-764
-765 Parameters
-766 ----------
-767 objects : list
-768 List of fit results to include in the analysis (optional).
-769 """
-770
-771 if objects is None:
-772 obs_list = []
-773 for obj in gc.get_objects():
-774 if isinstance(obj, Fit_result):
-775 obs_list.append(obj)
-776 else:
-777 obs_list = objects
-778
-779 p_values = [o.p_value for o in obs_list]
-780
-781 bins = len(p_values)
-782 x = np.arange(0, 1.001, 0.001)
-783 plt.plot(x, x, 'k', zorder=1)
-784 plt.xlim(0, 1)
-785 plt.ylim(0, 1)
-786 plt.xlabel('p-value')
-787 plt.ylabel('Cumulative probability')
-788 plt.title(str(bins) + ' p-values')
-789
-790 n = np.arange(1, bins + 1) / np.float64(bins)
-791 Xs = np.sort(p_values)
-792 plt.step(Xs, n)
-793 diffs = n - Xs
-794 loc_max_diff = np.argmax(np.abs(diffs))
-795 loc = Xs[loc_max_diff]
-796 plt.annotate('', xy=(loc, loc), xytext=(loc, loc + diffs[loc_max_diff]), arrowprops=dict(arrowstyle='<->', shrinkA=0, shrinkB=0))
-797 plt.draw()
-798
-799 print(scipy.stats.kstest(p_values, 'uniform'))
+ 792def ks_test(objects=None):
+793 """Performs a Kolmogorov–Smirnov test for the p-values of all fit object.
+794
+795 Parameters
+796 ----------
+797 objects : list
+798 List of fit results to include in the analysis (optional).
+799
+800 Returns
+801 -------
+802 None
+803 """
+804
+805 if objects is None:
+806 obs_list = []
+807 for obj in gc.get_objects():
+808 if isinstance(obj, Fit_result):
+809 obs_list.append(obj)
+810 else:
+811 obs_list = objects
+812
+813 p_values = [o.p_value for o in obs_list]
+814
+815 bins = len(p_values)
+816 x = np.arange(0, 1.001, 0.001)
+817 plt.plot(x, x, 'k', zorder=1)
+818 plt.xlim(0, 1)
+819 plt.ylim(0, 1)
+820 plt.xlabel('p-value')
+821 plt.ylabel('Cumulative probability')
+822 plt.title(str(bins) + ' p-values')
+823
+824 n = np.arange(1, bins + 1) / np.float64(bins)
+825 Xs = np.sort(p_values)
+826 plt.step(Xs, n)
+827 diffs = n - Xs
+828 loc_max_diff = np.argmax(np.abs(diffs))
+829 loc = Xs[loc_max_diff]
+830 plt.annotate('', xy=(loc, loc), xytext=(loc, loc + diffs[loc_max_diff]), arrowprops=dict(arrowstyle='<->', shrinkA=0, shrinkB=0))
+831 plt.draw()
+832
+833 print(scipy.stats.kstest(p_values, 'uniform'))
@@ -1713,6 +1821,12 @@ check if the residuals of the fit are gaussian distributed.
objects (list):
List of fit results to include in the analysis (optional).
+
+Returns
+
+
diff --git a/docs/pyerrors/input/bdio.html b/docs/pyerrors/input/bdio.html
index ebce901d..4de802e0 100644
--- a/docs/pyerrors/input/bdio.html
+++ b/docs/pyerrors/input/bdio.html
@@ -105,665 +105,680 @@
18 ----------
19 file_path -- path to the bdio file
20 bdio_path -- path to the shared bdio library libbdio.so (default ./libbdio.so)
- 21 """
- 22 bdio = ctypes.cdll.LoadLibrary(bdio_path)
- 23
- 24 bdio_open = bdio.bdio_open
- 25 bdio_open.restype = ctypes.c_void_p
- 26
- 27 bdio_close = bdio.bdio_close
- 28 bdio_close.restype = ctypes.c_int
- 29 bdio_close.argtypes = [ctypes.c_void_p]
- 30
- 31 bdio_seek_record = bdio.bdio_seek_record
- 32 bdio_seek_record.restype = ctypes.c_int
- 33 bdio_seek_record.argtypes = [ctypes.c_void_p]
- 34
- 35 bdio_get_rlen = bdio.bdio_get_rlen
- 36 bdio_get_rlen.restype = ctypes.c_int
- 37 bdio_get_rlen.argtypes = [ctypes.c_void_p]
- 38
- 39 bdio_get_ruinfo = bdio.bdio_get_ruinfo
- 40 bdio_get_ruinfo.restype = ctypes.c_int
- 41 bdio_get_ruinfo.argtypes = [ctypes.c_void_p]
- 42
- 43 bdio_read = bdio.bdio_read
- 44 bdio_read.restype = ctypes.c_size_t
- 45 bdio_read.argtypes = [ctypes.c_char_p, ctypes.c_size_t, ctypes.c_void_p]
- 46
- 47 bdio_read_f64 = bdio.bdio_read_f64
- 48 bdio_read_f64.restype = ctypes.c_size_t
- 49 bdio_read_f64.argtypes = [ctypes.c_void_p, ctypes.c_size_t, ctypes.c_void_p]
- 50
- 51 bdio_read_int32 = bdio.bdio_read_int32
- 52 bdio_read_int32.restype = ctypes.c_size_t
- 53 bdio_read_int32.argtypes = [ctypes.c_void_p, ctypes.c_size_t, ctypes.c_void_p]
- 54
- 55 b_path = file_path.encode('utf-8')
- 56 read = 'r'
- 57 b_read = read.encode('utf-8')
- 58
- 59 fbdio = bdio_open(ctypes.c_char_p(b_path), ctypes.c_char_p(b_read), None)
- 60
- 61 return_list = []
- 62
- 63 print('Reading of bdio file started')
- 64 while True:
- 65 bdio_seek_record(fbdio)
- 66 ruinfo = bdio_get_ruinfo(fbdio)
+ 21
+ 22 Returns
+ 23 -------
+ 24 data : List[Obs]
+ 25 Extracted data
+ 26 """
+ 27 bdio = ctypes.cdll.LoadLibrary(bdio_path)
+ 28
+ 29 bdio_open = bdio.bdio_open
+ 30 bdio_open.restype = ctypes.c_void_p
+ 31
+ 32 bdio_close = bdio.bdio_close
+ 33 bdio_close.restype = ctypes.c_int
+ 34 bdio_close.argtypes = [ctypes.c_void_p]
+ 35
+ 36 bdio_seek_record = bdio.bdio_seek_record
+ 37 bdio_seek_record.restype = ctypes.c_int
+ 38 bdio_seek_record.argtypes = [ctypes.c_void_p]
+ 39
+ 40 bdio_get_rlen = bdio.bdio_get_rlen
+ 41 bdio_get_rlen.restype = ctypes.c_int
+ 42 bdio_get_rlen.argtypes = [ctypes.c_void_p]
+ 43
+ 44 bdio_get_ruinfo = bdio.bdio_get_ruinfo
+ 45 bdio_get_ruinfo.restype = ctypes.c_int
+ 46 bdio_get_ruinfo.argtypes = [ctypes.c_void_p]
+ 47
+ 48 bdio_read = bdio.bdio_read
+ 49 bdio_read.restype = ctypes.c_size_t
+ 50 bdio_read.argtypes = [ctypes.c_char_p, ctypes.c_size_t, ctypes.c_void_p]
+ 51
+ 52 bdio_read_f64 = bdio.bdio_read_f64
+ 53 bdio_read_f64.restype = ctypes.c_size_t
+ 54 bdio_read_f64.argtypes = [ctypes.c_void_p, ctypes.c_size_t, ctypes.c_void_p]
+ 55
+ 56 bdio_read_int32 = bdio.bdio_read_int32
+ 57 bdio_read_int32.restype = ctypes.c_size_t
+ 58 bdio_read_int32.argtypes = [ctypes.c_void_p, ctypes.c_size_t, ctypes.c_void_p]
+ 59
+ 60 b_path = file_path.encode('utf-8')
+ 61 read = 'r'
+ 62 b_read = read.encode('utf-8')
+ 63
+ 64 fbdio = bdio_open(ctypes.c_char_p(b_path), ctypes.c_char_p(b_read), None)
+ 65
+ 66 return_list = []
67
- 68 if ruinfo == 7:
- 69 print('MD5sum found') # For now we just ignore these entries and do not perform any checks on them
- 70 continue
- 71
- 72 if ruinfo < 0:
- 73 # EOF reached
- 74 break
- 75 bdio_get_rlen(fbdio)
+ 68 print('Reading of bdio file started')
+ 69 while True:
+ 70 bdio_seek_record(fbdio)
+ 71 ruinfo = bdio_get_ruinfo(fbdio)
+ 72
+ 73 if ruinfo == 7:
+ 74 print('MD5sum found') # For now we just ignore these entries and do not perform any checks on them
+ 75 continue
76
- 77 def read_c_double():
- 78 d_buf = ctypes.c_double
- 79 pd_buf = d_buf()
- 80 ppd_buf = ctypes.c_void_p(ctypes.addressof(pd_buf))
- 81 bdio_read_f64(ppd_buf, ctypes.c_size_t(8), ctypes.c_void_p(fbdio))
- 82 return pd_buf.value
- 83
- 84 mean = read_c_double()
- 85 print('mean', mean)
- 86
- 87 def read_c_size_t():
- 88 d_buf = ctypes.c_size_t
- 89 pd_buf = d_buf()
- 90 ppd_buf = ctypes.c_void_p(ctypes.addressof(pd_buf))
- 91 bdio_read_int32(ppd_buf, ctypes.c_size_t(4), ctypes.c_void_p(fbdio))
- 92 return pd_buf.value
- 93
- 94 neid = read_c_size_t()
- 95 print('neid', neid)
- 96
- 97 ndata = []
- 98 for index in range(neid):
- 99 ndata.append(read_c_size_t())
-100 print('ndata', ndata)
+ 77 if ruinfo < 0:
+ 78 # EOF reached
+ 79 break
+ 80 bdio_get_rlen(fbdio)
+ 81
+ 82 def read_c_double():
+ 83 d_buf = ctypes.c_double
+ 84 pd_buf = d_buf()
+ 85 ppd_buf = ctypes.c_void_p(ctypes.addressof(pd_buf))
+ 86 bdio_read_f64(ppd_buf, ctypes.c_size_t(8), ctypes.c_void_p(fbdio))
+ 87 return pd_buf.value
+ 88
+ 89 mean = read_c_double()
+ 90 print('mean', mean)
+ 91
+ 92 def read_c_size_t():
+ 93 d_buf = ctypes.c_size_t
+ 94 pd_buf = d_buf()
+ 95 ppd_buf = ctypes.c_void_p(ctypes.addressof(pd_buf))
+ 96 bdio_read_int32(ppd_buf, ctypes.c_size_t(4), ctypes.c_void_p(fbdio))
+ 97 return pd_buf.value
+ 98
+ 99 neid = read_c_size_t()
+100 print('neid', neid)
101
-102 nrep = []
+102 ndata = []
103 for index in range(neid):
-104 nrep.append(read_c_size_t())
-105 print('nrep', nrep)
+104 ndata.append(read_c_size_t())
+105 print('ndata', ndata)
106
-107 vrep = []
+107 nrep = []
108 for index in range(neid):
-109 vrep.append([])
-110 for jndex in range(nrep[index]):
-111 vrep[-1].append(read_c_size_t())
-112 print('vrep', vrep)
-113
-114 ids = []
-115 for index in range(neid):
-116 ids.append(read_c_size_t())
-117 print('ids', ids)
+109 nrep.append(read_c_size_t())
+110 print('nrep', nrep)
+111
+112 vrep = []
+113 for index in range(neid):
+114 vrep.append([])
+115 for jndex in range(nrep[index]):
+116 vrep[-1].append(read_c_size_t())
+117 print('vrep', vrep)
118
-119 nt = []
+119 ids = []
120 for index in range(neid):
-121 nt.append(read_c_size_t())
-122 print('nt', nt)
+121 ids.append(read_c_size_t())
+122 print('ids', ids)
123
-124 zero = []
+124 nt = []
125 for index in range(neid):
-126 zero.append(read_c_double())
-127 print('zero', zero)
+126 nt.append(read_c_size_t())
+127 print('nt', nt)
128
-129 four = []
+129 zero = []
130 for index in range(neid):
-131 four.append(read_c_double())
-132 print('four', four)
+131 zero.append(read_c_double())
+132 print('zero', zero)
133
-134 d_buf = ctypes.c_double * np.sum(ndata)
-135 pd_buf = d_buf()
-136 ppd_buf = ctypes.c_void_p(ctypes.addressof(pd_buf))
-137 bdio_read_f64(ppd_buf, ctypes.c_size_t(8 * np.sum(ndata)), ctypes.c_void_p(fbdio))
-138 delta = pd_buf[:]
-139
-140 samples = np.split(np.asarray(delta) + mean, np.cumsum([a for su in vrep for a in su])[:-1])
-141 no_reps = [len(o) for o in vrep]
-142 assert len(ids) == len(no_reps)
-143 tmp_names = []
-144 ens_length = max([len(str(o)) for o in ids])
-145 for loc_id, reps in zip(ids, no_reps):
-146 for index in range(reps):
-147 missing_chars = ens_length - len(str(loc_id))
-148 tmp_names.append(str(loc_id) + ' ' * missing_chars + '|r' + '{0:03d}'.format(index))
-149
-150 return_list.append(Obs(samples, tmp_names))
-151
-152 bdio_close(fbdio)
-153 print()
-154 print(len(return_list), 'observable(s) extracted.')
-155 return return_list
+134 four = []
+135 for index in range(neid):
+136 four.append(read_c_double())
+137 print('four', four)
+138
+139 d_buf = ctypes.c_double * np.sum(ndata)
+140 pd_buf = d_buf()
+141 ppd_buf = ctypes.c_void_p(ctypes.addressof(pd_buf))
+142 bdio_read_f64(ppd_buf, ctypes.c_size_t(8 * np.sum(ndata)), ctypes.c_void_p(fbdio))
+143 delta = pd_buf[:]
+144
+145 samples = np.split(np.asarray(delta) + mean, np.cumsum([a for su in vrep for a in su])[:-1])
+146 no_reps = [len(o) for o in vrep]
+147 assert len(ids) == len(no_reps)
+148 tmp_names = []
+149 ens_length = max([len(str(o)) for o in ids])
+150 for loc_id, reps in zip(ids, no_reps):
+151 for index in range(reps):
+152 missing_chars = ens_length - len(str(loc_id))
+153 tmp_names.append(str(loc_id) + ' ' * missing_chars + '|r' + '{0:03d}'.format(index))
+154
+155 return_list.append(Obs(samples, tmp_names))
156
-157
-158def write_ADerrors(obs_list, file_path, bdio_path='./libbdio.so', **kwargs):
-159 """ Write Obs to a bdio file according to ADerrors conventions
-160
-161 read_mesons requires bdio to be compiled into a shared library. This can be achieved by
-162 adding the flag -fPIC to CC and changing the all target to
-163
-164 all: bdio.o $(LIBDIR)
-165 gcc -shared -Wl,-soname,libbdio.so -o $(BUILDDIR)/libbdio.so $(BUILDDIR)/bdio.o
-166 cp $(BUILDDIR)/libbdio.so $(LIBDIR)/
-167
-168 Parameters
-169 ----------
-170 file_path -- path to the bdio file
-171 bdio_path -- path to the shared bdio library libbdio.so (default ./libbdio.so)
-172 """
-173
-174 for obs in obs_list:
-175 if not hasattr(obs, 'e_names'):
-176 raise Exception('Run the gamma method first for all obs.')
+157 bdio_close(fbdio)
+158 print()
+159 print(len(return_list), 'observable(s) extracted.')
+160 return return_list
+161
+162
+163def write_ADerrors(obs_list, file_path, bdio_path='./libbdio.so', **kwargs):
+164 """ Write Obs to a bdio file according to ADerrors conventions
+165
+166 read_mesons requires bdio to be compiled into a shared library. This can be achieved by
+167 adding the flag -fPIC to CC and changing the all target to
+168
+169 all: bdio.o $(LIBDIR)
+170 gcc -shared -Wl,-soname,libbdio.so -o $(BUILDDIR)/libbdio.so $(BUILDDIR)/bdio.o
+171 cp $(BUILDDIR)/libbdio.so $(LIBDIR)/
+172
+173 Parameters
+174 ----------
+175 file_path -- path to the bdio file
+176 bdio_path -- path to the shared bdio library libbdio.so (default ./libbdio.so)
177
-178 bdio = ctypes.cdll.LoadLibrary(bdio_path)
-179
-180 bdio_open = bdio.bdio_open
-181 bdio_open.restype = ctypes.c_void_p
-182
-183 bdio_close = bdio.bdio_close
-184 bdio_close.restype = ctypes.c_int
-185 bdio_close.argtypes = [ctypes.c_void_p]
-186
-187 bdio_start_record = bdio.bdio_start_record
-188 bdio_start_record.restype = ctypes.c_int
-189 bdio_start_record.argtypes = [ctypes.c_size_t, ctypes.c_size_t, ctypes.c_void_p]
-190
-191 bdio_flush_record = bdio.bdio_flush_record
-192 bdio_flush_record.restype = ctypes.c_int
-193 bdio_flush_record.argytpes = [ctypes.c_void_p]
-194
-195 bdio_write_f64 = bdio.bdio_write_f64
-196 bdio_write_f64.restype = ctypes.c_size_t
-197 bdio_write_f64.argtypes = [ctypes.c_void_p, ctypes.c_size_t, ctypes.c_void_p]
-198
-199 bdio_write_int32 = bdio.bdio_write_int32
-200 bdio_write_int32.restype = ctypes.c_size_t
-201 bdio_write_int32.argtypes = [ctypes.c_void_p, ctypes.c_size_t, ctypes.c_void_p]
-202
-203 b_path = file_path.encode('utf-8')
-204 write = 'w'
-205 b_write = write.encode('utf-8')
-206 form = 'pyerrors ADerror export'
-207 b_form = form.encode('utf-8')
+178 Returns
+179 -------
+180 success : int
+181 returns 0 is successful
+182 """
+183
+184 for obs in obs_list:
+185 if not hasattr(obs, 'e_names'):
+186 raise Exception('Run the gamma method first for all obs.')
+187
+188 bdio = ctypes.cdll.LoadLibrary(bdio_path)
+189
+190 bdio_open = bdio.bdio_open
+191 bdio_open.restype = ctypes.c_void_p
+192
+193 bdio_close = bdio.bdio_close
+194 bdio_close.restype = ctypes.c_int
+195 bdio_close.argtypes = [ctypes.c_void_p]
+196
+197 bdio_start_record = bdio.bdio_start_record
+198 bdio_start_record.restype = ctypes.c_int
+199 bdio_start_record.argtypes = [ctypes.c_size_t, ctypes.c_size_t, ctypes.c_void_p]
+200
+201 bdio_flush_record = bdio.bdio_flush_record
+202 bdio_flush_record.restype = ctypes.c_int
+203 bdio_flush_record.argytpes = [ctypes.c_void_p]
+204
+205 bdio_write_f64 = bdio.bdio_write_f64
+206 bdio_write_f64.restype = ctypes.c_size_t
+207 bdio_write_f64.argtypes = [ctypes.c_void_p, ctypes.c_size_t, ctypes.c_void_p]
208
-209 fbdio = bdio_open(ctypes.c_char_p(b_path), ctypes.c_char_p(b_write), b_form)
-210
-211 for obs in obs_list:
-212 # mean = obs.value
-213 neid = len(obs.e_names)
-214 vrep = [[obs.shape[o] for o in sl] for sl in list(obs.e_content.values())]
-215 vrep_write = [item for sublist in vrep for item in sublist]
-216 ndata = [np.sum(o) for o in vrep]
-217 nrep = [len(o) for o in vrep]
-218 print('ndata', ndata)
-219 print('nrep', nrep)
-220 print('vrep', vrep)
-221 keys = list(obs.e_content.keys())
-222 ids = []
-223 for key in keys:
-224 try: # Try to convert key to integer
-225 ids.append(int(key))
-226 except Exception: # If not possible construct a hash
-227 ids.append(int(hashlib.sha256(key.encode('utf-8')).hexdigest(), 16) % 10 ** 8)
-228 print('ids', ids)
-229 nt = []
-230 for e, e_name in enumerate(obs.e_names):
-231
-232 r_length = []
-233 for r_name in obs.e_content[e_name]:
-234 r_length.append(len(obs.deltas[r_name]))
-235
-236 # e_N = np.sum(r_length)
-237 nt.append(max(r_length) // 2)
-238 print('nt', nt)
-239 zero = neid * [0.0]
-240 four = neid * [4.0]
-241 print('zero', zero)
-242 print('four', four)
-243 delta = np.concatenate([item for sublist in [[obs.deltas[o] for o in sl] for sl in list(obs.e_content.values())] for item in sublist])
-244
-245 bdio_start_record(0x00, 8, fbdio)
-246
-247 def write_c_double(double):
-248 pd_buf = ctypes.c_double(double)
-249 ppd_buf = ctypes.c_void_p(ctypes.addressof(pd_buf))
-250 bdio_write_f64(ppd_buf, ctypes.c_size_t(8), ctypes.c_void_p(fbdio))
-251
-252 def write_c_size_t(int32):
-253 pd_buf = ctypes.c_size_t(int32)
-254 ppd_buf = ctypes.c_void_p(ctypes.addressof(pd_buf))
-255 bdio_write_int32(ppd_buf, ctypes.c_size_t(4), ctypes.c_void_p(fbdio))
+209 bdio_write_int32 = bdio.bdio_write_int32
+210 bdio_write_int32.restype = ctypes.c_size_t
+211 bdio_write_int32.argtypes = [ctypes.c_void_p, ctypes.c_size_t, ctypes.c_void_p]
+212
+213 b_path = file_path.encode('utf-8')
+214 write = 'w'
+215 b_write = write.encode('utf-8')
+216 form = 'pyerrors ADerror export'
+217 b_form = form.encode('utf-8')
+218
+219 fbdio = bdio_open(ctypes.c_char_p(b_path), ctypes.c_char_p(b_write), b_form)
+220
+221 for obs in obs_list:
+222 # mean = obs.value
+223 neid = len(obs.e_names)
+224 vrep = [[obs.shape[o] for o in sl] for sl in list(obs.e_content.values())]
+225 vrep_write = [item for sublist in vrep for item in sublist]
+226 ndata = [np.sum(o) for o in vrep]
+227 nrep = [len(o) for o in vrep]
+228 print('ndata', ndata)
+229 print('nrep', nrep)
+230 print('vrep', vrep)
+231 keys = list(obs.e_content.keys())
+232 ids = []
+233 for key in keys:
+234 try: # Try to convert key to integer
+235 ids.append(int(key))
+236 except Exception: # If not possible construct a hash
+237 ids.append(int(hashlib.sha256(key.encode('utf-8')).hexdigest(), 16) % 10 ** 8)
+238 print('ids', ids)
+239 nt = []
+240 for e, e_name in enumerate(obs.e_names):
+241
+242 r_length = []
+243 for r_name in obs.e_content[e_name]:
+244 r_length.append(len(obs.deltas[r_name]))
+245
+246 # e_N = np.sum(r_length)
+247 nt.append(max(r_length) // 2)
+248 print('nt', nt)
+249 zero = neid * [0.0]
+250 four = neid * [4.0]
+251 print('zero', zero)
+252 print('four', four)
+253 delta = np.concatenate([item for sublist in [[obs.deltas[o] for o in sl] for sl in list(obs.e_content.values())] for item in sublist])
+254
+255 bdio_start_record(0x00, 8, fbdio)
256
-257 write_c_double(obs.value)
-258 write_c_size_t(neid)
-259
-260 for element in ndata:
-261 write_c_size_t(element)
-262 for element in nrep:
-263 write_c_size_t(element)
-264 for element in vrep_write:
-265 write_c_size_t(element)
-266 for element in ids:
-267 write_c_size_t(element)
-268 for element in nt:
-269 write_c_size_t(element)
-270
-271 for element in zero:
-272 write_c_double(element)
-273 for element in four:
-274 write_c_double(element)
-275
-276 for element in delta:
-277 write_c_double(element)
-278
-279 bdio_close(fbdio)
-280 return 0
-281
-282
-283def _get_kwd(string, key):
-284 return (string.split(key, 1)[1]).split(" ", 1)[0]
+257 def write_c_double(double):
+258 pd_buf = ctypes.c_double(double)
+259 ppd_buf = ctypes.c_void_p(ctypes.addressof(pd_buf))
+260 bdio_write_f64(ppd_buf, ctypes.c_size_t(8), ctypes.c_void_p(fbdio))
+261
+262 def write_c_size_t(int32):
+263 pd_buf = ctypes.c_size_t(int32)
+264 ppd_buf = ctypes.c_void_p(ctypes.addressof(pd_buf))
+265 bdio_write_int32(ppd_buf, ctypes.c_size_t(4), ctypes.c_void_p(fbdio))
+266
+267 write_c_double(obs.value)
+268 write_c_size_t(neid)
+269
+270 for element in ndata:
+271 write_c_size_t(element)
+272 for element in nrep:
+273 write_c_size_t(element)
+274 for element in vrep_write:
+275 write_c_size_t(element)
+276 for element in ids:
+277 write_c_size_t(element)
+278 for element in nt:
+279 write_c_size_t(element)
+280
+281 for element in zero:
+282 write_c_double(element)
+283 for element in four:
+284 write_c_double(element)
285
-286
-287def _get_corr_name(string, key):
-288 return (string.split(key, 1)[1]).split(' NDIM=', 1)[0]
-289
-290
-291def read_mesons(file_path, bdio_path='./libbdio.so', **kwargs):
-292 """ Extract mesons data from a bdio file and return it as a dictionary
-293
-294 The dictionary can be accessed with a tuple consisting of (type, source_position, kappa1, kappa2)
+286 for element in delta:
+287 write_c_double(element)
+288
+289 bdio_close(fbdio)
+290 return 0
+291
+292
+293def _get_kwd(string, key):
+294 return (string.split(key, 1)[1]).split(" ", 1)[0]
295
-296 read_mesons requires bdio to be compiled into a shared library. This can be achieved by
-297 adding the flag -fPIC to CC and changing the all target to
-298
-299 all: bdio.o $(LIBDIR)
-300 gcc -shared -Wl,-soname,libbdio.so -o $(BUILDDIR)/libbdio.so $(BUILDDIR)/bdio.o
-301 cp $(BUILDDIR)/libbdio.so $(LIBDIR)/
-302
-303 Parameters
-304 ----------
-305 file_path : str
-306 path to the bdio file
-307 bdio_path : str
-308 path to the shared bdio library libbdio.so (default ./libbdio.so)
-309 start : int
-310 The first configuration to be read (default 1)
-311 stop : int
-312 The last configuration to be read (default None)
-313 step : int
-314 Fixed step size between two measurements (default 1)
-315 alternative_ensemble_name : str
-316 Manually overwrite ensemble name
-317 """
-318
-319 start = kwargs.get('start', 1)
-320 stop = kwargs.get('stop', None)
-321 step = kwargs.get('step', 1)
-322
-323 bdio = ctypes.cdll.LoadLibrary(bdio_path)
-324
-325 bdio_open = bdio.bdio_open
-326 bdio_open.restype = ctypes.c_void_p
+296
+297def _get_corr_name(string, key):
+298 return (string.split(key, 1)[1]).split(' NDIM=', 1)[0]
+299
+300
+301def read_mesons(file_path, bdio_path='./libbdio.so', **kwargs):
+302 """ Extract mesons data from a bdio file and return it as a dictionary
+303
+304 The dictionary can be accessed with a tuple consisting of (type, source_position, kappa1, kappa2)
+305
+306 read_mesons requires bdio to be compiled into a shared library. This can be achieved by
+307 adding the flag -fPIC to CC and changing the all target to
+308
+309 all: bdio.o $(LIBDIR)
+310 gcc -shared -Wl,-soname,libbdio.so -o $(BUILDDIR)/libbdio.so $(BUILDDIR)/bdio.o
+311 cp $(BUILDDIR)/libbdio.so $(LIBDIR)/
+312
+313 Parameters
+314 ----------
+315 file_path : str
+316 path to the bdio file
+317 bdio_path : str
+318 path to the shared bdio library libbdio.so (default ./libbdio.so)
+319 start : int
+320 The first configuration to be read (default 1)
+321 stop : int
+322 The last configuration to be read (default None)
+323 step : int
+324 Fixed step size between two measurements (default 1)
+325 alternative_ensemble_name : str
+326 Manually overwrite ensemble name
327
-328 bdio_close = bdio.bdio_close
-329 bdio_close.restype = ctypes.c_int
-330 bdio_close.argtypes = [ctypes.c_void_p]
-331
-332 bdio_seek_record = bdio.bdio_seek_record
-333 bdio_seek_record.restype = ctypes.c_int
-334 bdio_seek_record.argtypes = [ctypes.c_void_p]
-335
-336 bdio_get_rlen = bdio.bdio_get_rlen
-337 bdio_get_rlen.restype = ctypes.c_int
-338 bdio_get_rlen.argtypes = [ctypes.c_void_p]
+328 Returns
+329 -------
+330 data : dict
+331 Extracted meson data
+332 """
+333
+334 start = kwargs.get('start', 1)
+335 stop = kwargs.get('stop', None)
+336 step = kwargs.get('step', 1)
+337
+338 bdio = ctypes.cdll.LoadLibrary(bdio_path)
339
-340 bdio_get_ruinfo = bdio.bdio_get_ruinfo
-341 bdio_get_ruinfo.restype = ctypes.c_int
-342 bdio_get_ruinfo.argtypes = [ctypes.c_void_p]
-343
-344 bdio_read = bdio.bdio_read
-345 bdio_read.restype = ctypes.c_size_t
-346 bdio_read.argtypes = [ctypes.c_char_p, ctypes.c_size_t, ctypes.c_void_p]
-347
-348 bdio_read_f64 = bdio.bdio_read_f64
-349 bdio_read_f64.restype = ctypes.c_size_t
-350 bdio_read_f64.argtypes = [ctypes.c_void_p, ctypes.c_size_t, ctypes.c_void_p]
-351
-352 b_path = file_path.encode('utf-8')
-353 read = 'r'
-354 b_read = read.encode('utf-8')
-355 form = 'Generic Correlator Format 1.0'
-356 b_form = form.encode('utf-8')
-357
-358 ensemble_name = ''
-359 volume = [] # lattice volume
-360 boundary_conditions = []
-361 corr_name = [] # Contains correlator names
-362 corr_type = [] # Contains correlator data type (important for reading out numerical data)
-363 corr_props = [] # Contanis propagator types (Component of corr_kappa)
-364 d0 = 0 # tvals
-365 d1 = 0 # nnoise
-366 prop_kappa = [] # Contains propagator kappas (Component of corr_kappa)
-367 prop_source = [] # Contains propagator source positions
-368 # Check noise type for multiple replica?
-369 corr_no = -1
-370 data = []
-371 idl = []
+340 bdio_open = bdio.bdio_open
+341 bdio_open.restype = ctypes.c_void_p
+342
+343 bdio_close = bdio.bdio_close
+344 bdio_close.restype = ctypes.c_int
+345 bdio_close.argtypes = [ctypes.c_void_p]
+346
+347 bdio_seek_record = bdio.bdio_seek_record
+348 bdio_seek_record.restype = ctypes.c_int
+349 bdio_seek_record.argtypes = [ctypes.c_void_p]
+350
+351 bdio_get_rlen = bdio.bdio_get_rlen
+352 bdio_get_rlen.restype = ctypes.c_int
+353 bdio_get_rlen.argtypes = [ctypes.c_void_p]
+354
+355 bdio_get_ruinfo = bdio.bdio_get_ruinfo
+356 bdio_get_ruinfo.restype = ctypes.c_int
+357 bdio_get_ruinfo.argtypes = [ctypes.c_void_p]
+358
+359 bdio_read = bdio.bdio_read
+360 bdio_read.restype = ctypes.c_size_t
+361 bdio_read.argtypes = [ctypes.c_char_p, ctypes.c_size_t, ctypes.c_void_p]
+362
+363 bdio_read_f64 = bdio.bdio_read_f64
+364 bdio_read_f64.restype = ctypes.c_size_t
+365 bdio_read_f64.argtypes = [ctypes.c_void_p, ctypes.c_size_t, ctypes.c_void_p]
+366
+367 b_path = file_path.encode('utf-8')
+368 read = 'r'
+369 b_read = read.encode('utf-8')
+370 form = 'Generic Correlator Format 1.0'
+371 b_form = form.encode('utf-8')
372
-373 fbdio = bdio_open(ctypes.c_char_p(b_path), ctypes.c_char_p(b_read), ctypes.c_char_p(b_form))
-374
-375 print('Reading of bdio file started')
-376 while True:
-377 bdio_seek_record(fbdio)
-378 ruinfo = bdio_get_ruinfo(fbdio)
-379 if ruinfo < 0:
-380 # EOF reached
-381 break
-382 rlen = bdio_get_rlen(fbdio)
-383 if ruinfo == 5:
-384 d_buf = ctypes.c_double * (2 + d0 * d1 * 2)
-385 pd_buf = d_buf()
-386 ppd_buf = ctypes.c_void_p(ctypes.addressof(pd_buf))
-387 bdio_read_f64(ppd_buf, ctypes.c_size_t(rlen), ctypes.c_void_p(fbdio))
-388 if corr_type[corr_no] == 'complex':
-389 tmp_mean = np.mean(np.asarray(np.split(np.asarray(pd_buf[2 + 2 * d1:-2 * d1:2]), d0 - 2)), axis=1)
-390 else:
-391 tmp_mean = np.mean(np.asarray(np.split(np.asarray(pd_buf[2 + d1:-d0 * d1 - d1]), d0 - 2)), axis=1)
-392
-393 data[corr_no].append(tmp_mean)
-394 corr_no += 1
-395 else:
-396 alt_buf = ctypes.create_string_buffer(1024)
-397 palt_buf = ctypes.c_char_p(ctypes.addressof(alt_buf))
-398 iread = bdio_read(palt_buf, ctypes.c_size_t(rlen), ctypes.c_void_p(fbdio))
-399 if rlen != iread:
-400 print('Error')
-401 for i, item in enumerate(alt_buf):
-402 if item == b'\x00':
-403 alt_buf[i] = b' '
-404 tmp_string = (alt_buf[:].decode("utf-8")).rstrip()
-405 if ruinfo == 0:
-406 ensemble_name = _get_kwd(tmp_string, 'ENSEMBLE=')
-407 volume.append(int(_get_kwd(tmp_string, 'L0=')))
-408 volume.append(int(_get_kwd(tmp_string, 'L1=')))
-409 volume.append(int(_get_kwd(tmp_string, 'L2=')))
-410 volume.append(int(_get_kwd(tmp_string, 'L3=')))
-411 boundary_conditions.append(_get_kwd(tmp_string, 'BC0='))
-412 boundary_conditions.append(_get_kwd(tmp_string, 'BC1='))
-413 boundary_conditions.append(_get_kwd(tmp_string, 'BC2='))
-414 boundary_conditions.append(_get_kwd(tmp_string, 'BC3='))
-415
-416 if ruinfo == 1:
-417 corr_name.append(_get_corr_name(tmp_string, 'CORR_NAME='))
-418 corr_type.append(_get_kwd(tmp_string, 'DATATYPE='))
-419 corr_props.append([_get_kwd(tmp_string, 'PROP0='), _get_kwd(tmp_string, 'PROP1=')])
-420 if d0 == 0:
-421 d0 = int(_get_kwd(tmp_string, 'D0='))
-422 else:
-423 if d0 != int(_get_kwd(tmp_string, 'D0=')):
-424 print('Error: Varying number of time values')
-425 if d1 == 0:
-426 d1 = int(_get_kwd(tmp_string, 'D1='))
-427 else:
-428 if d1 != int(_get_kwd(tmp_string, 'D1=')):
-429 print('Error: Varying number of random sources')
-430 if ruinfo == 2:
-431 prop_kappa.append(_get_kwd(tmp_string, 'KAPPA='))
-432 prop_source.append(_get_kwd(tmp_string, 'x0='))
-433 if ruinfo == 4:
-434 cnfg_no = int(_get_kwd(tmp_string, 'CNFG_ID='))
-435 if stop:
-436 if cnfg_no > kwargs.get('stop'):
-437 break
-438 idl.append(cnfg_no)
-439 print('\r%s %i' % ('Reading configuration', cnfg_no), end='\r')
-440 if len(idl) == 1:
-441 no_corrs = len(corr_name)
-442 data = []
-443 for c in range(no_corrs):
-444 data.append([])
-445
-446 corr_no = 0
-447
-448 bdio_close(fbdio)
-449
-450 print('\nEnsemble: ', ensemble_name)
-451 if 'alternative_ensemble_name' in kwargs:
-452 ensemble_name = kwargs.get('alternative_ensemble_name')
-453 print('Ensemble name overwritten to', ensemble_name)
-454 print('Lattice volume: ', volume)
-455 print('Boundary conditions: ', boundary_conditions)
-456 print('Number of time values: ', d0)
-457 print('Number of random sources: ', d1)
-458 print('Number of corrs: ', len(corr_name))
-459 print('Number of configurations: ', len(idl))
+373 ensemble_name = ''
+374 volume = [] # lattice volume
+375 boundary_conditions = []
+376 corr_name = [] # Contains correlator names
+377 corr_type = [] # Contains correlator data type (important for reading out numerical data)
+378 corr_props = [] # Contanis propagator types (Component of corr_kappa)
+379 d0 = 0 # tvals
+380 d1 = 0 # nnoise
+381 prop_kappa = [] # Contains propagator kappas (Component of corr_kappa)
+382 prop_source = [] # Contains propagator source positions
+383 # Check noise type for multiple replica?
+384 corr_no = -1
+385 data = []
+386 idl = []
+387
+388 fbdio = bdio_open(ctypes.c_char_p(b_path), ctypes.c_char_p(b_read), ctypes.c_char_p(b_form))
+389
+390 print('Reading of bdio file started')
+391 while True:
+392 bdio_seek_record(fbdio)
+393 ruinfo = bdio_get_ruinfo(fbdio)
+394 if ruinfo < 0:
+395 # EOF reached
+396 break
+397 rlen = bdio_get_rlen(fbdio)
+398 if ruinfo == 5:
+399 d_buf = ctypes.c_double * (2 + d0 * d1 * 2)
+400 pd_buf = d_buf()
+401 ppd_buf = ctypes.c_void_p(ctypes.addressof(pd_buf))
+402 bdio_read_f64(ppd_buf, ctypes.c_size_t(rlen), ctypes.c_void_p(fbdio))
+403 if corr_type[corr_no] == 'complex':
+404 tmp_mean = np.mean(np.asarray(np.split(np.asarray(pd_buf[2 + 2 * d1:-2 * d1:2]), d0 - 2)), axis=1)
+405 else:
+406 tmp_mean = np.mean(np.asarray(np.split(np.asarray(pd_buf[2 + d1:-d0 * d1 - d1]), d0 - 2)), axis=1)
+407
+408 data[corr_no].append(tmp_mean)
+409 corr_no += 1
+410 else:
+411 alt_buf = ctypes.create_string_buffer(1024)
+412 palt_buf = ctypes.c_char_p(ctypes.addressof(alt_buf))
+413 iread = bdio_read(palt_buf, ctypes.c_size_t(rlen), ctypes.c_void_p(fbdio))
+414 if rlen != iread:
+415 print('Error')
+416 for i, item in enumerate(alt_buf):
+417 if item == b'\x00':
+418 alt_buf[i] = b' '
+419 tmp_string = (alt_buf[:].decode("utf-8")).rstrip()
+420 if ruinfo == 0:
+421 ensemble_name = _get_kwd(tmp_string, 'ENSEMBLE=')
+422 volume.append(int(_get_kwd(tmp_string, 'L0=')))
+423 volume.append(int(_get_kwd(tmp_string, 'L1=')))
+424 volume.append(int(_get_kwd(tmp_string, 'L2=')))
+425 volume.append(int(_get_kwd(tmp_string, 'L3=')))
+426 boundary_conditions.append(_get_kwd(tmp_string, 'BC0='))
+427 boundary_conditions.append(_get_kwd(tmp_string, 'BC1='))
+428 boundary_conditions.append(_get_kwd(tmp_string, 'BC2='))
+429 boundary_conditions.append(_get_kwd(tmp_string, 'BC3='))
+430
+431 if ruinfo == 1:
+432 corr_name.append(_get_corr_name(tmp_string, 'CORR_NAME='))
+433 corr_type.append(_get_kwd(tmp_string, 'DATATYPE='))
+434 corr_props.append([_get_kwd(tmp_string, 'PROP0='), _get_kwd(tmp_string, 'PROP1=')])
+435 if d0 == 0:
+436 d0 = int(_get_kwd(tmp_string, 'D0='))
+437 else:
+438 if d0 != int(_get_kwd(tmp_string, 'D0=')):
+439 print('Error: Varying number of time values')
+440 if d1 == 0:
+441 d1 = int(_get_kwd(tmp_string, 'D1='))
+442 else:
+443 if d1 != int(_get_kwd(tmp_string, 'D1=')):
+444 print('Error: Varying number of random sources')
+445 if ruinfo == 2:
+446 prop_kappa.append(_get_kwd(tmp_string, 'KAPPA='))
+447 prop_source.append(_get_kwd(tmp_string, 'x0='))
+448 if ruinfo == 4:
+449 cnfg_no = int(_get_kwd(tmp_string, 'CNFG_ID='))
+450 if stop:
+451 if cnfg_no > kwargs.get('stop'):
+452 break
+453 idl.append(cnfg_no)
+454 print('\r%s %i' % ('Reading configuration', cnfg_no), end='\r')
+455 if len(idl) == 1:
+456 no_corrs = len(corr_name)
+457 data = []
+458 for c in range(no_corrs):
+459 data.append([])
460
-461 corr_kappa = [] # Contains kappa values for both propagators of given correlation function
-462 corr_source = []
-463 for item in corr_props:
-464 corr_kappa.append([float(prop_kappa[int(item[0])]), float(prop_kappa[int(item[1])])])
-465 if prop_source[int(item[0])] != prop_source[int(item[1])]:
-466 raise Exception('Source position do not match for correlator' + str(item))
-467 else:
-468 corr_source.append(int(prop_source[int(item[0])]))
-469
-470 if stop is None:
-471 stop = idl[-1]
-472 idl_target = range(start, stop + 1, step)
-473
-474 if set(idl) != set(idl_target):
-475 try:
-476 indices = [idl.index(i) for i in idl_target]
-477 except ValueError as err:
-478 raise Exception('Configurations in file do no match target list!', err)
-479 else:
-480 indices = None
-481
-482 result = {}
-483 for c in range(no_corrs):
-484 tmp_corr = []
-485 tmp_data = np.asarray(data[c])
-486 for t in range(d0 - 2):
-487 if indices:
-488 deltas = [tmp_data[:, t][index] for index in indices]
-489 else:
-490 deltas = tmp_data[:, t]
-491 tmp_corr.append(Obs([deltas], [ensemble_name], idl=[idl_target]))
-492 result[(corr_name[c], corr_source[c]) + tuple(corr_kappa[c])] = tmp_corr
-493
-494 # Check that all data entries have the same number of configurations
-495 if len(set([o[0].N for o in list(result.values())])) != 1:
-496 raise Exception('Error: Not all correlators have the same number of configurations. bdio file is possibly corrupted.')
-497
-498 return result
-499
-500
-501def read_dSdm(file_path, bdio_path='./libbdio.so', **kwargs):
-502 """ Extract dSdm data from a bdio file and return it as a dictionary
-503
-504 The dictionary can be accessed with a tuple consisting of (type, kappa)
-505
-506 read_dSdm requires bdio to be compiled into a shared library. This can be achieved by
-507 adding the flag -fPIC to CC and changing the all target to
+461 corr_no = 0
+462
+463 bdio_close(fbdio)
+464
+465 print('\nEnsemble: ', ensemble_name)
+466 if 'alternative_ensemble_name' in kwargs:
+467 ensemble_name = kwargs.get('alternative_ensemble_name')
+468 print('Ensemble name overwritten to', ensemble_name)
+469 print('Lattice volume: ', volume)
+470 print('Boundary conditions: ', boundary_conditions)
+471 print('Number of time values: ', d0)
+472 print('Number of random sources: ', d1)
+473 print('Number of corrs: ', len(corr_name))
+474 print('Number of configurations: ', len(idl))
+475
+476 corr_kappa = [] # Contains kappa values for both propagators of given correlation function
+477 corr_source = []
+478 for item in corr_props:
+479 corr_kappa.append([float(prop_kappa[int(item[0])]), float(prop_kappa[int(item[1])])])
+480 if prop_source[int(item[0])] != prop_source[int(item[1])]:
+481 raise Exception('Source position do not match for correlator' + str(item))
+482 else:
+483 corr_source.append(int(prop_source[int(item[0])]))
+484
+485 if stop is None:
+486 stop = idl[-1]
+487 idl_target = range(start, stop + 1, step)
+488
+489 if set(idl) != set(idl_target):
+490 try:
+491 indices = [idl.index(i) for i in idl_target]
+492 except ValueError as err:
+493 raise Exception('Configurations in file do no match target list!', err)
+494 else:
+495 indices = None
+496
+497 result = {}
+498 for c in range(no_corrs):
+499 tmp_corr = []
+500 tmp_data = np.asarray(data[c])
+501 for t in range(d0 - 2):
+502 if indices:
+503 deltas = [tmp_data[:, t][index] for index in indices]
+504 else:
+505 deltas = tmp_data[:, t]
+506 tmp_corr.append(Obs([deltas], [ensemble_name], idl=[idl_target]))
+507 result[(corr_name[c], corr_source[c]) + tuple(corr_kappa[c])] = tmp_corr
508
-509 all: bdio.o $(LIBDIR)
-510 gcc -shared -Wl,-soname,libbdio.so -o $(BUILDDIR)/libbdio.so $(BUILDDIR)/bdio.o
-511 cp $(BUILDDIR)/libbdio.so $(LIBDIR)/
+509 # Check that all data entries have the same number of configurations
+510 if len(set([o[0].N for o in list(result.values())])) != 1:
+511 raise Exception('Error: Not all correlators have the same number of configurations. bdio file is possibly corrupted.')
512
-513 Parameters
-514 ----------
-515 file_path : str
-516 path to the bdio file
-517 bdio_path : str
-518 path to the shared bdio library libbdio.so (default ./libbdio.so)
-519 start : int
-520 The first configuration to be read (default 1)
-521 stop : int
-522 The last configuration to be read (default None)
-523 step : int
-524 Fixed step size between two measurements (default 1)
-525 alternative_ensemble_name : str
-526 Manually overwrite ensemble name
-527 """
-528
-529 start = kwargs.get('start', 1)
-530 stop = kwargs.get('stop', None)
-531 step = kwargs.get('step', 1)
-532
-533 bdio = ctypes.cdll.LoadLibrary(bdio_path)
-534
-535 bdio_open = bdio.bdio_open
-536 bdio_open.restype = ctypes.c_void_p
-537
-538 bdio_close = bdio.bdio_close
-539 bdio_close.restype = ctypes.c_int
-540 bdio_close.argtypes = [ctypes.c_void_p]
-541
-542 bdio_seek_record = bdio.bdio_seek_record
-543 bdio_seek_record.restype = ctypes.c_int
-544 bdio_seek_record.argtypes = [ctypes.c_void_p]
-545
-546 bdio_get_rlen = bdio.bdio_get_rlen
-547 bdio_get_rlen.restype = ctypes.c_int
-548 bdio_get_rlen.argtypes = [ctypes.c_void_p]
+513 return result
+514
+515
+516def read_dSdm(file_path, bdio_path='./libbdio.so', **kwargs):
+517 """ Extract dSdm data from a bdio file and return it as a dictionary
+518
+519 The dictionary can be accessed with a tuple consisting of (type, kappa)
+520
+521 read_dSdm requires bdio to be compiled into a shared library. This can be achieved by
+522 adding the flag -fPIC to CC and changing the all target to
+523
+524 all: bdio.o $(LIBDIR)
+525 gcc -shared -Wl,-soname,libbdio.so -o $(BUILDDIR)/libbdio.so $(BUILDDIR)/bdio.o
+526 cp $(BUILDDIR)/libbdio.so $(LIBDIR)/
+527
+528 Parameters
+529 ----------
+530 file_path : str
+531 path to the bdio file
+532 bdio_path : str
+533 path to the shared bdio library libbdio.so (default ./libbdio.so)
+534 start : int
+535 The first configuration to be read (default 1)
+536 stop : int
+537 The last configuration to be read (default None)
+538 step : int
+539 Fixed step size between two measurements (default 1)
+540 alternative_ensemble_name : str
+541 Manually overwrite ensemble name
+542 """
+543
+544 start = kwargs.get('start', 1)
+545 stop = kwargs.get('stop', None)
+546 step = kwargs.get('step', 1)
+547
+548 bdio = ctypes.cdll.LoadLibrary(bdio_path)
549
-550 bdio_get_ruinfo = bdio.bdio_get_ruinfo
-551 bdio_get_ruinfo.restype = ctypes.c_int
-552 bdio_get_ruinfo.argtypes = [ctypes.c_void_p]
-553
-554 bdio_read = bdio.bdio_read
-555 bdio_read.restype = ctypes.c_size_t
-556 bdio_read.argtypes = [ctypes.c_char_p, ctypes.c_size_t, ctypes.c_void_p]
-557
-558 bdio_read_f64 = bdio.bdio_read_f64
-559 bdio_read_f64.restype = ctypes.c_size_t
-560 bdio_read_f64.argtypes = [ctypes.c_void_p, ctypes.c_size_t, ctypes.c_void_p]
-561
-562 b_path = file_path.encode('utf-8')
-563 read = 'r'
-564 b_read = read.encode('utf-8')
-565 form = 'Generic Correlator Format 1.0'
-566 b_form = form.encode('utf-8')
-567
-568 ensemble_name = ''
-569 volume = [] # lattice volume
-570 boundary_conditions = []
-571 corr_name = [] # Contains correlator names
-572 corr_type = [] # Contains correlator data type (important for reading out numerical data)
-573 corr_props = [] # Contains propagator types (Component of corr_kappa)
-574 d0 = 0 # tvals
-575 # d1 = 0 # nnoise
-576 prop_kappa = [] # Contains propagator kappas (Component of corr_kappa)
-577 # Check noise type for multiple replica?
-578 corr_no = -1
-579 data = []
-580 idl = []
-581
-582 fbdio = bdio_open(ctypes.c_char_p(b_path), ctypes.c_char_p(b_read), ctypes.c_char_p(b_form))
-583
-584 print('Reading of bdio file started')
-585 while True:
-586 bdio_seek_record(fbdio)
-587 ruinfo = bdio_get_ruinfo(fbdio)
-588 if ruinfo < 0:
-589 # EOF reached
-590 break
-591 rlen = bdio_get_rlen(fbdio)
-592 if ruinfo == 5:
-593 d_buf = ctypes.c_double * (2 + d0)
-594 pd_buf = d_buf()
-595 ppd_buf = ctypes.c_void_p(ctypes.addressof(pd_buf))
-596 bdio_read_f64(ppd_buf, ctypes.c_size_t(rlen), ctypes.c_void_p(fbdio))
-597 tmp_mean = np.mean(np.asarray(pd_buf[2:]))
+550 bdio_open = bdio.bdio_open
+551 bdio_open.restype = ctypes.c_void_p
+552
+553 bdio_close = bdio.bdio_close
+554 bdio_close.restype = ctypes.c_int
+555 bdio_close.argtypes = [ctypes.c_void_p]
+556
+557 bdio_seek_record = bdio.bdio_seek_record
+558 bdio_seek_record.restype = ctypes.c_int
+559 bdio_seek_record.argtypes = [ctypes.c_void_p]
+560
+561 bdio_get_rlen = bdio.bdio_get_rlen
+562 bdio_get_rlen.restype = ctypes.c_int
+563 bdio_get_rlen.argtypes = [ctypes.c_void_p]
+564
+565 bdio_get_ruinfo = bdio.bdio_get_ruinfo
+566 bdio_get_ruinfo.restype = ctypes.c_int
+567 bdio_get_ruinfo.argtypes = [ctypes.c_void_p]
+568
+569 bdio_read = bdio.bdio_read
+570 bdio_read.restype = ctypes.c_size_t
+571 bdio_read.argtypes = [ctypes.c_char_p, ctypes.c_size_t, ctypes.c_void_p]
+572
+573 bdio_read_f64 = bdio.bdio_read_f64
+574 bdio_read_f64.restype = ctypes.c_size_t
+575 bdio_read_f64.argtypes = [ctypes.c_void_p, ctypes.c_size_t, ctypes.c_void_p]
+576
+577 b_path = file_path.encode('utf-8')
+578 read = 'r'
+579 b_read = read.encode('utf-8')
+580 form = 'Generic Correlator Format 1.0'
+581 b_form = form.encode('utf-8')
+582
+583 ensemble_name = ''
+584 volume = [] # lattice volume
+585 boundary_conditions = []
+586 corr_name = [] # Contains correlator names
+587 corr_type = [] # Contains correlator data type (important for reading out numerical data)
+588 corr_props = [] # Contains propagator types (Component of corr_kappa)
+589 d0 = 0 # tvals
+590 # d1 = 0 # nnoise
+591 prop_kappa = [] # Contains propagator kappas (Component of corr_kappa)
+592 # Check noise type for multiple replica?
+593 corr_no = -1
+594 data = []
+595 idl = []
+596
+597 fbdio = bdio_open(ctypes.c_char_p(b_path), ctypes.c_char_p(b_read), ctypes.c_char_p(b_form))
598
-599 data[corr_no].append(tmp_mean)
-600 corr_no += 1
-601 else:
-602 alt_buf = ctypes.create_string_buffer(1024)
-603 palt_buf = ctypes.c_char_p(ctypes.addressof(alt_buf))
-604 iread = bdio_read(palt_buf, ctypes.c_size_t(rlen), ctypes.c_void_p(fbdio))
-605 if rlen != iread:
-606 print('Error')
-607 for i, item in enumerate(alt_buf):
-608 if item == b'\x00':
-609 alt_buf[i] = b' '
-610 tmp_string = (alt_buf[:].decode("utf-8")).rstrip()
-611 if ruinfo == 0:
-612 creator = _get_kwd(tmp_string, 'CREATOR=')
-613 ensemble_name = _get_kwd(tmp_string, 'ENSEMBLE=')
-614 volume.append(int(_get_kwd(tmp_string, 'L0=')))
-615 volume.append(int(_get_kwd(tmp_string, 'L1=')))
-616 volume.append(int(_get_kwd(tmp_string, 'L2=')))
-617 volume.append(int(_get_kwd(tmp_string, 'L3=')))
-618 boundary_conditions.append(_get_kwd(tmp_string, 'BC0='))
-619 boundary_conditions.append(_get_kwd(tmp_string, 'BC1='))
-620 boundary_conditions.append(_get_kwd(tmp_string, 'BC2='))
-621 boundary_conditions.append(_get_kwd(tmp_string, 'BC3='))
-622
-623 if ruinfo == 1:
-624 corr_name.append(_get_corr_name(tmp_string, 'CORR_NAME='))
-625 corr_type.append(_get_kwd(tmp_string, 'DATATYPE='))
-626 corr_props.append(_get_kwd(tmp_string, 'PROP0='))
-627 if d0 == 0:
-628 d0 = int(_get_kwd(tmp_string, 'D0='))
-629 else:
-630 if d0 != int(_get_kwd(tmp_string, 'D0=')):
-631 print('Error: Varying number of time values')
-632 if ruinfo == 2:
-633 prop_kappa.append(_get_kwd(tmp_string, 'KAPPA='))
-634 if ruinfo == 4:
-635 cnfg_no = int(_get_kwd(tmp_string, 'CNFG_ID='))
-636 if stop:
-637 if cnfg_no > kwargs.get('stop'):
-638 break
-639 idl.append(cnfg_no)
-640 print('\r%s %i' % ('Reading configuration', cnfg_no), end='\r')
-641 if len(idl) == 1:
-642 no_corrs = len(corr_name)
-643 data = []
-644 for c in range(no_corrs):
-645 data.append([])
-646
-647 corr_no = 0
-648 bdio_close(fbdio)
-649
-650 print('\nCreator: ', creator)
-651 print('Ensemble: ', ensemble_name)
-652 print('Lattice volume: ', volume)
-653 print('Boundary conditions: ', boundary_conditions)
-654 print('Number of random sources: ', d0)
-655 print('Number of corrs: ', len(corr_name))
-656 print('Number of configurations: ', cnfg_no + 1)
-657
-658 corr_kappa = [] # Contains kappa values for both propagators of given correlation function
-659 for item in corr_props:
-660 corr_kappa.append(float(prop_kappa[int(item)]))
+599 print('Reading of bdio file started')
+600 while True:
+601 bdio_seek_record(fbdio)
+602 ruinfo = bdio_get_ruinfo(fbdio)
+603 if ruinfo < 0:
+604 # EOF reached
+605 break
+606 rlen = bdio_get_rlen(fbdio)
+607 if ruinfo == 5:
+608 d_buf = ctypes.c_double * (2 + d0)
+609 pd_buf = d_buf()
+610 ppd_buf = ctypes.c_void_p(ctypes.addressof(pd_buf))
+611 bdio_read_f64(ppd_buf, ctypes.c_size_t(rlen), ctypes.c_void_p(fbdio))
+612 tmp_mean = np.mean(np.asarray(pd_buf[2:]))
+613
+614 data[corr_no].append(tmp_mean)
+615 corr_no += 1
+616 else:
+617 alt_buf = ctypes.create_string_buffer(1024)
+618 palt_buf = ctypes.c_char_p(ctypes.addressof(alt_buf))
+619 iread = bdio_read(palt_buf, ctypes.c_size_t(rlen), ctypes.c_void_p(fbdio))
+620 if rlen != iread:
+621 print('Error')
+622 for i, item in enumerate(alt_buf):
+623 if item == b'\x00':
+624 alt_buf[i] = b' '
+625 tmp_string = (alt_buf[:].decode("utf-8")).rstrip()
+626 if ruinfo == 0:
+627 creator = _get_kwd(tmp_string, 'CREATOR=')
+628 ensemble_name = _get_kwd(tmp_string, 'ENSEMBLE=')
+629 volume.append(int(_get_kwd(tmp_string, 'L0=')))
+630 volume.append(int(_get_kwd(tmp_string, 'L1=')))
+631 volume.append(int(_get_kwd(tmp_string, 'L2=')))
+632 volume.append(int(_get_kwd(tmp_string, 'L3=')))
+633 boundary_conditions.append(_get_kwd(tmp_string, 'BC0='))
+634 boundary_conditions.append(_get_kwd(tmp_string, 'BC1='))
+635 boundary_conditions.append(_get_kwd(tmp_string, 'BC2='))
+636 boundary_conditions.append(_get_kwd(tmp_string, 'BC3='))
+637
+638 if ruinfo == 1:
+639 corr_name.append(_get_corr_name(tmp_string, 'CORR_NAME='))
+640 corr_type.append(_get_kwd(tmp_string, 'DATATYPE='))
+641 corr_props.append(_get_kwd(tmp_string, 'PROP0='))
+642 if d0 == 0:
+643 d0 = int(_get_kwd(tmp_string, 'D0='))
+644 else:
+645 if d0 != int(_get_kwd(tmp_string, 'D0=')):
+646 print('Error: Varying number of time values')
+647 if ruinfo == 2:
+648 prop_kappa.append(_get_kwd(tmp_string, 'KAPPA='))
+649 if ruinfo == 4:
+650 cnfg_no = int(_get_kwd(tmp_string, 'CNFG_ID='))
+651 if stop:
+652 if cnfg_no > kwargs.get('stop'):
+653 break
+654 idl.append(cnfg_no)
+655 print('\r%s %i' % ('Reading configuration', cnfg_no), end='\r')
+656 if len(idl) == 1:
+657 no_corrs = len(corr_name)
+658 data = []
+659 for c in range(no_corrs):
+660 data.append([])
661
-662 if stop is None:
-663 stop = idl[-1]
-664 idl_target = range(start, stop + 1, step)
-665 try:
-666 indices = [idl.index(i) for i in idl_target]
-667 except ValueError as err:
-668 raise Exception('Configurations in file do no match target list!', err)
-669
-670 result = {}
-671 for c in range(no_corrs):
-672 deltas = [np.asarray(data[c])[index] for index in indices]
-673 result[(corr_name[c], str(corr_kappa[c]))] = Obs([deltas], [ensemble_name], idl=[idl_target])
-674
-675 # Check that all data entries have the same number of configurations
-676 if len(set([o.N for o in list(result.values())])) != 1:
-677 raise Exception('Error: Not all correlators have the same number of configurations. bdio file is possibly corrupted.')
-678
-679 return result
+662 corr_no = 0
+663 bdio_close(fbdio)
+664
+665 print('\nCreator: ', creator)
+666 print('Ensemble: ', ensemble_name)
+667 print('Lattice volume: ', volume)
+668 print('Boundary conditions: ', boundary_conditions)
+669 print('Number of random sources: ', d0)
+670 print('Number of corrs: ', len(corr_name))
+671 print('Number of configurations: ', cnfg_no + 1)
+672
+673 corr_kappa = [] # Contains kappa values for both propagators of given correlation function
+674 for item in corr_props:
+675 corr_kappa.append(float(prop_kappa[int(item)]))
+676
+677 if stop is None:
+678 stop = idl[-1]
+679 idl_target = range(start, stop + 1, step)
+680 try:
+681 indices = [idl.index(i) for i in idl_target]
+682 except ValueError as err:
+683 raise Exception('Configurations in file do no match target list!', err)
+684
+685 result = {}
+686 for c in range(no_corrs):
+687 deltas = [np.asarray(data[c])[index] for index in indices]
+688 result[(corr_name[c], str(corr_kappa[c]))] = Obs([deltas], [ensemble_name], idl=[idl_target])
+689
+690 # Check that all data entries have the same number of configurations
+691 if len(set([o.N for o in list(result.values())])) != 1:
+692 raise Exception('Error: Not all correlators have the same number of configurations. bdio file is possibly corrupted.')
+693
+694 return result
@@ -793,141 +808,146 @@
19 ----------
20 file_path -- path to the bdio file
21 bdio_path -- path to the shared bdio library libbdio.so (default ./libbdio.so)
- 22 """
- 23 bdio = ctypes.cdll.LoadLibrary(bdio_path)
- 24
- 25 bdio_open = bdio.bdio_open
- 26 bdio_open.restype = ctypes.c_void_p
- 27
- 28 bdio_close = bdio.bdio_close
- 29 bdio_close.restype = ctypes.c_int
- 30 bdio_close.argtypes = [ctypes.c_void_p]
- 31
- 32 bdio_seek_record = bdio.bdio_seek_record
- 33 bdio_seek_record.restype = ctypes.c_int
- 34 bdio_seek_record.argtypes = [ctypes.c_void_p]
- 35
- 36 bdio_get_rlen = bdio.bdio_get_rlen
- 37 bdio_get_rlen.restype = ctypes.c_int
- 38 bdio_get_rlen.argtypes = [ctypes.c_void_p]
- 39
- 40 bdio_get_ruinfo = bdio.bdio_get_ruinfo
- 41 bdio_get_ruinfo.restype = ctypes.c_int
- 42 bdio_get_ruinfo.argtypes = [ctypes.c_void_p]
- 43
- 44 bdio_read = bdio.bdio_read
- 45 bdio_read.restype = ctypes.c_size_t
- 46 bdio_read.argtypes = [ctypes.c_char_p, ctypes.c_size_t, ctypes.c_void_p]
- 47
- 48 bdio_read_f64 = bdio.bdio_read_f64
- 49 bdio_read_f64.restype = ctypes.c_size_t
- 50 bdio_read_f64.argtypes = [ctypes.c_void_p, ctypes.c_size_t, ctypes.c_void_p]
- 51
- 52 bdio_read_int32 = bdio.bdio_read_int32
- 53 bdio_read_int32.restype = ctypes.c_size_t
- 54 bdio_read_int32.argtypes = [ctypes.c_void_p, ctypes.c_size_t, ctypes.c_void_p]
- 55
- 56 b_path = file_path.encode('utf-8')
- 57 read = 'r'
- 58 b_read = read.encode('utf-8')
- 59
- 60 fbdio = bdio_open(ctypes.c_char_p(b_path), ctypes.c_char_p(b_read), None)
- 61
- 62 return_list = []
- 63
- 64 print('Reading of bdio file started')
- 65 while True:
- 66 bdio_seek_record(fbdio)
- 67 ruinfo = bdio_get_ruinfo(fbdio)
+ 22
+ 23 Returns
+ 24 -------
+ 25 data : List[Obs]
+ 26 Extracted data
+ 27 """
+ 28 bdio = ctypes.cdll.LoadLibrary(bdio_path)
+ 29
+ 30 bdio_open = bdio.bdio_open
+ 31 bdio_open.restype = ctypes.c_void_p
+ 32
+ 33 bdio_close = bdio.bdio_close
+ 34 bdio_close.restype = ctypes.c_int
+ 35 bdio_close.argtypes = [ctypes.c_void_p]
+ 36
+ 37 bdio_seek_record = bdio.bdio_seek_record
+ 38 bdio_seek_record.restype = ctypes.c_int
+ 39 bdio_seek_record.argtypes = [ctypes.c_void_p]
+ 40
+ 41 bdio_get_rlen = bdio.bdio_get_rlen
+ 42 bdio_get_rlen.restype = ctypes.c_int
+ 43 bdio_get_rlen.argtypes = [ctypes.c_void_p]
+ 44
+ 45 bdio_get_ruinfo = bdio.bdio_get_ruinfo
+ 46 bdio_get_ruinfo.restype = ctypes.c_int
+ 47 bdio_get_ruinfo.argtypes = [ctypes.c_void_p]
+ 48
+ 49 bdio_read = bdio.bdio_read
+ 50 bdio_read.restype = ctypes.c_size_t
+ 51 bdio_read.argtypes = [ctypes.c_char_p, ctypes.c_size_t, ctypes.c_void_p]
+ 52
+ 53 bdio_read_f64 = bdio.bdio_read_f64
+ 54 bdio_read_f64.restype = ctypes.c_size_t
+ 55 bdio_read_f64.argtypes = [ctypes.c_void_p, ctypes.c_size_t, ctypes.c_void_p]
+ 56
+ 57 bdio_read_int32 = bdio.bdio_read_int32
+ 58 bdio_read_int32.restype = ctypes.c_size_t
+ 59 bdio_read_int32.argtypes = [ctypes.c_void_p, ctypes.c_size_t, ctypes.c_void_p]
+ 60
+ 61 b_path = file_path.encode('utf-8')
+ 62 read = 'r'
+ 63 b_read = read.encode('utf-8')
+ 64
+ 65 fbdio = bdio_open(ctypes.c_char_p(b_path), ctypes.c_char_p(b_read), None)
+ 66
+ 67 return_list = []
68
- 69 if ruinfo == 7:
- 70 print('MD5sum found') # For now we just ignore these entries and do not perform any checks on them
- 71 continue
- 72
- 73 if ruinfo < 0:
- 74 # EOF reached
- 75 break
- 76 bdio_get_rlen(fbdio)
+ 69 print('Reading of bdio file started')
+ 70 while True:
+ 71 bdio_seek_record(fbdio)
+ 72 ruinfo = bdio_get_ruinfo(fbdio)
+ 73
+ 74 if ruinfo == 7:
+ 75 print('MD5sum found') # For now we just ignore these entries and do not perform any checks on them
+ 76 continue
77
- 78 def read_c_double():
- 79 d_buf = ctypes.c_double
- 80 pd_buf = d_buf()
- 81 ppd_buf = ctypes.c_void_p(ctypes.addressof(pd_buf))
- 82 bdio_read_f64(ppd_buf, ctypes.c_size_t(8), ctypes.c_void_p(fbdio))
- 83 return pd_buf.value
- 84
- 85 mean = read_c_double()
- 86 print('mean', mean)
- 87
- 88 def read_c_size_t():
- 89 d_buf = ctypes.c_size_t
- 90 pd_buf = d_buf()
- 91 ppd_buf = ctypes.c_void_p(ctypes.addressof(pd_buf))
- 92 bdio_read_int32(ppd_buf, ctypes.c_size_t(4), ctypes.c_void_p(fbdio))
- 93 return pd_buf.value
- 94
- 95 neid = read_c_size_t()
- 96 print('neid', neid)
- 97
- 98 ndata = []
- 99 for index in range(neid):
-100 ndata.append(read_c_size_t())
-101 print('ndata', ndata)
+ 78 if ruinfo < 0:
+ 79 # EOF reached
+ 80 break
+ 81 bdio_get_rlen(fbdio)
+ 82
+ 83 def read_c_double():
+ 84 d_buf = ctypes.c_double
+ 85 pd_buf = d_buf()
+ 86 ppd_buf = ctypes.c_void_p(ctypes.addressof(pd_buf))
+ 87 bdio_read_f64(ppd_buf, ctypes.c_size_t(8), ctypes.c_void_p(fbdio))
+ 88 return pd_buf.value
+ 89
+ 90 mean = read_c_double()
+ 91 print('mean', mean)
+ 92
+ 93 def read_c_size_t():
+ 94 d_buf = ctypes.c_size_t
+ 95 pd_buf = d_buf()
+ 96 ppd_buf = ctypes.c_void_p(ctypes.addressof(pd_buf))
+ 97 bdio_read_int32(ppd_buf, ctypes.c_size_t(4), ctypes.c_void_p(fbdio))
+ 98 return pd_buf.value
+ 99
+100 neid = read_c_size_t()
+101 print('neid', neid)
102
-103 nrep = []
+103 ndata = []
104 for index in range(neid):
-105 nrep.append(read_c_size_t())
-106 print('nrep', nrep)
+105 ndata.append(read_c_size_t())
+106 print('ndata', ndata)
107
-108 vrep = []
+108 nrep = []
109 for index in range(neid):
-110 vrep.append([])
-111 for jndex in range(nrep[index]):
-112 vrep[-1].append(read_c_size_t())
-113 print('vrep', vrep)
-114
-115 ids = []
-116 for index in range(neid):
-117 ids.append(read_c_size_t())
-118 print('ids', ids)
+110 nrep.append(read_c_size_t())
+111 print('nrep', nrep)
+112
+113 vrep = []
+114 for index in range(neid):
+115 vrep.append([])
+116 for jndex in range(nrep[index]):
+117 vrep[-1].append(read_c_size_t())
+118 print('vrep', vrep)
119
-120 nt = []
+120 ids = []
121 for index in range(neid):
-122 nt.append(read_c_size_t())
-123 print('nt', nt)
+122 ids.append(read_c_size_t())
+123 print('ids', ids)
124
-125 zero = []
+125 nt = []
126 for index in range(neid):
-127 zero.append(read_c_double())
-128 print('zero', zero)
+127 nt.append(read_c_size_t())
+128 print('nt', nt)
129
-130 four = []
+130 zero = []
131 for index in range(neid):
-132 four.append(read_c_double())
-133 print('four', four)
+132 zero.append(read_c_double())
+133 print('zero', zero)
134
-135 d_buf = ctypes.c_double * np.sum(ndata)
-136 pd_buf = d_buf()
-137 ppd_buf = ctypes.c_void_p(ctypes.addressof(pd_buf))
-138 bdio_read_f64(ppd_buf, ctypes.c_size_t(8 * np.sum(ndata)), ctypes.c_void_p(fbdio))
-139 delta = pd_buf[:]
-140
-141 samples = np.split(np.asarray(delta) + mean, np.cumsum([a for su in vrep for a in su])[:-1])
-142 no_reps = [len(o) for o in vrep]
-143 assert len(ids) == len(no_reps)
-144 tmp_names = []
-145 ens_length = max([len(str(o)) for o in ids])
-146 for loc_id, reps in zip(ids, no_reps):
-147 for index in range(reps):
-148 missing_chars = ens_length - len(str(loc_id))
-149 tmp_names.append(str(loc_id) + ' ' * missing_chars + '|r' + '{0:03d}'.format(index))
-150
-151 return_list.append(Obs(samples, tmp_names))
-152
-153 bdio_close(fbdio)
-154 print()
-155 print(len(return_list), 'observable(s) extracted.')
-156 return return_list
+135 four = []
+136 for index in range(neid):
+137 four.append(read_c_double())
+138 print('four', four)
+139
+140 d_buf = ctypes.c_double * np.sum(ndata)
+141 pd_buf = d_buf()
+142 ppd_buf = ctypes.c_void_p(ctypes.addressof(pd_buf))
+143 bdio_read_f64(ppd_buf, ctypes.c_size_t(8 * np.sum(ndata)), ctypes.c_void_p(fbdio))
+144 delta = pd_buf[:]
+145
+146 samples = np.split(np.asarray(delta) + mean, np.cumsum([a for su in vrep for a in su])[:-1])
+147 no_reps = [len(o) for o in vrep]
+148 assert len(ids) == len(no_reps)
+149 tmp_names = []
+150 ens_length = max([len(str(o)) for o in ids])
+151 for loc_id, reps in zip(ids, no_reps):
+152 for index in range(reps):
+153 missing_chars = ens_length - len(str(loc_id))
+154 tmp_names.append(str(loc_id) + ' ' * missing_chars + '|r' + '{0:03d}'.format(index))
+155
+156 return_list.append(Obs(samples, tmp_names))
+157
+158 bdio_close(fbdio)
+159 print()
+160 print(len(return_list), 'observable(s) extracted.')
+161 return return_list
@@ -946,6 +966,13 @@ adding the flag -fPIC to CC and changing the all target to
file_path -- path to the bdio file
bdio_path -- path to the shared bdio library libbdio.so (default ./libbdio.so)
+
+Returns
+
+
+- data (List[Obs]):
+Extracted data
+
@@ -961,129 +988,134 @@ adding the flag -fPIC to CC and changing the all target to
- 159def write_ADerrors(obs_list, file_path, bdio_path='./libbdio.so', **kwargs):
-160 """ Write Obs to a bdio file according to ADerrors conventions
-161
-162 read_mesons requires bdio to be compiled into a shared library. This can be achieved by
-163 adding the flag -fPIC to CC and changing the all target to
-164
-165 all: bdio.o $(LIBDIR)
-166 gcc -shared -Wl,-soname,libbdio.so -o $(BUILDDIR)/libbdio.so $(BUILDDIR)/bdio.o
-167 cp $(BUILDDIR)/libbdio.so $(LIBDIR)/
-168
-169 Parameters
-170 ----------
-171 file_path -- path to the bdio file
-172 bdio_path -- path to the shared bdio library libbdio.so (default ./libbdio.so)
-173 """
-174
-175 for obs in obs_list:
-176 if not hasattr(obs, 'e_names'):
-177 raise Exception('Run the gamma method first for all obs.')
+ 164def write_ADerrors(obs_list, file_path, bdio_path='./libbdio.so', **kwargs):
+165 """ Write Obs to a bdio file according to ADerrors conventions
+166
+167 read_mesons requires bdio to be compiled into a shared library. This can be achieved by
+168 adding the flag -fPIC to CC and changing the all target to
+169
+170 all: bdio.o $(LIBDIR)
+171 gcc -shared -Wl,-soname,libbdio.so -o $(BUILDDIR)/libbdio.so $(BUILDDIR)/bdio.o
+172 cp $(BUILDDIR)/libbdio.so $(LIBDIR)/
+173
+174 Parameters
+175 ----------
+176 file_path -- path to the bdio file
+177 bdio_path -- path to the shared bdio library libbdio.so (default ./libbdio.so)
178
-179 bdio = ctypes.cdll.LoadLibrary(bdio_path)
-180
-181 bdio_open = bdio.bdio_open
-182 bdio_open.restype = ctypes.c_void_p
-183
-184 bdio_close = bdio.bdio_close
-185 bdio_close.restype = ctypes.c_int
-186 bdio_close.argtypes = [ctypes.c_void_p]
-187
-188 bdio_start_record = bdio.bdio_start_record
-189 bdio_start_record.restype = ctypes.c_int
-190 bdio_start_record.argtypes = [ctypes.c_size_t, ctypes.c_size_t, ctypes.c_void_p]
-191
-192 bdio_flush_record = bdio.bdio_flush_record
-193 bdio_flush_record.restype = ctypes.c_int
-194 bdio_flush_record.argytpes = [ctypes.c_void_p]
-195
-196 bdio_write_f64 = bdio.bdio_write_f64
-197 bdio_write_f64.restype = ctypes.c_size_t
-198 bdio_write_f64.argtypes = [ctypes.c_void_p, ctypes.c_size_t, ctypes.c_void_p]
-199
-200 bdio_write_int32 = bdio.bdio_write_int32
-201 bdio_write_int32.restype = ctypes.c_size_t
-202 bdio_write_int32.argtypes = [ctypes.c_void_p, ctypes.c_size_t, ctypes.c_void_p]
-203
-204 b_path = file_path.encode('utf-8')
-205 write = 'w'
-206 b_write = write.encode('utf-8')
-207 form = 'pyerrors ADerror export'
-208 b_form = form.encode('utf-8')
+179 Returns
+180 -------
+181 success : int
+182 returns 0 is successful
+183 """
+184
+185 for obs in obs_list:
+186 if not hasattr(obs, 'e_names'):
+187 raise Exception('Run the gamma method first for all obs.')
+188
+189 bdio = ctypes.cdll.LoadLibrary(bdio_path)
+190
+191 bdio_open = bdio.bdio_open
+192 bdio_open.restype = ctypes.c_void_p
+193
+194 bdio_close = bdio.bdio_close
+195 bdio_close.restype = ctypes.c_int
+196 bdio_close.argtypes = [ctypes.c_void_p]
+197
+198 bdio_start_record = bdio.bdio_start_record
+199 bdio_start_record.restype = ctypes.c_int
+200 bdio_start_record.argtypes = [ctypes.c_size_t, ctypes.c_size_t, ctypes.c_void_p]
+201
+202 bdio_flush_record = bdio.bdio_flush_record
+203 bdio_flush_record.restype = ctypes.c_int
+204 bdio_flush_record.argytpes = [ctypes.c_void_p]
+205
+206 bdio_write_f64 = bdio.bdio_write_f64
+207 bdio_write_f64.restype = ctypes.c_size_t
+208 bdio_write_f64.argtypes = [ctypes.c_void_p, ctypes.c_size_t, ctypes.c_void_p]
209
-210 fbdio = bdio_open(ctypes.c_char_p(b_path), ctypes.c_char_p(b_write), b_form)
-211
-212 for obs in obs_list:
-213 # mean = obs.value
-214 neid = len(obs.e_names)
-215 vrep = [[obs.shape[o] for o in sl] for sl in list(obs.e_content.values())]
-216 vrep_write = [item for sublist in vrep for item in sublist]
-217 ndata = [np.sum(o) for o in vrep]
-218 nrep = [len(o) for o in vrep]
-219 print('ndata', ndata)
-220 print('nrep', nrep)
-221 print('vrep', vrep)
-222 keys = list(obs.e_content.keys())
-223 ids = []
-224 for key in keys:
-225 try: # Try to convert key to integer
-226 ids.append(int(key))
-227 except Exception: # If not possible construct a hash
-228 ids.append(int(hashlib.sha256(key.encode('utf-8')).hexdigest(), 16) % 10 ** 8)
-229 print('ids', ids)
-230 nt = []
-231 for e, e_name in enumerate(obs.e_names):
-232
-233 r_length = []
-234 for r_name in obs.e_content[e_name]:
-235 r_length.append(len(obs.deltas[r_name]))
-236
-237 # e_N = np.sum(r_length)
-238 nt.append(max(r_length) // 2)
-239 print('nt', nt)
-240 zero = neid * [0.0]
-241 four = neid * [4.0]
-242 print('zero', zero)
-243 print('four', four)
-244 delta = np.concatenate([item for sublist in [[obs.deltas[o] for o in sl] for sl in list(obs.e_content.values())] for item in sublist])
-245
-246 bdio_start_record(0x00, 8, fbdio)
-247
-248 def write_c_double(double):
-249 pd_buf = ctypes.c_double(double)
-250 ppd_buf = ctypes.c_void_p(ctypes.addressof(pd_buf))
-251 bdio_write_f64(ppd_buf, ctypes.c_size_t(8), ctypes.c_void_p(fbdio))
-252
-253 def write_c_size_t(int32):
-254 pd_buf = ctypes.c_size_t(int32)
-255 ppd_buf = ctypes.c_void_p(ctypes.addressof(pd_buf))
-256 bdio_write_int32(ppd_buf, ctypes.c_size_t(4), ctypes.c_void_p(fbdio))
+210 bdio_write_int32 = bdio.bdio_write_int32
+211 bdio_write_int32.restype = ctypes.c_size_t
+212 bdio_write_int32.argtypes = [ctypes.c_void_p, ctypes.c_size_t, ctypes.c_void_p]
+213
+214 b_path = file_path.encode('utf-8')
+215 write = 'w'
+216 b_write = write.encode('utf-8')
+217 form = 'pyerrors ADerror export'
+218 b_form = form.encode('utf-8')
+219
+220 fbdio = bdio_open(ctypes.c_char_p(b_path), ctypes.c_char_p(b_write), b_form)
+221
+222 for obs in obs_list:
+223 # mean = obs.value
+224 neid = len(obs.e_names)
+225 vrep = [[obs.shape[o] for o in sl] for sl in list(obs.e_content.values())]
+226 vrep_write = [item for sublist in vrep for item in sublist]
+227 ndata = [np.sum(o) for o in vrep]
+228 nrep = [len(o) for o in vrep]
+229 print('ndata', ndata)
+230 print('nrep', nrep)
+231 print('vrep', vrep)
+232 keys = list(obs.e_content.keys())
+233 ids = []
+234 for key in keys:
+235 try: # Try to convert key to integer
+236 ids.append(int(key))
+237 except Exception: # If not possible construct a hash
+238 ids.append(int(hashlib.sha256(key.encode('utf-8')).hexdigest(), 16) % 10 ** 8)
+239 print('ids', ids)
+240 nt = []
+241 for e, e_name in enumerate(obs.e_names):
+242
+243 r_length = []
+244 for r_name in obs.e_content[e_name]:
+245 r_length.append(len(obs.deltas[r_name]))
+246
+247 # e_N = np.sum(r_length)
+248 nt.append(max(r_length) // 2)
+249 print('nt', nt)
+250 zero = neid * [0.0]
+251 four = neid * [4.0]
+252 print('zero', zero)
+253 print('four', four)
+254 delta = np.concatenate([item for sublist in [[obs.deltas[o] for o in sl] for sl in list(obs.e_content.values())] for item in sublist])
+255
+256 bdio_start_record(0x00, 8, fbdio)
257
-258 write_c_double(obs.value)
-259 write_c_size_t(neid)
-260
-261 for element in ndata:
-262 write_c_size_t(element)
-263 for element in nrep:
-264 write_c_size_t(element)
-265 for element in vrep_write:
-266 write_c_size_t(element)
-267 for element in ids:
-268 write_c_size_t(element)
-269 for element in nt:
-270 write_c_size_t(element)
-271
-272 for element in zero:
-273 write_c_double(element)
-274 for element in four:
-275 write_c_double(element)
-276
-277 for element in delta:
-278 write_c_double(element)
-279
-280 bdio_close(fbdio)
-281 return 0
+258 def write_c_double(double):
+259 pd_buf = ctypes.c_double(double)
+260 ppd_buf = ctypes.c_void_p(ctypes.addressof(pd_buf))
+261 bdio_write_f64(ppd_buf, ctypes.c_size_t(8), ctypes.c_void_p(fbdio))
+262
+263 def write_c_size_t(int32):
+264 pd_buf = ctypes.c_size_t(int32)
+265 ppd_buf = ctypes.c_void_p(ctypes.addressof(pd_buf))
+266 bdio_write_int32(ppd_buf, ctypes.c_size_t(4), ctypes.c_void_p(fbdio))
+267
+268 write_c_double(obs.value)
+269 write_c_size_t(neid)
+270
+271 for element in ndata:
+272 write_c_size_t(element)
+273 for element in nrep:
+274 write_c_size_t(element)
+275 for element in vrep_write:
+276 write_c_size_t(element)
+277 for element in ids:
+278 write_c_size_t(element)
+279 for element in nt:
+280 write_c_size_t(element)
+281
+282 for element in zero:
+283 write_c_double(element)
+284 for element in four:
+285 write_c_double(element)
+286
+287 for element in delta:
+288 write_c_double(element)
+289
+290 bdio_close(fbdio)
+291 return 0
@@ -1102,6 +1134,13 @@ adding the flag -fPIC to CC and changing the all target to
file_path -- path to the bdio file
bdio_path -- path to the shared bdio library libbdio.so (default ./libbdio.so)
+
+Returns
+
+
+- success (int):
+returns 0 is successful
+
@@ -1117,214 +1156,219 @@ adding the flag -fPIC to CC and changing the all target to
- 292def read_mesons(file_path, bdio_path='./libbdio.so', **kwargs):
-293 """ Extract mesons data from a bdio file and return it as a dictionary
-294
-295 The dictionary can be accessed with a tuple consisting of (type, source_position, kappa1, kappa2)
-296
-297 read_mesons requires bdio to be compiled into a shared library. This can be achieved by
-298 adding the flag -fPIC to CC and changing the all target to
-299
-300 all: bdio.o $(LIBDIR)
-301 gcc -shared -Wl,-soname,libbdio.so -o $(BUILDDIR)/libbdio.so $(BUILDDIR)/bdio.o
-302 cp $(BUILDDIR)/libbdio.so $(LIBDIR)/
-303
-304 Parameters
-305 ----------
-306 file_path : str
-307 path to the bdio file
-308 bdio_path : str
-309 path to the shared bdio library libbdio.so (default ./libbdio.so)
-310 start : int
-311 The first configuration to be read (default 1)
-312 stop : int
-313 The last configuration to be read (default None)
-314 step : int
-315 Fixed step size between two measurements (default 1)
-316 alternative_ensemble_name : str
-317 Manually overwrite ensemble name
-318 """
-319
-320 start = kwargs.get('start', 1)
-321 stop = kwargs.get('stop', None)
-322 step = kwargs.get('step', 1)
-323
-324 bdio = ctypes.cdll.LoadLibrary(bdio_path)
-325
-326 bdio_open = bdio.bdio_open
-327 bdio_open.restype = ctypes.c_void_p
+ 302def read_mesons(file_path, bdio_path='./libbdio.so', **kwargs):
+303 """ Extract mesons data from a bdio file and return it as a dictionary
+304
+305 The dictionary can be accessed with a tuple consisting of (type, source_position, kappa1, kappa2)
+306
+307 read_mesons requires bdio to be compiled into a shared library. This can be achieved by
+308 adding the flag -fPIC to CC and changing the all target to
+309
+310 all: bdio.o $(LIBDIR)
+311 gcc -shared -Wl,-soname,libbdio.so -o $(BUILDDIR)/libbdio.so $(BUILDDIR)/bdio.o
+312 cp $(BUILDDIR)/libbdio.so $(LIBDIR)/
+313
+314 Parameters
+315 ----------
+316 file_path : str
+317 path to the bdio file
+318 bdio_path : str
+319 path to the shared bdio library libbdio.so (default ./libbdio.so)
+320 start : int
+321 The first configuration to be read (default 1)
+322 stop : int
+323 The last configuration to be read (default None)
+324 step : int
+325 Fixed step size between two measurements (default 1)
+326 alternative_ensemble_name : str
+327 Manually overwrite ensemble name
328
-329 bdio_close = bdio.bdio_close
-330 bdio_close.restype = ctypes.c_int
-331 bdio_close.argtypes = [ctypes.c_void_p]
-332
-333 bdio_seek_record = bdio.bdio_seek_record
-334 bdio_seek_record.restype = ctypes.c_int
-335 bdio_seek_record.argtypes = [ctypes.c_void_p]
-336
-337 bdio_get_rlen = bdio.bdio_get_rlen
-338 bdio_get_rlen.restype = ctypes.c_int
-339 bdio_get_rlen.argtypes = [ctypes.c_void_p]
+329 Returns
+330 -------
+331 data : dict
+332 Extracted meson data
+333 """
+334
+335 start = kwargs.get('start', 1)
+336 stop = kwargs.get('stop', None)
+337 step = kwargs.get('step', 1)
+338
+339 bdio = ctypes.cdll.LoadLibrary(bdio_path)
340
-341 bdio_get_ruinfo = bdio.bdio_get_ruinfo
-342 bdio_get_ruinfo.restype = ctypes.c_int
-343 bdio_get_ruinfo.argtypes = [ctypes.c_void_p]
-344
-345 bdio_read = bdio.bdio_read
-346 bdio_read.restype = ctypes.c_size_t
-347 bdio_read.argtypes = [ctypes.c_char_p, ctypes.c_size_t, ctypes.c_void_p]
-348
-349 bdio_read_f64 = bdio.bdio_read_f64
-350 bdio_read_f64.restype = ctypes.c_size_t
-351 bdio_read_f64.argtypes = [ctypes.c_void_p, ctypes.c_size_t, ctypes.c_void_p]
-352
-353 b_path = file_path.encode('utf-8')
-354 read = 'r'
-355 b_read = read.encode('utf-8')
-356 form = 'Generic Correlator Format 1.0'
-357 b_form = form.encode('utf-8')
-358
-359 ensemble_name = ''
-360 volume = [] # lattice volume
-361 boundary_conditions = []
-362 corr_name = [] # Contains correlator names
-363 corr_type = [] # Contains correlator data type (important for reading out numerical data)
-364 corr_props = [] # Contanis propagator types (Component of corr_kappa)
-365 d0 = 0 # tvals
-366 d1 = 0 # nnoise
-367 prop_kappa = [] # Contains propagator kappas (Component of corr_kappa)
-368 prop_source = [] # Contains propagator source positions
-369 # Check noise type for multiple replica?
-370 corr_no = -1
-371 data = []
-372 idl = []
+341 bdio_open = bdio.bdio_open
+342 bdio_open.restype = ctypes.c_void_p
+343
+344 bdio_close = bdio.bdio_close
+345 bdio_close.restype = ctypes.c_int
+346 bdio_close.argtypes = [ctypes.c_void_p]
+347
+348 bdio_seek_record = bdio.bdio_seek_record
+349 bdio_seek_record.restype = ctypes.c_int
+350 bdio_seek_record.argtypes = [ctypes.c_void_p]
+351
+352 bdio_get_rlen = bdio.bdio_get_rlen
+353 bdio_get_rlen.restype = ctypes.c_int
+354 bdio_get_rlen.argtypes = [ctypes.c_void_p]
+355
+356 bdio_get_ruinfo = bdio.bdio_get_ruinfo
+357 bdio_get_ruinfo.restype = ctypes.c_int
+358 bdio_get_ruinfo.argtypes = [ctypes.c_void_p]
+359
+360 bdio_read = bdio.bdio_read
+361 bdio_read.restype = ctypes.c_size_t
+362 bdio_read.argtypes = [ctypes.c_char_p, ctypes.c_size_t, ctypes.c_void_p]
+363
+364 bdio_read_f64 = bdio.bdio_read_f64
+365 bdio_read_f64.restype = ctypes.c_size_t
+366 bdio_read_f64.argtypes = [ctypes.c_void_p, ctypes.c_size_t, ctypes.c_void_p]
+367
+368 b_path = file_path.encode('utf-8')
+369 read = 'r'
+370 b_read = read.encode('utf-8')
+371 form = 'Generic Correlator Format 1.0'
+372 b_form = form.encode('utf-8')
373
-374 fbdio = bdio_open(ctypes.c_char_p(b_path), ctypes.c_char_p(b_read), ctypes.c_char_p(b_form))
-375
-376 print('Reading of bdio file started')
-377 while True:
-378 bdio_seek_record(fbdio)
-379 ruinfo = bdio_get_ruinfo(fbdio)
-380 if ruinfo < 0:
-381 # EOF reached
-382 break
-383 rlen = bdio_get_rlen(fbdio)
-384 if ruinfo == 5:
-385 d_buf = ctypes.c_double * (2 + d0 * d1 * 2)
-386 pd_buf = d_buf()
-387 ppd_buf = ctypes.c_void_p(ctypes.addressof(pd_buf))
-388 bdio_read_f64(ppd_buf, ctypes.c_size_t(rlen), ctypes.c_void_p(fbdio))
-389 if corr_type[corr_no] == 'complex':
-390 tmp_mean = np.mean(np.asarray(np.split(np.asarray(pd_buf[2 + 2 * d1:-2 * d1:2]), d0 - 2)), axis=1)
-391 else:
-392 tmp_mean = np.mean(np.asarray(np.split(np.asarray(pd_buf[2 + d1:-d0 * d1 - d1]), d0 - 2)), axis=1)
-393
-394 data[corr_no].append(tmp_mean)
-395 corr_no += 1
-396 else:
-397 alt_buf = ctypes.create_string_buffer(1024)
-398 palt_buf = ctypes.c_char_p(ctypes.addressof(alt_buf))
-399 iread = bdio_read(palt_buf, ctypes.c_size_t(rlen), ctypes.c_void_p(fbdio))
-400 if rlen != iread:
-401 print('Error')
-402 for i, item in enumerate(alt_buf):
-403 if item == b'\x00':
-404 alt_buf[i] = b' '
-405 tmp_string = (alt_buf[:].decode("utf-8")).rstrip()
-406 if ruinfo == 0:
-407 ensemble_name = _get_kwd(tmp_string, 'ENSEMBLE=')
-408 volume.append(int(_get_kwd(tmp_string, 'L0=')))
-409 volume.append(int(_get_kwd(tmp_string, 'L1=')))
-410 volume.append(int(_get_kwd(tmp_string, 'L2=')))
-411 volume.append(int(_get_kwd(tmp_string, 'L3=')))
-412 boundary_conditions.append(_get_kwd(tmp_string, 'BC0='))
-413 boundary_conditions.append(_get_kwd(tmp_string, 'BC1='))
-414 boundary_conditions.append(_get_kwd(tmp_string, 'BC2='))
-415 boundary_conditions.append(_get_kwd(tmp_string, 'BC3='))
-416
-417 if ruinfo == 1:
-418 corr_name.append(_get_corr_name(tmp_string, 'CORR_NAME='))
-419 corr_type.append(_get_kwd(tmp_string, 'DATATYPE='))
-420 corr_props.append([_get_kwd(tmp_string, 'PROP0='), _get_kwd(tmp_string, 'PROP1=')])
-421 if d0 == 0:
-422 d0 = int(_get_kwd(tmp_string, 'D0='))
-423 else:
-424 if d0 != int(_get_kwd(tmp_string, 'D0=')):
-425 print('Error: Varying number of time values')
-426 if d1 == 0:
-427 d1 = int(_get_kwd(tmp_string, 'D1='))
-428 else:
-429 if d1 != int(_get_kwd(tmp_string, 'D1=')):
-430 print('Error: Varying number of random sources')
-431 if ruinfo == 2:
-432 prop_kappa.append(_get_kwd(tmp_string, 'KAPPA='))
-433 prop_source.append(_get_kwd(tmp_string, 'x0='))
-434 if ruinfo == 4:
-435 cnfg_no = int(_get_kwd(tmp_string, 'CNFG_ID='))
-436 if stop:
-437 if cnfg_no > kwargs.get('stop'):
-438 break
-439 idl.append(cnfg_no)
-440 print('\r%s %i' % ('Reading configuration', cnfg_no), end='\r')
-441 if len(idl) == 1:
-442 no_corrs = len(corr_name)
-443 data = []
-444 for c in range(no_corrs):
-445 data.append([])
-446
-447 corr_no = 0
-448
-449 bdio_close(fbdio)
-450
-451 print('\nEnsemble: ', ensemble_name)
-452 if 'alternative_ensemble_name' in kwargs:
-453 ensemble_name = kwargs.get('alternative_ensemble_name')
-454 print('Ensemble name overwritten to', ensemble_name)
-455 print('Lattice volume: ', volume)
-456 print('Boundary conditions: ', boundary_conditions)
-457 print('Number of time values: ', d0)
-458 print('Number of random sources: ', d1)
-459 print('Number of corrs: ', len(corr_name))
-460 print('Number of configurations: ', len(idl))
+374 ensemble_name = ''
+375 volume = [] # lattice volume
+376 boundary_conditions = []
+377 corr_name = [] # Contains correlator names
+378 corr_type = [] # Contains correlator data type (important for reading out numerical data)
+379 corr_props = [] # Contanis propagator types (Component of corr_kappa)
+380 d0 = 0 # tvals
+381 d1 = 0 # nnoise
+382 prop_kappa = [] # Contains propagator kappas (Component of corr_kappa)
+383 prop_source = [] # Contains propagator source positions
+384 # Check noise type for multiple replica?
+385 corr_no = -1
+386 data = []
+387 idl = []
+388
+389 fbdio = bdio_open(ctypes.c_char_p(b_path), ctypes.c_char_p(b_read), ctypes.c_char_p(b_form))
+390
+391 print('Reading of bdio file started')
+392 while True:
+393 bdio_seek_record(fbdio)
+394 ruinfo = bdio_get_ruinfo(fbdio)
+395 if ruinfo < 0:
+396 # EOF reached
+397 break
+398 rlen = bdio_get_rlen(fbdio)
+399 if ruinfo == 5:
+400 d_buf = ctypes.c_double * (2 + d0 * d1 * 2)
+401 pd_buf = d_buf()
+402 ppd_buf = ctypes.c_void_p(ctypes.addressof(pd_buf))
+403 bdio_read_f64(ppd_buf, ctypes.c_size_t(rlen), ctypes.c_void_p(fbdio))
+404 if corr_type[corr_no] == 'complex':
+405 tmp_mean = np.mean(np.asarray(np.split(np.asarray(pd_buf[2 + 2 * d1:-2 * d1:2]), d0 - 2)), axis=1)
+406 else:
+407 tmp_mean = np.mean(np.asarray(np.split(np.asarray(pd_buf[2 + d1:-d0 * d1 - d1]), d0 - 2)), axis=1)
+408
+409 data[corr_no].append(tmp_mean)
+410 corr_no += 1
+411 else:
+412 alt_buf = ctypes.create_string_buffer(1024)
+413 palt_buf = ctypes.c_char_p(ctypes.addressof(alt_buf))
+414 iread = bdio_read(palt_buf, ctypes.c_size_t(rlen), ctypes.c_void_p(fbdio))
+415 if rlen != iread:
+416 print('Error')
+417 for i, item in enumerate(alt_buf):
+418 if item == b'\x00':
+419 alt_buf[i] = b' '
+420 tmp_string = (alt_buf[:].decode("utf-8")).rstrip()
+421 if ruinfo == 0:
+422 ensemble_name = _get_kwd(tmp_string, 'ENSEMBLE=')
+423 volume.append(int(_get_kwd(tmp_string, 'L0=')))
+424 volume.append(int(_get_kwd(tmp_string, 'L1=')))
+425 volume.append(int(_get_kwd(tmp_string, 'L2=')))
+426 volume.append(int(_get_kwd(tmp_string, 'L3=')))
+427 boundary_conditions.append(_get_kwd(tmp_string, 'BC0='))
+428 boundary_conditions.append(_get_kwd(tmp_string, 'BC1='))
+429 boundary_conditions.append(_get_kwd(tmp_string, 'BC2='))
+430 boundary_conditions.append(_get_kwd(tmp_string, 'BC3='))
+431
+432 if ruinfo == 1:
+433 corr_name.append(_get_corr_name(tmp_string, 'CORR_NAME='))
+434 corr_type.append(_get_kwd(tmp_string, 'DATATYPE='))
+435 corr_props.append([_get_kwd(tmp_string, 'PROP0='), _get_kwd(tmp_string, 'PROP1=')])
+436 if d0 == 0:
+437 d0 = int(_get_kwd(tmp_string, 'D0='))
+438 else:
+439 if d0 != int(_get_kwd(tmp_string, 'D0=')):
+440 print('Error: Varying number of time values')
+441 if d1 == 0:
+442 d1 = int(_get_kwd(tmp_string, 'D1='))
+443 else:
+444 if d1 != int(_get_kwd(tmp_string, 'D1=')):
+445 print('Error: Varying number of random sources')
+446 if ruinfo == 2:
+447 prop_kappa.append(_get_kwd(tmp_string, 'KAPPA='))
+448 prop_source.append(_get_kwd(tmp_string, 'x0='))
+449 if ruinfo == 4:
+450 cnfg_no = int(_get_kwd(tmp_string, 'CNFG_ID='))
+451 if stop:
+452 if cnfg_no > kwargs.get('stop'):
+453 break
+454 idl.append(cnfg_no)
+455 print('\r%s %i' % ('Reading configuration', cnfg_no), end='\r')
+456 if len(idl) == 1:
+457 no_corrs = len(corr_name)
+458 data = []
+459 for c in range(no_corrs):
+460 data.append([])
461
-462 corr_kappa = [] # Contains kappa values for both propagators of given correlation function
-463 corr_source = []
-464 for item in corr_props:
-465 corr_kappa.append([float(prop_kappa[int(item[0])]), float(prop_kappa[int(item[1])])])
-466 if prop_source[int(item[0])] != prop_source[int(item[1])]:
-467 raise Exception('Source position do not match for correlator' + str(item))
-468 else:
-469 corr_source.append(int(prop_source[int(item[0])]))
-470
-471 if stop is None:
-472 stop = idl[-1]
-473 idl_target = range(start, stop + 1, step)
-474
-475 if set(idl) != set(idl_target):
-476 try:
-477 indices = [idl.index(i) for i in idl_target]
-478 except ValueError as err:
-479 raise Exception('Configurations in file do no match target list!', err)
-480 else:
-481 indices = None
-482
-483 result = {}
-484 for c in range(no_corrs):
-485 tmp_corr = []
-486 tmp_data = np.asarray(data[c])
-487 for t in range(d0 - 2):
-488 if indices:
-489 deltas = [tmp_data[:, t][index] for index in indices]
-490 else:
-491 deltas = tmp_data[:, t]
-492 tmp_corr.append(Obs([deltas], [ensemble_name], idl=[idl_target]))
-493 result[(corr_name[c], corr_source[c]) + tuple(corr_kappa[c])] = tmp_corr
-494
-495 # Check that all data entries have the same number of configurations
-496 if len(set([o[0].N for o in list(result.values())])) != 1:
-497 raise Exception('Error: Not all correlators have the same number of configurations. bdio file is possibly corrupted.')
-498
-499 return result
+462 corr_no = 0
+463
+464 bdio_close(fbdio)
+465
+466 print('\nEnsemble: ', ensemble_name)
+467 if 'alternative_ensemble_name' in kwargs:
+468 ensemble_name = kwargs.get('alternative_ensemble_name')
+469 print('Ensemble name overwritten to', ensemble_name)
+470 print('Lattice volume: ', volume)
+471 print('Boundary conditions: ', boundary_conditions)
+472 print('Number of time values: ', d0)
+473 print('Number of random sources: ', d1)
+474 print('Number of corrs: ', len(corr_name))
+475 print('Number of configurations: ', len(idl))
+476
+477 corr_kappa = [] # Contains kappa values for both propagators of given correlation function
+478 corr_source = []
+479 for item in corr_props:
+480 corr_kappa.append([float(prop_kappa[int(item[0])]), float(prop_kappa[int(item[1])])])
+481 if prop_source[int(item[0])] != prop_source[int(item[1])]:
+482 raise Exception('Source position do not match for correlator' + str(item))
+483 else:
+484 corr_source.append(int(prop_source[int(item[0])]))
+485
+486 if stop is None:
+487 stop = idl[-1]
+488 idl_target = range(start, stop + 1, step)
+489
+490 if set(idl) != set(idl_target):
+491 try:
+492 indices = [idl.index(i) for i in idl_target]
+493 except ValueError as err:
+494 raise Exception('Configurations in file do no match target list!', err)
+495 else:
+496 indices = None
+497
+498 result = {}
+499 for c in range(no_corrs):
+500 tmp_corr = []
+501 tmp_data = np.asarray(data[c])
+502 for t in range(d0 - 2):
+503 if indices:
+504 deltas = [tmp_data[:, t][index] for index in indices]
+505 else:
+506 deltas = tmp_data[:, t]
+507 tmp_corr.append(Obs([deltas], [ensemble_name], idl=[idl_target]))
+508 result[(corr_name[c], corr_source[c]) + tuple(corr_kappa[c])] = tmp_corr
+509
+510 # Check that all data entries have the same number of configurations
+511 if len(set([o[0].N for o in list(result.values())])) != 1:
+512 raise Exception('Error: Not all correlators have the same number of configurations. bdio file is possibly corrupted.')
+513
+514 return result
@@ -1355,6 +1399,13 @@ Fixed step size between two measurements (default 1)
alternative_ensemble_name (str):
Manually overwrite ensemble name
+
+Returns
+
+
+- data (dict):
+Extracted meson data
+
@@ -1370,185 +1421,185 @@ Manually overwrite ensemble name
- 502def read_dSdm(file_path, bdio_path='./libbdio.so', **kwargs):
-503 """ Extract dSdm data from a bdio file and return it as a dictionary
-504
-505 The dictionary can be accessed with a tuple consisting of (type, kappa)
-506
-507 read_dSdm requires bdio to be compiled into a shared library. This can be achieved by
-508 adding the flag -fPIC to CC and changing the all target to
-509
-510 all: bdio.o $(LIBDIR)
-511 gcc -shared -Wl,-soname,libbdio.so -o $(BUILDDIR)/libbdio.so $(BUILDDIR)/bdio.o
-512 cp $(BUILDDIR)/libbdio.so $(LIBDIR)/
-513
-514 Parameters
-515 ----------
-516 file_path : str
-517 path to the bdio file
-518 bdio_path : str
-519 path to the shared bdio library libbdio.so (default ./libbdio.so)
-520 start : int
-521 The first configuration to be read (default 1)
-522 stop : int
-523 The last configuration to be read (default None)
-524 step : int
-525 Fixed step size between two measurements (default 1)
-526 alternative_ensemble_name : str
-527 Manually overwrite ensemble name
-528 """
-529
-530 start = kwargs.get('start', 1)
-531 stop = kwargs.get('stop', None)
-532 step = kwargs.get('step', 1)
-533
-534 bdio = ctypes.cdll.LoadLibrary(bdio_path)
-535
-536 bdio_open = bdio.bdio_open
-537 bdio_open.restype = ctypes.c_void_p
-538
-539 bdio_close = bdio.bdio_close
-540 bdio_close.restype = ctypes.c_int
-541 bdio_close.argtypes = [ctypes.c_void_p]
-542
-543 bdio_seek_record = bdio.bdio_seek_record
-544 bdio_seek_record.restype = ctypes.c_int
-545 bdio_seek_record.argtypes = [ctypes.c_void_p]
-546
-547 bdio_get_rlen = bdio.bdio_get_rlen
-548 bdio_get_rlen.restype = ctypes.c_int
-549 bdio_get_rlen.argtypes = [ctypes.c_void_p]
+ 517def read_dSdm(file_path, bdio_path='./libbdio.so', **kwargs):
+518 """ Extract dSdm data from a bdio file and return it as a dictionary
+519
+520 The dictionary can be accessed with a tuple consisting of (type, kappa)
+521
+522 read_dSdm requires bdio to be compiled into a shared library. This can be achieved by
+523 adding the flag -fPIC to CC and changing the all target to
+524
+525 all: bdio.o $(LIBDIR)
+526 gcc -shared -Wl,-soname,libbdio.so -o $(BUILDDIR)/libbdio.so $(BUILDDIR)/bdio.o
+527 cp $(BUILDDIR)/libbdio.so $(LIBDIR)/
+528
+529 Parameters
+530 ----------
+531 file_path : str
+532 path to the bdio file
+533 bdio_path : str
+534 path to the shared bdio library libbdio.so (default ./libbdio.so)
+535 start : int
+536 The first configuration to be read (default 1)
+537 stop : int
+538 The last configuration to be read (default None)
+539 step : int
+540 Fixed step size between two measurements (default 1)
+541 alternative_ensemble_name : str
+542 Manually overwrite ensemble name
+543 """
+544
+545 start = kwargs.get('start', 1)
+546 stop = kwargs.get('stop', None)
+547 step = kwargs.get('step', 1)
+548
+549 bdio = ctypes.cdll.LoadLibrary(bdio_path)
550
-551 bdio_get_ruinfo = bdio.bdio_get_ruinfo
-552 bdio_get_ruinfo.restype = ctypes.c_int
-553 bdio_get_ruinfo.argtypes = [ctypes.c_void_p]
-554
-555 bdio_read = bdio.bdio_read
-556 bdio_read.restype = ctypes.c_size_t
-557 bdio_read.argtypes = [ctypes.c_char_p, ctypes.c_size_t, ctypes.c_void_p]
-558
-559 bdio_read_f64 = bdio.bdio_read_f64
-560 bdio_read_f64.restype = ctypes.c_size_t
-561 bdio_read_f64.argtypes = [ctypes.c_void_p, ctypes.c_size_t, ctypes.c_void_p]
-562
-563 b_path = file_path.encode('utf-8')
-564 read = 'r'
-565 b_read = read.encode('utf-8')
-566 form = 'Generic Correlator Format 1.0'
-567 b_form = form.encode('utf-8')
-568
-569 ensemble_name = ''
-570 volume = [] # lattice volume
-571 boundary_conditions = []
-572 corr_name = [] # Contains correlator names
-573 corr_type = [] # Contains correlator data type (important for reading out numerical data)
-574 corr_props = [] # Contains propagator types (Component of corr_kappa)
-575 d0 = 0 # tvals
-576 # d1 = 0 # nnoise
-577 prop_kappa = [] # Contains propagator kappas (Component of corr_kappa)
-578 # Check noise type for multiple replica?
-579 corr_no = -1
-580 data = []
-581 idl = []
-582
-583 fbdio = bdio_open(ctypes.c_char_p(b_path), ctypes.c_char_p(b_read), ctypes.c_char_p(b_form))
-584
-585 print('Reading of bdio file started')
-586 while True:
-587 bdio_seek_record(fbdio)
-588 ruinfo = bdio_get_ruinfo(fbdio)
-589 if ruinfo < 0:
-590 # EOF reached
-591 break
-592 rlen = bdio_get_rlen(fbdio)
-593 if ruinfo == 5:
-594 d_buf = ctypes.c_double * (2 + d0)
-595 pd_buf = d_buf()
-596 ppd_buf = ctypes.c_void_p(ctypes.addressof(pd_buf))
-597 bdio_read_f64(ppd_buf, ctypes.c_size_t(rlen), ctypes.c_void_p(fbdio))
-598 tmp_mean = np.mean(np.asarray(pd_buf[2:]))
+551 bdio_open = bdio.bdio_open
+552 bdio_open.restype = ctypes.c_void_p
+553
+554 bdio_close = bdio.bdio_close
+555 bdio_close.restype = ctypes.c_int
+556 bdio_close.argtypes = [ctypes.c_void_p]
+557
+558 bdio_seek_record = bdio.bdio_seek_record
+559 bdio_seek_record.restype = ctypes.c_int
+560 bdio_seek_record.argtypes = [ctypes.c_void_p]
+561
+562 bdio_get_rlen = bdio.bdio_get_rlen
+563 bdio_get_rlen.restype = ctypes.c_int
+564 bdio_get_rlen.argtypes = [ctypes.c_void_p]
+565
+566 bdio_get_ruinfo = bdio.bdio_get_ruinfo
+567 bdio_get_ruinfo.restype = ctypes.c_int
+568 bdio_get_ruinfo.argtypes = [ctypes.c_void_p]
+569
+570 bdio_read = bdio.bdio_read
+571 bdio_read.restype = ctypes.c_size_t
+572 bdio_read.argtypes = [ctypes.c_char_p, ctypes.c_size_t, ctypes.c_void_p]
+573
+574 bdio_read_f64 = bdio.bdio_read_f64
+575 bdio_read_f64.restype = ctypes.c_size_t
+576 bdio_read_f64.argtypes = [ctypes.c_void_p, ctypes.c_size_t, ctypes.c_void_p]
+577
+578 b_path = file_path.encode('utf-8')
+579 read = 'r'
+580 b_read = read.encode('utf-8')
+581 form = 'Generic Correlator Format 1.0'
+582 b_form = form.encode('utf-8')
+583
+584 ensemble_name = ''
+585 volume = [] # lattice volume
+586 boundary_conditions = []
+587 corr_name = [] # Contains correlator names
+588 corr_type = [] # Contains correlator data type (important for reading out numerical data)
+589 corr_props = [] # Contains propagator types (Component of corr_kappa)
+590 d0 = 0 # tvals
+591 # d1 = 0 # nnoise
+592 prop_kappa = [] # Contains propagator kappas (Component of corr_kappa)
+593 # Check noise type for multiple replica?
+594 corr_no = -1
+595 data = []
+596 idl = []
+597
+598 fbdio = bdio_open(ctypes.c_char_p(b_path), ctypes.c_char_p(b_read), ctypes.c_char_p(b_form))
599
-600 data[corr_no].append(tmp_mean)
-601 corr_no += 1
-602 else:
-603 alt_buf = ctypes.create_string_buffer(1024)
-604 palt_buf = ctypes.c_char_p(ctypes.addressof(alt_buf))
-605 iread = bdio_read(palt_buf, ctypes.c_size_t(rlen), ctypes.c_void_p(fbdio))
-606 if rlen != iread:
-607 print('Error')
-608 for i, item in enumerate(alt_buf):
-609 if item == b'\x00':
-610 alt_buf[i] = b' '
-611 tmp_string = (alt_buf[:].decode("utf-8")).rstrip()
-612 if ruinfo == 0:
-613 creator = _get_kwd(tmp_string, 'CREATOR=')
-614 ensemble_name = _get_kwd(tmp_string, 'ENSEMBLE=')
-615 volume.append(int(_get_kwd(tmp_string, 'L0=')))
-616 volume.append(int(_get_kwd(tmp_string, 'L1=')))
-617 volume.append(int(_get_kwd(tmp_string, 'L2=')))
-618 volume.append(int(_get_kwd(tmp_string, 'L3=')))
-619 boundary_conditions.append(_get_kwd(tmp_string, 'BC0='))
-620 boundary_conditions.append(_get_kwd(tmp_string, 'BC1='))
-621 boundary_conditions.append(_get_kwd(tmp_string, 'BC2='))
-622 boundary_conditions.append(_get_kwd(tmp_string, 'BC3='))
-623
-624 if ruinfo == 1:
-625 corr_name.append(_get_corr_name(tmp_string, 'CORR_NAME='))
-626 corr_type.append(_get_kwd(tmp_string, 'DATATYPE='))
-627 corr_props.append(_get_kwd(tmp_string, 'PROP0='))
-628 if d0 == 0:
-629 d0 = int(_get_kwd(tmp_string, 'D0='))
-630 else:
-631 if d0 != int(_get_kwd(tmp_string, 'D0=')):
-632 print('Error: Varying number of time values')
-633 if ruinfo == 2:
-634 prop_kappa.append(_get_kwd(tmp_string, 'KAPPA='))
-635 if ruinfo == 4:
-636 cnfg_no = int(_get_kwd(tmp_string, 'CNFG_ID='))
-637 if stop:
-638 if cnfg_no > kwargs.get('stop'):
-639 break
-640 idl.append(cnfg_no)
-641 print('\r%s %i' % ('Reading configuration', cnfg_no), end='\r')
-642 if len(idl) == 1:
-643 no_corrs = len(corr_name)
-644 data = []
-645 for c in range(no_corrs):
-646 data.append([])
-647
-648 corr_no = 0
-649 bdio_close(fbdio)
-650
-651 print('\nCreator: ', creator)
-652 print('Ensemble: ', ensemble_name)
-653 print('Lattice volume: ', volume)
-654 print('Boundary conditions: ', boundary_conditions)
-655 print('Number of random sources: ', d0)
-656 print('Number of corrs: ', len(corr_name))
-657 print('Number of configurations: ', cnfg_no + 1)
-658
-659 corr_kappa = [] # Contains kappa values for both propagators of given correlation function
-660 for item in corr_props:
-661 corr_kappa.append(float(prop_kappa[int(item)]))
+600 print('Reading of bdio file started')
+601 while True:
+602 bdio_seek_record(fbdio)
+603 ruinfo = bdio_get_ruinfo(fbdio)
+604 if ruinfo < 0:
+605 # EOF reached
+606 break
+607 rlen = bdio_get_rlen(fbdio)
+608 if ruinfo == 5:
+609 d_buf = ctypes.c_double * (2 + d0)
+610 pd_buf = d_buf()
+611 ppd_buf = ctypes.c_void_p(ctypes.addressof(pd_buf))
+612 bdio_read_f64(ppd_buf, ctypes.c_size_t(rlen), ctypes.c_void_p(fbdio))
+613 tmp_mean = np.mean(np.asarray(pd_buf[2:]))
+614
+615 data[corr_no].append(tmp_mean)
+616 corr_no += 1
+617 else:
+618 alt_buf = ctypes.create_string_buffer(1024)
+619 palt_buf = ctypes.c_char_p(ctypes.addressof(alt_buf))
+620 iread = bdio_read(palt_buf, ctypes.c_size_t(rlen), ctypes.c_void_p(fbdio))
+621 if rlen != iread:
+622 print('Error')
+623 for i, item in enumerate(alt_buf):
+624 if item == b'\x00':
+625 alt_buf[i] = b' '
+626 tmp_string = (alt_buf[:].decode("utf-8")).rstrip()
+627 if ruinfo == 0:
+628 creator = _get_kwd(tmp_string, 'CREATOR=')
+629 ensemble_name = _get_kwd(tmp_string, 'ENSEMBLE=')
+630 volume.append(int(_get_kwd(tmp_string, 'L0=')))
+631 volume.append(int(_get_kwd(tmp_string, 'L1=')))
+632 volume.append(int(_get_kwd(tmp_string, 'L2=')))
+633 volume.append(int(_get_kwd(tmp_string, 'L3=')))
+634 boundary_conditions.append(_get_kwd(tmp_string, 'BC0='))
+635 boundary_conditions.append(_get_kwd(tmp_string, 'BC1='))
+636 boundary_conditions.append(_get_kwd(tmp_string, 'BC2='))
+637 boundary_conditions.append(_get_kwd(tmp_string, 'BC3='))
+638
+639 if ruinfo == 1:
+640 corr_name.append(_get_corr_name(tmp_string, 'CORR_NAME='))
+641 corr_type.append(_get_kwd(tmp_string, 'DATATYPE='))
+642 corr_props.append(_get_kwd(tmp_string, 'PROP0='))
+643 if d0 == 0:
+644 d0 = int(_get_kwd(tmp_string, 'D0='))
+645 else:
+646 if d0 != int(_get_kwd(tmp_string, 'D0=')):
+647 print('Error: Varying number of time values')
+648 if ruinfo == 2:
+649 prop_kappa.append(_get_kwd(tmp_string, 'KAPPA='))
+650 if ruinfo == 4:
+651 cnfg_no = int(_get_kwd(tmp_string, 'CNFG_ID='))
+652 if stop:
+653 if cnfg_no > kwargs.get('stop'):
+654 break
+655 idl.append(cnfg_no)
+656 print('\r%s %i' % ('Reading configuration', cnfg_no), end='\r')
+657 if len(idl) == 1:
+658 no_corrs = len(corr_name)
+659 data = []
+660 for c in range(no_corrs):
+661 data.append([])
662
-663 if stop is None:
-664 stop = idl[-1]
-665 idl_target = range(start, stop + 1, step)
-666 try:
-667 indices = [idl.index(i) for i in idl_target]
-668 except ValueError as err:
-669 raise Exception('Configurations in file do no match target list!', err)
-670
-671 result = {}
-672 for c in range(no_corrs):
-673 deltas = [np.asarray(data[c])[index] for index in indices]
-674 result[(corr_name[c], str(corr_kappa[c]))] = Obs([deltas], [ensemble_name], idl=[idl_target])
-675
-676 # Check that all data entries have the same number of configurations
-677 if len(set([o.N for o in list(result.values())])) != 1:
-678 raise Exception('Error: Not all correlators have the same number of configurations. bdio file is possibly corrupted.')
-679
-680 return result
+663 corr_no = 0
+664 bdio_close(fbdio)
+665
+666 print('\nCreator: ', creator)
+667 print('Ensemble: ', ensemble_name)
+668 print('Lattice volume: ', volume)
+669 print('Boundary conditions: ', boundary_conditions)
+670 print('Number of random sources: ', d0)
+671 print('Number of corrs: ', len(corr_name))
+672 print('Number of configurations: ', cnfg_no + 1)
+673
+674 corr_kappa = [] # Contains kappa values for both propagators of given correlation function
+675 for item in corr_props:
+676 corr_kappa.append(float(prop_kappa[int(item)]))
+677
+678 if stop is None:
+679 stop = idl[-1]
+680 idl_target = range(start, stop + 1, step)
+681 try:
+682 indices = [idl.index(i) for i in idl_target]
+683 except ValueError as err:
+684 raise Exception('Configurations in file do no match target list!', err)
+685
+686 result = {}
+687 for c in range(no_corrs):
+688 deltas = [np.asarray(data[c])[index] for index in indices]
+689 result[(corr_name[c], str(corr_kappa[c]))] = Obs([deltas], [ensemble_name], idl=[idl_target])
+690
+691 # Check that all data entries have the same number of configurations
+692 if len(set([o.N for o in list(result.values())])) != 1:
+693 raise Exception('Error: Not all correlators have the same number of configurations. bdio file is possibly corrupted.')
+694
+695 return result
diff --git a/docs/pyerrors/input/dobs.html b/docs/pyerrors/input/dobs.html
index 0531c98c..3c37ca33 100644
--- a/docs/pyerrors/input/dobs.html
+++ b/docs/pyerrors/input/dobs.html
@@ -202,776 +202,818 @@
106 A list of symbols that describe the observables to be written. May be empty.
107 enstag : str
108 Enstag that is written to pobs. If None, the ensemble name is used.
-109 """
-110
-111 od = {}
-112 ename = obsl[0].e_names[0]
-113 names = list(obsl[0].deltas.keys())
-114 nr = len(names)
-115 onames = [name.replace('|', '') for name in names]
-116 for o in obsl:
-117 if len(o.e_names) != 1:
-118 raise Exception('You try to export dobs to obs!')
-119 if o.e_names[0] != ename:
-120 raise Exception('You try to export dobs to obs!')
-121 if len(o.deltas.keys()) != nr:
-122 raise Exception('Incompatible obses in list')
-123 od['observables'] = {}
-124 od['observables']['schema'] = {'name': 'lattobs', 'version': '1.0'}
-125 od['observables']['origin'] = {
-126 'who': getpass.getuser(),
-127 'date': str(datetime.datetime.now())[:-7],
-128 'host': socket.gethostname(),
-129 'tool': {'name': 'pyerrors', 'version': pyerrorsversion.__version__}}
-130 od['observables']['pobs'] = {}
-131 pd = od['observables']['pobs']
-132 pd['spec'] = spec
-133 pd['origin'] = origin
-134 pd['name'] = name
-135 if enstag:
-136 if not isinstance(enstag, str):
-137 raise Exception('enstag has to be a string!')
-138 pd['enstag'] = enstag
-139 else:
-140 pd['enstag'] = ename
-141 pd['nr'] = '%d' % (nr)
-142 pd['array'] = []
-143 osymbol = 'cfg'
-144 if not isinstance(symbol, list):
-145 raise Exception('Symbol has to be a list!')
-146 if not (len(symbol) == 0 or len(symbol) == len(obsl)):
-147 raise Exception('Symbol has to be a list of lenght 0 or %d!' % (len(obsl)))
-148 for s in symbol:
-149 osymbol += ' %s' % s
-150 for r in range(nr):
-151 ad = {}
-152 ad['id'] = onames[r]
-153 Nconf = len(obsl[0].deltas[names[r]])
-154 layout = '%d i f%d' % (Nconf, len(obsl))
-155 ad['layout'] = layout
-156 ad['symbol'] = osymbol
-157 data = ''
-158 for c in range(Nconf):
-159 data += '%d ' % obsl[0].idl[names[r]][c]
-160 for o in obsl:
-161 num = o.deltas[names[r]][c] + o.r_values[names[r]]
-162 if num == 0:
-163 data += '0 '
-164 else:
-165 data += '%1.16e ' % (num)
-166 data += '\n'
-167 ad['#data'] = data
-168 pd['array'].append(ad)
-169
-170 rs = '<?xml version="1.0" encoding="utf-8"?>\n' + _dict_to_xmlstring_spaces(od)
-171 return rs
-172
-173
-174def write_pobs(obsl, fname, name, spec='', origin='', symbol=[], enstag=None, gz=True):
-175 """Export a list of Obs or structures containing Obs to a .xml.gz file
-176 according to the Zeuthen pobs format.
+109
+110 Returns
+111 -------
+112 xml_str : str
+113 XML formatted string of the input data
+114 """
+115
+116 od = {}
+117 ename = obsl[0].e_names[0]
+118 names = list(obsl[0].deltas.keys())
+119 nr = len(names)
+120 onames = [name.replace('|', '') for name in names]
+121 for o in obsl:
+122 if len(o.e_names) != 1:
+123 raise Exception('You try to export dobs to obs!')
+124 if o.e_names[0] != ename:
+125 raise Exception('You try to export dobs to obs!')
+126 if len(o.deltas.keys()) != nr:
+127 raise Exception('Incompatible obses in list')
+128 od['observables'] = {}
+129 od['observables']['schema'] = {'name': 'lattobs', 'version': '1.0'}
+130 od['observables']['origin'] = {
+131 'who': getpass.getuser(),
+132 'date': str(datetime.datetime.now())[:-7],
+133 'host': socket.gethostname(),
+134 'tool': {'name': 'pyerrors', 'version': pyerrorsversion.__version__}}
+135 od['observables']['pobs'] = {}
+136 pd = od['observables']['pobs']
+137 pd['spec'] = spec
+138 pd['origin'] = origin
+139 pd['name'] = name
+140 if enstag:
+141 if not isinstance(enstag, str):
+142 raise Exception('enstag has to be a string!')
+143 pd['enstag'] = enstag
+144 else:
+145 pd['enstag'] = ename
+146 pd['nr'] = '%d' % (nr)
+147 pd['array'] = []
+148 osymbol = 'cfg'
+149 if not isinstance(symbol, list):
+150 raise Exception('Symbol has to be a list!')
+151 if not (len(symbol) == 0 or len(symbol) == len(obsl)):
+152 raise Exception('Symbol has to be a list of lenght 0 or %d!' % (len(obsl)))
+153 for s in symbol:
+154 osymbol += ' %s' % s
+155 for r in range(nr):
+156 ad = {}
+157 ad['id'] = onames[r]
+158 Nconf = len(obsl[0].deltas[names[r]])
+159 layout = '%d i f%d' % (Nconf, len(obsl))
+160 ad['layout'] = layout
+161 ad['symbol'] = osymbol
+162 data = ''
+163 for c in range(Nconf):
+164 data += '%d ' % obsl[0].idl[names[r]][c]
+165 for o in obsl:
+166 num = o.deltas[names[r]][c] + o.r_values[names[r]]
+167 if num == 0:
+168 data += '0 '
+169 else:
+170 data += '%1.16e ' % (num)
+171 data += '\n'
+172 ad['#data'] = data
+173 pd['array'].append(ad)
+174
+175 rs = '<?xml version="1.0" encoding="utf-8"?>\n' + _dict_to_xmlstring_spaces(od)
+176 return rs
177
-178 Tags are not written or recovered automatically. The separator | is removed from the replica names.
-179
-180 Parameters
-181 ----------
-182 obsl : list
-183 List of Obs that will be exported.
-184 The Obs inside a structure have to be defined on the same ensemble.
-185 fname : str
-186 Filename of the output file.
-187 name : str
-188 The name of the observable.
-189 spec : str
-190 Optional string that describes the contents of the file.
-191 origin : str
-192 Specify where the data has its origin.
-193 symbol : list
-194 A list of symbols that describe the observables to be written. May be empty.
-195 enstag : str
-196 Enstag that is written to pobs. If None, the ensemble name is used.
-197 gz : bool
-198 If True, the output is a gzipped xml. If False, the output is an xml file.
-199 """
-200 pobsstring = create_pobs_string(obsl, name, spec, origin, symbol, enstag)
-201
-202 if not fname.endswith('.xml') and not fname.endswith('.gz'):
-203 fname += '.xml'
+178
+179def write_pobs(obsl, fname, name, spec='', origin='', symbol=[], enstag=None, gz=True):
+180 """Export a list of Obs or structures containing Obs to a .xml.gz file
+181 according to the Zeuthen pobs format.
+182
+183 Tags are not written or recovered automatically. The separator | is removed from the replica names.
+184
+185 Parameters
+186 ----------
+187 obsl : list
+188 List of Obs that will be exported.
+189 The Obs inside a structure have to be defined on the same ensemble.
+190 fname : str
+191 Filename of the output file.
+192 name : str
+193 The name of the observable.
+194 spec : str
+195 Optional string that describes the contents of the file.
+196 origin : str
+197 Specify where the data has its origin.
+198 symbol : list
+199 A list of symbols that describe the observables to be written. May be empty.
+200 enstag : str
+201 Enstag that is written to pobs. If None, the ensemble name is used.
+202 gz : bool
+203 If True, the output is a gzipped xml. If False, the output is an xml file.
204
-205 if gz:
-206 if not fname.endswith('.gz'):
-207 fname += '.gz'
-208
-209 fp = gzip.open(fname, 'wb')
-210 fp.write(pobsstring.encode('utf-8'))
-211 else:
-212 fp = open(fname, 'w', encoding='utf-8')
-213 fp.write(pobsstring)
-214 fp.close()
-215
-216
-217def _import_data(string):
-218 return json.loads("[" + ",".join(string.replace(' +', ' ').split()) + "]")
-219
-220
-221def _check(condition):
-222 if not condition:
-223 raise Exception("XML file format not supported")
+205 Returns
+206 -------
+207 None
+208 """
+209 pobsstring = create_pobs_string(obsl, name, spec, origin, symbol, enstag)
+210
+211 if not fname.endswith('.xml') and not fname.endswith('.gz'):
+212 fname += '.xml'
+213
+214 if gz:
+215 if not fname.endswith('.gz'):
+216 fname += '.gz'
+217
+218 fp = gzip.open(fname, 'wb')
+219 fp.write(pobsstring.encode('utf-8'))
+220 else:
+221 fp = open(fname, 'w', encoding='utf-8')
+222 fp.write(pobsstring)
+223 fp.close()
224
225
-226class _NoTagInDataError(Exception):
-227 """Raised when tag is not in data"""
-228 def __init__(self, tag):
-229 self.tag = tag
-230 super().__init__('Tag %s not in data!' % (self.tag))
-231
-232
-233def _find_tag(dat, tag):
-234 for i in range(len(dat)):
-235 if dat[i].tag == tag:
-236 return i
-237 raise _NoTagInDataError(tag)
-238
-239
-240def _import_array(arr):
-241 name = arr[_find_tag(arr, 'id')].text.strip()
-242 index = _find_tag(arr, 'layout')
-243 try:
-244 sindex = _find_tag(arr, 'symbol')
-245 except _NoTagInDataError:
-246 sindex = 0
-247 if sindex > index:
-248 tmp = _import_data(arr[sindex].tail)
-249 else:
-250 tmp = _import_data(arr[index].tail)
-251
-252 li = arr[index].text.strip()
-253 m = li.split()
-254 if m[1] == "i" and m[2][0] == "f":
-255 nc = int(m[0])
-256 na = int(m[2].lstrip('f'))
-257 _dat = []
-258 mask = []
-259 for a in range(na):
-260 mask += [a]
-261 _dat += [np.array(tmp[1 + a:: na + 1])]
-262 _check(len(tmp[0:: na + 1]) == nc)
-263 return [name, tmp[0:: na + 1], mask, _dat]
-264 elif m[1][0] == 'f' and len(m) < 3:
-265 sh = (int(m[0]), int(m[1].lstrip('f')))
-266 return np.reshape(tmp, sh)
-267 elif any(['f' in s for s in m]):
-268 for si in range(len(m)):
-269 if m[si] == 'f':
-270 break
-271 sh = [int(m[i]) for i in range(si)]
-272 return np.reshape(tmp, sh)
-273 else:
-274 print(name, m)
-275 _check(False)
-276
-277
-278def _import_rdata(rd):
-279 name, idx, mask, deltas = _import_array(rd)
-280 return deltas, name, idx
-281
-282
-283def _import_cdata(cd):
-284 _check(cd[0].tag == "id")
-285 _check(cd[1][0].text.strip() == "cov")
-286 cov = _import_array(cd[1])
-287 grad = _import_array(cd[2])
-288 return cd[0].text.strip(), cov, grad
-289
+226def _import_data(string):
+227 return json.loads("[" + ",".join(string.replace(' +', ' ').split()) + "]")
+228
+229
+230def _check(condition):
+231 if not condition:
+232 raise Exception("XML file format not supported")
+233
+234
+235class _NoTagInDataError(Exception):
+236 """Raised when tag is not in data"""
+237 def __init__(self, tag):
+238 self.tag = tag
+239 super().__init__('Tag %s not in data!' % (self.tag))
+240
+241
+242def _find_tag(dat, tag):
+243 for i in range(len(dat)):
+244 if dat[i].tag == tag:
+245 return i
+246 raise _NoTagInDataError(tag)
+247
+248
+249def _import_array(arr):
+250 name = arr[_find_tag(arr, 'id')].text.strip()
+251 index = _find_tag(arr, 'layout')
+252 try:
+253 sindex = _find_tag(arr, 'symbol')
+254 except _NoTagInDataError:
+255 sindex = 0
+256 if sindex > index:
+257 tmp = _import_data(arr[sindex].tail)
+258 else:
+259 tmp = _import_data(arr[index].tail)
+260
+261 li = arr[index].text.strip()
+262 m = li.split()
+263 if m[1] == "i" and m[2][0] == "f":
+264 nc = int(m[0])
+265 na = int(m[2].lstrip('f'))
+266 _dat = []
+267 mask = []
+268 for a in range(na):
+269 mask += [a]
+270 _dat += [np.array(tmp[1 + a:: na + 1])]
+271 _check(len(tmp[0:: na + 1]) == nc)
+272 return [name, tmp[0:: na + 1], mask, _dat]
+273 elif m[1][0] == 'f' and len(m) < 3:
+274 sh = (int(m[0]), int(m[1].lstrip('f')))
+275 return np.reshape(tmp, sh)
+276 elif any(['f' in s for s in m]):
+277 for si in range(len(m)):
+278 if m[si] == 'f':
+279 break
+280 sh = [int(m[i]) for i in range(si)]
+281 return np.reshape(tmp, sh)
+282 else:
+283 print(name, m)
+284 _check(False)
+285
+286
+287def _import_rdata(rd):
+288 name, idx, mask, deltas = _import_array(rd)
+289 return deltas, name, idx
290
-291def read_pobs(fname, full_output=False, gz=True, separator_insertion=None):
-292 """Import a list of Obs from an xml.gz file in the Zeuthen pobs format.
-293
-294 Tags are not written or recovered automatically.
-295
-296 Parameters
-297 ----------
-298 fname : str
-299 Filename of the input file.
-300 full_output : bool
-301 If True, a dict containing auxiliary information and the data is returned.
-302 If False, only the data is returned as list.
-303 separatior_insertion: str or int
-304 str: replace all occurences of "separator_insertion" within the replica names
-305 by "|%s" % (separator_insertion) when constructing the names of the replica.
-306 int: Insert the separator "|" at the position given by separator_insertion.
-307 None (default): Replica names remain unchanged.
-308 """
-309
-310 if not fname.endswith('.xml') and not fname.endswith('.gz'):
-311 fname += '.xml'
-312 if gz:
-313 if not fname.endswith('.gz'):
-314 fname += '.gz'
-315 with gzip.open(fname, 'r') as fin:
-316 content = fin.read()
-317 else:
-318 if fname.endswith('.gz'):
-319 warnings.warn("Trying to read from %s without unzipping!" % fname, UserWarning)
-320 with open(fname, 'r') as fin:
-321 content = fin.read()
-322
-323 # parse xml file content
-324 root = et.fromstring(content)
-325
-326 _check(root[2].tag == 'pobs')
-327 pobs = root[2]
-328
-329 version = root[0][1].text.strip()
-330
-331 _check(root[1].tag == 'origin')
-332 file_origin = _etree_to_dict(root[1])['origin']
-333
-334 deltas = []
-335 names = []
-336 idl = []
-337 for i in range(5, len(pobs)):
-338 delta, name, idx = _import_rdata(pobs[i])
-339 deltas.append(delta)
-340 if separator_insertion is None:
-341 pass
-342 elif isinstance(separator_insertion, int):
-343 name = name[:separator_insertion] + '|' + name[separator_insertion:]
-344 elif isinstance(separator_insertion, str):
-345 name = name.replace(separator_insertion, "|%s" % (separator_insertion))
-346 else:
-347 raise Exception("separator_insertion has to be string or int, is ", type(separator_insertion))
-348 names.append(name)
-349 idl.append(idx)
-350 res = [Obs([d[i] for d in deltas], names, idl=idl) for i in range(len(deltas[0]))]
-351
-352 descriptiond = {}
-353 for i in range(4):
-354 descriptiond[pobs[i].tag] = pobs[i].text.strip()
-355
-356 _check(pobs[4].tag == "nr")
-357
-358 _check(pobs[5].tag == 'array')
-359 if pobs[5][1].tag == 'symbol':
-360 symbol = pobs[5][1].text.strip()
-361 descriptiond['symbol'] = symbol
-362
-363 if full_output:
-364 retd = {}
-365 tool = file_origin.get('tool', None)
-366 if tool:
-367 program = tool['name'] + ' ' + tool['version']
-368 else:
-369 program = ''
-370 retd['program'] = program
-371 retd['version'] = version
-372 retd['who'] = file_origin['who']
-373 retd['date'] = file_origin['date']
-374 retd['host'] = file_origin['host']
-375 retd['description'] = descriptiond
-376 retd['obsdata'] = res
-377 return retd
-378 else:
-379 return res
-380
-381
-382# this is based on Mattia Bruno's implementation at https://github.com/mbruno46/pyobs/blob/master/pyobs/IO/xml.py
-383def import_dobs_string(content, noempty=False, full_output=False, separator_insertion=True):
-384 """Import a list of Obs from a string in the Zeuthen dobs format.
-385
-386 Tags are not written or recovered automatically.
-387
-388 Parameters
-389 ----------
-390 content : str
-391 XML string containing the data
-392 noemtpy : bool
-393 If True, ensembles with no contribution to the Obs are not included.
-394 If False, ensembles are included as written in the file, possibly with vanishing entries.
-395 full_output : bool
-396 If True, a dict containing auxiliary information and the data is returned.
-397 If False, only the data is returned as list.
-398 separatior_insertion: str, int or bool
-399 str: replace all occurences of "separator_insertion" within the replica names
-400 by "|%s" % (separator_insertion) when constructing the names of the replica.
-401 int: Insert the separator "|" at the position given by separator_insertion.
-402 True (default): separator "|" is inserted after len(ensname), assuming that the
-403 ensemble name is a prefix to the replica name.
-404 None or False: No separator is inserted.
-405 """
-406
-407 root = et.fromstring(content)
-408
-409 _check(root.tag == 'OBSERVABLES')
-410 _check(root[0].tag == 'SCHEMA')
-411 version = root[0][1].text.strip()
-412
-413 _check(root[1].tag == 'origin')
-414 file_origin = _etree_to_dict(root[1])['origin']
-415
-416 _check(root[2].tag == 'dobs')
-417
-418 dobs = root[2]
-419
-420 descriptiond = {}
-421 for i in range(3):
-422 descriptiond[dobs[i].tag] = dobs[i].text.strip()
-423
-424 _check(dobs[3].tag == 'array')
-425
-426 symbol = []
-427 if dobs[3][1].tag == 'symbol':
-428 symbol = dobs[3][1].text.strip()
-429 descriptiond['symbol'] = symbol
-430 mean = _import_array(dobs[3])[0]
+291
+292def _import_cdata(cd):
+293 _check(cd[0].tag == "id")
+294 _check(cd[1][0].text.strip() == "cov")
+295 cov = _import_array(cd[1])
+296 grad = _import_array(cd[2])
+297 return cd[0].text.strip(), cov, grad
+298
+299
+300def read_pobs(fname, full_output=False, gz=True, separator_insertion=None):
+301 """Import a list of Obs from an xml.gz file in the Zeuthen pobs format.
+302
+303 Tags are not written or recovered automatically.
+304
+305 Parameters
+306 ----------
+307 fname : str
+308 Filename of the input file.
+309 full_output : bool
+310 If True, a dict containing auxiliary information and the data is returned.
+311 If False, only the data is returned as list.
+312 separatior_insertion: str or int
+313 str: replace all occurences of "separator_insertion" within the replica names
+314 by "|%s" % (separator_insertion) when constructing the names of the replica.
+315 int: Insert the separator "|" at the position given by separator_insertion.
+316 None (default): Replica names remain unchanged.
+317
+318 Returns
+319 -------
+320 res : list[Obs]
+321 Imported data
+322 or
+323 res : dict
+324 Imported data and meta-data
+325 """
+326
+327 if not fname.endswith('.xml') and not fname.endswith('.gz'):
+328 fname += '.xml'
+329 if gz:
+330 if not fname.endswith('.gz'):
+331 fname += '.gz'
+332 with gzip.open(fname, 'r') as fin:
+333 content = fin.read()
+334 else:
+335 if fname.endswith('.gz'):
+336 warnings.warn("Trying to read from %s without unzipping!" % fname, UserWarning)
+337 with open(fname, 'r') as fin:
+338 content = fin.read()
+339
+340 # parse xml file content
+341 root = et.fromstring(content)
+342
+343 _check(root[2].tag == 'pobs')
+344 pobs = root[2]
+345
+346 version = root[0][1].text.strip()
+347
+348 _check(root[1].tag == 'origin')
+349 file_origin = _etree_to_dict(root[1])['origin']
+350
+351 deltas = []
+352 names = []
+353 idl = []
+354 for i in range(5, len(pobs)):
+355 delta, name, idx = _import_rdata(pobs[i])
+356 deltas.append(delta)
+357 if separator_insertion is None:
+358 pass
+359 elif isinstance(separator_insertion, int):
+360 name = name[:separator_insertion] + '|' + name[separator_insertion:]
+361 elif isinstance(separator_insertion, str):
+362 name = name.replace(separator_insertion, "|%s" % (separator_insertion))
+363 else:
+364 raise Exception("separator_insertion has to be string or int, is ", type(separator_insertion))
+365 names.append(name)
+366 idl.append(idx)
+367 res = [Obs([d[i] for d in deltas], names, idl=idl) for i in range(len(deltas[0]))]
+368
+369 descriptiond = {}
+370 for i in range(4):
+371 descriptiond[pobs[i].tag] = pobs[i].text.strip()
+372
+373 _check(pobs[4].tag == "nr")
+374
+375 _check(pobs[5].tag == 'array')
+376 if pobs[5][1].tag == 'symbol':
+377 symbol = pobs[5][1].text.strip()
+378 descriptiond['symbol'] = symbol
+379
+380 if full_output:
+381 retd = {}
+382 tool = file_origin.get('tool', None)
+383 if tool:
+384 program = tool['name'] + ' ' + tool['version']
+385 else:
+386 program = ''
+387 retd['program'] = program
+388 retd['version'] = version
+389 retd['who'] = file_origin['who']
+390 retd['date'] = file_origin['date']
+391 retd['host'] = file_origin['host']
+392 retd['description'] = descriptiond
+393 retd['obsdata'] = res
+394 return retd
+395 else:
+396 return res
+397
+398
+399# this is based on Mattia Bruno's implementation at https://github.com/mbruno46/pyobs/blob/master/pyobs/IO/xml.py
+400def import_dobs_string(content, noempty=False, full_output=False, separator_insertion=True):
+401 """Import a list of Obs from a string in the Zeuthen dobs format.
+402
+403 Tags are not written or recovered automatically.
+404
+405 Parameters
+406 ----------
+407 content : str
+408 XML string containing the data
+409 noemtpy : bool
+410 If True, ensembles with no contribution to the Obs are not included.
+411 If False, ensembles are included as written in the file, possibly with vanishing entries.
+412 full_output : bool
+413 If True, a dict containing auxiliary information and the data is returned.
+414 If False, only the data is returned as list.
+415 separatior_insertion: str, int or bool
+416 str: replace all occurences of "separator_insertion" within the replica names
+417 by "|%s" % (separator_insertion) when constructing the names of the replica.
+418 int: Insert the separator "|" at the position given by separator_insertion.
+419 True (default): separator "|" is inserted after len(ensname), assuming that the
+420 ensemble name is a prefix to the replica name.
+421 None or False: No separator is inserted.
+422
+423 Returns
+424 -------
+425 res : list[Obs]
+426 Imported data
+427 or
+428 res : dict
+429 Imported data and meta-data
+430 """
431
-432 _check(dobs[4].tag == "ne")
-433 ne = int(dobs[4].text.strip())
-434 _check(dobs[5].tag == "nc")
-435 nc = int(dobs[5].text.strip())
-436
-437 idld = {}
-438 deltad = {}
-439 covd = {}
-440 gradd = {}
-441 names = []
-442 e_names = []
-443 enstags = {}
-444 for k in range(6, len(list(dobs))):
-445 if dobs[k].tag == "edata":
-446 _check(dobs[k][0].tag == "enstag")
-447 ename = dobs[k][0].text.strip()
-448 e_names.append(ename)
-449 _check(dobs[k][1].tag == "nr")
-450 R = int(dobs[k][1].text.strip())
-451 for i in range(2, 2 + R):
-452 deltas, rname, idx = _import_rdata(dobs[k][i])
-453 if separator_insertion is None or False:
-454 pass
-455 elif separator_insertion is True:
-456 if rname.startswith(ename):
-457 rname = rname[:len(ename)] + '|' + rname[len(ename):]
-458 elif isinstance(separator_insertion, int):
-459 rname = rname[:separator_insertion] + '|' + rname[separator_insertion:]
-460 elif isinstance(separator_insertion, str):
-461 rname = rname.replace(separator_insertion, "|%s" % (separator_insertion))
-462 else:
-463 raise Exception("separator_insertion has to be string or int, is ", type(separator_insertion))
-464 if '|' in rname:
-465 new_ename = rname[:rname.index('|')]
-466 else:
-467 new_ename = ename
-468 enstags[new_ename] = ename
-469 idld[rname] = idx
-470 deltad[rname] = deltas
-471 names.append(rname)
-472 elif dobs[k].tag == "cdata":
-473 cname, cov, grad = _import_cdata(dobs[k])
-474 covd[cname] = cov
-475 if grad.shape[1] == 1:
-476 gradd[cname] = [grad for i in range(len(mean))]
-477 else:
-478 gradd[cname] = grad.T
-479 else:
-480 _check(False)
-481 names = list(set(names))
-482
-483 for name in names:
-484 for i in range(len(deltad[name])):
-485 deltad[name][i] = np.array(deltad[name][i]) + mean[i]
-486
-487 res = []
-488 for i in range(len(mean)):
-489 deltas = []
-490 idl = []
-491 obs_names = []
-492 for name in names:
-493 h = np.unique(deltad[name][i])
-494 if len(h) == 1 and np.all(h == mean[i]) and noempty:
-495 continue
-496 deltas.append(deltad[name][i])
-497 obs_names.append(name)
-498 idl.append(idld[name])
-499 res.append(Obs(deltas, obs_names, idl=idl))
-500 res[-1]._value = mean[i]
-501 _check(len(e_names) == ne)
-502
-503 cnames = list(covd.keys())
-504 for i in range(len(res)):
-505 new_covobs = {name: Covobs(0, covd[name], name, grad=gradd[name][i]) for name in cnames}
-506 if noempty:
-507 for name in cnames:
-508 if np.all(new_covobs[name].grad == 0):
-509 del new_covobs[name]
-510 cnames_loc = list(new_covobs.keys())
-511 else:
-512 cnames_loc = cnames
-513 for name in cnames_loc:
-514 res[i].names.append(name)
-515 res[i].shape[name] = 1
-516 res[i].idl[name] = []
-517 res[i]._covobs = new_covobs
-518
-519 if symbol:
-520 for i in range(len(res)):
-521 res[i].tag = symbol[i]
-522 if res[i].tag == 'None':
-523 res[i].tag = None
-524 if not noempty:
-525 _check(len(res[0].covobs.keys()) == nc)
-526 if full_output:
-527 retd = {}
-528 tool = file_origin.get('tool', None)
-529 if tool:
-530 program = tool['name'] + ' ' + tool['version']
-531 else:
-532 program = ''
-533 retd['program'] = program
-534 retd['version'] = version
-535 retd['who'] = file_origin['who']
-536 retd['date'] = file_origin['date']
-537 retd['host'] = file_origin['host']
-538 retd['description'] = descriptiond
-539 retd['enstags'] = enstags
-540 retd['obsdata'] = res
-541 return retd
-542 else:
-543 return res
-544
-545
-546def read_dobs(fname, noempty=False, full_output=False, gz=True, separator_insertion=True):
-547 """Import a list of Obs from an xml.gz file in the Zeuthen dobs format.
-548
-549 Tags are not written or recovered automatically.
-550
-551 Parameters
-552 ----------
-553 fname : str
-554 Filename of the input file.
-555 noemtpy : bool
-556 If True, ensembles with no contribution to the Obs are not included.
-557 If False, ensembles are included as written in the file.
-558 full_output : bool
-559 If True, a dict containing auxiliary information and the data is returned.
-560 If False, only the data is returned as list.
-561 gz : bool
-562 If True, assumes that data is gzipped. If False, assumes XML file.
-563 separatior_insertion: str, int or bool
-564 str: replace all occurences of "separator_insertion" within the replica names
-565 by "|%s" % (separator_insertion) when constructing the names of the replica.
-566 int: Insert the separator "|" at the position given by separator_insertion.
-567 True (default): separator "|" is inserted after len(ensname), assuming that the
-568 ensemble name is a prefix to the replica name.
-569 None or False: No separator is inserted.
-570 """
-571
-572 if not fname.endswith('.xml') and not fname.endswith('.gz'):
-573 fname += '.xml'
-574 if gz:
-575 if not fname.endswith('.gz'):
-576 fname += '.gz'
-577 with gzip.open(fname, 'r') as fin:
-578 content = fin.read()
-579 else:
-580 if fname.endswith('.gz'):
-581 warnings.warn("Trying to read from %s without unzipping!" % fname, UserWarning)
-582 with open(fname, 'r') as fin:
-583 content = fin.read()
-584
-585 return import_dobs_string(content, noempty, full_output, separator_insertion=separator_insertion)
-586
-587
-588def _dobsdict_to_xmlstring(d):
-589 if isinstance(d, dict):
-590 iters = ''
-591 for k in d:
-592 if k.startswith('#value'):
-593 for li in d[k]:
-594 iters += li
-595 return iters + '\n'
-596 elif k.startswith('#'):
-597 for li in d[k]:
-598 iters += li
-599 iters = '<array>\n' + iters + '<%sarray>\n' % ('/')
-600 return iters
-601 if isinstance(d[k], dict):
-602 iters += '<%s>\n' % (k) + _dobsdict_to_xmlstring(d[k]) + '<%s%s>\n' % ('/', k)
-603 elif isinstance(d[k], str):
-604 if len(d[k]) > 100:
-605 iters += '<%s>\n ' % (k) + d[k] + ' \n<%s%s>\n' % ('/', k)
-606 else:
-607 iters += '<%s> ' % (k) + d[k] + ' <%s%s>\n' % ('/', k)
-608 elif isinstance(d[k], list):
-609 tmps = ''
-610 if k in ['edata', 'cdata']:
-611 for i in range(len(d[k])):
-612 tmps += '<%s>\n' % (k) + _dobsdict_to_xmlstring(d[k][i]) + '</%s>\n' % (k)
-613 else:
-614 for i in range(len(d[k])):
-615 tmps += _dobsdict_to_xmlstring(d[k][i])
-616 iters += tmps
-617 elif isinstance(d[k], (int, float)):
-618 iters += '<%s> ' % (k) + str(d[k]) + ' <%s%s>\n' % ('/', k)
-619 elif not d[k]:
-620 return '\n'
-621 else:
-622 raise Exception('Type', type(d[k]), 'not supported in export!')
-623 else:
-624 raise Exception('Type', type(d), 'not supported in export!')
-625 return iters
-626
-627
-628def _dobsdict_to_xmlstring_spaces(d, space=' '):
-629 s = _dobsdict_to_xmlstring(d)
-630 o = ''
-631 c = 0
-632 cm = False
-633 for li in s.split('\n'):
-634 if li.startswith('<%s' % ('/')):
-635 c -= 1
-636 cm = True
-637 for i in range(c):
-638 o += space
-639 o += li + '\n'
-640 if li.startswith('<') and not cm:
-641 if not '<%s' % ('/') in li:
-642 c += 1
-643 cm = False
-644 return o
-645
-646
-647def create_dobs_string(obsl, name, spec='dobs v1.0', origin='', symbol=[], who=None, enstags=None):
-648 """Generate the string for the export of a list of Obs or structures containing Obs
-649 to a .xml.gz file according to the Zeuthen dobs format.
-650
-651 Tags are not written or recovered automatically. The separator |is removed from the replica names.
-652
-653 Parameters
-654 ----------
-655 obsl : list
-656 List of Obs that will be exported.
-657 The Obs inside a structure do not have to be defined on the same set of configurations,
-658 but the storage requirement is increased, if this is not the case.
-659 name : str
-660 The name of the observable.
-661 spec : str
-662 Optional string that describes the contents of the file.
-663 origin : str
-664 Specify where the data has its origin.
-665 symbol : list
-666 A list of symbols that describe the observables to be written. May be empty.
-667 who : str
-668 Provide the name of the person that exports the data.
-669 enstags : dict
-670 Provide alternative enstag for ensembles in the form enstags = {ename: enstag}
-671 Otherwise, the ensemble name is used.
-672 """
-673 if enstags is None:
-674 enstags = {}
-675 od = {}
-676 r_names = []
-677 for o in obsl:
-678 r_names += [name for name in o.names if name.split('|')[0] in o.mc_names]
-679 r_names = sorted(set(r_names))
-680 mc_names = sorted(set([n.split('|')[0] for n in r_names]))
-681 for tmpname in mc_names:
-682 if tmpname not in enstags:
-683 enstags[tmpname] = tmpname
-684 ne = len(set(mc_names))
-685 cov_names = []
-686 for o in obsl:
-687 cov_names += list(o.cov_names)
-688 cov_names = sorted(set(cov_names))
-689 nc = len(set(cov_names))
-690 od['OBSERVABLES'] = {}
-691 od['OBSERVABLES']['SCHEMA'] = {'NAME': 'lattobs', 'VERSION': '1.0'}
-692 if who is None:
-693 who = getpass.getuser()
-694 od['OBSERVABLES']['origin'] = {
-695 'who': who,
-696 'date': str(datetime.datetime.now())[:-7],
-697 'host': socket.gethostname(),
-698 'tool': {'name': 'pyerrors', 'version': pyerrorsversion.__version__}}
-699 od['OBSERVABLES']['dobs'] = {}
-700 pd = od['OBSERVABLES']['dobs']
-701 pd['spec'] = spec
-702 pd['origin'] = origin
-703 pd['name'] = name
-704 pd['array'] = {}
-705 pd['array']['id'] = 'val'
-706 pd['array']['layout'] = '1 f%d' % (len(obsl))
-707 osymbol = ''
-708 if symbol:
-709 if not isinstance(symbol, list):
-710 raise Exception('Symbol has to be a list!')
-711 if not (len(symbol) == 0 or len(symbol) == len(obsl)):
-712 raise Exception('Symbol has to be a list of lenght 0 or %d!' % (len(obsl)))
-713 osymbol = symbol[0]
-714 for s in symbol[1:]:
-715 osymbol += ' %s' % s
-716 pd['array']['symbol'] = osymbol
-717
-718 pd['array']['#values'] = [' '.join(['%1.16e' % o.value for o in obsl])]
-719 pd['ne'] = '%d' % (ne)
-720 pd['nc'] = '%d' % (nc)
-721 pd['edata'] = []
-722 for name in mc_names:
-723 ed = {}
-724 ed['enstag'] = enstags[name]
-725 onames = sorted([n for n in r_names if (n.startswith(name + '|') or n == name)])
-726 nr = len(onames)
-727 ed['nr'] = nr
-728 ed[''] = []
-729
-730 for r in range(nr):
-731 ad = {}
-732 repname = onames[r]
-733 ad['id'] = repname.replace('|', '')
-734 idx = _merge_idx([o.idl.get(repname, []) for o in obsl])
-735 Nconf = len(idx)
-736 layout = '%d i f%d' % (Nconf, len(obsl))
-737 ad['layout'] = layout
-738 data = ''
-739 counters = [0 for o in obsl]
-740 offsets = [o.r_values[repname] - o.value if repname in o.r_values else 0 for o in obsl]
-741 for ci in idx:
-742 data += '%d ' % ci
-743 for oi in range(len(obsl)):
-744 o = obsl[oi]
-745 if repname in o.idl:
-746 if counters[oi] < 0:
-747 num = offsets[oi]
-748 if num == 0:
-749 data += '0 '
-750 else:
-751 data += '%1.16e ' % (num)
-752 continue
-753 if o.idl[repname][counters[oi]] == ci:
-754 num = o.deltas[repname][counters[oi]] + offsets[oi]
-755 if num == 0:
-756 data += '0 '
-757 else:
-758 data += '%1.16e ' % (num)
-759 counters[oi] += 1
-760 if counters[oi] >= len(o.idl[repname]):
-761 counters[oi] = -1
-762 else:
-763 num = offsets[oi]
-764 if num == 0:
-765 data += '0 '
-766 else:
-767 data += '%1.16e ' % (num)
-768 else:
-769 data += '0 '
-770 data += '\n'
-771 ad['#data'] = data
-772 ed[''].append(ad)
-773 pd['edata'].append(ed)
-774
-775 allcov = {}
-776 for o in obsl:
-777 for cname in o.cov_names:
-778 if cname in allcov:
-779 if not np.array_equal(allcov[cname], o.covobs[cname].cov):
-780 raise Exception('Inconsistent covariance matrices for %s!' % (cname))
-781 else:
-782 allcov[cname] = o.covobs[cname].cov
-783 pd['cdata'] = []
-784 for cname in cov_names:
-785 cd = {}
-786 cd['id'] = cname
-787
-788 covd = {'id': 'cov'}
-789 if allcov[cname].shape == ():
-790 ncov = 1
-791 covd['layout'] = '1 1 f'
-792 covd['#data'] = '%1.14e' % (allcov[cname])
-793 else:
-794 shape = allcov[cname].shape
-795 assert (shape[0] == shape[1])
-796 ncov = shape[0]
-797 covd['layout'] = '%d %d f' % (ncov, ncov)
-798 ds = ''
-799 for i in range(ncov):
-800 for j in range(ncov):
-801 val = allcov[cname][i][j]
-802 if val == 0:
-803 ds += '0 '
-804 else:
-805 ds += '%1.14e ' % (val)
-806 ds += '\n'
-807 covd['#data'] = ds
-808
-809 gradd = {'id': 'grad'}
-810 gradd['layout'] = '%d f%d' % (ncov, len(obsl))
-811 ds = ''
-812 for i in range(ncov):
-813 for o in obsl:
-814 if cname in o.covobs:
-815 val = o.covobs[cname].grad[i]
-816 if val != 0:
-817 ds += '%1.14e ' % (val)
-818 else:
-819 ds += '0 '
-820 else:
-821 ds += '0 '
-822 gradd['#data'] = ds
-823 cd['array'] = [covd, gradd]
-824 pd['cdata'].append(cd)
+432 root = et.fromstring(content)
+433
+434 _check(root.tag == 'OBSERVABLES')
+435 _check(root[0].tag == 'SCHEMA')
+436 version = root[0][1].text.strip()
+437
+438 _check(root[1].tag == 'origin')
+439 file_origin = _etree_to_dict(root[1])['origin']
+440
+441 _check(root[2].tag == 'dobs')
+442
+443 dobs = root[2]
+444
+445 descriptiond = {}
+446 for i in range(3):
+447 descriptiond[dobs[i].tag] = dobs[i].text.strip()
+448
+449 _check(dobs[3].tag == 'array')
+450
+451 symbol = []
+452 if dobs[3][1].tag == 'symbol':
+453 symbol = dobs[3][1].text.strip()
+454 descriptiond['symbol'] = symbol
+455 mean = _import_array(dobs[3])[0]
+456
+457 _check(dobs[4].tag == "ne")
+458 ne = int(dobs[4].text.strip())
+459 _check(dobs[5].tag == "nc")
+460 nc = int(dobs[5].text.strip())
+461
+462 idld = {}
+463 deltad = {}
+464 covd = {}
+465 gradd = {}
+466 names = []
+467 e_names = []
+468 enstags = {}
+469 for k in range(6, len(list(dobs))):
+470 if dobs[k].tag == "edata":
+471 _check(dobs[k][0].tag == "enstag")
+472 ename = dobs[k][0].text.strip()
+473 e_names.append(ename)
+474 _check(dobs[k][1].tag == "nr")
+475 R = int(dobs[k][1].text.strip())
+476 for i in range(2, 2 + R):
+477 deltas, rname, idx = _import_rdata(dobs[k][i])
+478 if separator_insertion is None or False:
+479 pass
+480 elif separator_insertion is True:
+481 if rname.startswith(ename):
+482 rname = rname[:len(ename)] + '|' + rname[len(ename):]
+483 elif isinstance(separator_insertion, int):
+484 rname = rname[:separator_insertion] + '|' + rname[separator_insertion:]
+485 elif isinstance(separator_insertion, str):
+486 rname = rname.replace(separator_insertion, "|%s" % (separator_insertion))
+487 else:
+488 raise Exception("separator_insertion has to be string or int, is ", type(separator_insertion))
+489 if '|' in rname:
+490 new_ename = rname[:rname.index('|')]
+491 else:
+492 new_ename = ename
+493 enstags[new_ename] = ename
+494 idld[rname] = idx
+495 deltad[rname] = deltas
+496 names.append(rname)
+497 elif dobs[k].tag == "cdata":
+498 cname, cov, grad = _import_cdata(dobs[k])
+499 covd[cname] = cov
+500 if grad.shape[1] == 1:
+501 gradd[cname] = [grad for i in range(len(mean))]
+502 else:
+503 gradd[cname] = grad.T
+504 else:
+505 _check(False)
+506 names = list(set(names))
+507
+508 for name in names:
+509 for i in range(len(deltad[name])):
+510 deltad[name][i] = np.array(deltad[name][i]) + mean[i]
+511
+512 res = []
+513 for i in range(len(mean)):
+514 deltas = []
+515 idl = []
+516 obs_names = []
+517 for name in names:
+518 h = np.unique(deltad[name][i])
+519 if len(h) == 1 and np.all(h == mean[i]) and noempty:
+520 continue
+521 deltas.append(deltad[name][i])
+522 obs_names.append(name)
+523 idl.append(idld[name])
+524 res.append(Obs(deltas, obs_names, idl=idl))
+525 res[-1]._value = mean[i]
+526 _check(len(e_names) == ne)
+527
+528 cnames = list(covd.keys())
+529 for i in range(len(res)):
+530 new_covobs = {name: Covobs(0, covd[name], name, grad=gradd[name][i]) for name in cnames}
+531 if noempty:
+532 for name in cnames:
+533 if np.all(new_covobs[name].grad == 0):
+534 del new_covobs[name]
+535 cnames_loc = list(new_covobs.keys())
+536 else:
+537 cnames_loc = cnames
+538 for name in cnames_loc:
+539 res[i].names.append(name)
+540 res[i].shape[name] = 1
+541 res[i].idl[name] = []
+542 res[i]._covobs = new_covobs
+543
+544 if symbol:
+545 for i in range(len(res)):
+546 res[i].tag = symbol[i]
+547 if res[i].tag == 'None':
+548 res[i].tag = None
+549 if not noempty:
+550 _check(len(res[0].covobs.keys()) == nc)
+551 if full_output:
+552 retd = {}
+553 tool = file_origin.get('tool', None)
+554 if tool:
+555 program = tool['name'] + ' ' + tool['version']
+556 else:
+557 program = ''
+558 retd['program'] = program
+559 retd['version'] = version
+560 retd['who'] = file_origin['who']
+561 retd['date'] = file_origin['date']
+562 retd['host'] = file_origin['host']
+563 retd['description'] = descriptiond
+564 retd['enstags'] = enstags
+565 retd['obsdata'] = res
+566 return retd
+567 else:
+568 return res
+569
+570
+571def read_dobs(fname, noempty=False, full_output=False, gz=True, separator_insertion=True):
+572 """Import a list of Obs from an xml.gz file in the Zeuthen dobs format.
+573
+574 Tags are not written or recovered automatically.
+575
+576 Parameters
+577 ----------
+578 fname : str
+579 Filename of the input file.
+580 noemtpy : bool
+581 If True, ensembles with no contribution to the Obs are not included.
+582 If False, ensembles are included as written in the file.
+583 full_output : bool
+584 If True, a dict containing auxiliary information and the data is returned.
+585 If False, only the data is returned as list.
+586 gz : bool
+587 If True, assumes that data is gzipped. If False, assumes XML file.
+588 separatior_insertion: str, int or bool
+589 str: replace all occurences of "separator_insertion" within the replica names
+590 by "|%s" % (separator_insertion) when constructing the names of the replica.
+591 int: Insert the separator "|" at the position given by separator_insertion.
+592 True (default): separator "|" is inserted after len(ensname), assuming that the
+593 ensemble name is a prefix to the replica name.
+594 None or False: No separator is inserted.
+595
+596 Returns
+597 -------
+598 res : list[Obs]
+599 Imported data
+600 or
+601 res : dict
+602 Imported data and meta-data
+603 """
+604
+605 if not fname.endswith('.xml') and not fname.endswith('.gz'):
+606 fname += '.xml'
+607 if gz:
+608 if not fname.endswith('.gz'):
+609 fname += '.gz'
+610 with gzip.open(fname, 'r') as fin:
+611 content = fin.read()
+612 else:
+613 if fname.endswith('.gz'):
+614 warnings.warn("Trying to read from %s without unzipping!" % fname, UserWarning)
+615 with open(fname, 'r') as fin:
+616 content = fin.read()
+617
+618 return import_dobs_string(content, noempty, full_output, separator_insertion=separator_insertion)
+619
+620
+621def _dobsdict_to_xmlstring(d):
+622 if isinstance(d, dict):
+623 iters = ''
+624 for k in d:
+625 if k.startswith('#value'):
+626 for li in d[k]:
+627 iters += li
+628 return iters + '\n'
+629 elif k.startswith('#'):
+630 for li in d[k]:
+631 iters += li
+632 iters = '<array>\n' + iters + '<%sarray>\n' % ('/')
+633 return iters
+634 if isinstance(d[k], dict):
+635 iters += '<%s>\n' % (k) + _dobsdict_to_xmlstring(d[k]) + '<%s%s>\n' % ('/', k)
+636 elif isinstance(d[k], str):
+637 if len(d[k]) > 100:
+638 iters += '<%s>\n ' % (k) + d[k] + ' \n<%s%s>\n' % ('/', k)
+639 else:
+640 iters += '<%s> ' % (k) + d[k] + ' <%s%s>\n' % ('/', k)
+641 elif isinstance(d[k], list):
+642 tmps = ''
+643 if k in ['edata', 'cdata']:
+644 for i in range(len(d[k])):
+645 tmps += '<%s>\n' % (k) + _dobsdict_to_xmlstring(d[k][i]) + '</%s>\n' % (k)
+646 else:
+647 for i in range(len(d[k])):
+648 tmps += _dobsdict_to_xmlstring(d[k][i])
+649 iters += tmps
+650 elif isinstance(d[k], (int, float)):
+651 iters += '<%s> ' % (k) + str(d[k]) + ' <%s%s>\n' % ('/', k)
+652 elif not d[k]:
+653 return '\n'
+654 else:
+655 raise Exception('Type', type(d[k]), 'not supported in export!')
+656 else:
+657 raise Exception('Type', type(d), 'not supported in export!')
+658 return iters
+659
+660
+661def _dobsdict_to_xmlstring_spaces(d, space=' '):
+662 s = _dobsdict_to_xmlstring(d)
+663 o = ''
+664 c = 0
+665 cm = False
+666 for li in s.split('\n'):
+667 if li.startswith('<%s' % ('/')):
+668 c -= 1
+669 cm = True
+670 for i in range(c):
+671 o += space
+672 o += li + '\n'
+673 if li.startswith('<') and not cm:
+674 if not '<%s' % ('/') in li:
+675 c += 1
+676 cm = False
+677 return o
+678
+679
+680def create_dobs_string(obsl, name, spec='dobs v1.0', origin='', symbol=[], who=None, enstags=None):
+681 """Generate the string for the export of a list of Obs or structures containing Obs
+682 to a .xml.gz file according to the Zeuthen dobs format.
+683
+684 Tags are not written or recovered automatically. The separator |is removed from the replica names.
+685
+686 Parameters
+687 ----------
+688 obsl : list
+689 List of Obs that will be exported.
+690 The Obs inside a structure do not have to be defined on the same set of configurations,
+691 but the storage requirement is increased, if this is not the case.
+692 name : str
+693 The name of the observable.
+694 spec : str
+695 Optional string that describes the contents of the file.
+696 origin : str
+697 Specify where the data has its origin.
+698 symbol : list
+699 A list of symbols that describe the observables to be written. May be empty.
+700 who : str
+701 Provide the name of the person that exports the data.
+702 enstags : dict
+703 Provide alternative enstag for ensembles in the form enstags = {ename: enstag}
+704 Otherwise, the ensemble name is used.
+705
+706 Returns
+707 -------
+708 xml_str : str
+709 XML string generated from the data
+710 """
+711 if enstags is None:
+712 enstags = {}
+713 od = {}
+714 r_names = []
+715 for o in obsl:
+716 r_names += [name for name in o.names if name.split('|')[0] in o.mc_names]
+717 r_names = sorted(set(r_names))
+718 mc_names = sorted(set([n.split('|')[0] for n in r_names]))
+719 for tmpname in mc_names:
+720 if tmpname not in enstags:
+721 enstags[tmpname] = tmpname
+722 ne = len(set(mc_names))
+723 cov_names = []
+724 for o in obsl:
+725 cov_names += list(o.cov_names)
+726 cov_names = sorted(set(cov_names))
+727 nc = len(set(cov_names))
+728 od['OBSERVABLES'] = {}
+729 od['OBSERVABLES']['SCHEMA'] = {'NAME': 'lattobs', 'VERSION': '1.0'}
+730 if who is None:
+731 who = getpass.getuser()
+732 od['OBSERVABLES']['origin'] = {
+733 'who': who,
+734 'date': str(datetime.datetime.now())[:-7],
+735 'host': socket.gethostname(),
+736 'tool': {'name': 'pyerrors', 'version': pyerrorsversion.__version__}}
+737 od['OBSERVABLES']['dobs'] = {}
+738 pd = od['OBSERVABLES']['dobs']
+739 pd['spec'] = spec
+740 pd['origin'] = origin
+741 pd['name'] = name
+742 pd['array'] = {}
+743 pd['array']['id'] = 'val'
+744 pd['array']['layout'] = '1 f%d' % (len(obsl))
+745 osymbol = ''
+746 if symbol:
+747 if not isinstance(symbol, list):
+748 raise Exception('Symbol has to be a list!')
+749 if not (len(symbol) == 0 or len(symbol) == len(obsl)):
+750 raise Exception('Symbol has to be a list of lenght 0 or %d!' % (len(obsl)))
+751 osymbol = symbol[0]
+752 for s in symbol[1:]:
+753 osymbol += ' %s' % s
+754 pd['array']['symbol'] = osymbol
+755
+756 pd['array']['#values'] = [' '.join(['%1.16e' % o.value for o in obsl])]
+757 pd['ne'] = '%d' % (ne)
+758 pd['nc'] = '%d' % (nc)
+759 pd['edata'] = []
+760 for name in mc_names:
+761 ed = {}
+762 ed['enstag'] = enstags[name]
+763 onames = sorted([n for n in r_names if (n.startswith(name + '|') or n == name)])
+764 nr = len(onames)
+765 ed['nr'] = nr
+766 ed[''] = []
+767
+768 for r in range(nr):
+769 ad = {}
+770 repname = onames[r]
+771 ad['id'] = repname.replace('|', '')
+772 idx = _merge_idx([o.idl.get(repname, []) for o in obsl])
+773 Nconf = len(idx)
+774 layout = '%d i f%d' % (Nconf, len(obsl))
+775 ad['layout'] = layout
+776 data = ''
+777 counters = [0 for o in obsl]
+778 offsets = [o.r_values[repname] - o.value if repname in o.r_values else 0 for o in obsl]
+779 for ci in idx:
+780 data += '%d ' % ci
+781 for oi in range(len(obsl)):
+782 o = obsl[oi]
+783 if repname in o.idl:
+784 if counters[oi] < 0:
+785 num = offsets[oi]
+786 if num == 0:
+787 data += '0 '
+788 else:
+789 data += '%1.16e ' % (num)
+790 continue
+791 if o.idl[repname][counters[oi]] == ci:
+792 num = o.deltas[repname][counters[oi]] + offsets[oi]
+793 if num == 0:
+794 data += '0 '
+795 else:
+796 data += '%1.16e ' % (num)
+797 counters[oi] += 1
+798 if counters[oi] >= len(o.idl[repname]):
+799 counters[oi] = -1
+800 else:
+801 num = offsets[oi]
+802 if num == 0:
+803 data += '0 '
+804 else:
+805 data += '%1.16e ' % (num)
+806 else:
+807 data += '0 '
+808 data += '\n'
+809 ad['#data'] = data
+810 ed[''].append(ad)
+811 pd['edata'].append(ed)
+812
+813 allcov = {}
+814 for o in obsl:
+815 for cname in o.cov_names:
+816 if cname in allcov:
+817 if not np.array_equal(allcov[cname], o.covobs[cname].cov):
+818 raise Exception('Inconsistent covariance matrices for %s!' % (cname))
+819 else:
+820 allcov[cname] = o.covobs[cname].cov
+821 pd['cdata'] = []
+822 for cname in cov_names:
+823 cd = {}
+824 cd['id'] = cname
825
-826 rs = '<?xml version="1.0" encoding="utf-8"?>\n' + _dobsdict_to_xmlstring_spaces(od)
-827
-828 return rs
-829
-830
-831def write_dobs(obsl, fname, name, spec='dobs v1.0', origin='', symbol=[], who=None, enstags=None, gz=True):
-832 """Export a list of Obs or structures containing Obs to a .xml.gz file
-833 according to the Zeuthen dobs format.
-834
-835 Tags are not written or recovered automatically. The separator | is removed from the replica names.
-836
-837 Parameters
-838 ----------
-839 obsl : list
-840 List of Obs that will be exported.
-841 The Obs inside a structure do not have to be defined on the same set of configurations,
-842 but the storage requirement is increased, if this is not the case.
-843 fname : str
-844 Filename of the output file.
-845 name : str
-846 The name of the observable.
-847 spec : str
-848 Optional string that describes the contents of the file.
-849 origin : str
-850 Specify where the data has its origin.
-851 symbol : list
-852 A list of symbols that describe the observables to be written. May be empty.
-853 who : str
-854 Provide the name of the person that exports the data.
-855 enstags : dict
-856 Provide alternative enstag for ensembles in the form enstags = {ename: enstag}
-857 Otherwise, the ensemble name is used.
-858 gz : bool
-859 If True, the output is a gzipped XML. If False, the output is a XML file.
-860 """
-861 if enstags is None:
-862 enstags = {}
+826 covd = {'id': 'cov'}
+827 if allcov[cname].shape == ():
+828 ncov = 1
+829 covd['layout'] = '1 1 f'
+830 covd['#data'] = '%1.14e' % (allcov[cname])
+831 else:
+832 shape = allcov[cname].shape
+833 assert (shape[0] == shape[1])
+834 ncov = shape[0]
+835 covd['layout'] = '%d %d f' % (ncov, ncov)
+836 ds = ''
+837 for i in range(ncov):
+838 for j in range(ncov):
+839 val = allcov[cname][i][j]
+840 if val == 0:
+841 ds += '0 '
+842 else:
+843 ds += '%1.14e ' % (val)
+844 ds += '\n'
+845 covd['#data'] = ds
+846
+847 gradd = {'id': 'grad'}
+848 gradd['layout'] = '%d f%d' % (ncov, len(obsl))
+849 ds = ''
+850 for i in range(ncov):
+851 for o in obsl:
+852 if cname in o.covobs:
+853 val = o.covobs[cname].grad[i]
+854 if val != 0:
+855 ds += '%1.14e ' % (val)
+856 else:
+857 ds += '0 '
+858 else:
+859 ds += '0 '
+860 gradd['#data'] = ds
+861 cd['array'] = [covd, gradd]
+862 pd['cdata'].append(cd)
863
-864 dobsstring = create_dobs_string(obsl, name, spec, origin, symbol, who, enstags=enstags)
+864 rs = '<?xml version="1.0" encoding="utf-8"?>\n' + _dobsdict_to_xmlstring_spaces(od)
865
-866 if not fname.endswith('.xml') and not fname.endswith('.gz'):
-867 fname += '.xml'
+866 return rs
+867
868
-869 if gz:
-870 if not fname.endswith('.gz'):
-871 fname += '.gz'
+869def write_dobs(obsl, fname, name, spec='dobs v1.0', origin='', symbol=[], who=None, enstags=None, gz=True):
+870 """Export a list of Obs or structures containing Obs to a .xml.gz file
+871 according to the Zeuthen dobs format.
872
-873 fp = gzip.open(fname, 'wb')
-874 fp.write(dobsstring.encode('utf-8'))
-875 else:
-876 fp = open(fname, 'w', encoding='utf-8')
-877 fp.write(dobsstring)
-878 fp.close()
+873 Tags are not written or recovered automatically. The separator | is removed from the replica names.
+874
+875 Parameters
+876 ----------
+877 obsl : list
+878 List of Obs that will be exported.
+879 The Obs inside a structure do not have to be defined on the same set of configurations,
+880 but the storage requirement is increased, if this is not the case.
+881 fname : str
+882 Filename of the output file.
+883 name : str
+884 The name of the observable.
+885 spec : str
+886 Optional string that describes the contents of the file.
+887 origin : str
+888 Specify where the data has its origin.
+889 symbol : list
+890 A list of symbols that describe the observables to be written. May be empty.
+891 who : str
+892 Provide the name of the person that exports the data.
+893 enstags : dict
+894 Provide alternative enstag for ensembles in the form enstags = {ename: enstag}
+895 Otherwise, the ensemble name is used.
+896 gz : bool
+897 If True, the output is a gzipped XML. If False, the output is a XML file.
+898
+899 Returns
+900 -------
+901 None
+902 """
+903 if enstags is None:
+904 enstags = {}
+905
+906 dobsstring = create_dobs_string(obsl, name, spec, origin, symbol, who, enstags=enstags)
+907
+908 if not fname.endswith('.xml') and not fname.endswith('.gz'):
+909 fname += '.xml'
+910
+911 if gz:
+912 if not fname.endswith('.gz'):
+913 fname += '.gz'
+914
+915 fp = gzip.open(fname, 'wb')
+916 fp.write(dobsstring.encode('utf-8'))
+917 else:
+918 fp = open(fname, 'w', encoding='utf-8')
+919 fp.write(dobsstring)
+920 fp.close()
@@ -1008,69 +1050,74 @@
107 A list of symbols that describe the observables to be written. May be empty.
108 enstag : str
109 Enstag that is written to pobs. If None, the ensemble name is used.
-110 """
-111
-112 od = {}
-113 ename = obsl[0].e_names[0]
-114 names = list(obsl[0].deltas.keys())
-115 nr = len(names)
-116 onames = [name.replace('|', '') for name in names]
-117 for o in obsl:
-118 if len(o.e_names) != 1:
-119 raise Exception('You try to export dobs to obs!')
-120 if o.e_names[0] != ename:
-121 raise Exception('You try to export dobs to obs!')
-122 if len(o.deltas.keys()) != nr:
-123 raise Exception('Incompatible obses in list')
-124 od['observables'] = {}
-125 od['observables']['schema'] = {'name': 'lattobs', 'version': '1.0'}
-126 od['observables']['origin'] = {
-127 'who': getpass.getuser(),
-128 'date': str(datetime.datetime.now())[:-7],
-129 'host': socket.gethostname(),
-130 'tool': {'name': 'pyerrors', 'version': pyerrorsversion.__version__}}
-131 od['observables']['pobs'] = {}
-132 pd = od['observables']['pobs']
-133 pd['spec'] = spec
-134 pd['origin'] = origin
-135 pd['name'] = name
-136 if enstag:
-137 if not isinstance(enstag, str):
-138 raise Exception('enstag has to be a string!')
-139 pd['enstag'] = enstag
-140 else:
-141 pd['enstag'] = ename
-142 pd['nr'] = '%d' % (nr)
-143 pd['array'] = []
-144 osymbol = 'cfg'
-145 if not isinstance(symbol, list):
-146 raise Exception('Symbol has to be a list!')
-147 if not (len(symbol) == 0 or len(symbol) == len(obsl)):
-148 raise Exception('Symbol has to be a list of lenght 0 or %d!' % (len(obsl)))
-149 for s in symbol:
-150 osymbol += ' %s' % s
-151 for r in range(nr):
-152 ad = {}
-153 ad['id'] = onames[r]
-154 Nconf = len(obsl[0].deltas[names[r]])
-155 layout = '%d i f%d' % (Nconf, len(obsl))
-156 ad['layout'] = layout
-157 ad['symbol'] = osymbol
-158 data = ''
-159 for c in range(Nconf):
-160 data += '%d ' % obsl[0].idl[names[r]][c]
-161 for o in obsl:
-162 num = o.deltas[names[r]][c] + o.r_values[names[r]]
-163 if num == 0:
-164 data += '0 '
-165 else:
-166 data += '%1.16e ' % (num)
-167 data += '\n'
-168 ad['#data'] = data
-169 pd['array'].append(ad)
-170
-171 rs = '<?xml version="1.0" encoding="utf-8"?>\n' + _dict_to_xmlstring_spaces(od)
-172 return rs
+110
+111 Returns
+112 -------
+113 xml_str : str
+114 XML formatted string of the input data
+115 """
+116
+117 od = {}
+118 ename = obsl[0].e_names[0]
+119 names = list(obsl[0].deltas.keys())
+120 nr = len(names)
+121 onames = [name.replace('|', '') for name in names]
+122 for o in obsl:
+123 if len(o.e_names) != 1:
+124 raise Exception('You try to export dobs to obs!')
+125 if o.e_names[0] != ename:
+126 raise Exception('You try to export dobs to obs!')
+127 if len(o.deltas.keys()) != nr:
+128 raise Exception('Incompatible obses in list')
+129 od['observables'] = {}
+130 od['observables']['schema'] = {'name': 'lattobs', 'version': '1.0'}
+131 od['observables']['origin'] = {
+132 'who': getpass.getuser(),
+133 'date': str(datetime.datetime.now())[:-7],
+134 'host': socket.gethostname(),
+135 'tool': {'name': 'pyerrors', 'version': pyerrorsversion.__version__}}
+136 od['observables']['pobs'] = {}
+137 pd = od['observables']['pobs']
+138 pd['spec'] = spec
+139 pd['origin'] = origin
+140 pd['name'] = name
+141 if enstag:
+142 if not isinstance(enstag, str):
+143 raise Exception('enstag has to be a string!')
+144 pd['enstag'] = enstag
+145 else:
+146 pd['enstag'] = ename
+147 pd['nr'] = '%d' % (nr)
+148 pd['array'] = []
+149 osymbol = 'cfg'
+150 if not isinstance(symbol, list):
+151 raise Exception('Symbol has to be a list!')
+152 if not (len(symbol) == 0 or len(symbol) == len(obsl)):
+153 raise Exception('Symbol has to be a list of lenght 0 or %d!' % (len(obsl)))
+154 for s in symbol:
+155 osymbol += ' %s' % s
+156 for r in range(nr):
+157 ad = {}
+158 ad['id'] = onames[r]
+159 Nconf = len(obsl[0].deltas[names[r]])
+160 layout = '%d i f%d' % (Nconf, len(obsl))
+161 ad['layout'] = layout
+162 ad['symbol'] = osymbol
+163 data = ''
+164 for c in range(Nconf):
+165 data += '%d ' % obsl[0].idl[names[r]][c]
+166 for o in obsl:
+167 num = o.deltas[names[r]][c] + o.r_values[names[r]]
+168 if num == 0:
+169 data += '0 '
+170 else:
+171 data += '%1.16e ' % (num)
+172 data += '\n'
+173 ad['#data'] = data
+174 pd['array'].append(ad)
+175
+176 rs = '<?xml version="1.0" encoding="utf-8"?>\n' + _dict_to_xmlstring_spaces(od)
+177 return rs
@@ -1096,6 +1143,13 @@ A list of symbols that describe the observables to be written. May be empty.enstag (str):
Enstag that is written to pobs. If None, the ensemble name is used.
+
+Returns
+
+
+- xml_str (str):
+XML formatted string of the input data
+
@@ -1111,47 +1165,51 @@ Enstag that is written to pobs. If None, the ensemble name is used.
- 175def write_pobs(obsl, fname, name, spec='', origin='', symbol=[], enstag=None, gz=True):
-176 """Export a list of Obs or structures containing Obs to a .xml.gz file
-177 according to the Zeuthen pobs format.
-178
-179 Tags are not written or recovered automatically. The separator | is removed from the replica names.
-180
-181 Parameters
-182 ----------
-183 obsl : list
-184 List of Obs that will be exported.
-185 The Obs inside a structure have to be defined on the same ensemble.
-186 fname : str
-187 Filename of the output file.
-188 name : str
-189 The name of the observable.
-190 spec : str
-191 Optional string that describes the contents of the file.
-192 origin : str
-193 Specify where the data has its origin.
-194 symbol : list
-195 A list of symbols that describe the observables to be written. May be empty.
-196 enstag : str
-197 Enstag that is written to pobs. If None, the ensemble name is used.
-198 gz : bool
-199 If True, the output is a gzipped xml. If False, the output is an xml file.
-200 """
-201 pobsstring = create_pobs_string(obsl, name, spec, origin, symbol, enstag)
-202
-203 if not fname.endswith('.xml') and not fname.endswith('.gz'):
-204 fname += '.xml'
+ 180def write_pobs(obsl, fname, name, spec='', origin='', symbol=[], enstag=None, gz=True):
+181 """Export a list of Obs or structures containing Obs to a .xml.gz file
+182 according to the Zeuthen pobs format.
+183
+184 Tags are not written or recovered automatically. The separator | is removed from the replica names.
+185
+186 Parameters
+187 ----------
+188 obsl : list
+189 List of Obs that will be exported.
+190 The Obs inside a structure have to be defined on the same ensemble.
+191 fname : str
+192 Filename of the output file.
+193 name : str
+194 The name of the observable.
+195 spec : str
+196 Optional string that describes the contents of the file.
+197 origin : str
+198 Specify where the data has its origin.
+199 symbol : list
+200 A list of symbols that describe the observables to be written. May be empty.
+201 enstag : str
+202 Enstag that is written to pobs. If None, the ensemble name is used.
+203 gz : bool
+204 If True, the output is a gzipped xml. If False, the output is an xml file.
205
-206 if gz:
-207 if not fname.endswith('.gz'):
-208 fname += '.gz'
-209
-210 fp = gzip.open(fname, 'wb')
-211 fp.write(pobsstring.encode('utf-8'))
-212 else:
-213 fp = open(fname, 'w', encoding='utf-8')
-214 fp.write(pobsstring)
-215 fp.close()
+206 Returns
+207 -------
+208 None
+209 """
+210 pobsstring = create_pobs_string(obsl, name, spec, origin, symbol, enstag)
+211
+212 if not fname.endswith('.xml') and not fname.endswith('.gz'):
+213 fname += '.xml'
+214
+215 if gz:
+216 if not fname.endswith('.gz'):
+217 fname += '.gz'
+218
+219 fp = gzip.open(fname, 'wb')
+220 fp.write(pobsstring.encode('utf-8'))
+221 else:
+222 fp = open(fname, 'w', encoding='utf-8')
+223 fp.write(pobsstring)
+224 fp.close()
@@ -1181,6 +1239,12 @@ Enstag that is written to pobs. If None, the ensemble name is used.
gz (bool):
If True, the output is a gzipped xml. If False, the output is an xml file.
+
+Returns
+
+
@@ -1196,95 +1260,103 @@ If True, the output is a gzipped xml. If False, the output is an xml file.
- 292def read_pobs(fname, full_output=False, gz=True, separator_insertion=None):
-293 """Import a list of Obs from an xml.gz file in the Zeuthen pobs format.
-294
-295 Tags are not written or recovered automatically.
-296
-297 Parameters
-298 ----------
-299 fname : str
-300 Filename of the input file.
-301 full_output : bool
-302 If True, a dict containing auxiliary information and the data is returned.
-303 If False, only the data is returned as list.
-304 separatior_insertion: str or int
-305 str: replace all occurences of "separator_insertion" within the replica names
-306 by "|%s" % (separator_insertion) when constructing the names of the replica.
-307 int: Insert the separator "|" at the position given by separator_insertion.
-308 None (default): Replica names remain unchanged.
-309 """
-310
-311 if not fname.endswith('.xml') and not fname.endswith('.gz'):
-312 fname += '.xml'
-313 if gz:
-314 if not fname.endswith('.gz'):
-315 fname += '.gz'
-316 with gzip.open(fname, 'r') as fin:
-317 content = fin.read()
-318 else:
-319 if fname.endswith('.gz'):
-320 warnings.warn("Trying to read from %s without unzipping!" % fname, UserWarning)
-321 with open(fname, 'r') as fin:
-322 content = fin.read()
-323
-324 # parse xml file content
-325 root = et.fromstring(content)
-326
-327 _check(root[2].tag == 'pobs')
-328 pobs = root[2]
-329
-330 version = root[0][1].text.strip()
-331
-332 _check(root[1].tag == 'origin')
-333 file_origin = _etree_to_dict(root[1])['origin']
-334
-335 deltas = []
-336 names = []
-337 idl = []
-338 for i in range(5, len(pobs)):
-339 delta, name, idx = _import_rdata(pobs[i])
-340 deltas.append(delta)
-341 if separator_insertion is None:
-342 pass
-343 elif isinstance(separator_insertion, int):
-344 name = name[:separator_insertion] + '|' + name[separator_insertion:]
-345 elif isinstance(separator_insertion, str):
-346 name = name.replace(separator_insertion, "|%s" % (separator_insertion))
-347 else:
-348 raise Exception("separator_insertion has to be string or int, is ", type(separator_insertion))
-349 names.append(name)
-350 idl.append(idx)
-351 res = [Obs([d[i] for d in deltas], names, idl=idl) for i in range(len(deltas[0]))]
-352
-353 descriptiond = {}
-354 for i in range(4):
-355 descriptiond[pobs[i].tag] = pobs[i].text.strip()
-356
-357 _check(pobs[4].tag == "nr")
-358
-359 _check(pobs[5].tag == 'array')
-360 if pobs[5][1].tag == 'symbol':
-361 symbol = pobs[5][1].text.strip()
-362 descriptiond['symbol'] = symbol
-363
-364 if full_output:
-365 retd = {}
-366 tool = file_origin.get('tool', None)
-367 if tool:
-368 program = tool['name'] + ' ' + tool['version']
-369 else:
-370 program = ''
-371 retd['program'] = program
-372 retd['version'] = version
-373 retd['who'] = file_origin['who']
-374 retd['date'] = file_origin['date']
-375 retd['host'] = file_origin['host']
-376 retd['description'] = descriptiond
-377 retd['obsdata'] = res
-378 return retd
-379 else:
-380 return res
+ 301def read_pobs(fname, full_output=False, gz=True, separator_insertion=None):
+302 """Import a list of Obs from an xml.gz file in the Zeuthen pobs format.
+303
+304 Tags are not written or recovered automatically.
+305
+306 Parameters
+307 ----------
+308 fname : str
+309 Filename of the input file.
+310 full_output : bool
+311 If True, a dict containing auxiliary information and the data is returned.
+312 If False, only the data is returned as list.
+313 separatior_insertion: str or int
+314 str: replace all occurences of "separator_insertion" within the replica names
+315 by "|%s" % (separator_insertion) when constructing the names of the replica.
+316 int: Insert the separator "|" at the position given by separator_insertion.
+317 None (default): Replica names remain unchanged.
+318
+319 Returns
+320 -------
+321 res : list[Obs]
+322 Imported data
+323 or
+324 res : dict
+325 Imported data and meta-data
+326 """
+327
+328 if not fname.endswith('.xml') and not fname.endswith('.gz'):
+329 fname += '.xml'
+330 if gz:
+331 if not fname.endswith('.gz'):
+332 fname += '.gz'
+333 with gzip.open(fname, 'r') as fin:
+334 content = fin.read()
+335 else:
+336 if fname.endswith('.gz'):
+337 warnings.warn("Trying to read from %s without unzipping!" % fname, UserWarning)
+338 with open(fname, 'r') as fin:
+339 content = fin.read()
+340
+341 # parse xml file content
+342 root = et.fromstring(content)
+343
+344 _check(root[2].tag == 'pobs')
+345 pobs = root[2]
+346
+347 version = root[0][1].text.strip()
+348
+349 _check(root[1].tag == 'origin')
+350 file_origin = _etree_to_dict(root[1])['origin']
+351
+352 deltas = []
+353 names = []
+354 idl = []
+355 for i in range(5, len(pobs)):
+356 delta, name, idx = _import_rdata(pobs[i])
+357 deltas.append(delta)
+358 if separator_insertion is None:
+359 pass
+360 elif isinstance(separator_insertion, int):
+361 name = name[:separator_insertion] + '|' + name[separator_insertion:]
+362 elif isinstance(separator_insertion, str):
+363 name = name.replace(separator_insertion, "|%s" % (separator_insertion))
+364 else:
+365 raise Exception("separator_insertion has to be string or int, is ", type(separator_insertion))
+366 names.append(name)
+367 idl.append(idx)
+368 res = [Obs([d[i] for d in deltas], names, idl=idl) for i in range(len(deltas[0]))]
+369
+370 descriptiond = {}
+371 for i in range(4):
+372 descriptiond[pobs[i].tag] = pobs[i].text.strip()
+373
+374 _check(pobs[4].tag == "nr")
+375
+376 _check(pobs[5].tag == 'array')
+377 if pobs[5][1].tag == 'symbol':
+378 symbol = pobs[5][1].text.strip()
+379 descriptiond['symbol'] = symbol
+380
+381 if full_output:
+382 retd = {}
+383 tool = file_origin.get('tool', None)
+384 if tool:
+385 program = tool['name'] + ' ' + tool['version']
+386 else:
+387 program = ''
+388 retd['program'] = program
+389 retd['version'] = version
+390 retd['who'] = file_origin['who']
+391 retd['date'] = file_origin['date']
+392 retd['host'] = file_origin['host']
+393 retd['description'] = descriptiond
+394 retd['obsdata'] = res
+395 return retd
+396 else:
+397 return res
@@ -1306,6 +1378,16 @@ by "|%s" % (separator_insertion) when constructing the names of the replica.
int: Insert the separator "|" at the position given by separator_insertion.
None (default): Replica names remain unchanged.
+
+Returns
+
+
+- res (list[Obs]):
+Imported data
+- or
+- res (dict):
+Imported data and meta-data
+
@@ -1321,167 +1403,175 @@ None (default): Replica names remain unchanged.
- 384def import_dobs_string(content, noempty=False, full_output=False, separator_insertion=True):
-385 """Import a list of Obs from a string in the Zeuthen dobs format.
-386
-387 Tags are not written or recovered automatically.
-388
-389 Parameters
-390 ----------
-391 content : str
-392 XML string containing the data
-393 noemtpy : bool
-394 If True, ensembles with no contribution to the Obs are not included.
-395 If False, ensembles are included as written in the file, possibly with vanishing entries.
-396 full_output : bool
-397 If True, a dict containing auxiliary information and the data is returned.
-398 If False, only the data is returned as list.
-399 separatior_insertion: str, int or bool
-400 str: replace all occurences of "separator_insertion" within the replica names
-401 by "|%s" % (separator_insertion) when constructing the names of the replica.
-402 int: Insert the separator "|" at the position given by separator_insertion.
-403 True (default): separator "|" is inserted after len(ensname), assuming that the
-404 ensemble name is a prefix to the replica name.
-405 None or False: No separator is inserted.
-406 """
-407
-408 root = et.fromstring(content)
-409
-410 _check(root.tag == 'OBSERVABLES')
-411 _check(root[0].tag == 'SCHEMA')
-412 version = root[0][1].text.strip()
-413
-414 _check(root[1].tag == 'origin')
-415 file_origin = _etree_to_dict(root[1])['origin']
-416
-417 _check(root[2].tag == 'dobs')
-418
-419 dobs = root[2]
-420
-421 descriptiond = {}
-422 for i in range(3):
-423 descriptiond[dobs[i].tag] = dobs[i].text.strip()
-424
-425 _check(dobs[3].tag == 'array')
-426
-427 symbol = []
-428 if dobs[3][1].tag == 'symbol':
-429 symbol = dobs[3][1].text.strip()
-430 descriptiond['symbol'] = symbol
-431 mean = _import_array(dobs[3])[0]
+ 401def import_dobs_string(content, noempty=False, full_output=False, separator_insertion=True):
+402 """Import a list of Obs from a string in the Zeuthen dobs format.
+403
+404 Tags are not written or recovered automatically.
+405
+406 Parameters
+407 ----------
+408 content : str
+409 XML string containing the data
+410 noemtpy : bool
+411 If True, ensembles with no contribution to the Obs are not included.
+412 If False, ensembles are included as written in the file, possibly with vanishing entries.
+413 full_output : bool
+414 If True, a dict containing auxiliary information and the data is returned.
+415 If False, only the data is returned as list.
+416 separatior_insertion: str, int or bool
+417 str: replace all occurences of "separator_insertion" within the replica names
+418 by "|%s" % (separator_insertion) when constructing the names of the replica.
+419 int: Insert the separator "|" at the position given by separator_insertion.
+420 True (default): separator "|" is inserted after len(ensname), assuming that the
+421 ensemble name is a prefix to the replica name.
+422 None or False: No separator is inserted.
+423
+424 Returns
+425 -------
+426 res : list[Obs]
+427 Imported data
+428 or
+429 res : dict
+430 Imported data and meta-data
+431 """
432
-433 _check(dobs[4].tag == "ne")
-434 ne = int(dobs[4].text.strip())
-435 _check(dobs[5].tag == "nc")
-436 nc = int(dobs[5].text.strip())
-437
-438 idld = {}
-439 deltad = {}
-440 covd = {}
-441 gradd = {}
-442 names = []
-443 e_names = []
-444 enstags = {}
-445 for k in range(6, len(list(dobs))):
-446 if dobs[k].tag == "edata":
-447 _check(dobs[k][0].tag == "enstag")
-448 ename = dobs[k][0].text.strip()
-449 e_names.append(ename)
-450 _check(dobs[k][1].tag == "nr")
-451 R = int(dobs[k][1].text.strip())
-452 for i in range(2, 2 + R):
-453 deltas, rname, idx = _import_rdata(dobs[k][i])
-454 if separator_insertion is None or False:
-455 pass
-456 elif separator_insertion is True:
-457 if rname.startswith(ename):
-458 rname = rname[:len(ename)] + '|' + rname[len(ename):]
-459 elif isinstance(separator_insertion, int):
-460 rname = rname[:separator_insertion] + '|' + rname[separator_insertion:]
-461 elif isinstance(separator_insertion, str):
-462 rname = rname.replace(separator_insertion, "|%s" % (separator_insertion))
-463 else:
-464 raise Exception("separator_insertion has to be string or int, is ", type(separator_insertion))
-465 if '|' in rname:
-466 new_ename = rname[:rname.index('|')]
-467 else:
-468 new_ename = ename
-469 enstags[new_ename] = ename
-470 idld[rname] = idx
-471 deltad[rname] = deltas
-472 names.append(rname)
-473 elif dobs[k].tag == "cdata":
-474 cname, cov, grad = _import_cdata(dobs[k])
-475 covd[cname] = cov
-476 if grad.shape[1] == 1:
-477 gradd[cname] = [grad for i in range(len(mean))]
-478 else:
-479 gradd[cname] = grad.T
-480 else:
-481 _check(False)
-482 names = list(set(names))
-483
-484 for name in names:
-485 for i in range(len(deltad[name])):
-486 deltad[name][i] = np.array(deltad[name][i]) + mean[i]
-487
-488 res = []
-489 for i in range(len(mean)):
-490 deltas = []
-491 idl = []
-492 obs_names = []
-493 for name in names:
-494 h = np.unique(deltad[name][i])
-495 if len(h) == 1 and np.all(h == mean[i]) and noempty:
-496 continue
-497 deltas.append(deltad[name][i])
-498 obs_names.append(name)
-499 idl.append(idld[name])
-500 res.append(Obs(deltas, obs_names, idl=idl))
-501 res[-1]._value = mean[i]
-502 _check(len(e_names) == ne)
-503
-504 cnames = list(covd.keys())
-505 for i in range(len(res)):
-506 new_covobs = {name: Covobs(0, covd[name], name, grad=gradd[name][i]) for name in cnames}
-507 if noempty:
-508 for name in cnames:
-509 if np.all(new_covobs[name].grad == 0):
-510 del new_covobs[name]
-511 cnames_loc = list(new_covobs.keys())
-512 else:
-513 cnames_loc = cnames
-514 for name in cnames_loc:
-515 res[i].names.append(name)
-516 res[i].shape[name] = 1
-517 res[i].idl[name] = []
-518 res[i]._covobs = new_covobs
-519
-520 if symbol:
-521 for i in range(len(res)):
-522 res[i].tag = symbol[i]
-523 if res[i].tag == 'None':
-524 res[i].tag = None
-525 if not noempty:
-526 _check(len(res[0].covobs.keys()) == nc)
-527 if full_output:
-528 retd = {}
-529 tool = file_origin.get('tool', None)
-530 if tool:
-531 program = tool['name'] + ' ' + tool['version']
-532 else:
-533 program = ''
-534 retd['program'] = program
-535 retd['version'] = version
-536 retd['who'] = file_origin['who']
-537 retd['date'] = file_origin['date']
-538 retd['host'] = file_origin['host']
-539 retd['description'] = descriptiond
-540 retd['enstags'] = enstags
-541 retd['obsdata'] = res
-542 return retd
-543 else:
-544 return res
+433 root = et.fromstring(content)
+434
+435 _check(root.tag == 'OBSERVABLES')
+436 _check(root[0].tag == 'SCHEMA')
+437 version = root[0][1].text.strip()
+438
+439 _check(root[1].tag == 'origin')
+440 file_origin = _etree_to_dict(root[1])['origin']
+441
+442 _check(root[2].tag == 'dobs')
+443
+444 dobs = root[2]
+445
+446 descriptiond = {}
+447 for i in range(3):
+448 descriptiond[dobs[i].tag] = dobs[i].text.strip()
+449
+450 _check(dobs[3].tag == 'array')
+451
+452 symbol = []
+453 if dobs[3][1].tag == 'symbol':
+454 symbol = dobs[3][1].text.strip()
+455 descriptiond['symbol'] = symbol
+456 mean = _import_array(dobs[3])[0]
+457
+458 _check(dobs[4].tag == "ne")
+459 ne = int(dobs[4].text.strip())
+460 _check(dobs[5].tag == "nc")
+461 nc = int(dobs[5].text.strip())
+462
+463 idld = {}
+464 deltad = {}
+465 covd = {}
+466 gradd = {}
+467 names = []
+468 e_names = []
+469 enstags = {}
+470 for k in range(6, len(list(dobs))):
+471 if dobs[k].tag == "edata":
+472 _check(dobs[k][0].tag == "enstag")
+473 ename = dobs[k][0].text.strip()
+474 e_names.append(ename)
+475 _check(dobs[k][1].tag == "nr")
+476 R = int(dobs[k][1].text.strip())
+477 for i in range(2, 2 + R):
+478 deltas, rname, idx = _import_rdata(dobs[k][i])
+479 if separator_insertion is None or False:
+480 pass
+481 elif separator_insertion is True:
+482 if rname.startswith(ename):
+483 rname = rname[:len(ename)] + '|' + rname[len(ename):]
+484 elif isinstance(separator_insertion, int):
+485 rname = rname[:separator_insertion] + '|' + rname[separator_insertion:]
+486 elif isinstance(separator_insertion, str):
+487 rname = rname.replace(separator_insertion, "|%s" % (separator_insertion))
+488 else:
+489 raise Exception("separator_insertion has to be string or int, is ", type(separator_insertion))
+490 if '|' in rname:
+491 new_ename = rname[:rname.index('|')]
+492 else:
+493 new_ename = ename
+494 enstags[new_ename] = ename
+495 idld[rname] = idx
+496 deltad[rname] = deltas
+497 names.append(rname)
+498 elif dobs[k].tag == "cdata":
+499 cname, cov, grad = _import_cdata(dobs[k])
+500 covd[cname] = cov
+501 if grad.shape[1] == 1:
+502 gradd[cname] = [grad for i in range(len(mean))]
+503 else:
+504 gradd[cname] = grad.T
+505 else:
+506 _check(False)
+507 names = list(set(names))
+508
+509 for name in names:
+510 for i in range(len(deltad[name])):
+511 deltad[name][i] = np.array(deltad[name][i]) + mean[i]
+512
+513 res = []
+514 for i in range(len(mean)):
+515 deltas = []
+516 idl = []
+517 obs_names = []
+518 for name in names:
+519 h = np.unique(deltad[name][i])
+520 if len(h) == 1 and np.all(h == mean[i]) and noempty:
+521 continue
+522 deltas.append(deltad[name][i])
+523 obs_names.append(name)
+524 idl.append(idld[name])
+525 res.append(Obs(deltas, obs_names, idl=idl))
+526 res[-1]._value = mean[i]
+527 _check(len(e_names) == ne)
+528
+529 cnames = list(covd.keys())
+530 for i in range(len(res)):
+531 new_covobs = {name: Covobs(0, covd[name], name, grad=gradd[name][i]) for name in cnames}
+532 if noempty:
+533 for name in cnames:
+534 if np.all(new_covobs[name].grad == 0):
+535 del new_covobs[name]
+536 cnames_loc = list(new_covobs.keys())
+537 else:
+538 cnames_loc = cnames
+539 for name in cnames_loc:
+540 res[i].names.append(name)
+541 res[i].shape[name] = 1
+542 res[i].idl[name] = []
+543 res[i]._covobs = new_covobs
+544
+545 if symbol:
+546 for i in range(len(res)):
+547 res[i].tag = symbol[i]
+548 if res[i].tag == 'None':
+549 res[i].tag = None
+550 if not noempty:
+551 _check(len(res[0].covobs.keys()) == nc)
+552 if full_output:
+553 retd = {}
+554 tool = file_origin.get('tool', None)
+555 if tool:
+556 program = tool['name'] + ' ' + tool['version']
+557 else:
+558 program = ''
+559 retd['program'] = program
+560 retd['version'] = version
+561 retd['who'] = file_origin['who']
+562 retd['date'] = file_origin['date']
+563 retd['host'] = file_origin['host']
+564 retd['description'] = descriptiond
+565 retd['enstags'] = enstags
+566 retd['obsdata'] = res
+567 return retd
+568 else:
+569 return res
@@ -1508,6 +1598,16 @@ True (default): separator "|" is inserted after len(ensname), assuming that the
ensemble name is a prefix to the replica name.
None or False: No separator is inserted.
+
+Returns
+
+
+- res (list[Obs]):
+Imported data
+- or
+- res (dict):
+Imported data and meta-data
+
@@ -1523,46 +1623,54 @@ None or False: No separator is inserted.
- 547def read_dobs(fname, noempty=False, full_output=False, gz=True, separator_insertion=True):
-548 """Import a list of Obs from an xml.gz file in the Zeuthen dobs format.
-549
-550 Tags are not written or recovered automatically.
-551
-552 Parameters
-553 ----------
-554 fname : str
-555 Filename of the input file.
-556 noemtpy : bool
-557 If True, ensembles with no contribution to the Obs are not included.
-558 If False, ensembles are included as written in the file.
-559 full_output : bool
-560 If True, a dict containing auxiliary information and the data is returned.
-561 If False, only the data is returned as list.
-562 gz : bool
-563 If True, assumes that data is gzipped. If False, assumes XML file.
-564 separatior_insertion: str, int or bool
-565 str: replace all occurences of "separator_insertion" within the replica names
-566 by "|%s" % (separator_insertion) when constructing the names of the replica.
-567 int: Insert the separator "|" at the position given by separator_insertion.
-568 True (default): separator "|" is inserted after len(ensname), assuming that the
-569 ensemble name is a prefix to the replica name.
-570 None or False: No separator is inserted.
-571 """
-572
-573 if not fname.endswith('.xml') and not fname.endswith('.gz'):
-574 fname += '.xml'
-575 if gz:
-576 if not fname.endswith('.gz'):
-577 fname += '.gz'
-578 with gzip.open(fname, 'r') as fin:
-579 content = fin.read()
-580 else:
-581 if fname.endswith('.gz'):
-582 warnings.warn("Trying to read from %s without unzipping!" % fname, UserWarning)
-583 with open(fname, 'r') as fin:
-584 content = fin.read()
-585
-586 return import_dobs_string(content, noempty, full_output, separator_insertion=separator_insertion)
+ 572def read_dobs(fname, noempty=False, full_output=False, gz=True, separator_insertion=True):
+573 """Import a list of Obs from an xml.gz file in the Zeuthen dobs format.
+574
+575 Tags are not written or recovered automatically.
+576
+577 Parameters
+578 ----------
+579 fname : str
+580 Filename of the input file.
+581 noemtpy : bool
+582 If True, ensembles with no contribution to the Obs are not included.
+583 If False, ensembles are included as written in the file.
+584 full_output : bool
+585 If True, a dict containing auxiliary information and the data is returned.
+586 If False, only the data is returned as list.
+587 gz : bool
+588 If True, assumes that data is gzipped. If False, assumes XML file.
+589 separatior_insertion: str, int or bool
+590 str: replace all occurences of "separator_insertion" within the replica names
+591 by "|%s" % (separator_insertion) when constructing the names of the replica.
+592 int: Insert the separator "|" at the position given by separator_insertion.
+593 True (default): separator "|" is inserted after len(ensname), assuming that the
+594 ensemble name is a prefix to the replica name.
+595 None or False: No separator is inserted.
+596
+597 Returns
+598 -------
+599 res : list[Obs]
+600 Imported data
+601 or
+602 res : dict
+603 Imported data and meta-data
+604 """
+605
+606 if not fname.endswith('.xml') and not fname.endswith('.gz'):
+607 fname += '.xml'
+608 if gz:
+609 if not fname.endswith('.gz'):
+610 fname += '.gz'
+611 with gzip.open(fname, 'r') as fin:
+612 content = fin.read()
+613 else:
+614 if fname.endswith('.gz'):
+615 warnings.warn("Trying to read from %s without unzipping!" % fname, UserWarning)
+616 with open(fname, 'r') as fin:
+617 content = fin.read()
+618
+619 return import_dobs_string(content, noempty, full_output, separator_insertion=separator_insertion)
@@ -1591,6 +1699,16 @@ True (default): separator "|" is inserted after len(ensname), assuming that the
ensemble name is a prefix to the replica name.
None or False: No separator is inserted.
+
+Returns
+
+
+- res (list[Obs]):
+Imported data
+- or
+- res (dict):
+Imported data and meta-data
+
@@ -1606,188 +1724,193 @@ None or False: No separator is inserted.
- 648def create_dobs_string(obsl, name, spec='dobs v1.0', origin='', symbol=[], who=None, enstags=None):
-649 """Generate the string for the export of a list of Obs or structures containing Obs
-650 to a .xml.gz file according to the Zeuthen dobs format.
-651
-652 Tags are not written or recovered automatically. The separator |is removed from the replica names.
-653
-654 Parameters
-655 ----------
-656 obsl : list
-657 List of Obs that will be exported.
-658 The Obs inside a structure do not have to be defined on the same set of configurations,
-659 but the storage requirement is increased, if this is not the case.
-660 name : str
-661 The name of the observable.
-662 spec : str
-663 Optional string that describes the contents of the file.
-664 origin : str
-665 Specify where the data has its origin.
-666 symbol : list
-667 A list of symbols that describe the observables to be written. May be empty.
-668 who : str
-669 Provide the name of the person that exports the data.
-670 enstags : dict
-671 Provide alternative enstag for ensembles in the form enstags = {ename: enstag}
-672 Otherwise, the ensemble name is used.
-673 """
-674 if enstags is None:
-675 enstags = {}
-676 od = {}
-677 r_names = []
-678 for o in obsl:
-679 r_names += [name for name in o.names if name.split('|')[0] in o.mc_names]
-680 r_names = sorted(set(r_names))
-681 mc_names = sorted(set([n.split('|')[0] for n in r_names]))
-682 for tmpname in mc_names:
-683 if tmpname not in enstags:
-684 enstags[tmpname] = tmpname
-685 ne = len(set(mc_names))
-686 cov_names = []
-687 for o in obsl:
-688 cov_names += list(o.cov_names)
-689 cov_names = sorted(set(cov_names))
-690 nc = len(set(cov_names))
-691 od['OBSERVABLES'] = {}
-692 od['OBSERVABLES']['SCHEMA'] = {'NAME': 'lattobs', 'VERSION': '1.0'}
-693 if who is None:
-694 who = getpass.getuser()
-695 od['OBSERVABLES']['origin'] = {
-696 'who': who,
-697 'date': str(datetime.datetime.now())[:-7],
-698 'host': socket.gethostname(),
-699 'tool': {'name': 'pyerrors', 'version': pyerrorsversion.__version__}}
-700 od['OBSERVABLES']['dobs'] = {}
-701 pd = od['OBSERVABLES']['dobs']
-702 pd['spec'] = spec
-703 pd['origin'] = origin
-704 pd['name'] = name
-705 pd['array'] = {}
-706 pd['array']['id'] = 'val'
-707 pd['array']['layout'] = '1 f%d' % (len(obsl))
-708 osymbol = ''
-709 if symbol:
-710 if not isinstance(symbol, list):
-711 raise Exception('Symbol has to be a list!')
-712 if not (len(symbol) == 0 or len(symbol) == len(obsl)):
-713 raise Exception('Symbol has to be a list of lenght 0 or %d!' % (len(obsl)))
-714 osymbol = symbol[0]
-715 for s in symbol[1:]:
-716 osymbol += ' %s' % s
-717 pd['array']['symbol'] = osymbol
-718
-719 pd['array']['#values'] = [' '.join(['%1.16e' % o.value for o in obsl])]
-720 pd['ne'] = '%d' % (ne)
-721 pd['nc'] = '%d' % (nc)
-722 pd['edata'] = []
-723 for name in mc_names:
-724 ed = {}
-725 ed['enstag'] = enstags[name]
-726 onames = sorted([n for n in r_names if (n.startswith(name + '|') or n == name)])
-727 nr = len(onames)
-728 ed['nr'] = nr
-729 ed[''] = []
-730
-731 for r in range(nr):
-732 ad = {}
-733 repname = onames[r]
-734 ad['id'] = repname.replace('|', '')
-735 idx = _merge_idx([o.idl.get(repname, []) for o in obsl])
-736 Nconf = len(idx)
-737 layout = '%d i f%d' % (Nconf, len(obsl))
-738 ad['layout'] = layout
-739 data = ''
-740 counters = [0 for o in obsl]
-741 offsets = [o.r_values[repname] - o.value if repname in o.r_values else 0 for o in obsl]
-742 for ci in idx:
-743 data += '%d ' % ci
-744 for oi in range(len(obsl)):
-745 o = obsl[oi]
-746 if repname in o.idl:
-747 if counters[oi] < 0:
-748 num = offsets[oi]
-749 if num == 0:
-750 data += '0 '
-751 else:
-752 data += '%1.16e ' % (num)
-753 continue
-754 if o.idl[repname][counters[oi]] == ci:
-755 num = o.deltas[repname][counters[oi]] + offsets[oi]
-756 if num == 0:
-757 data += '0 '
-758 else:
-759 data += '%1.16e ' % (num)
-760 counters[oi] += 1
-761 if counters[oi] >= len(o.idl[repname]):
-762 counters[oi] = -1
-763 else:
-764 num = offsets[oi]
-765 if num == 0:
-766 data += '0 '
-767 else:
-768 data += '%1.16e ' % (num)
-769 else:
-770 data += '0 '
-771 data += '\n'
-772 ad['#data'] = data
-773 ed[''].append(ad)
-774 pd['edata'].append(ed)
-775
-776 allcov = {}
-777 for o in obsl:
-778 for cname in o.cov_names:
-779 if cname in allcov:
-780 if not np.array_equal(allcov[cname], o.covobs[cname].cov):
-781 raise Exception('Inconsistent covariance matrices for %s!' % (cname))
-782 else:
-783 allcov[cname] = o.covobs[cname].cov
-784 pd['cdata'] = []
-785 for cname in cov_names:
-786 cd = {}
-787 cd['id'] = cname
-788
-789 covd = {'id': 'cov'}
-790 if allcov[cname].shape == ():
-791 ncov = 1
-792 covd['layout'] = '1 1 f'
-793 covd['#data'] = '%1.14e' % (allcov[cname])
-794 else:
-795 shape = allcov[cname].shape
-796 assert (shape[0] == shape[1])
-797 ncov = shape[0]
-798 covd['layout'] = '%d %d f' % (ncov, ncov)
-799 ds = ''
-800 for i in range(ncov):
-801 for j in range(ncov):
-802 val = allcov[cname][i][j]
-803 if val == 0:
-804 ds += '0 '
-805 else:
-806 ds += '%1.14e ' % (val)
-807 ds += '\n'
-808 covd['#data'] = ds
-809
-810 gradd = {'id': 'grad'}
-811 gradd['layout'] = '%d f%d' % (ncov, len(obsl))
-812 ds = ''
-813 for i in range(ncov):
-814 for o in obsl:
-815 if cname in o.covobs:
-816 val = o.covobs[cname].grad[i]
-817 if val != 0:
-818 ds += '%1.14e ' % (val)
-819 else:
-820 ds += '0 '
-821 else:
-822 ds += '0 '
-823 gradd['#data'] = ds
-824 cd['array'] = [covd, gradd]
-825 pd['cdata'].append(cd)
+ 681def create_dobs_string(obsl, name, spec='dobs v1.0', origin='', symbol=[], who=None, enstags=None):
+682 """Generate the string for the export of a list of Obs or structures containing Obs
+683 to a .xml.gz file according to the Zeuthen dobs format.
+684
+685 Tags are not written or recovered automatically. The separator |is removed from the replica names.
+686
+687 Parameters
+688 ----------
+689 obsl : list
+690 List of Obs that will be exported.
+691 The Obs inside a structure do not have to be defined on the same set of configurations,
+692 but the storage requirement is increased, if this is not the case.
+693 name : str
+694 The name of the observable.
+695 spec : str
+696 Optional string that describes the contents of the file.
+697 origin : str
+698 Specify where the data has its origin.
+699 symbol : list
+700 A list of symbols that describe the observables to be written. May be empty.
+701 who : str
+702 Provide the name of the person that exports the data.
+703 enstags : dict
+704 Provide alternative enstag for ensembles in the form enstags = {ename: enstag}
+705 Otherwise, the ensemble name is used.
+706
+707 Returns
+708 -------
+709 xml_str : str
+710 XML string generated from the data
+711 """
+712 if enstags is None:
+713 enstags = {}
+714 od = {}
+715 r_names = []
+716 for o in obsl:
+717 r_names += [name for name in o.names if name.split('|')[0] in o.mc_names]
+718 r_names = sorted(set(r_names))
+719 mc_names = sorted(set([n.split('|')[0] for n in r_names]))
+720 for tmpname in mc_names:
+721 if tmpname not in enstags:
+722 enstags[tmpname] = tmpname
+723 ne = len(set(mc_names))
+724 cov_names = []
+725 for o in obsl:
+726 cov_names += list(o.cov_names)
+727 cov_names = sorted(set(cov_names))
+728 nc = len(set(cov_names))
+729 od['OBSERVABLES'] = {}
+730 od['OBSERVABLES']['SCHEMA'] = {'NAME': 'lattobs', 'VERSION': '1.0'}
+731 if who is None:
+732 who = getpass.getuser()
+733 od['OBSERVABLES']['origin'] = {
+734 'who': who,
+735 'date': str(datetime.datetime.now())[:-7],
+736 'host': socket.gethostname(),
+737 'tool': {'name': 'pyerrors', 'version': pyerrorsversion.__version__}}
+738 od['OBSERVABLES']['dobs'] = {}
+739 pd = od['OBSERVABLES']['dobs']
+740 pd['spec'] = spec
+741 pd['origin'] = origin
+742 pd['name'] = name
+743 pd['array'] = {}
+744 pd['array']['id'] = 'val'
+745 pd['array']['layout'] = '1 f%d' % (len(obsl))
+746 osymbol = ''
+747 if symbol:
+748 if not isinstance(symbol, list):
+749 raise Exception('Symbol has to be a list!')
+750 if not (len(symbol) == 0 or len(symbol) == len(obsl)):
+751 raise Exception('Symbol has to be a list of lenght 0 or %d!' % (len(obsl)))
+752 osymbol = symbol[0]
+753 for s in symbol[1:]:
+754 osymbol += ' %s' % s
+755 pd['array']['symbol'] = osymbol
+756
+757 pd['array']['#values'] = [' '.join(['%1.16e' % o.value for o in obsl])]
+758 pd['ne'] = '%d' % (ne)
+759 pd['nc'] = '%d' % (nc)
+760 pd['edata'] = []
+761 for name in mc_names:
+762 ed = {}
+763 ed['enstag'] = enstags[name]
+764 onames = sorted([n for n in r_names if (n.startswith(name + '|') or n == name)])
+765 nr = len(onames)
+766 ed['nr'] = nr
+767 ed[''] = []
+768
+769 for r in range(nr):
+770 ad = {}
+771 repname = onames[r]
+772 ad['id'] = repname.replace('|', '')
+773 idx = _merge_idx([o.idl.get(repname, []) for o in obsl])
+774 Nconf = len(idx)
+775 layout = '%d i f%d' % (Nconf, len(obsl))
+776 ad['layout'] = layout
+777 data = ''
+778 counters = [0 for o in obsl]
+779 offsets = [o.r_values[repname] - o.value if repname in o.r_values else 0 for o in obsl]
+780 for ci in idx:
+781 data += '%d ' % ci
+782 for oi in range(len(obsl)):
+783 o = obsl[oi]
+784 if repname in o.idl:
+785 if counters[oi] < 0:
+786 num = offsets[oi]
+787 if num == 0:
+788 data += '0 '
+789 else:
+790 data += '%1.16e ' % (num)
+791 continue
+792 if o.idl[repname][counters[oi]] == ci:
+793 num = o.deltas[repname][counters[oi]] + offsets[oi]
+794 if num == 0:
+795 data += '0 '
+796 else:
+797 data += '%1.16e ' % (num)
+798 counters[oi] += 1
+799 if counters[oi] >= len(o.idl[repname]):
+800 counters[oi] = -1
+801 else:
+802 num = offsets[oi]
+803 if num == 0:
+804 data += '0 '
+805 else:
+806 data += '%1.16e ' % (num)
+807 else:
+808 data += '0 '
+809 data += '\n'
+810 ad['#data'] = data
+811 ed[''].append(ad)
+812 pd['edata'].append(ed)
+813
+814 allcov = {}
+815 for o in obsl:
+816 for cname in o.cov_names:
+817 if cname in allcov:
+818 if not np.array_equal(allcov[cname], o.covobs[cname].cov):
+819 raise Exception('Inconsistent covariance matrices for %s!' % (cname))
+820 else:
+821 allcov[cname] = o.covobs[cname].cov
+822 pd['cdata'] = []
+823 for cname in cov_names:
+824 cd = {}
+825 cd['id'] = cname
826
-827 rs = '<?xml version="1.0" encoding="utf-8"?>\n' + _dobsdict_to_xmlstring_spaces(od)
-828
-829 return rs
+827 covd = {'id': 'cov'}
+828 if allcov[cname].shape == ():
+829 ncov = 1
+830 covd['layout'] = '1 1 f'
+831 covd['#data'] = '%1.14e' % (allcov[cname])
+832 else:
+833 shape = allcov[cname].shape
+834 assert (shape[0] == shape[1])
+835 ncov = shape[0]
+836 covd['layout'] = '%d %d f' % (ncov, ncov)
+837 ds = ''
+838 for i in range(ncov):
+839 for j in range(ncov):
+840 val = allcov[cname][i][j]
+841 if val == 0:
+842 ds += '0 '
+843 else:
+844 ds += '%1.14e ' % (val)
+845 ds += '\n'
+846 covd['#data'] = ds
+847
+848 gradd = {'id': 'grad'}
+849 gradd['layout'] = '%d f%d' % (ncov, len(obsl))
+850 ds = ''
+851 for i in range(ncov):
+852 for o in obsl:
+853 if cname in o.covobs:
+854 val = o.covobs[cname].grad[i]
+855 if val != 0:
+856 ds += '%1.14e ' % (val)
+857 else:
+858 ds += '0 '
+859 else:
+860 ds += '0 '
+861 gradd['#data'] = ds
+862 cd['array'] = [covd, gradd]
+863 pd['cdata'].append(cd)
+864
+865 rs = '<?xml version="1.0" encoding="utf-8"?>\n' + _dobsdict_to_xmlstring_spaces(od)
+866
+867 return rs
@@ -1817,6 +1940,13 @@ Provide the name of the person that exports the data.
Provide alternative enstag for ensembles in the form enstags = {ename: enstag}
Otherwise, the ensemble name is used.
+
+Returns
+
+
+- xml_str (str):
+XML string generated from the data
+
@@ -1832,54 +1962,58 @@ Otherwise, the ensemble name is used.
- 832def write_dobs(obsl, fname, name, spec='dobs v1.0', origin='', symbol=[], who=None, enstags=None, gz=True):
-833 """Export a list of Obs or structures containing Obs to a .xml.gz file
-834 according to the Zeuthen dobs format.
-835
-836 Tags are not written or recovered automatically. The separator | is removed from the replica names.
-837
-838 Parameters
-839 ----------
-840 obsl : list
-841 List of Obs that will be exported.
-842 The Obs inside a structure do not have to be defined on the same set of configurations,
-843 but the storage requirement is increased, if this is not the case.
-844 fname : str
-845 Filename of the output file.
-846 name : str
-847 The name of the observable.
-848 spec : str
-849 Optional string that describes the contents of the file.
-850 origin : str
-851 Specify where the data has its origin.
-852 symbol : list
-853 A list of symbols that describe the observables to be written. May be empty.
-854 who : str
-855 Provide the name of the person that exports the data.
-856 enstags : dict
-857 Provide alternative enstag for ensembles in the form enstags = {ename: enstag}
-858 Otherwise, the ensemble name is used.
-859 gz : bool
-860 If True, the output is a gzipped XML. If False, the output is a XML file.
-861 """
-862 if enstags is None:
-863 enstags = {}
-864
-865 dobsstring = create_dobs_string(obsl, name, spec, origin, symbol, who, enstags=enstags)
-866
-867 if not fname.endswith('.xml') and not fname.endswith('.gz'):
-868 fname += '.xml'
-869
-870 if gz:
-871 if not fname.endswith('.gz'):
-872 fname += '.gz'
+ 870def write_dobs(obsl, fname, name, spec='dobs v1.0', origin='', symbol=[], who=None, enstags=None, gz=True):
+871 """Export a list of Obs or structures containing Obs to a .xml.gz file
+872 according to the Zeuthen dobs format.
873
-874 fp = gzip.open(fname, 'wb')
-875 fp.write(dobsstring.encode('utf-8'))
-876 else:
-877 fp = open(fname, 'w', encoding='utf-8')
-878 fp.write(dobsstring)
-879 fp.close()
+874 Tags are not written or recovered automatically. The separator | is removed from the replica names.
+875
+876 Parameters
+877 ----------
+878 obsl : list
+879 List of Obs that will be exported.
+880 The Obs inside a structure do not have to be defined on the same set of configurations,
+881 but the storage requirement is increased, if this is not the case.
+882 fname : str
+883 Filename of the output file.
+884 name : str
+885 The name of the observable.
+886 spec : str
+887 Optional string that describes the contents of the file.
+888 origin : str
+889 Specify where the data has its origin.
+890 symbol : list
+891 A list of symbols that describe the observables to be written. May be empty.
+892 who : str
+893 Provide the name of the person that exports the data.
+894 enstags : dict
+895 Provide alternative enstag for ensembles in the form enstags = {ename: enstag}
+896 Otherwise, the ensemble name is used.
+897 gz : bool
+898 If True, the output is a gzipped XML. If False, the output is a XML file.
+899
+900 Returns
+901 -------
+902 None
+903 """
+904 if enstags is None:
+905 enstags = {}
+906
+907 dobsstring = create_dobs_string(obsl, name, spec, origin, symbol, who, enstags=enstags)
+908
+909 if not fname.endswith('.xml') and not fname.endswith('.gz'):
+910 fname += '.xml'
+911
+912 if gz:
+913 if not fname.endswith('.gz'):
+914 fname += '.gz'
+915
+916 fp = gzip.open(fname, 'wb')
+917 fp.write(dobsstring.encode('utf-8'))
+918 else:
+919 fp = open(fname, 'w', encoding='utf-8')
+920 fp.write(dobsstring)
+921 fp.close()
@@ -1913,6 +2047,12 @@ Otherwise, the ensemble name is used.
gz (bool):
If True, the output is a gzipped XML. If False, the output is a XML file.
+
+Returns
+
+
diff --git a/docs/pyerrors/input/hadrons.html b/docs/pyerrors/input/hadrons.html
index 77d311cd..1db5836f 100644
--- a/docs/pyerrors/input/hadrons.html
+++ b/docs/pyerrors/input/hadrons.html
@@ -176,393 +176,418 @@
74 two-point function. The gammas argument dominateds over meson.
75 idl : range
76 If specified only configurations in the given range are read in.
- 77 '''
- 78
- 79 files, idx = _get_files(path, filestem, idl)
- 80
- 81 tree = meson.rsplit('_')[0]
- 82 if gammas is not None:
- 83 h5file = h5py.File(path + '/' + files[0], "r")
- 84 found_meson = None
- 85 for key in h5file[tree].keys():
- 86 if gammas[0] == h5file[tree][key].attrs["gamma_snk"][0].decode() and h5file[tree][key].attrs["gamma_src"][0].decode() == gammas[1]:
- 87 found_meson = key
- 88 break
- 89 h5file.close()
- 90 if found_meson:
- 91 meson = found_meson
- 92 else:
- 93 raise Exception("Source Sink combination " + str(gammas) + " not found.")
- 94
- 95 corr_data = []
- 96 infos = []
- 97 for hd5_file in files:
- 98 h5file = h5py.File(path + '/' + hd5_file, "r")
- 99 if not tree + '/' + meson in h5file:
-100 raise Exception("Entry '" + meson + "' not contained in the files.")
-101 raw_data = h5file[tree + '/' + meson + '/corr']
-102 real_data = raw_data[:]["re"].astype(np.double)
-103 corr_data.append(real_data)
-104 if not infos:
-105 for k, i in h5file[tree + '/' + meson].attrs.items():
-106 infos.append(k + ': ' + i[0].decode())
-107 h5file.close()
-108 corr_data = np.array(corr_data)
-109
-110 l_obs = []
-111 for c in corr_data.T:
-112 l_obs.append(Obs([c], [ens_id], idl=[idx]))
-113
-114 corr = Corr(l_obs)
-115 corr.tag = r", ".join(infos)
-116 return corr
-117
+ 77
+ 78 Returns
+ 79 -------
+ 80 corr : Corr
+ 81 Correlator of the source sink combination in question.
+ 82 '''
+ 83
+ 84 files, idx = _get_files(path, filestem, idl)
+ 85
+ 86 tree = meson.rsplit('_')[0]
+ 87 if gammas is not None:
+ 88 h5file = h5py.File(path + '/' + files[0], "r")
+ 89 found_meson = None
+ 90 for key in h5file[tree].keys():
+ 91 if gammas[0] == h5file[tree][key].attrs["gamma_snk"][0].decode() and h5file[tree][key].attrs["gamma_src"][0].decode() == gammas[1]:
+ 92 found_meson = key
+ 93 break
+ 94 h5file.close()
+ 95 if found_meson:
+ 96 meson = found_meson
+ 97 else:
+ 98 raise Exception("Source Sink combination " + str(gammas) + " not found.")
+ 99
+100 corr_data = []
+101 infos = []
+102 for hd5_file in files:
+103 h5file = h5py.File(path + '/' + hd5_file, "r")
+104 if not tree + '/' + meson in h5file:
+105 raise Exception("Entry '" + meson + "' not contained in the files.")
+106 raw_data = h5file[tree + '/' + meson + '/corr']
+107 real_data = raw_data[:]["re"].astype(np.double)
+108 corr_data.append(real_data)
+109 if not infos:
+110 for k, i in h5file[tree + '/' + meson].attrs.items():
+111 infos.append(k + ': ' + i[0].decode())
+112 h5file.close()
+113 corr_data = np.array(corr_data)
+114
+115 l_obs = []
+116 for c in corr_data.T:
+117 l_obs.append(Obs([c], [ens_id], idl=[idx]))
118
-119def read_DistillationContraction_hd5(path, ens_id, diagrams=["direct"], idl=None):
-120 """Read hadrons DistillationContraction hdf5 files in given directory structure
-121
-122 Parameters
-123 -----------------
-124 path : str
-125 path to the directories to read
-126 ens_id : str
-127 name of the ensemble, required for internal bookkeeping
-128 diagrams : list
-129 List of strings of the diagrams to extract, e.g. ["direct", "box", "cross"].
-130 idl : range
-131 If specified only configurations in the given range are read in.
-132 """
-133
-134 res_dict = {}
-135
-136 directories, idx = _get_files(path, "data", idl)
+119 corr = Corr(l_obs)
+120 corr.tag = r", ".join(infos)
+121 return corr
+122
+123
+124def read_DistillationContraction_hd5(path, ens_id, diagrams=["direct"], idl=None):
+125 """Read hadrons DistillationContraction hdf5 files in given directory structure
+126
+127 Parameters
+128 -----------------
+129 path : str
+130 path to the directories to read
+131 ens_id : str
+132 name of the ensemble, required for internal bookkeeping
+133 diagrams : list
+134 List of strings of the diagrams to extract, e.g. ["direct", "box", "cross"].
+135 idl : range
+136 If specified only configurations in the given range are read in.
137
-138 explore_path = Path(path + "/" + directories[0])
-139
-140 for explore_file in explore_path.iterdir():
-141 if explore_file.is_file():
-142 stem = explore_file.with_suffix("").with_suffix("").as_posix().split("/")[-1]
-143 else:
-144 continue
+138 Returns
+139 -------
+140 result : dict
+141 extracted DistillationContration data
+142 """
+143
+144 res_dict = {}
145
-146 file_list = []
-147 for dir in directories:
-148 tmp_path = Path(path + "/" + dir)
-149 file_list.append((tmp_path / stem).as_posix() + tmp_path.suffix + ".h5")
-150
-151 corr_data = {}
-152
-153 for diagram in diagrams:
-154 corr_data[diagram] = []
+146 directories, idx = _get_files(path, "data", idl)
+147
+148 explore_path = Path(path + "/" + directories[0])
+149
+150 for explore_file in explore_path.iterdir():
+151 if explore_file.is_file():
+152 stem = explore_file.with_suffix("").with_suffix("").as_posix().split("/")[-1]
+153 else:
+154 continue
155
-156 for n_file, (hd5_file, n_traj) in enumerate(zip(file_list, list(idx))):
-157 h5file = h5py.File(hd5_file)
-158
-159 if n_file == 0:
-160 if h5file["DistillationContraction/Metadata"].attrs.get("TimeSources")[0].decode() != "0...":
-161 raise Exception("Routine is only implemented for files containing inversions on all timeslices.")
+156 file_list = []
+157 for dir in directories:
+158 tmp_path = Path(path + "/" + dir)
+159 file_list.append((tmp_path / stem).as_posix() + tmp_path.suffix + ".h5")
+160
+161 corr_data = {}
162
-163 Nt = h5file["DistillationContraction/Metadata"].attrs.get("Nt")[0]
-164
-165 identifier = []
-166 for in_file in range(len(h5file["DistillationContraction/Metadata/DmfInputFiles"].attrs.keys()) - 1):
-167 encoded_info = h5file["DistillationContraction/Metadata/DmfInputFiles"].attrs.get("DmfInputFiles_" + str(in_file))
-168 full_info = encoded_info[0].decode().split("/")[-1].replace(".h5", "").split("_")
-169 my_tuple = (full_info[0], full_info[1][1:], full_info[2], full_info[3])
-170 identifier.append(my_tuple)
-171 identifier = tuple(identifier)
-172 # "DistillationContraction/Metadata/DmfSuffix" contains info about different quarks, irrelevant in the SU(3) case.
-173
-174 for diagram in diagrams:
-175
-176 if diagram == "triangle" and "Identity" not in str(identifier):
-177 part = "im"
-178 else:
-179 part = "re"
-180
-181 real_data = np.zeros(Nt)
-182 for x0 in range(Nt):
-183 raw_data = h5file["DistillationContraction/Correlators/" + diagram + "/" + str(x0)][:][part].astype(np.double)
-184 real_data += np.roll(raw_data, -x0)
-185 real_data /= Nt
-186
-187 corr_data[diagram].append(real_data)
-188 h5file.close()
-189
-190 res_dict[str(identifier)] = {}
-191
-192 for diagram in diagrams:
-193
-194 tmp_data = np.array(corr_data[diagram])
-195
-196 l_obs = []
-197 for c in tmp_data.T:
-198 l_obs.append(Obs([c], [ens_id], idl=[idx]))
+163 for diagram in diagrams:
+164 corr_data[diagram] = []
+165
+166 for n_file, (hd5_file, n_traj) in enumerate(zip(file_list, list(idx))):
+167 h5file = h5py.File(hd5_file)
+168
+169 if n_file == 0:
+170 if h5file["DistillationContraction/Metadata"].attrs.get("TimeSources")[0].decode() != "0...":
+171 raise Exception("Routine is only implemented for files containing inversions on all timeslices.")
+172
+173 Nt = h5file["DistillationContraction/Metadata"].attrs.get("Nt")[0]
+174
+175 identifier = []
+176 for in_file in range(len(h5file["DistillationContraction/Metadata/DmfInputFiles"].attrs.keys()) - 1):
+177 encoded_info = h5file["DistillationContraction/Metadata/DmfInputFiles"].attrs.get("DmfInputFiles_" + str(in_file))
+178 full_info = encoded_info[0].decode().split("/")[-1].replace(".h5", "").split("_")
+179 my_tuple = (full_info[0], full_info[1][1:], full_info[2], full_info[3])
+180 identifier.append(my_tuple)
+181 identifier = tuple(identifier)
+182 # "DistillationContraction/Metadata/DmfSuffix" contains info about different quarks, irrelevant in the SU(3) case.
+183
+184 for diagram in diagrams:
+185
+186 if diagram == "triangle" and "Identity" not in str(identifier):
+187 part = "im"
+188 else:
+189 part = "re"
+190
+191 real_data = np.zeros(Nt)
+192 for x0 in range(Nt):
+193 raw_data = h5file["DistillationContraction/Correlators/" + diagram + "/" + str(x0)][:][part].astype(np.double)
+194 real_data += np.roll(raw_data, -x0)
+195 real_data /= Nt
+196
+197 corr_data[diagram].append(real_data)
+198 h5file.close()
199
-200 corr = Corr(l_obs)
-201 corr.tag = str(identifier)
-202
-203 res_dict[str(identifier)][diagram] = corr
-204
-205 return res_dict
-206
-207
-208class Npr_matrix(np.ndarray):
+200 res_dict[str(identifier)] = {}
+201
+202 for diagram in diagrams:
+203
+204 tmp_data = np.array(corr_data[diagram])
+205
+206 l_obs = []
+207 for c in tmp_data.T:
+208 l_obs.append(Obs([c], [ens_id], idl=[idx]))
209
-210 def __new__(cls, input_array, mom_in=None, mom_out=None):
-211 obj = np.asarray(input_array).view(cls)
-212 obj.mom_in = mom_in
-213 obj.mom_out = mom_out
-214 return obj
-215
-216 @property
-217 def g5H(self):
-218 """Gamma_5 hermitean conjugate
+210 corr = Corr(l_obs)
+211 corr.tag = str(identifier)
+212
+213 res_dict[str(identifier)][diagram] = corr
+214
+215 return res_dict
+216
+217
+218class Npr_matrix(np.ndarray):
219
-220 Uses the fact that the propagator is gamma5 hermitean, so just the
-221 in and out momenta of the propagator are exchanged.
-222 """
-223 return Npr_matrix(self,
-224 mom_in=self.mom_out,
-225 mom_out=self.mom_in)
-226
-227 def _propagate_mom(self, other, name):
-228 s_mom = getattr(self, name, None)
-229 o_mom = getattr(other, name, None)
-230 if s_mom is not None and o_mom is not None:
-231 if not np.allclose(s_mom, o_mom):
-232 raise Exception(name + ' does not match.')
-233 return o_mom if o_mom is not None else s_mom
-234
-235 def __matmul__(self, other):
-236 return self.__new__(Npr_matrix,
-237 super().__matmul__(other),
-238 self._propagate_mom(other, 'mom_in'),
-239 self._propagate_mom(other, 'mom_out'))
-240
-241 def __array_finalize__(self, obj):
-242 if obj is None:
-243 return
-244 self.mom_in = getattr(obj, 'mom_in', None)
-245 self.mom_out = getattr(obj, 'mom_out', None)
-246
-247
-248def read_ExternalLeg_hd5(path, filestem, ens_id, idl=None):
-249 """Read hadrons ExternalLeg hdf5 file and output an array of CObs
+220 def __new__(cls, input_array, mom_in=None, mom_out=None):
+221 obj = np.asarray(input_array).view(cls)
+222 obj.mom_in = mom_in
+223 obj.mom_out = mom_out
+224 return obj
+225
+226 @property
+227 def g5H(self):
+228 """Gamma_5 hermitean conjugate
+229
+230 Uses the fact that the propagator is gamma5 hermitean, so just the
+231 in and out momenta of the propagator are exchanged.
+232 """
+233 return Npr_matrix(self,
+234 mom_in=self.mom_out,
+235 mom_out=self.mom_in)
+236
+237 def _propagate_mom(self, other, name):
+238 s_mom = getattr(self, name, None)
+239 o_mom = getattr(other, name, None)
+240 if s_mom is not None and o_mom is not None:
+241 if not np.allclose(s_mom, o_mom):
+242 raise Exception(name + ' does not match.')
+243 return o_mom if o_mom is not None else s_mom
+244
+245 def __matmul__(self, other):
+246 return self.__new__(Npr_matrix,
+247 super().__matmul__(other),
+248 self._propagate_mom(other, 'mom_in'),
+249 self._propagate_mom(other, 'mom_out'))
250
-251 Parameters
-252 ----------
-253 path : str
-254 path to the files to read
-255 filestem : str
-256 namestem of the files to read
-257 ens_id : str
-258 name of the ensemble, required for internal bookkeeping
-259 idl : range
-260 If specified only configurations in the given range are read in.
-261 """
-262
-263 files, idx = _get_files(path, filestem, idl)
-264
-265 mom = None
-266
-267 corr_data = []
-268 for hd5_file in files:
-269 file = h5py.File(path + '/' + hd5_file, "r")
-270 raw_data = file['ExternalLeg/corr'][0][0].view('complex')
-271 corr_data.append(raw_data)
-272 if mom is None:
-273 mom = np.array(str(file['ExternalLeg/info'].attrs['pIn'])[3:-2].strip().split(), dtype=float)
-274 file.close()
-275 corr_data = np.array(corr_data)
-276
-277 rolled_array = np.rollaxis(corr_data, 0, 5)
-278
-279 matrix = np.empty((rolled_array.shape[:-1]), dtype=object)
-280 for si, sj, ci, cj in np.ndindex(rolled_array.shape[:-1]):
-281 real = Obs([rolled_array[si, sj, ci, cj].real], [ens_id], idl=[idx])
-282 imag = Obs([rolled_array[si, sj, ci, cj].imag], [ens_id], idl=[idx])
-283 matrix[si, sj, ci, cj] = CObs(real, imag)
-284
-285 return Npr_matrix(matrix, mom_in=mom)
-286
-287
-288def read_Bilinear_hd5(path, filestem, ens_id, idl=None):
-289 """Read hadrons Bilinear hdf5 file and output an array of CObs
-290
-291 Parameters
-292 ----------
-293 path : str
-294 path to the files to read
-295 filestem : str
-296 namestem of the files to read
-297 ens_id : str
-298 name of the ensemble, required for internal bookkeeping
-299 idl : range
-300 If specified only configurations in the given range are read in.
-301 """
+251 def __array_finalize__(self, obj):
+252 if obj is None:
+253 return
+254 self.mom_in = getattr(obj, 'mom_in', None)
+255 self.mom_out = getattr(obj, 'mom_out', None)
+256
+257
+258def read_ExternalLeg_hd5(path, filestem, ens_id, idl=None):
+259 """Read hadrons ExternalLeg hdf5 file and output an array of CObs
+260
+261 Parameters
+262 ----------
+263 path : str
+264 path to the files to read
+265 filestem : str
+266 namestem of the files to read
+267 ens_id : str
+268 name of the ensemble, required for internal bookkeeping
+269 idl : range
+270 If specified only configurations in the given range are read in.
+271
+272 Returns
+273 -------
+274 result : Npr_matrix
+275 read Cobs-matrix
+276 """
+277
+278 files, idx = _get_files(path, filestem, idl)
+279
+280 mom = None
+281
+282 corr_data = []
+283 for hd5_file in files:
+284 file = h5py.File(path + '/' + hd5_file, "r")
+285 raw_data = file['ExternalLeg/corr'][0][0].view('complex')
+286 corr_data.append(raw_data)
+287 if mom is None:
+288 mom = np.array(str(file['ExternalLeg/info'].attrs['pIn'])[3:-2].strip().split(), dtype=float)
+289 file.close()
+290 corr_data = np.array(corr_data)
+291
+292 rolled_array = np.rollaxis(corr_data, 0, 5)
+293
+294 matrix = np.empty((rolled_array.shape[:-1]), dtype=object)
+295 for si, sj, ci, cj in np.ndindex(rolled_array.shape[:-1]):
+296 real = Obs([rolled_array[si, sj, ci, cj].real], [ens_id], idl=[idx])
+297 imag = Obs([rolled_array[si, sj, ci, cj].imag], [ens_id], idl=[idx])
+298 matrix[si, sj, ci, cj] = CObs(real, imag)
+299
+300 return Npr_matrix(matrix, mom_in=mom)
+301
302
-303 files, idx = _get_files(path, filestem, idl)
-304
-305 mom_in = None
-306 mom_out = None
-307
-308 corr_data = {}
-309 for hd5_file in files:
-310 file = h5py.File(path + '/' + hd5_file, "r")
-311 for i in range(16):
-312 name = file['Bilinear/Bilinear_' + str(i) + '/info'].attrs['gamma'][0].decode('UTF-8')
-313 if name not in corr_data:
-314 corr_data[name] = []
-315 raw_data = file['Bilinear/Bilinear_' + str(i) + '/corr'][0][0].view('complex')
-316 corr_data[name].append(raw_data)
-317 if mom_in is None:
-318 mom_in = np.array(str(file['Bilinear/Bilinear_' + str(i) + '/info'].attrs['pIn'])[3:-2].strip().split(), dtype=float)
-319 if mom_out is None:
-320 mom_out = np.array(str(file['Bilinear/Bilinear_' + str(i) + '/info'].attrs['pOut'])[3:-2].strip().split(), dtype=float)
-321
-322 file.close()
-323
-324 result_dict = {}
-325
-326 for key, data in corr_data.items():
-327 local_data = np.array(data)
-328
-329 rolled_array = np.rollaxis(local_data, 0, 5)
-330
-331 matrix = np.empty((rolled_array.shape[:-1]), dtype=object)
-332 for si, sj, ci, cj in np.ndindex(rolled_array.shape[:-1]):
-333 real = Obs([rolled_array[si, sj, ci, cj].real], [ens_id], idl=[idx])
-334 imag = Obs([rolled_array[si, sj, ci, cj].imag], [ens_id], idl=[idx])
-335 matrix[si, sj, ci, cj] = CObs(real, imag)
-336
-337 result_dict[key] = Npr_matrix(matrix, mom_in=mom_in, mom_out=mom_out)
-338
-339 return result_dict
-340
+303def read_Bilinear_hd5(path, filestem, ens_id, idl=None):
+304 """Read hadrons Bilinear hdf5 file and output an array of CObs
+305
+306 Parameters
+307 ----------
+308 path : str
+309 path to the files to read
+310 filestem : str
+311 namestem of the files to read
+312 ens_id : str
+313 name of the ensemble, required for internal bookkeeping
+314 idl : range
+315 If specified only configurations in the given range are read in.
+316
+317 Returns
+318 -------
+319 result_dict: dict[Npr_matrix]
+320 extracted Bilinears
+321 """
+322
+323 files, idx = _get_files(path, filestem, idl)
+324
+325 mom_in = None
+326 mom_out = None
+327
+328 corr_data = {}
+329 for hd5_file in files:
+330 file = h5py.File(path + '/' + hd5_file, "r")
+331 for i in range(16):
+332 name = file['Bilinear/Bilinear_' + str(i) + '/info'].attrs['gamma'][0].decode('UTF-8')
+333 if name not in corr_data:
+334 corr_data[name] = []
+335 raw_data = file['Bilinear/Bilinear_' + str(i) + '/corr'][0][0].view('complex')
+336 corr_data[name].append(raw_data)
+337 if mom_in is None:
+338 mom_in = np.array(str(file['Bilinear/Bilinear_' + str(i) + '/info'].attrs['pIn'])[3:-2].strip().split(), dtype=float)
+339 if mom_out is None:
+340 mom_out = np.array(str(file['Bilinear/Bilinear_' + str(i) + '/info'].attrs['pOut'])[3:-2].strip().split(), dtype=float)
341
-342def read_Fourquark_hd5(path, filestem, ens_id, idl=None, vertices=["VA", "AV"]):
-343 """Read hadrons FourquarkFullyConnected hdf5 file and output an array of CObs
-344
-345 Parameters
-346 ----------
-347 path : str
-348 path to the files to read
-349 filestem : str
-350 namestem of the files to read
-351 ens_id : str
-352 name of the ensemble, required for internal bookkeeping
-353 idl : range
-354 If specified only configurations in the given range are read in.
-355 vertices : list
-356 Vertex functions to be extracted.
-357 """
+342 file.close()
+343
+344 result_dict = {}
+345
+346 for key, data in corr_data.items():
+347 local_data = np.array(data)
+348
+349 rolled_array = np.rollaxis(local_data, 0, 5)
+350
+351 matrix = np.empty((rolled_array.shape[:-1]), dtype=object)
+352 for si, sj, ci, cj in np.ndindex(rolled_array.shape[:-1]):
+353 real = Obs([rolled_array[si, sj, ci, cj].real], [ens_id], idl=[idx])
+354 imag = Obs([rolled_array[si, sj, ci, cj].imag], [ens_id], idl=[idx])
+355 matrix[si, sj, ci, cj] = CObs(real, imag)
+356
+357 result_dict[key] = Npr_matrix(matrix, mom_in=mom_in, mom_out=mom_out)
358
-359 files, idx = _get_files(path, filestem, idl)
+359 return result_dict
360
-361 mom_in = None
-362 mom_out = None
-363
-364 vertex_names = []
-365 for vertex in vertices:
-366 vertex_names += _get_lorentz_names(vertex)
-367
-368 corr_data = {}
-369
-370 tree = 'FourQuarkFullyConnected/FourQuarkFullyConnected_'
-371
-372 for hd5_file in files:
-373 file = h5py.File(path + '/' + hd5_file, "r")
-374
-375 for i in range(32):
-376 name = (file[tree + str(i) + '/info'].attrs['gammaA'][0].decode('UTF-8'), file[tree + str(i) + '/info'].attrs['gammaB'][0].decode('UTF-8'))
-377 if name in vertex_names:
-378 if name not in corr_data:
-379 corr_data[name] = []
-380 raw_data = file[tree + str(i) + '/corr'][0][0].view('complex')
-381 corr_data[name].append(raw_data)
-382 if mom_in is None:
-383 mom_in = np.array(str(file[tree + str(i) + '/info'].attrs['pIn'])[3:-2].strip().split(), dtype=float)
-384 if mom_out is None:
-385 mom_out = np.array(str(file[tree + str(i) + '/info'].attrs['pOut'])[3:-2].strip().split(), dtype=float)
-386
-387 file.close()
+361
+362def read_Fourquark_hd5(path, filestem, ens_id, idl=None, vertices=["VA", "AV"]):
+363 """Read hadrons FourquarkFullyConnected hdf5 file and output an array of CObs
+364
+365 Parameters
+366 ----------
+367 path : str
+368 path to the files to read
+369 filestem : str
+370 namestem of the files to read
+371 ens_id : str
+372 name of the ensemble, required for internal bookkeeping
+373 idl : range
+374 If specified only configurations in the given range are read in.
+375 vertices : list
+376 Vertex functions to be extracted.
+377
+378 Returns
+379 -------
+380 result_dict : dict
+381 extracted fourquark matrizes
+382 """
+383
+384 files, idx = _get_files(path, filestem, idl)
+385
+386 mom_in = None
+387 mom_out = None
388
-389 intermediate_dict = {}
-390
-391 for vertex in vertices:
-392 lorentz_names = _get_lorentz_names(vertex)
-393 for v_name in lorentz_names:
-394 if v_name in [('SigmaXY', 'SigmaZT'),
-395 ('SigmaXT', 'SigmaYZ'),
-396 ('SigmaYZ', 'SigmaXT'),
-397 ('SigmaZT', 'SigmaXY')]:
-398 sign = -1
-399 else:
-400 sign = 1
-401 if vertex not in intermediate_dict:
-402 intermediate_dict[vertex] = sign * np.array(corr_data[v_name])
-403 else:
-404 intermediate_dict[vertex] += sign * np.array(corr_data[v_name])
-405
-406 result_dict = {}
-407
-408 for key, data in intermediate_dict.items():
-409
-410 rolled_array = np.moveaxis(data, 0, 8)
+389 vertex_names = []
+390 for vertex in vertices:
+391 vertex_names += _get_lorentz_names(vertex)
+392
+393 corr_data = {}
+394
+395 tree = 'FourQuarkFullyConnected/FourQuarkFullyConnected_'
+396
+397 for hd5_file in files:
+398 file = h5py.File(path + '/' + hd5_file, "r")
+399
+400 for i in range(32):
+401 name = (file[tree + str(i) + '/info'].attrs['gammaA'][0].decode('UTF-8'), file[tree + str(i) + '/info'].attrs['gammaB'][0].decode('UTF-8'))
+402 if name in vertex_names:
+403 if name not in corr_data:
+404 corr_data[name] = []
+405 raw_data = file[tree + str(i) + '/corr'][0][0].view('complex')
+406 corr_data[name].append(raw_data)
+407 if mom_in is None:
+408 mom_in = np.array(str(file[tree + str(i) + '/info'].attrs['pIn'])[3:-2].strip().split(), dtype=float)
+409 if mom_out is None:
+410 mom_out = np.array(str(file[tree + str(i) + '/info'].attrs['pOut'])[3:-2].strip().split(), dtype=float)
411
-412 matrix = np.empty((rolled_array.shape[:-1]), dtype=object)
-413 for index in np.ndindex(rolled_array.shape[:-1]):
-414 real = Obs([rolled_array[index].real], [ens_id], idl=[idx])
-415 imag = Obs([rolled_array[index].imag], [ens_id], idl=[idx])
-416 matrix[index] = CObs(real, imag)
-417
-418 result_dict[key] = Npr_matrix(matrix, mom_in=mom_in, mom_out=mom_out)
-419
-420 return result_dict
-421
-422
-423def _get_lorentz_names(name):
-424 lorentz_index = ['X', 'Y', 'Z', 'T']
-425
-426 res = []
-427
-428 if name == "TT":
-429 for i in range(4):
-430 for j in range(i + 1, 4):
-431 res.append(("Sigma" + lorentz_index[i] + lorentz_index[j], "Sigma" + lorentz_index[i] + lorentz_index[j]))
-432 return res
-433
-434 if name == "TTtilde":
-435 for i in range(4):
-436 for j in range(i + 1, 4):
-437 for k in range(4):
-438 for o in range(k + 1, 4):
-439 fac = epsilon_tensor_rank4(i, j, k, o)
-440 if not np.isclose(fac, 0.0):
-441 res.append(("Sigma" + lorentz_index[i] + lorentz_index[j], "Sigma" + lorentz_index[k] + lorentz_index[o]))
-442 return res
-443
-444 assert len(name) == 2
-445
-446 if 'S' in name or 'P' in name:
-447 if not set(name) <= set(['S', 'P']):
-448 raise Exception("'" + name + "' is not a Lorentz scalar")
-449
-450 g_names = {'S': 'Identity',
-451 'P': 'Gamma5'}
+412 file.close()
+413
+414 intermediate_dict = {}
+415
+416 for vertex in vertices:
+417 lorentz_names = _get_lorentz_names(vertex)
+418 for v_name in lorentz_names:
+419 if v_name in [('SigmaXY', 'SigmaZT'),
+420 ('SigmaXT', 'SigmaYZ'),
+421 ('SigmaYZ', 'SigmaXT'),
+422 ('SigmaZT', 'SigmaXY')]:
+423 sign = -1
+424 else:
+425 sign = 1
+426 if vertex not in intermediate_dict:
+427 intermediate_dict[vertex] = sign * np.array(corr_data[v_name])
+428 else:
+429 intermediate_dict[vertex] += sign * np.array(corr_data[v_name])
+430
+431 result_dict = {}
+432
+433 for key, data in intermediate_dict.items():
+434
+435 rolled_array = np.moveaxis(data, 0, 8)
+436
+437 matrix = np.empty((rolled_array.shape[:-1]), dtype=object)
+438 for index in np.ndindex(rolled_array.shape[:-1]):
+439 real = Obs([rolled_array[index].real], [ens_id], idl=[idx])
+440 imag = Obs([rolled_array[index].imag], [ens_id], idl=[idx])
+441 matrix[index] = CObs(real, imag)
+442
+443 result_dict[key] = Npr_matrix(matrix, mom_in=mom_in, mom_out=mom_out)
+444
+445 return result_dict
+446
+447
+448def _get_lorentz_names(name):
+449 lorentz_index = ['X', 'Y', 'Z', 'T']
+450
+451 res = []
452
-453 res.append((g_names[name[0]], g_names[name[1]]))
-454
-455 else:
-456 if not set(name) <= set(['V', 'A']):
-457 raise Exception("'" + name + "' is not a Lorentz scalar")
+453 if name == "TT":
+454 for i in range(4):
+455 for j in range(i + 1, 4):
+456 res.append(("Sigma" + lorentz_index[i] + lorentz_index[j], "Sigma" + lorentz_index[i] + lorentz_index[j]))
+457 return res
458
-459 for ind in lorentz_index:
-460 res.append(('Gamma' + ind + (name[0] == 'A') * 'Gamma5',
-461 'Gamma' + ind + (name[1] == 'A') * 'Gamma5'))
-462
-463 return res
+459 if name == "TTtilde":
+460 for i in range(4):
+461 for j in range(i + 1, 4):
+462 for k in range(4):
+463 for o in range(k + 1, 4):
+464 fac = epsilon_tensor_rank4(i, j, k, o)
+465 if not np.isclose(fac, 0.0):
+466 res.append(("Sigma" + lorentz_index[i] + lorentz_index[j], "Sigma" + lorentz_index[k] + lorentz_index[o]))
+467 return res
+468
+469 assert len(name) == 2
+470
+471 if 'S' in name or 'P' in name:
+472 if not set(name) <= set(['S', 'P']):
+473 raise Exception("'" + name + "' is not a Lorentz scalar")
+474
+475 g_names = {'S': 'Identity',
+476 'P': 'Gamma5'}
+477
+478 res.append((g_names[name[0]], g_names[name[1]]))
+479
+480 else:
+481 if not set(name) <= set(['V', 'A']):
+482 raise Exception("'" + name + "' is not a Lorentz scalar")
+483
+484 for ind in lorentz_index:
+485 res.append(('Gamma' + ind + (name[0] == 'A') * 'Gamma5',
+486 'Gamma' + ind + (name[1] == 'A') * 'Gamma5'))
+487
+488 return res
@@ -599,46 +624,51 @@
75 two-point function. The gammas argument dominateds over meson.
76 idl : range
77 If specified only configurations in the given range are read in.
- 78 '''
- 79
- 80 files, idx = _get_files(path, filestem, idl)
- 81
- 82 tree = meson.rsplit('_')[0]
- 83 if gammas is not None:
- 84 h5file = h5py.File(path + '/' + files[0], "r")
- 85 found_meson = None
- 86 for key in h5file[tree].keys():
- 87 if gammas[0] == h5file[tree][key].attrs["gamma_snk"][0].decode() and h5file[tree][key].attrs["gamma_src"][0].decode() == gammas[1]:
- 88 found_meson = key
- 89 break
- 90 h5file.close()
- 91 if found_meson:
- 92 meson = found_meson
- 93 else:
- 94 raise Exception("Source Sink combination " + str(gammas) + " not found.")
- 95
- 96 corr_data = []
- 97 infos = []
- 98 for hd5_file in files:
- 99 h5file = h5py.File(path + '/' + hd5_file, "r")
-100 if not tree + '/' + meson in h5file:
-101 raise Exception("Entry '" + meson + "' not contained in the files.")
-102 raw_data = h5file[tree + '/' + meson + '/corr']
-103 real_data = raw_data[:]["re"].astype(np.double)
-104 corr_data.append(real_data)
-105 if not infos:
-106 for k, i in h5file[tree + '/' + meson].attrs.items():
-107 infos.append(k + ': ' + i[0].decode())
-108 h5file.close()
-109 corr_data = np.array(corr_data)
-110
-111 l_obs = []
-112 for c in corr_data.T:
-113 l_obs.append(Obs([c], [ens_id], idl=[idx]))
-114
-115 corr = Corr(l_obs)
-116 corr.tag = r", ".join(infos)
-117 return corr
+ 78
+ 79 Returns
+ 80 -------
+ 81 corr : Corr
+ 82 Correlator of the source sink combination in question.
+ 83 '''
+ 84
+ 85 files, idx = _get_files(path, filestem, idl)
+ 86
+ 87 tree = meson.rsplit('_')[0]
+ 88 if gammas is not None:
+ 89 h5file = h5py.File(path + '/' + files[0], "r")
+ 90 found_meson = None
+ 91 for key in h5file[tree].keys():
+ 92 if gammas[0] == h5file[tree][key].attrs["gamma_snk"][0].decode() and h5file[tree][key].attrs["gamma_src"][0].decode() == gammas[1]:
+ 93 found_meson = key
+ 94 break
+ 95 h5file.close()
+ 96 if found_meson:
+ 97 meson = found_meson
+ 98 else:
+ 99 raise Exception("Source Sink combination " + str(gammas) + " not found.")
+100
+101 corr_data = []
+102 infos = []
+103 for hd5_file in files:
+104 h5file = h5py.File(path + '/' + hd5_file, "r")
+105 if not tree + '/' + meson in h5file:
+106 raise Exception("Entry '" + meson + "' not contained in the files.")
+107 raw_data = h5file[tree + '/' + meson + '/corr']
+108 real_data = raw_data[:]["re"].astype(np.double)
+109 corr_data.append(real_data)
+110 if not infos:
+111 for k, i in h5file[tree + '/' + meson].attrs.items():
+112 infos.append(k + ': ' + i[0].decode())
+113 h5file.close()
+114 corr_data = np.array(corr_data)
+115
+116 l_obs = []
+117 for c in corr_data.T:
+118 l_obs.append(Obs([c], [ens_id], idl=[idx]))
+119
+120 corr = Corr(l_obs)
+121 corr.tag = r", ".join(infos)
+122 return corr
@@ -664,6 +694,13 @@ two-point function. The gammas argument dominateds over meson.
idl (range):
If specified only configurations in the given range are read in.
+
+Returns
+
+
+- corr (Corr):
+Correlator of the source sink combination in question.
+
@@ -679,93 +716,98 @@ If specified only configurations in the given range are read in.
- 120def read_DistillationContraction_hd5(path, ens_id, diagrams=["direct"], idl=None):
-121 """Read hadrons DistillationContraction hdf5 files in given directory structure
-122
-123 Parameters
-124 -----------------
-125 path : str
-126 path to the directories to read
-127 ens_id : str
-128 name of the ensemble, required for internal bookkeeping
-129 diagrams : list
-130 List of strings of the diagrams to extract, e.g. ["direct", "box", "cross"].
-131 idl : range
-132 If specified only configurations in the given range are read in.
-133 """
-134
-135 res_dict = {}
-136
-137 directories, idx = _get_files(path, "data", idl)
+ 125def read_DistillationContraction_hd5(path, ens_id, diagrams=["direct"], idl=None):
+126 """Read hadrons DistillationContraction hdf5 files in given directory structure
+127
+128 Parameters
+129 -----------------
+130 path : str
+131 path to the directories to read
+132 ens_id : str
+133 name of the ensemble, required for internal bookkeeping
+134 diagrams : list
+135 List of strings of the diagrams to extract, e.g. ["direct", "box", "cross"].
+136 idl : range
+137 If specified only configurations in the given range are read in.
138
-139 explore_path = Path(path + "/" + directories[0])
-140
-141 for explore_file in explore_path.iterdir():
-142 if explore_file.is_file():
-143 stem = explore_file.with_suffix("").with_suffix("").as_posix().split("/")[-1]
-144 else:
-145 continue
+139 Returns
+140 -------
+141 result : dict
+142 extracted DistillationContration data
+143 """
+144
+145 res_dict = {}
146
-147 file_list = []
-148 for dir in directories:
-149 tmp_path = Path(path + "/" + dir)
-150 file_list.append((tmp_path / stem).as_posix() + tmp_path.suffix + ".h5")
-151
-152 corr_data = {}
-153
-154 for diagram in diagrams:
-155 corr_data[diagram] = []
+147 directories, idx = _get_files(path, "data", idl)
+148
+149 explore_path = Path(path + "/" + directories[0])
+150
+151 for explore_file in explore_path.iterdir():
+152 if explore_file.is_file():
+153 stem = explore_file.with_suffix("").with_suffix("").as_posix().split("/")[-1]
+154 else:
+155 continue
156
-157 for n_file, (hd5_file, n_traj) in enumerate(zip(file_list, list(idx))):
-158 h5file = h5py.File(hd5_file)
-159
-160 if n_file == 0:
-161 if h5file["DistillationContraction/Metadata"].attrs.get("TimeSources")[0].decode() != "0...":
-162 raise Exception("Routine is only implemented for files containing inversions on all timeslices.")
+157 file_list = []
+158 for dir in directories:
+159 tmp_path = Path(path + "/" + dir)
+160 file_list.append((tmp_path / stem).as_posix() + tmp_path.suffix + ".h5")
+161
+162 corr_data = {}
163
-164 Nt = h5file["DistillationContraction/Metadata"].attrs.get("Nt")[0]
-165
-166 identifier = []
-167 for in_file in range(len(h5file["DistillationContraction/Metadata/DmfInputFiles"].attrs.keys()) - 1):
-168 encoded_info = h5file["DistillationContraction/Metadata/DmfInputFiles"].attrs.get("DmfInputFiles_" + str(in_file))
-169 full_info = encoded_info[0].decode().split("/")[-1].replace(".h5", "").split("_")
-170 my_tuple = (full_info[0], full_info[1][1:], full_info[2], full_info[3])
-171 identifier.append(my_tuple)
-172 identifier = tuple(identifier)
-173 # "DistillationContraction/Metadata/DmfSuffix" contains info about different quarks, irrelevant in the SU(3) case.
-174
-175 for diagram in diagrams:
-176
-177 if diagram == "triangle" and "Identity" not in str(identifier):
-178 part = "im"
-179 else:
-180 part = "re"
-181
-182 real_data = np.zeros(Nt)
-183 for x0 in range(Nt):
-184 raw_data = h5file["DistillationContraction/Correlators/" + diagram + "/" + str(x0)][:][part].astype(np.double)
-185 real_data += np.roll(raw_data, -x0)
-186 real_data /= Nt
-187
-188 corr_data[diagram].append(real_data)
-189 h5file.close()
-190
-191 res_dict[str(identifier)] = {}
-192
-193 for diagram in diagrams:
-194
-195 tmp_data = np.array(corr_data[diagram])
-196
-197 l_obs = []
-198 for c in tmp_data.T:
-199 l_obs.append(Obs([c], [ens_id], idl=[idx]))
+164 for diagram in diagrams:
+165 corr_data[diagram] = []
+166
+167 for n_file, (hd5_file, n_traj) in enumerate(zip(file_list, list(idx))):
+168 h5file = h5py.File(hd5_file)
+169
+170 if n_file == 0:
+171 if h5file["DistillationContraction/Metadata"].attrs.get("TimeSources")[0].decode() != "0...":
+172 raise Exception("Routine is only implemented for files containing inversions on all timeslices.")
+173
+174 Nt = h5file["DistillationContraction/Metadata"].attrs.get("Nt")[0]
+175
+176 identifier = []
+177 for in_file in range(len(h5file["DistillationContraction/Metadata/DmfInputFiles"].attrs.keys()) - 1):
+178 encoded_info = h5file["DistillationContraction/Metadata/DmfInputFiles"].attrs.get("DmfInputFiles_" + str(in_file))
+179 full_info = encoded_info[0].decode().split("/")[-1].replace(".h5", "").split("_")
+180 my_tuple = (full_info[0], full_info[1][1:], full_info[2], full_info[3])
+181 identifier.append(my_tuple)
+182 identifier = tuple(identifier)
+183 # "DistillationContraction/Metadata/DmfSuffix" contains info about different quarks, irrelevant in the SU(3) case.
+184
+185 for diagram in diagrams:
+186
+187 if diagram == "triangle" and "Identity" not in str(identifier):
+188 part = "im"
+189 else:
+190 part = "re"
+191
+192 real_data = np.zeros(Nt)
+193 for x0 in range(Nt):
+194 raw_data = h5file["DistillationContraction/Correlators/" + diagram + "/" + str(x0)][:][part].astype(np.double)
+195 real_data += np.roll(raw_data, -x0)
+196 real_data /= Nt
+197
+198 corr_data[diagram].append(real_data)
+199 h5file.close()
200
-201 corr = Corr(l_obs)
-202 corr.tag = str(identifier)
-203
-204 res_dict[str(identifier)][diagram] = corr
-205
-206 return res_dict
+201 res_dict[str(identifier)] = {}
+202
+203 for diagram in diagrams:
+204
+205 tmp_data = np.array(corr_data[diagram])
+206
+207 l_obs = []
+208 for c in tmp_data.T:
+209 l_obs.append(Obs([c], [ens_id], idl=[idx]))
+210
+211 corr = Corr(l_obs)
+212 corr.tag = str(identifier)
+213
+214 res_dict[str(identifier)][diagram] = corr
+215
+216 return res_dict
@@ -783,6 +825,13 @@ List of strings of the diagrams to extract, e.g. ["direct", "box", "cross"].idl (range):
If specified only configurations in the given range are read in.
+
+Returns
+
+
+- result (dict):
+extracted DistillationContration data
+
@@ -798,44 +847,44 @@ If specified only configurations in the given range are read in.
- 209class Npr_matrix(np.ndarray):
-210
-211 def __new__(cls, input_array, mom_in=None, mom_out=None):
-212 obj = np.asarray(input_array).view(cls)
-213 obj.mom_in = mom_in
-214 obj.mom_out = mom_out
-215 return obj
-216
-217 @property
-218 def g5H(self):
-219 """Gamma_5 hermitean conjugate
+ 219class Npr_matrix(np.ndarray):
220
-221 Uses the fact that the propagator is gamma5 hermitean, so just the
-222 in and out momenta of the propagator are exchanged.
-223 """
-224 return Npr_matrix(self,
-225 mom_in=self.mom_out,
-226 mom_out=self.mom_in)
-227
-228 def _propagate_mom(self, other, name):
-229 s_mom = getattr(self, name, None)
-230 o_mom = getattr(other, name, None)
-231 if s_mom is not None and o_mom is not None:
-232 if not np.allclose(s_mom, o_mom):
-233 raise Exception(name + ' does not match.')
-234 return o_mom if o_mom is not None else s_mom
-235
-236 def __matmul__(self, other):
-237 return self.__new__(Npr_matrix,
-238 super().__matmul__(other),
-239 self._propagate_mom(other, 'mom_in'),
-240 self._propagate_mom(other, 'mom_out'))
-241
-242 def __array_finalize__(self, obj):
-243 if obj is None:
-244 return
-245 self.mom_in = getattr(obj, 'mom_in', None)
-246 self.mom_out = getattr(obj, 'mom_out', None)
+221 def __new__(cls, input_array, mom_in=None, mom_out=None):
+222 obj = np.asarray(input_array).view(cls)
+223 obj.mom_in = mom_in
+224 obj.mom_out = mom_out
+225 return obj
+226
+227 @property
+228 def g5H(self):
+229 """Gamma_5 hermitean conjugate
+230
+231 Uses the fact that the propagator is gamma5 hermitean, so just the
+232 in and out momenta of the propagator are exchanged.
+233 """
+234 return Npr_matrix(self,
+235 mom_in=self.mom_out,
+236 mom_out=self.mom_in)
+237
+238 def _propagate_mom(self, other, name):
+239 s_mom = getattr(self, name, None)
+240 o_mom = getattr(other, name, None)
+241 if s_mom is not None and o_mom is not None:
+242 if not np.allclose(s_mom, o_mom):
+243 raise Exception(name + ' does not match.')
+244 return o_mom if o_mom is not None else s_mom
+245
+246 def __matmul__(self, other):
+247 return self.__new__(Npr_matrix,
+248 super().__matmul__(other),
+249 self._propagate_mom(other, 'mom_in'),
+250 self._propagate_mom(other, 'mom_out'))
+251
+252 def __array_finalize__(self, obj):
+253 if obj is None:
+254 return
+255 self.mom_in = getattr(obj, 'mom_in', None)
+256 self.mom_out = getattr(obj, 'mom_out', None)
@@ -1090,44 +1139,49 @@ in and out momenta of the propagator are exchanged.
- 249def read_ExternalLeg_hd5(path, filestem, ens_id, idl=None):
-250 """Read hadrons ExternalLeg hdf5 file and output an array of CObs
-251
-252 Parameters
-253 ----------
-254 path : str
-255 path to the files to read
-256 filestem : str
-257 namestem of the files to read
-258 ens_id : str
-259 name of the ensemble, required for internal bookkeeping
-260 idl : range
-261 If specified only configurations in the given range are read in.
-262 """
-263
-264 files, idx = _get_files(path, filestem, idl)
-265
-266 mom = None
-267
-268 corr_data = []
-269 for hd5_file in files:
-270 file = h5py.File(path + '/' + hd5_file, "r")
-271 raw_data = file['ExternalLeg/corr'][0][0].view('complex')
-272 corr_data.append(raw_data)
-273 if mom is None:
-274 mom = np.array(str(file['ExternalLeg/info'].attrs['pIn'])[3:-2].strip().split(), dtype=float)
-275 file.close()
-276 corr_data = np.array(corr_data)
-277
-278 rolled_array = np.rollaxis(corr_data, 0, 5)
-279
-280 matrix = np.empty((rolled_array.shape[:-1]), dtype=object)
-281 for si, sj, ci, cj in np.ndindex(rolled_array.shape[:-1]):
-282 real = Obs([rolled_array[si, sj, ci, cj].real], [ens_id], idl=[idx])
-283 imag = Obs([rolled_array[si, sj, ci, cj].imag], [ens_id], idl=[idx])
-284 matrix[si, sj, ci, cj] = CObs(real, imag)
-285
-286 return Npr_matrix(matrix, mom_in=mom)
+ 259def read_ExternalLeg_hd5(path, filestem, ens_id, idl=None):
+260 """Read hadrons ExternalLeg hdf5 file and output an array of CObs
+261
+262 Parameters
+263 ----------
+264 path : str
+265 path to the files to read
+266 filestem : str
+267 namestem of the files to read
+268 ens_id : str
+269 name of the ensemble, required for internal bookkeeping
+270 idl : range
+271 If specified only configurations in the given range are read in.
+272
+273 Returns
+274 -------
+275 result : Npr_matrix
+276 read Cobs-matrix
+277 """
+278
+279 files, idx = _get_files(path, filestem, idl)
+280
+281 mom = None
+282
+283 corr_data = []
+284 for hd5_file in files:
+285 file = h5py.File(path + '/' + hd5_file, "r")
+286 raw_data = file['ExternalLeg/corr'][0][0].view('complex')
+287 corr_data.append(raw_data)
+288 if mom is None:
+289 mom = np.array(str(file['ExternalLeg/info'].attrs['pIn'])[3:-2].strip().split(), dtype=float)
+290 file.close()
+291 corr_data = np.array(corr_data)
+292
+293 rolled_array = np.rollaxis(corr_data, 0, 5)
+294
+295 matrix = np.empty((rolled_array.shape[:-1]), dtype=object)
+296 for si, sj, ci, cj in np.ndindex(rolled_array.shape[:-1]):
+297 real = Obs([rolled_array[si, sj, ci, cj].real], [ens_id], idl=[idx])
+298 imag = Obs([rolled_array[si, sj, ci, cj].imag], [ens_id], idl=[idx])
+299 matrix[si, sj, ci, cj] = CObs(real, imag)
+300
+301 return Npr_matrix(matrix, mom_in=mom)
@@ -1145,6 +1199,13 @@ name of the ensemble, required for internal bookkeeping
idl (range):
If specified only configurations in the given range are read in.
+
+Returns
+
+
+- result (Npr_matrix):
+read Cobs-matrix
+
@@ -1160,58 +1221,63 @@ If specified only configurations in the given range are read in.
- 289def read_Bilinear_hd5(path, filestem, ens_id, idl=None):
-290 """Read hadrons Bilinear hdf5 file and output an array of CObs
-291
-292 Parameters
-293 ----------
-294 path : str
-295 path to the files to read
-296 filestem : str
-297 namestem of the files to read
-298 ens_id : str
-299 name of the ensemble, required for internal bookkeeping
-300 idl : range
-301 If specified only configurations in the given range are read in.
-302 """
-303
-304 files, idx = _get_files(path, filestem, idl)
-305
-306 mom_in = None
-307 mom_out = None
-308
-309 corr_data = {}
-310 for hd5_file in files:
-311 file = h5py.File(path + '/' + hd5_file, "r")
-312 for i in range(16):
-313 name = file['Bilinear/Bilinear_' + str(i) + '/info'].attrs['gamma'][0].decode('UTF-8')
-314 if name not in corr_data:
-315 corr_data[name] = []
-316 raw_data = file['Bilinear/Bilinear_' + str(i) + '/corr'][0][0].view('complex')
-317 corr_data[name].append(raw_data)
-318 if mom_in is None:
-319 mom_in = np.array(str(file['Bilinear/Bilinear_' + str(i) + '/info'].attrs['pIn'])[3:-2].strip().split(), dtype=float)
-320 if mom_out is None:
-321 mom_out = np.array(str(file['Bilinear/Bilinear_' + str(i) + '/info'].attrs['pOut'])[3:-2].strip().split(), dtype=float)
-322
-323 file.close()
-324
-325 result_dict = {}
-326
-327 for key, data in corr_data.items():
-328 local_data = np.array(data)
-329
-330 rolled_array = np.rollaxis(local_data, 0, 5)
-331
-332 matrix = np.empty((rolled_array.shape[:-1]), dtype=object)
-333 for si, sj, ci, cj in np.ndindex(rolled_array.shape[:-1]):
-334 real = Obs([rolled_array[si, sj, ci, cj].real], [ens_id], idl=[idx])
-335 imag = Obs([rolled_array[si, sj, ci, cj].imag], [ens_id], idl=[idx])
-336 matrix[si, sj, ci, cj] = CObs(real, imag)
-337
-338 result_dict[key] = Npr_matrix(matrix, mom_in=mom_in, mom_out=mom_out)
-339
-340 return result_dict
+ 304def read_Bilinear_hd5(path, filestem, ens_id, idl=None):
+305 """Read hadrons Bilinear hdf5 file and output an array of CObs
+306
+307 Parameters
+308 ----------
+309 path : str
+310 path to the files to read
+311 filestem : str
+312 namestem of the files to read
+313 ens_id : str
+314 name of the ensemble, required for internal bookkeeping
+315 idl : range
+316 If specified only configurations in the given range are read in.
+317
+318 Returns
+319 -------
+320 result_dict: dict[Npr_matrix]
+321 extracted Bilinears
+322 """
+323
+324 files, idx = _get_files(path, filestem, idl)
+325
+326 mom_in = None
+327 mom_out = None
+328
+329 corr_data = {}
+330 for hd5_file in files:
+331 file = h5py.File(path + '/' + hd5_file, "r")
+332 for i in range(16):
+333 name = file['Bilinear/Bilinear_' + str(i) + '/info'].attrs['gamma'][0].decode('UTF-8')
+334 if name not in corr_data:
+335 corr_data[name] = []
+336 raw_data = file['Bilinear/Bilinear_' + str(i) + '/corr'][0][0].view('complex')
+337 corr_data[name].append(raw_data)
+338 if mom_in is None:
+339 mom_in = np.array(str(file['Bilinear/Bilinear_' + str(i) + '/info'].attrs['pIn'])[3:-2].strip().split(), dtype=float)
+340 if mom_out is None:
+341 mom_out = np.array(str(file['Bilinear/Bilinear_' + str(i) + '/info'].attrs['pOut'])[3:-2].strip().split(), dtype=float)
+342
+343 file.close()
+344
+345 result_dict = {}
+346
+347 for key, data in corr_data.items():
+348 local_data = np.array(data)
+349
+350 rolled_array = np.rollaxis(local_data, 0, 5)
+351
+352 matrix = np.empty((rolled_array.shape[:-1]), dtype=object)
+353 for si, sj, ci, cj in np.ndindex(rolled_array.shape[:-1]):
+354 real = Obs([rolled_array[si, sj, ci, cj].real], [ens_id], idl=[idx])
+355 imag = Obs([rolled_array[si, sj, ci, cj].imag], [ens_id], idl=[idx])
+356 matrix[si, sj, ci, cj] = CObs(real, imag)
+357
+358 result_dict[key] = Npr_matrix(matrix, mom_in=mom_in, mom_out=mom_out)
+359
+360 return result_dict
@@ -1229,6 +1295,13 @@ name of the ensemble, required for internal bookkeeping
idl (range):
If specified only configurations in the given range are read in.
+
+Returns
+
+
+- result_dict (dict[Npr_matrix]):
+extracted Bilinears
+
@@ -1244,85 +1317,90 @@ If specified only configurations in the given range are read in.
- 343def read_Fourquark_hd5(path, filestem, ens_id, idl=None, vertices=["VA", "AV"]):
-344 """Read hadrons FourquarkFullyConnected hdf5 file and output an array of CObs
-345
-346 Parameters
-347 ----------
-348 path : str
-349 path to the files to read
-350 filestem : str
-351 namestem of the files to read
-352 ens_id : str
-353 name of the ensemble, required for internal bookkeeping
-354 idl : range
-355 If specified only configurations in the given range are read in.
-356 vertices : list
-357 Vertex functions to be extracted.
-358 """
-359
-360 files, idx = _get_files(path, filestem, idl)
-361
-362 mom_in = None
-363 mom_out = None
-364
-365 vertex_names = []
-366 for vertex in vertices:
-367 vertex_names += _get_lorentz_names(vertex)
-368
-369 corr_data = {}
-370
-371 tree = 'FourQuarkFullyConnected/FourQuarkFullyConnected_'
-372
-373 for hd5_file in files:
-374 file = h5py.File(path + '/' + hd5_file, "r")
-375
-376 for i in range(32):
-377 name = (file[tree + str(i) + '/info'].attrs['gammaA'][0].decode('UTF-8'), file[tree + str(i) + '/info'].attrs['gammaB'][0].decode('UTF-8'))
-378 if name in vertex_names:
-379 if name not in corr_data:
-380 corr_data[name] = []
-381 raw_data = file[tree + str(i) + '/corr'][0][0].view('complex')
-382 corr_data[name].append(raw_data)
-383 if mom_in is None:
-384 mom_in = np.array(str(file[tree + str(i) + '/info'].attrs['pIn'])[3:-2].strip().split(), dtype=float)
-385 if mom_out is None:
-386 mom_out = np.array(str(file[tree + str(i) + '/info'].attrs['pOut'])[3:-2].strip().split(), dtype=float)
-387
-388 file.close()
+ 363def read_Fourquark_hd5(path, filestem, ens_id, idl=None, vertices=["VA", "AV"]):
+364 """Read hadrons FourquarkFullyConnected hdf5 file and output an array of CObs
+365
+366 Parameters
+367 ----------
+368 path : str
+369 path to the files to read
+370 filestem : str
+371 namestem of the files to read
+372 ens_id : str
+373 name of the ensemble, required for internal bookkeeping
+374 idl : range
+375 If specified only configurations in the given range are read in.
+376 vertices : list
+377 Vertex functions to be extracted.
+378
+379 Returns
+380 -------
+381 result_dict : dict
+382 extracted fourquark matrizes
+383 """
+384
+385 files, idx = _get_files(path, filestem, idl)
+386
+387 mom_in = None
+388 mom_out = None
389
-390 intermediate_dict = {}
-391
-392 for vertex in vertices:
-393 lorentz_names = _get_lorentz_names(vertex)
-394 for v_name in lorentz_names:
-395 if v_name in [('SigmaXY', 'SigmaZT'),
-396 ('SigmaXT', 'SigmaYZ'),
-397 ('SigmaYZ', 'SigmaXT'),
-398 ('SigmaZT', 'SigmaXY')]:
-399 sign = -1
-400 else:
-401 sign = 1
-402 if vertex not in intermediate_dict:
-403 intermediate_dict[vertex] = sign * np.array(corr_data[v_name])
-404 else:
-405 intermediate_dict[vertex] += sign * np.array(corr_data[v_name])
-406
-407 result_dict = {}
-408
-409 for key, data in intermediate_dict.items():
-410
-411 rolled_array = np.moveaxis(data, 0, 8)
+390 vertex_names = []
+391 for vertex in vertices:
+392 vertex_names += _get_lorentz_names(vertex)
+393
+394 corr_data = {}
+395
+396 tree = 'FourQuarkFullyConnected/FourQuarkFullyConnected_'
+397
+398 for hd5_file in files:
+399 file = h5py.File(path + '/' + hd5_file, "r")
+400
+401 for i in range(32):
+402 name = (file[tree + str(i) + '/info'].attrs['gammaA'][0].decode('UTF-8'), file[tree + str(i) + '/info'].attrs['gammaB'][0].decode('UTF-8'))
+403 if name in vertex_names:
+404 if name not in corr_data:
+405 corr_data[name] = []
+406 raw_data = file[tree + str(i) + '/corr'][0][0].view('complex')
+407 corr_data[name].append(raw_data)
+408 if mom_in is None:
+409 mom_in = np.array(str(file[tree + str(i) + '/info'].attrs['pIn'])[3:-2].strip().split(), dtype=float)
+410 if mom_out is None:
+411 mom_out = np.array(str(file[tree + str(i) + '/info'].attrs['pOut'])[3:-2].strip().split(), dtype=float)
412
-413 matrix = np.empty((rolled_array.shape[:-1]), dtype=object)
-414 for index in np.ndindex(rolled_array.shape[:-1]):
-415 real = Obs([rolled_array[index].real], [ens_id], idl=[idx])
-416 imag = Obs([rolled_array[index].imag], [ens_id], idl=[idx])
-417 matrix[index] = CObs(real, imag)
-418
-419 result_dict[key] = Npr_matrix(matrix, mom_in=mom_in, mom_out=mom_out)
-420
-421 return result_dict
+413 file.close()
+414
+415 intermediate_dict = {}
+416
+417 for vertex in vertices:
+418 lorentz_names = _get_lorentz_names(vertex)
+419 for v_name in lorentz_names:
+420 if v_name in [('SigmaXY', 'SigmaZT'),
+421 ('SigmaXT', 'SigmaYZ'),
+422 ('SigmaYZ', 'SigmaXT'),
+423 ('SigmaZT', 'SigmaXY')]:
+424 sign = -1
+425 else:
+426 sign = 1
+427 if vertex not in intermediate_dict:
+428 intermediate_dict[vertex] = sign * np.array(corr_data[v_name])
+429 else:
+430 intermediate_dict[vertex] += sign * np.array(corr_data[v_name])
+431
+432 result_dict = {}
+433
+434 for key, data in intermediate_dict.items():
+435
+436 rolled_array = np.moveaxis(data, 0, 8)
+437
+438 matrix = np.empty((rolled_array.shape[:-1]), dtype=object)
+439 for index in np.ndindex(rolled_array.shape[:-1]):
+440 real = Obs([rolled_array[index].real], [ens_id], idl=[idx])
+441 imag = Obs([rolled_array[index].imag], [ens_id], idl=[idx])
+442 matrix[index] = CObs(real, imag)
+443
+444 result_dict[key] = Npr_matrix(matrix, mom_in=mom_in, mom_out=mom_out)
+445
+446 return result_dict
@@ -1342,6 +1420,13 @@ If specified only configurations in the given range are read in.
vertices (list):
Vertex functions to be extracted.
+
+Returns
+
+
+- result_dict (dict):
+extracted fourquark matrizes
+
diff --git a/docs/pyerrors/input/json.html b/docs/pyerrors/input/json.html
index 469291f5..7766dd8f 100644
--- a/docs/pyerrors/input/json.html
+++ b/docs/pyerrors/input/json.html
@@ -122,673 +122,727 @@
29 indent : int
30 Specify the indentation level of the json file. None or 0 is permissible and
31 saves disk space.
- 32 """
- 33
- 34 def _gen_data_d_from_list(ol):
- 35 dl = []
- 36 No = len(ol)
- 37 for name in ol[0].mc_names:
- 38 ed = {}
- 39 ed['id'] = name
- 40 ed['replica'] = []
- 41 for r_name in ol[0].e_content[name]:
- 42 rd = {}
- 43 rd['name'] = r_name
- 44 rd['deltas'] = []
- 45 offsets = [o.r_values[r_name] - o.value for o in ol]
- 46 deltas = np.column_stack([ol[oi].deltas[r_name] + offsets[oi] for oi in range(No)])
- 47 for i in range(len(ol[0].idl[r_name])):
- 48 rd['deltas'].append([ol[0].idl[r_name][i]])
- 49 rd['deltas'][-1] += deltas[i].tolist()
- 50 ed['replica'].append(rd)
- 51 dl.append(ed)
- 52 return dl
- 53
- 54 def _gen_cdata_d_from_list(ol):
- 55 dl = []
- 56 for name in ol[0].cov_names:
- 57 ed = {}
- 58 ed['id'] = name
- 59 ed['layout'] = str(ol[0].covobs[name].cov.shape).lstrip('(').rstrip(')').rstrip(',')
- 60 ed['cov'] = list(np.ravel(ol[0].covobs[name].cov))
- 61 ncov = ol[0].covobs[name].cov.shape[0]
- 62 ed['grad'] = []
- 63 for i in range(ncov):
- 64 ed['grad'].append([])
- 65 for o in ol:
- 66 ed['grad'][-1].append(o.covobs[name].grad[i][0])
- 67 dl.append(ed)
- 68 return dl
- 69
- 70 def write_Obs_to_dict(o):
- 71 d = {}
- 72 d['type'] = 'Obs'
- 73 d['layout'] = '1'
- 74 if o.tag:
- 75 d['tag'] = [o.tag]
- 76 if o.reweighted:
- 77 d['reweighted'] = o.reweighted
- 78 d['value'] = [o.value]
- 79 data = _gen_data_d_from_list([o])
- 80 if len(data) > 0:
- 81 d['data'] = data
- 82 cdata = _gen_cdata_d_from_list([o])
- 83 if len(cdata) > 0:
- 84 d['cdata'] = cdata
- 85 return d
- 86
- 87 def write_List_to_dict(ol):
- 88 _assert_equal_properties(ol)
- 89 d = {}
- 90 d['type'] = 'List'
- 91 d['layout'] = '%d' % len(ol)
- 92 taglist = [o.tag for o in ol]
- 93 if np.any([tag is not None for tag in taglist]):
- 94 d['tag'] = taglist
- 95 if ol[0].reweighted:
- 96 d['reweighted'] = ol[0].reweighted
- 97 d['value'] = [o.value for o in ol]
- 98 data = _gen_data_d_from_list(ol)
- 99 if len(data) > 0:
-100 d['data'] = data
-101 cdata = _gen_cdata_d_from_list(ol)
-102 if len(cdata) > 0:
-103 d['cdata'] = cdata
-104 return d
-105
-106 def write_Array_to_dict(oa):
-107 ol = np.ravel(oa)
-108 _assert_equal_properties(ol)
-109 d = {}
-110 d['type'] = 'Array'
-111 d['layout'] = str(oa.shape).lstrip('(').rstrip(')').rstrip(',')
-112 taglist = [o.tag for o in ol]
-113 if np.any([tag is not None for tag in taglist]):
-114 d['tag'] = taglist
-115 if ol[0].reweighted:
-116 d['reweighted'] = ol[0].reweighted
-117 d['value'] = [o.value for o in ol]
-118 data = _gen_data_d_from_list(ol)
-119 if len(data) > 0:
-120 d['data'] = data
-121 cdata = _gen_cdata_d_from_list(ol)
-122 if len(cdata) > 0:
-123 d['cdata'] = cdata
-124 return d
-125
-126 def _nan_Obs_like(obs):
-127 samples = []
-128 names = []
-129 idl = []
-130 for key, value in obs.idl.items():
-131 samples.append([np.nan] * len(value))
-132 names.append(key)
-133 idl.append(value)
-134 my_obs = Obs(samples, names, idl)
-135 my_obs._covobs = obs._covobs
-136 for name in obs._covobs:
-137 my_obs.names.append(name)
-138 my_obs.reweighted = obs.reweighted
-139 return my_obs
-140
-141 def write_Corr_to_dict(my_corr):
-142 first_not_none = next(i for i, j in enumerate(my_corr.content) if np.all(j))
-143 dummy_array = np.empty((my_corr.N, my_corr.N), dtype=object)
-144 dummy_array[:] = _nan_Obs_like(my_corr.content[first_not_none].ravel()[0])
-145 content = [o if o is not None else dummy_array for o in my_corr.content]
-146 dat = write_Array_to_dict(np.array(content, dtype=object))
-147 dat['type'] = 'Corr'
-148 corr_meta_data = str(my_corr.tag)
-149 if 'tag' in dat.keys():
-150 dat['tag'].append(corr_meta_data)
-151 else:
-152 dat['tag'] = [corr_meta_data]
-153 taglist = dat['tag']
-154 dat['tag'] = {} # tag is now a dictionary, that contains the previous taglist in the key "tag"
-155 dat['tag']['tag'] = taglist
-156 if my_corr.prange is not None:
-157 dat['tag']['prange'] = my_corr.prange
-158 return dat
-159
-160 if not isinstance(ol, list):
-161 ol = [ol]
-162
-163 d = {}
-164 d['program'] = 'pyerrors %s' % (pyerrorsversion.__version__)
-165 d['version'] = '1.1'
-166 d['who'] = getpass.getuser()
-167 d['date'] = datetime.datetime.now().astimezone().strftime('%Y-%m-%d %H:%M:%S %z')
-168 d['host'] = socket.gethostname() + ', ' + platform.platform()
-169
-170 if description:
-171 d['description'] = description
-172
-173 d['obsdata'] = []
-174 for io in ol:
-175 if isinstance(io, Obs):
-176 d['obsdata'].append(write_Obs_to_dict(io))
-177 elif isinstance(io, list):
-178 d['obsdata'].append(write_List_to_dict(io))
-179 elif isinstance(io, np.ndarray):
-180 d['obsdata'].append(write_Array_to_dict(io))
-181 elif isinstance(io, Corr):
-182 d['obsdata'].append(write_Corr_to_dict(io))
-183 else:
-184 raise Exception("Unkown datatype.")
-185
-186 def _jsonifier(o):
-187 if isinstance(o, np.int64):
-188 return int(o)
-189 raise TypeError('%r is not JSON serializable' % o)
+ 32
+ 33 Returns
+ 34 -------
+ 35 json_string : str
+ 36 String for export to .json(.gz) file
+ 37 """
+ 38
+ 39 def _gen_data_d_from_list(ol):
+ 40 dl = []
+ 41 No = len(ol)
+ 42 for name in ol[0].mc_names:
+ 43 ed = {}
+ 44 ed['id'] = name
+ 45 ed['replica'] = []
+ 46 for r_name in ol[0].e_content[name]:
+ 47 rd = {}
+ 48 rd['name'] = r_name
+ 49 rd['deltas'] = []
+ 50 offsets = [o.r_values[r_name] - o.value for o in ol]
+ 51 deltas = np.column_stack([ol[oi].deltas[r_name] + offsets[oi] for oi in range(No)])
+ 52 for i in range(len(ol[0].idl[r_name])):
+ 53 rd['deltas'].append([ol[0].idl[r_name][i]])
+ 54 rd['deltas'][-1] += deltas[i].tolist()
+ 55 ed['replica'].append(rd)
+ 56 dl.append(ed)
+ 57 return dl
+ 58
+ 59 def _gen_cdata_d_from_list(ol):
+ 60 dl = []
+ 61 for name in ol[0].cov_names:
+ 62 ed = {}
+ 63 ed['id'] = name
+ 64 ed['layout'] = str(ol[0].covobs[name].cov.shape).lstrip('(').rstrip(')').rstrip(',')
+ 65 ed['cov'] = list(np.ravel(ol[0].covobs[name].cov))
+ 66 ncov = ol[0].covobs[name].cov.shape[0]
+ 67 ed['grad'] = []
+ 68 for i in range(ncov):
+ 69 ed['grad'].append([])
+ 70 for o in ol:
+ 71 ed['grad'][-1].append(o.covobs[name].grad[i][0])
+ 72 dl.append(ed)
+ 73 return dl
+ 74
+ 75 def write_Obs_to_dict(o):
+ 76 d = {}
+ 77 d['type'] = 'Obs'
+ 78 d['layout'] = '1'
+ 79 if o.tag:
+ 80 d['tag'] = [o.tag]
+ 81 if o.reweighted:
+ 82 d['reweighted'] = o.reweighted
+ 83 d['value'] = [o.value]
+ 84 data = _gen_data_d_from_list([o])
+ 85 if len(data) > 0:
+ 86 d['data'] = data
+ 87 cdata = _gen_cdata_d_from_list([o])
+ 88 if len(cdata) > 0:
+ 89 d['cdata'] = cdata
+ 90 return d
+ 91
+ 92 def write_List_to_dict(ol):
+ 93 _assert_equal_properties(ol)
+ 94 d = {}
+ 95 d['type'] = 'List'
+ 96 d['layout'] = '%d' % len(ol)
+ 97 taglist = [o.tag for o in ol]
+ 98 if np.any([tag is not None for tag in taglist]):
+ 99 d['tag'] = taglist
+100 if ol[0].reweighted:
+101 d['reweighted'] = ol[0].reweighted
+102 d['value'] = [o.value for o in ol]
+103 data = _gen_data_d_from_list(ol)
+104 if len(data) > 0:
+105 d['data'] = data
+106 cdata = _gen_cdata_d_from_list(ol)
+107 if len(cdata) > 0:
+108 d['cdata'] = cdata
+109 return d
+110
+111 def write_Array_to_dict(oa):
+112 ol = np.ravel(oa)
+113 _assert_equal_properties(ol)
+114 d = {}
+115 d['type'] = 'Array'
+116 d['layout'] = str(oa.shape).lstrip('(').rstrip(')').rstrip(',')
+117 taglist = [o.tag for o in ol]
+118 if np.any([tag is not None for tag in taglist]):
+119 d['tag'] = taglist
+120 if ol[0].reweighted:
+121 d['reweighted'] = ol[0].reweighted
+122 d['value'] = [o.value for o in ol]
+123 data = _gen_data_d_from_list(ol)
+124 if len(data) > 0:
+125 d['data'] = data
+126 cdata = _gen_cdata_d_from_list(ol)
+127 if len(cdata) > 0:
+128 d['cdata'] = cdata
+129 return d
+130
+131 def _nan_Obs_like(obs):
+132 samples = []
+133 names = []
+134 idl = []
+135 for key, value in obs.idl.items():
+136 samples.append([np.nan] * len(value))
+137 names.append(key)
+138 idl.append(value)
+139 my_obs = Obs(samples, names, idl)
+140 my_obs._covobs = obs._covobs
+141 for name in obs._covobs:
+142 my_obs.names.append(name)
+143 my_obs.reweighted = obs.reweighted
+144 return my_obs
+145
+146 def write_Corr_to_dict(my_corr):
+147 first_not_none = next(i for i, j in enumerate(my_corr.content) if np.all(j))
+148 dummy_array = np.empty((my_corr.N, my_corr.N), dtype=object)
+149 dummy_array[:] = _nan_Obs_like(my_corr.content[first_not_none].ravel()[0])
+150 content = [o if o is not None else dummy_array for o in my_corr.content]
+151 dat = write_Array_to_dict(np.array(content, dtype=object))
+152 dat['type'] = 'Corr'
+153 corr_meta_data = str(my_corr.tag)
+154 if 'tag' in dat.keys():
+155 dat['tag'].append(corr_meta_data)
+156 else:
+157 dat['tag'] = [corr_meta_data]
+158 taglist = dat['tag']
+159 dat['tag'] = {} # tag is now a dictionary, that contains the previous taglist in the key "tag"
+160 dat['tag']['tag'] = taglist
+161 if my_corr.prange is not None:
+162 dat['tag']['prange'] = my_corr.prange
+163 return dat
+164
+165 if not isinstance(ol, list):
+166 ol = [ol]
+167
+168 d = {}
+169 d['program'] = 'pyerrors %s' % (pyerrorsversion.__version__)
+170 d['version'] = '1.1'
+171 d['who'] = getpass.getuser()
+172 d['date'] = datetime.datetime.now().astimezone().strftime('%Y-%m-%d %H:%M:%S %z')
+173 d['host'] = socket.gethostname() + ', ' + platform.platform()
+174
+175 if description:
+176 d['description'] = description
+177
+178 d['obsdata'] = []
+179 for io in ol:
+180 if isinstance(io, Obs):
+181 d['obsdata'].append(write_Obs_to_dict(io))
+182 elif isinstance(io, list):
+183 d['obsdata'].append(write_List_to_dict(io))
+184 elif isinstance(io, np.ndarray):
+185 d['obsdata'].append(write_Array_to_dict(io))
+186 elif isinstance(io, Corr):
+187 d['obsdata'].append(write_Corr_to_dict(io))
+188 else:
+189 raise Exception("Unkown datatype.")
190
-191 if indent:
-192 return json.dumps(d, indent=indent, ensure_ascii=False, default=_jsonifier, write_mode=json.WM_SINGLE_LINE_ARRAY)
-193 else:
-194 return json.dumps(d, indent=indent, ensure_ascii=False, default=_jsonifier, write_mode=json.WM_COMPACT)
+191 def _jsonifier(o):
+192 if isinstance(o, np.int64):
+193 return int(o)
+194 raise TypeError('%r is not JSON serializable' % o)
195
-196
-197def dump_to_json(ol, fname, description='', indent=1, gz=True):
-198 """Export a list of Obs or structures containing Obs to a .json(.gz) file
-199
-200 Parameters
-201 ----------
-202 ol : list
-203 List of objects that will be exported. At the moment, these objects can be
-204 either of: Obs, list, numpy.ndarray, Corr.
-205 All Obs inside a structure have to be defined on the same set of configurations.
-206 fname : str
-207 Filename of the output file.
-208 description : str
-209 Optional string that describes the contents of the json file.
-210 indent : int
-211 Specify the indentation level of the json file. None or 0 is permissible and
-212 saves disk space.
-213 gz : bool
-214 If True, the output is a gzipped json. If False, the output is a json file.
-215 """
-216
-217 jsonstring = create_json_string(ol, description, indent)
-218
-219 if not fname.endswith('.json') and not fname.endswith('.gz'):
-220 fname += '.json'
-221
-222 if gz:
-223 if not fname.endswith('.gz'):
-224 fname += '.gz'
+196 if indent:
+197 return json.dumps(d, indent=indent, ensure_ascii=False, default=_jsonifier, write_mode=json.WM_SINGLE_LINE_ARRAY)
+198 else:
+199 return json.dumps(d, indent=indent, ensure_ascii=False, default=_jsonifier, write_mode=json.WM_COMPACT)
+200
+201
+202def dump_to_json(ol, fname, description='', indent=1, gz=True):
+203 """Export a list of Obs or structures containing Obs to a .json(.gz) file
+204
+205 Parameters
+206 ----------
+207 ol : list
+208 List of objects that will be exported. At the moment, these objects can be
+209 either of: Obs, list, numpy.ndarray, Corr.
+210 All Obs inside a structure have to be defined on the same set of configurations.
+211 fname : str
+212 Filename of the output file.
+213 description : str
+214 Optional string that describes the contents of the json file.
+215 indent : int
+216 Specify the indentation level of the json file. None or 0 is permissible and
+217 saves disk space.
+218 gz : bool
+219 If True, the output is a gzipped json. If False, the output is a json file.
+220
+221 Returns
+222 -------
+223 Null
+224 """
225
-226 fp = gzip.open(fname, 'wb')
-227 fp.write(jsonstring.encode('utf-8'))
-228 else:
-229 fp = open(fname, 'w', encoding='utf-8')
-230 fp.write(jsonstring)
-231 fp.close()
-232
-233
-234def _parse_json_dict(json_dict, verbose=True, full_output=False):
-235 """Reconstruct a list of Obs or structures containing Obs from a dict that
-236 was built out of a json string.
-237
-238 The following structures are supported: Obs, list, numpy.ndarray, Corr
-239 If the list contains only one element, it is unpacked from the list.
-240
-241 Parameters
-242 ----------
-243 json_string : str
-244 json string containing the data.
-245 verbose : bool
-246 Print additional information that was written to the file.
-247 full_output : bool
-248 If True, a dict containing auxiliary information and the data is returned.
-249 If False, only the data is returned.
-250 """
-251
-252 def _gen_obsd_from_datad(d):
-253 retd = {}
-254 if d:
-255 retd['names'] = []
-256 retd['idl'] = []
-257 retd['deltas'] = []
-258 for ens in d:
-259 for rep in ens['replica']:
-260 rep_name = rep['name']
-261 if len(rep_name) > len(ens["id"]):
-262 if rep_name[len(ens["id"])] != "|":
-263 tmp_list = list(rep_name)
-264 tmp_list = tmp_list[:len(ens["id"])] + ["|"] + tmp_list[len(ens["id"]):]
-265 rep_name = ''.join(tmp_list)
-266 retd['names'].append(rep_name)
-267 retd['idl'].append([di[0] for di in rep['deltas']])
-268 retd['deltas'].append(np.array([di[1:] for di in rep['deltas']]))
-269 return retd
-270
-271 def _gen_covobsd_from_cdatad(d):
-272 retd = {}
-273 for ens in d:
-274 retl = []
-275 name = ens['id']
-276 layouts = ens.get('layout', '1').strip()
-277 layout = [int(ls.strip()) for ls in layouts.split(',') if len(ls) > 0]
-278 cov = np.reshape(ens['cov'], layout)
-279 grad = ens['grad']
-280 nobs = len(grad[0])
-281 for i in range(nobs):
-282 retl.append({'name': name, 'cov': cov, 'grad': [g[i] for g in grad]})
-283 retd[name] = retl
-284 return retd
-285
-286 def get_Obs_from_dict(o):
-287 layouts = o.get('layout', '1').strip()
-288 if layouts != '1':
-289 raise Exception("layout is %s has to be 1 for type Obs." % (layouts), RuntimeWarning)
+226 jsonstring = create_json_string(ol, description, indent)
+227
+228 if not fname.endswith('.json') and not fname.endswith('.gz'):
+229 fname += '.json'
+230
+231 if gz:
+232 if not fname.endswith('.gz'):
+233 fname += '.gz'
+234
+235 fp = gzip.open(fname, 'wb')
+236 fp.write(jsonstring.encode('utf-8'))
+237 else:
+238 fp = open(fname, 'w', encoding='utf-8')
+239 fp.write(jsonstring)
+240 fp.close()
+241
+242
+243def _parse_json_dict(json_dict, verbose=True, full_output=False):
+244 """Reconstruct a list of Obs or structures containing Obs from a dict that
+245 was built out of a json string.
+246
+247 The following structures are supported: Obs, list, numpy.ndarray, Corr
+248 If the list contains only one element, it is unpacked from the list.
+249
+250 Parameters
+251 ----------
+252 json_string : str
+253 json string containing the data.
+254 verbose : bool
+255 Print additional information that was written to the file.
+256 full_output : bool
+257 If True, a dict containing auxiliary information and the data is returned.
+258 If False, only the data is returned.
+259
+260 Returns
+261 -------
+262 result : list[Obs]
+263 reconstructed list of observables from the json string
+264 or
+265 result : Obs
+266 only one observable if the list only has one entry
+267 or
+268 result : dict
+269 if full_output=True
+270 """
+271
+272 def _gen_obsd_from_datad(d):
+273 retd = {}
+274 if d:
+275 retd['names'] = []
+276 retd['idl'] = []
+277 retd['deltas'] = []
+278 for ens in d:
+279 for rep in ens['replica']:
+280 rep_name = rep['name']
+281 if len(rep_name) > len(ens["id"]):
+282 if rep_name[len(ens["id"])] != "|":
+283 tmp_list = list(rep_name)
+284 tmp_list = tmp_list[:len(ens["id"])] + ["|"] + tmp_list[len(ens["id"]):]
+285 rep_name = ''.join(tmp_list)
+286 retd['names'].append(rep_name)
+287 retd['idl'].append([di[0] for di in rep['deltas']])
+288 retd['deltas'].append(np.array([di[1:] for di in rep['deltas']]))
+289 return retd
290
-291 values = o['value']
-292 od = _gen_obsd_from_datad(o.get('data', {}))
-293 cd = _gen_covobsd_from_cdatad(o.get('cdata', {}))
-294
-295 if od:
-296 ret = Obs([[ddi[0] + values[0] for ddi in di] for di in od['deltas']], od['names'], idl=od['idl'])
-297 ret._value = values[0]
-298 else:
-299 ret = Obs([], [], means=[])
-300 ret._value = values[0]
-301 for name in cd:
-302 co = cd[name][0]
-303 ret._covobs[name] = Covobs(None, co['cov'], co['name'], grad=co['grad'])
-304 ret.names.append(co['name'])
+291 def _gen_covobsd_from_cdatad(d):
+292 retd = {}
+293 for ens in d:
+294 retl = []
+295 name = ens['id']
+296 layouts = ens.get('layout', '1').strip()
+297 layout = [int(ls.strip()) for ls in layouts.split(',') if len(ls) > 0]
+298 cov = np.reshape(ens['cov'], layout)
+299 grad = ens['grad']
+300 nobs = len(grad[0])
+301 for i in range(nobs):
+302 retl.append({'name': name, 'cov': cov, 'grad': [g[i] for g in grad]})
+303 retd[name] = retl
+304 return retd
305
-306 ret.reweighted = o.get('reweighted', False)
-307 ret.tag = o.get('tag', [None])[0]
-308 return ret
-309
-310 def get_List_from_dict(o):
-311 layouts = o.get('layout', '1').strip()
-312 layout = int(layouts)
-313 values = o['value']
-314 od = _gen_obsd_from_datad(o.get('data', {}))
-315 cd = _gen_covobsd_from_cdatad(o.get('cdata', {}))
-316
-317 ret = []
-318 taglist = o.get('tag', layout * [None])
-319 for i in range(layout):
-320 if od:
-321 ret.append(Obs([list(di[:, i] + values[i]) for di in od['deltas']], od['names'], idl=od['idl']))
-322 ret[-1]._value = values[i]
-323 else:
-324 ret.append(Obs([], [], means=[]))
-325 ret[-1]._value = values[i]
-326 print('Created Obs with means= ', values[i])
-327 for name in cd:
-328 co = cd[name][i]
-329 ret[-1]._covobs[name] = Covobs(None, co['cov'], co['name'], grad=co['grad'])
-330 ret[-1].names.append(co['name'])
-331
-332 ret[-1].reweighted = o.get('reweighted', False)
-333 ret[-1].tag = taglist[i]
-334 return ret
-335
-336 def get_Array_from_dict(o):
-337 layouts = o.get('layout', '1').strip()
-338 layout = [int(ls.strip()) for ls in layouts.split(',') if len(ls) > 0]
-339 N = np.prod(layout)
-340 values = o['value']
-341 od = _gen_obsd_from_datad(o.get('data', {}))
-342 cd = _gen_covobsd_from_cdatad(o.get('cdata', {}))
-343
-344 ret = []
-345 taglist = o.get('tag', N * [None])
-346 for i in range(N):
-347 if od:
-348 ret.append(Obs([di[:, i] + values[i] for di in od['deltas']], od['names'], idl=od['idl']))
-349 ret[-1]._value = values[i]
-350 else:
-351 ret.append(Obs([], [], means=[]))
-352 ret[-1]._value = values[i]
-353 for name in cd:
-354 co = cd[name][i]
-355 ret[-1]._covobs[name] = Covobs(None, co['cov'], co['name'], grad=co['grad'])
-356 ret[-1].names.append(co['name'])
-357 ret[-1].reweighted = o.get('reweighted', False)
-358 ret[-1].tag = taglist[i]
-359 return np.reshape(ret, layout)
-360
-361 def get_Corr_from_dict(o):
-362 if isinstance(o.get('tag'), list): # supports the old way
-363 taglist = o.get('tag') # This had to be modified to get the taglist from the dictionary
-364 temp_prange = None
-365 elif isinstance(o.get('tag'), dict):
-366 tagdic = o.get('tag')
-367 taglist = tagdic['tag']
-368 if 'prange' in tagdic:
-369 temp_prange = tagdic['prange']
+306 def get_Obs_from_dict(o):
+307 layouts = o.get('layout', '1').strip()
+308 if layouts != '1':
+309 raise Exception("layout is %s has to be 1 for type Obs." % (layouts), RuntimeWarning)
+310
+311 values = o['value']
+312 od = _gen_obsd_from_datad(o.get('data', {}))
+313 cd = _gen_covobsd_from_cdatad(o.get('cdata', {}))
+314
+315 if od:
+316 ret = Obs([[ddi[0] + values[0] for ddi in di] for di in od['deltas']], od['names'], idl=od['idl'])
+317 ret._value = values[0]
+318 else:
+319 ret = Obs([], [], means=[])
+320 ret._value = values[0]
+321 for name in cd:
+322 co = cd[name][0]
+323 ret._covobs[name] = Covobs(None, co['cov'], co['name'], grad=co['grad'])
+324 ret.names.append(co['name'])
+325
+326 ret.reweighted = o.get('reweighted', False)
+327 ret.tag = o.get('tag', [None])[0]
+328 return ret
+329
+330 def get_List_from_dict(o):
+331 layouts = o.get('layout', '1').strip()
+332 layout = int(layouts)
+333 values = o['value']
+334 od = _gen_obsd_from_datad(o.get('data', {}))
+335 cd = _gen_covobsd_from_cdatad(o.get('cdata', {}))
+336
+337 ret = []
+338 taglist = o.get('tag', layout * [None])
+339 for i in range(layout):
+340 if od:
+341 ret.append(Obs([list(di[:, i] + values[i]) for di in od['deltas']], od['names'], idl=od['idl']))
+342 ret[-1]._value = values[i]
+343 else:
+344 ret.append(Obs([], [], means=[]))
+345 ret[-1]._value = values[i]
+346 print('Created Obs with means= ', values[i])
+347 for name in cd:
+348 co = cd[name][i]
+349 ret[-1]._covobs[name] = Covobs(None, co['cov'], co['name'], grad=co['grad'])
+350 ret[-1].names.append(co['name'])
+351
+352 ret[-1].reweighted = o.get('reweighted', False)
+353 ret[-1].tag = taglist[i]
+354 return ret
+355
+356 def get_Array_from_dict(o):
+357 layouts = o.get('layout', '1').strip()
+358 layout = [int(ls.strip()) for ls in layouts.split(',') if len(ls) > 0]
+359 N = np.prod(layout)
+360 values = o['value']
+361 od = _gen_obsd_from_datad(o.get('data', {}))
+362 cd = _gen_covobsd_from_cdatad(o.get('cdata', {}))
+363
+364 ret = []
+365 taglist = o.get('tag', N * [None])
+366 for i in range(N):
+367 if od:
+368 ret.append(Obs([di[:, i] + values[i] for di in od['deltas']], od['names'], idl=od['idl']))
+369 ret[-1]._value = values[i]
370 else:
-371 temp_prange = None
-372 else:
-373 raise Exception("The tag is not a list or dict")
-374
-375 corr_tag = taglist[-1]
-376 tmp_o = o
-377 tmp_o['tag'] = taglist[:-1]
-378 if len(tmp_o['tag']) == 0:
-379 del tmp_o['tag']
-380 dat = get_Array_from_dict(tmp_o)
-381 my_corr = Corr([None if np.isnan(o.ravel()[0].value) else o for o in list(dat)])
-382 if corr_tag != 'None':
-383 my_corr.tag = corr_tag
-384
-385 my_corr.prange = temp_prange
-386 return my_corr
-387
-388 prog = json_dict.get('program', '')
-389 version = json_dict.get('version', '')
-390 who = json_dict.get('who', '')
-391 date = json_dict.get('date', '')
-392 host = json_dict.get('host', '')
-393 if prog and verbose:
-394 print('Data has been written using %s.' % (prog))
-395 if version and verbose:
-396 print('Format version %s' % (version))
-397 if np.any([who, date, host] and verbose):
-398 print('Written by %s on %s on host %s' % (who, date, host))
-399 description = json_dict.get('description', '')
-400 if description and verbose:
-401 print()
-402 print('Description: ', description)
-403 obsdata = json_dict['obsdata']
-404 ol = []
-405 for io in obsdata:
-406 if io['type'] == 'Obs':
-407 ol.append(get_Obs_from_dict(io))
-408 elif io['type'] == 'List':
-409 ol.append(get_List_from_dict(io))
-410 elif io['type'] == 'Array':
-411 ol.append(get_Array_from_dict(io))
-412 elif io['type'] == 'Corr':
-413 ol.append(get_Corr_from_dict(io))
-414 else:
-415 raise Exception("Unknown datatype.")
-416
-417 if full_output:
-418 retd = {}
-419 retd['program'] = prog
-420 retd['version'] = version
-421 retd['who'] = who
-422 retd['date'] = date
-423 retd['host'] = host
-424 retd['description'] = description
-425 retd['obsdata'] = ol
-426
-427 return retd
-428 else:
-429 if len(obsdata) == 1:
-430 ol = ol[0]
-431
-432 return ol
-433
-434
-435def import_json_string(json_string, verbose=True, full_output=False):
-436 """Reconstruct a list of Obs or structures containing Obs from a json string.
-437
-438 The following structures are supported: Obs, list, numpy.ndarray, Corr
-439 If the list contains only one element, it is unpacked from the list.
-440
-441 Parameters
-442 ----------
-443 json_string : str
-444 json string containing the data.
-445 verbose : bool
-446 Print additional information that was written to the file.
-447 full_output : bool
-448 If True, a dict containing auxiliary information and the data is returned.
-449 If False, only the data is returned.
-450 """
+371 ret.append(Obs([], [], means=[]))
+372 ret[-1]._value = values[i]
+373 for name in cd:
+374 co = cd[name][i]
+375 ret[-1]._covobs[name] = Covobs(None, co['cov'], co['name'], grad=co['grad'])
+376 ret[-1].names.append(co['name'])
+377 ret[-1].reweighted = o.get('reweighted', False)
+378 ret[-1].tag = taglist[i]
+379 return np.reshape(ret, layout)
+380
+381 def get_Corr_from_dict(o):
+382 if isinstance(o.get('tag'), list): # supports the old way
+383 taglist = o.get('tag') # This had to be modified to get the taglist from the dictionary
+384 temp_prange = None
+385 elif isinstance(o.get('tag'), dict):
+386 tagdic = o.get('tag')
+387 taglist = tagdic['tag']
+388 if 'prange' in tagdic:
+389 temp_prange = tagdic['prange']
+390 else:
+391 temp_prange = None
+392 else:
+393 raise Exception("The tag is not a list or dict")
+394
+395 corr_tag = taglist[-1]
+396 tmp_o = o
+397 tmp_o['tag'] = taglist[:-1]
+398 if len(tmp_o['tag']) == 0:
+399 del tmp_o['tag']
+400 dat = get_Array_from_dict(tmp_o)
+401 my_corr = Corr([None if np.isnan(o.ravel()[0].value) else o for o in list(dat)])
+402 if corr_tag != 'None':
+403 my_corr.tag = corr_tag
+404
+405 my_corr.prange = temp_prange
+406 return my_corr
+407
+408 prog = json_dict.get('program', '')
+409 version = json_dict.get('version', '')
+410 who = json_dict.get('who', '')
+411 date = json_dict.get('date', '')
+412 host = json_dict.get('host', '')
+413 if prog and verbose:
+414 print('Data has been written using %s.' % (prog))
+415 if version and verbose:
+416 print('Format version %s' % (version))
+417 if np.any([who, date, host] and verbose):
+418 print('Written by %s on %s on host %s' % (who, date, host))
+419 description = json_dict.get('description', '')
+420 if description and verbose:
+421 print()
+422 print('Description: ', description)
+423 obsdata = json_dict['obsdata']
+424 ol = []
+425 for io in obsdata:
+426 if io['type'] == 'Obs':
+427 ol.append(get_Obs_from_dict(io))
+428 elif io['type'] == 'List':
+429 ol.append(get_List_from_dict(io))
+430 elif io['type'] == 'Array':
+431 ol.append(get_Array_from_dict(io))
+432 elif io['type'] == 'Corr':
+433 ol.append(get_Corr_from_dict(io))
+434 else:
+435 raise Exception("Unknown datatype.")
+436
+437 if full_output:
+438 retd = {}
+439 retd['program'] = prog
+440 retd['version'] = version
+441 retd['who'] = who
+442 retd['date'] = date
+443 retd['host'] = host
+444 retd['description'] = description
+445 retd['obsdata'] = ol
+446
+447 return retd
+448 else:
+449 if len(obsdata) == 1:
+450 ol = ol[0]
451
-452 return _parse_json_dict(json.loads(json_string), verbose, full_output)
+452 return ol
453
454
-455def load_json(fname, verbose=True, gz=True, full_output=False):
-456 """Import a list of Obs or structures containing Obs from a .json(.gz) file.
+455def import_json_string(json_string, verbose=True, full_output=False):
+456 """Reconstruct a list of Obs or structures containing Obs from a json string.
457
458 The following structures are supported: Obs, list, numpy.ndarray, Corr
459 If the list contains only one element, it is unpacked from the list.
460
461 Parameters
462 ----------
-463 fname : str
-464 Filename of the input file.
+463 json_string : str
+464 json string containing the data.
465 verbose : bool
466 Print additional information that was written to the file.
-467 gz : bool
-468 If True, assumes that data is gzipped. If False, assumes JSON file.
-469 full_output : bool
-470 If True, a dict containing auxiliary information and the data is returned.
-471 If False, only the data is returned.
-472 """
-473 if not fname.endswith('.json') and not fname.endswith('.gz'):
-474 fname += '.json'
-475 if gz:
-476 if not fname.endswith('.gz'):
-477 fname += '.gz'
-478 with gzip.open(fname, 'r') as fin:
-479 d = json.load(fin)
-480 else:
-481 if fname.endswith('.gz'):
-482 warnings.warn("Trying to read from %s without unzipping!" % fname, UserWarning)
-483 with open(fname, 'r', encoding='utf-8') as fin:
-484 d = json.loads(fin.read())
+467 full_output : bool
+468 If True, a dict containing auxiliary information and the data is returned.
+469 If False, only the data is returned.
+470
+471 Returns
+472 -------
+473 result : list[Obs]
+474 reconstructed list of observables from the json string
+475 or
+476 result : Obs
+477 only one observable if the list only has one entry
+478 or
+479 result : dict
+480 if full_output=True
+481 """
+482
+483 return _parse_json_dict(json.loads(json_string), verbose, full_output)
+484
485
-486 return _parse_json_dict(d, verbose, full_output)
-487
+486def load_json(fname, verbose=True, gz=True, full_output=False):
+487 """Import a list of Obs or structures containing Obs from a .json(.gz) file.
488
-489def _ol_from_dict(ind, reps='DICTOBS'):
-490 """Convert a dictionary of Obs objects to a list and a dictionary that contains
-491 placeholders instead of the Obs objects.
-492
-493 Parameters
-494 ----------
-495 ind : dict
-496 Dict of JSON valid structures and objects that will be exported.
-497 At the moment, these object can be either of: Obs, list, numpy.ndarray, Corr.
-498 All Obs inside a structure have to be defined on the same set of configurations.
-499 reps : str
-500 Specify the structure of the placeholder in exported dict to be reps[0-9]+.
-501 """
-502
-503 obstypes = (Obs, Corr, np.ndarray)
-504
-505 if not reps.isalnum():
-506 raise Exception('Placeholder string has to be alphanumeric!')
-507 ol = []
-508 counter = 0
-509
-510 def dict_replace_obs(d):
-511 nonlocal ol
-512 nonlocal counter
-513 x = {}
-514 for k, v in d.items():
-515 if isinstance(v, dict):
-516 v = dict_replace_obs(v)
-517 elif isinstance(v, list) and all([isinstance(o, Obs) for o in v]):
-518 v = obslist_replace_obs(v)
-519 elif isinstance(v, list):
-520 v = list_replace_obs(v)
-521 elif isinstance(v, obstypes):
-522 ol.append(v)
-523 v = reps + '%d' % (counter)
-524 counter += 1
-525 elif isinstance(v, str):
-526 if bool(re.match(r'%s[0-9]+' % (reps), v)):
-527 raise Exception('Dict contains string %s that matches the placeholder! %s Cannot be savely exported.' % (v, reps))
-528 x[k] = v
-529 return x
+489 The following structures are supported: Obs, list, numpy.ndarray, Corr
+490 If the list contains only one element, it is unpacked from the list.
+491
+492 Parameters
+493 ----------
+494 fname : str
+495 Filename of the input file.
+496 verbose : bool
+497 Print additional information that was written to the file.
+498 gz : bool
+499 If True, assumes that data is gzipped. If False, assumes JSON file.
+500 full_output : bool
+501 If True, a dict containing auxiliary information and the data is returned.
+502 If False, only the data is returned.
+503
+504 Returns
+505 -------
+506 result : list[Obs]
+507 reconstructed list of observables from the json string
+508 or
+509 result : Obs
+510 only one observable if the list only has one entry
+511 or
+512 result : dict
+513 if full_output=True
+514 """
+515 if not fname.endswith('.json') and not fname.endswith('.gz'):
+516 fname += '.json'
+517 if gz:
+518 if not fname.endswith('.gz'):
+519 fname += '.gz'
+520 with gzip.open(fname, 'r') as fin:
+521 d = json.load(fin)
+522 else:
+523 if fname.endswith('.gz'):
+524 warnings.warn("Trying to read from %s without unzipping!" % fname, UserWarning)
+525 with open(fname, 'r', encoding='utf-8') as fin:
+526 d = json.loads(fin.read())
+527
+528 return _parse_json_dict(d, verbose, full_output)
+529
530
-531 def list_replace_obs(li):
-532 nonlocal ol
-533 nonlocal counter
-534 x = []
-535 for e in li:
-536 if isinstance(e, list):
-537 e = list_replace_obs(e)
-538 elif isinstance(e, list) and all([isinstance(o, Obs) for o in e]):
-539 e = obslist_replace_obs(e)
-540 elif isinstance(e, dict):
-541 e = dict_replace_obs(e)
-542 elif isinstance(e, obstypes):
-543 ol.append(e)
-544 e = reps + '%d' % (counter)
-545 counter += 1
-546 elif isinstance(e, str):
-547 if bool(re.match(r'%s[0-9]+' % (reps), e)):
-548 raise Exception('Dict contains string %s that matches the placeholder! %s Cannot be savely exported.' % (e, reps))
-549 x.append(e)
-550 return x
+531def _ol_from_dict(ind, reps='DICTOBS'):
+532 """Convert a dictionary of Obs objects to a list and a dictionary that contains
+533 placeholders instead of the Obs objects.
+534
+535 Parameters
+536 ----------
+537 ind : dict
+538 Dict of JSON valid structures and objects that will be exported.
+539 At the moment, these object can be either of: Obs, list, numpy.ndarray, Corr.
+540 All Obs inside a structure have to be defined on the same set of configurations.
+541 reps : str
+542 Specify the structure of the placeholder in exported dict to be reps[0-9]+.
+543 """
+544
+545 obstypes = (Obs, Corr, np.ndarray)
+546
+547 if not reps.isalnum():
+548 raise Exception('Placeholder string has to be alphanumeric!')
+549 ol = []
+550 counter = 0
551
-552 def obslist_replace_obs(li):
+552 def dict_replace_obs(d):
553 nonlocal ol
554 nonlocal counter
-555 il = []
-556 for e in li:
-557 il.append(e)
-558
-559 ol.append(il)
-560 x = reps + '%d' % (counter)
-561 counter += 1
-562 return x
-563
-564 nd = dict_replace_obs(ind)
-565
-566 return ol, nd
-567
-568
-569def dump_dict_to_json(od, fname, description='', indent=1, reps='DICTOBS', gz=True):
-570 """Export a dict of Obs or structures containing Obs to a .json(.gz) file
-571
-572 Parameters
-573 ----------
-574 od : dict
-575 Dict of JSON valid structures and objects that will be exported.
-576 At the moment, these objects can be either of: Obs, list, numpy.ndarray, Corr.
-577 All Obs inside a structure have to be defined on the same set of configurations.
-578 fname : str
-579 Filename of the output file.
-580 description : str
-581 Optional string that describes the contents of the json file.
-582 indent : int
-583 Specify the indentation level of the json file. None or 0 is permissible and
-584 saves disk space.
-585 reps : str
-586 Specify the structure of the placeholder in exported dict to be reps[0-9]+.
-587 gz : bool
-588 If True, the output is a gzipped json. If False, the output is a json file.
-589 """
-590
-591 if not isinstance(od, dict):
-592 raise Exception('od has to be a dictionary. Did you want to use dump_to_json?')
+555 x = {}
+556 for k, v in d.items():
+557 if isinstance(v, dict):
+558 v = dict_replace_obs(v)
+559 elif isinstance(v, list) and all([isinstance(o, Obs) for o in v]):
+560 v = obslist_replace_obs(v)
+561 elif isinstance(v, list):
+562 v = list_replace_obs(v)
+563 elif isinstance(v, obstypes):
+564 ol.append(v)
+565 v = reps + '%d' % (counter)
+566 counter += 1
+567 elif isinstance(v, str):
+568 if bool(re.match(r'%s[0-9]+' % (reps), v)):
+569 raise Exception('Dict contains string %s that matches the placeholder! %s Cannot be savely exported.' % (v, reps))
+570 x[k] = v
+571 return x
+572
+573 def list_replace_obs(li):
+574 nonlocal ol
+575 nonlocal counter
+576 x = []
+577 for e in li:
+578 if isinstance(e, list):
+579 e = list_replace_obs(e)
+580 elif isinstance(e, list) and all([isinstance(o, Obs) for o in e]):
+581 e = obslist_replace_obs(e)
+582 elif isinstance(e, dict):
+583 e = dict_replace_obs(e)
+584 elif isinstance(e, obstypes):
+585 ol.append(e)
+586 e = reps + '%d' % (counter)
+587 counter += 1
+588 elif isinstance(e, str):
+589 if bool(re.match(r'%s[0-9]+' % (reps), e)):
+590 raise Exception('Dict contains string %s that matches the placeholder! %s Cannot be savely exported.' % (e, reps))
+591 x.append(e)
+592 return x
593
-594 infostring = ('This JSON file contains a python dictionary that has been parsed to a list of structures. '
-595 'OBSDICT contains the dictionary, where Obs or other structures have been replaced by '
-596 '' + reps + '[0-9]+. The field description contains the additional description of this JSON file. '
-597 'This file may be parsed to a dict with the pyerrors routine load_json_dict.')
-598
-599 desc_dict = {'INFO': infostring, 'OBSDICT': {}, 'description': description}
-600 ol, desc_dict['OBSDICT'] = _ol_from_dict(od, reps=reps)
-601
-602 dump_to_json(ol, fname, description=desc_dict, indent=indent, gz=gz)
-603
-604
-605def _od_from_list_and_dict(ol, ind, reps='DICTOBS'):
-606 """Parse a list of Obs or structures containing Obs and an accompanying
-607 dict, where the structures have been replaced by placeholders to a
-608 dict that contains the structures.
+594 def obslist_replace_obs(li):
+595 nonlocal ol
+596 nonlocal counter
+597 il = []
+598 for e in li:
+599 il.append(e)
+600
+601 ol.append(il)
+602 x = reps + '%d' % (counter)
+603 counter += 1
+604 return x
+605
+606 nd = dict_replace_obs(ind)
+607
+608 return ol, nd
609
-610 The following structures are supported: Obs, list, numpy.ndarray, Corr
-611
-612 Parameters
-613 ----------
-614 ol : list
-615 List of objects -
-616 At the moment, these objects can be either of: Obs, list, numpy.ndarray, Corr.
-617 All Obs inside a structure have to be defined on the same set of configurations.
-618 ind : dict
-619 Dict that defines the structure of the resulting dict and contains placeholders
-620 reps : str
-621 Specify the structure of the placeholder in imported dict to be reps[0-9]+.
-622 """
-623 if not reps.isalnum():
-624 raise Exception('Placeholder string has to be alphanumeric!')
-625
-626 counter = 0
-627
-628 def dict_replace_string(d):
-629 nonlocal counter
-630 nonlocal ol
-631 x = {}
-632 for k, v in d.items():
-633 if isinstance(v, dict):
-634 v = dict_replace_string(v)
-635 elif isinstance(v, list):
-636 v = list_replace_string(v)
-637 elif isinstance(v, str) and bool(re.match(r'%s[0-9]+' % (reps), v)):
-638 index = int(v[len(reps):])
-639 v = ol[index]
-640 counter += 1
-641 x[k] = v
-642 return x
-643
-644 def list_replace_string(li):
-645 nonlocal counter
-646 nonlocal ol
-647 x = []
-648 for e in li:
-649 if isinstance(e, list):
-650 e = list_replace_string(e)
-651 elif isinstance(e, dict):
-652 e = dict_replace_string(e)
-653 elif isinstance(e, str) and bool(re.match(r'%s[0-9]+' % (reps), e)):
-654 index = int(e[len(reps):])
-655 e = ol[index]
-656 counter += 1
-657 x.append(e)
-658 return x
-659
-660 nd = dict_replace_string(ind)
-661
-662 if counter == 0:
-663 raise Exception('No placeholder has been replaced! Check if reps is set correctly.')
-664
-665 return nd
-666
-667
-668def load_json_dict(fname, verbose=True, gz=True, full_output=False, reps='DICTOBS'):
-669 """Import a dict of Obs or structures containing Obs from a .json(.gz) file.
-670
-671 The following structures are supported: Obs, list, numpy.ndarray, Corr
-672
-673 Parameters
-674 ----------
-675 fname : str
-676 Filename of the input file.
-677 verbose : bool
-678 Print additional information that was written to the file.
-679 gz : bool
-680 If True, assumes that data is gzipped. If False, assumes JSON file.
-681 full_output : bool
-682 If True, a dict containing auxiliary information and the data is returned.
-683 If False, only the data is returned.
-684 reps : str
-685 Specify the structure of the placeholder in imported dict to be reps[0-9]+.
-686 """
-687 indata = load_json(fname, verbose=verbose, gz=gz, full_output=True)
-688 description = indata['description']['description']
-689 indict = indata['description']['OBSDICT']
-690 ol = indata['obsdata']
-691 od = _od_from_list_and_dict(ol, indict, reps=reps)
-692
-693 if full_output:
-694 indata['description'] = description
-695 indata['obsdata'] = od
-696 return indata
-697 else:
-698 return od
+610
+611def dump_dict_to_json(od, fname, description='', indent=1, reps='DICTOBS', gz=True):
+612 """Export a dict of Obs or structures containing Obs to a .json(.gz) file
+613
+614 Parameters
+615 ----------
+616 od : dict
+617 Dict of JSON valid structures and objects that will be exported.
+618 At the moment, these objects can be either of: Obs, list, numpy.ndarray, Corr.
+619 All Obs inside a structure have to be defined on the same set of configurations.
+620 fname : str
+621 Filename of the output file.
+622 description : str
+623 Optional string that describes the contents of the json file.
+624 indent : int
+625 Specify the indentation level of the json file. None or 0 is permissible and
+626 saves disk space.
+627 reps : str
+628 Specify the structure of the placeholder in exported dict to be reps[0-9]+.
+629 gz : bool
+630 If True, the output is a gzipped json. If False, the output is a json file.
+631
+632 Returns
+633 -------
+634 None
+635 """
+636
+637 if not isinstance(od, dict):
+638 raise Exception('od has to be a dictionary. Did you want to use dump_to_json?')
+639
+640 infostring = ('This JSON file contains a python dictionary that has been parsed to a list of structures. '
+641 'OBSDICT contains the dictionary, where Obs or other structures have been replaced by '
+642 '' + reps + '[0-9]+. The field description contains the additional description of this JSON file. '
+643 'This file may be parsed to a dict with the pyerrors routine load_json_dict.')
+644
+645 desc_dict = {'INFO': infostring, 'OBSDICT': {}, 'description': description}
+646 ol, desc_dict['OBSDICT'] = _ol_from_dict(od, reps=reps)
+647
+648 dump_to_json(ol, fname, description=desc_dict, indent=indent, gz=gz)
+649
+650
+651def _od_from_list_and_dict(ol, ind, reps='DICTOBS'):
+652 """Parse a list of Obs or structures containing Obs and an accompanying
+653 dict, where the structures have been replaced by placeholders to a
+654 dict that contains the structures.
+655
+656 The following structures are supported: Obs, list, numpy.ndarray, Corr
+657
+658 Parameters
+659 ----------
+660 ol : list
+661 List of objects -
+662 At the moment, these objects can be either of: Obs, list, numpy.ndarray, Corr.
+663 All Obs inside a structure have to be defined on the same set of configurations.
+664 ind : dict
+665 Dict that defines the structure of the resulting dict and contains placeholders
+666 reps : str
+667 Specify the structure of the placeholder in imported dict to be reps[0-9]+.
+668 """
+669 if not reps.isalnum():
+670 raise Exception('Placeholder string has to be alphanumeric!')
+671
+672 counter = 0
+673
+674 def dict_replace_string(d):
+675 nonlocal counter
+676 nonlocal ol
+677 x = {}
+678 for k, v in d.items():
+679 if isinstance(v, dict):
+680 v = dict_replace_string(v)
+681 elif isinstance(v, list):
+682 v = list_replace_string(v)
+683 elif isinstance(v, str) and bool(re.match(r'%s[0-9]+' % (reps), v)):
+684 index = int(v[len(reps):])
+685 v = ol[index]
+686 counter += 1
+687 x[k] = v
+688 return x
+689
+690 def list_replace_string(li):
+691 nonlocal counter
+692 nonlocal ol
+693 x = []
+694 for e in li:
+695 if isinstance(e, list):
+696 e = list_replace_string(e)
+697 elif isinstance(e, dict):
+698 e = dict_replace_string(e)
+699 elif isinstance(e, str) and bool(re.match(r'%s[0-9]+' % (reps), e)):
+700 index = int(e[len(reps):])
+701 e = ol[index]
+702 counter += 1
+703 x.append(e)
+704 return x
+705
+706 nd = dict_replace_string(ind)
+707
+708 if counter == 0:
+709 raise Exception('No placeholder has been replaced! Check if reps is set correctly.')
+710
+711 return nd
+712
+713
+714def load_json_dict(fname, verbose=True, gz=True, full_output=False, reps='DICTOBS'):
+715 """Import a dict of Obs or structures containing Obs from a .json(.gz) file.
+716
+717 The following structures are supported: Obs, list, numpy.ndarray, Corr
+718
+719 Parameters
+720 ----------
+721 fname : str
+722 Filename of the input file.
+723 verbose : bool
+724 Print additional information that was written to the file.
+725 gz : bool
+726 If True, assumes that data is gzipped. If False, assumes JSON file.
+727 full_output : bool
+728 If True, a dict containing auxiliary information and the data is returned.
+729 If False, only the data is returned.
+730 reps : str
+731 Specify the structure of the placeholder in imported dict to be reps[0-9]+.
+732
+733 Returns
+734 -------
+735 data : Obs / list / Corr
+736 Read data
+737 or
+738 data : dict
+739 Read data and meta-data
+740 """
+741 indata = load_json(fname, verbose=verbose, gz=gz, full_output=True)
+742 description = indata['description']['description']
+743 indict = indata['description']['OBSDICT']
+744 ol = indata['obsdata']
+745 od = _od_from_list_and_dict(ol, indict, reps=reps)
+746
+747 if full_output:
+748 indata['description'] = description
+749 indata['obsdata'] = od
+750 return indata
+751 else:
+752 return od
@@ -819,169 +873,174 @@
30 indent : int
31 Specify the indentation level of the json file. None or 0 is permissible and
32 saves disk space.
- 33 """
- 34
- 35 def _gen_data_d_from_list(ol):
- 36 dl = []
- 37 No = len(ol)
- 38 for name in ol[0].mc_names:
- 39 ed = {}
- 40 ed['id'] = name
- 41 ed['replica'] = []
- 42 for r_name in ol[0].e_content[name]:
- 43 rd = {}
- 44 rd['name'] = r_name
- 45 rd['deltas'] = []
- 46 offsets = [o.r_values[r_name] - o.value for o in ol]
- 47 deltas = np.column_stack([ol[oi].deltas[r_name] + offsets[oi] for oi in range(No)])
- 48 for i in range(len(ol[0].idl[r_name])):
- 49 rd['deltas'].append([ol[0].idl[r_name][i]])
- 50 rd['deltas'][-1] += deltas[i].tolist()
- 51 ed['replica'].append(rd)
- 52 dl.append(ed)
- 53 return dl
- 54
- 55 def _gen_cdata_d_from_list(ol):
- 56 dl = []
- 57 for name in ol[0].cov_names:
- 58 ed = {}
- 59 ed['id'] = name
- 60 ed['layout'] = str(ol[0].covobs[name].cov.shape).lstrip('(').rstrip(')').rstrip(',')
- 61 ed['cov'] = list(np.ravel(ol[0].covobs[name].cov))
- 62 ncov = ol[0].covobs[name].cov.shape[0]
- 63 ed['grad'] = []
- 64 for i in range(ncov):
- 65 ed['grad'].append([])
- 66 for o in ol:
- 67 ed['grad'][-1].append(o.covobs[name].grad[i][0])
- 68 dl.append(ed)
- 69 return dl
- 70
- 71 def write_Obs_to_dict(o):
- 72 d = {}
- 73 d['type'] = 'Obs'
- 74 d['layout'] = '1'
- 75 if o.tag:
- 76 d['tag'] = [o.tag]
- 77 if o.reweighted:
- 78 d['reweighted'] = o.reweighted
- 79 d['value'] = [o.value]
- 80 data = _gen_data_d_from_list([o])
- 81 if len(data) > 0:
- 82 d['data'] = data
- 83 cdata = _gen_cdata_d_from_list([o])
- 84 if len(cdata) > 0:
- 85 d['cdata'] = cdata
- 86 return d
- 87
- 88 def write_List_to_dict(ol):
- 89 _assert_equal_properties(ol)
- 90 d = {}
- 91 d['type'] = 'List'
- 92 d['layout'] = '%d' % len(ol)
- 93 taglist = [o.tag for o in ol]
- 94 if np.any([tag is not None for tag in taglist]):
- 95 d['tag'] = taglist
- 96 if ol[0].reweighted:
- 97 d['reweighted'] = ol[0].reweighted
- 98 d['value'] = [o.value for o in ol]
- 99 data = _gen_data_d_from_list(ol)
-100 if len(data) > 0:
-101 d['data'] = data
-102 cdata = _gen_cdata_d_from_list(ol)
-103 if len(cdata) > 0:
-104 d['cdata'] = cdata
-105 return d
-106
-107 def write_Array_to_dict(oa):
-108 ol = np.ravel(oa)
-109 _assert_equal_properties(ol)
-110 d = {}
-111 d['type'] = 'Array'
-112 d['layout'] = str(oa.shape).lstrip('(').rstrip(')').rstrip(',')
-113 taglist = [o.tag for o in ol]
-114 if np.any([tag is not None for tag in taglist]):
-115 d['tag'] = taglist
-116 if ol[0].reweighted:
-117 d['reweighted'] = ol[0].reweighted
-118 d['value'] = [o.value for o in ol]
-119 data = _gen_data_d_from_list(ol)
-120 if len(data) > 0:
-121 d['data'] = data
-122 cdata = _gen_cdata_d_from_list(ol)
-123 if len(cdata) > 0:
-124 d['cdata'] = cdata
-125 return d
-126
-127 def _nan_Obs_like(obs):
-128 samples = []
-129 names = []
-130 idl = []
-131 for key, value in obs.idl.items():
-132 samples.append([np.nan] * len(value))
-133 names.append(key)
-134 idl.append(value)
-135 my_obs = Obs(samples, names, idl)
-136 my_obs._covobs = obs._covobs
-137 for name in obs._covobs:
-138 my_obs.names.append(name)
-139 my_obs.reweighted = obs.reweighted
-140 return my_obs
-141
-142 def write_Corr_to_dict(my_corr):
-143 first_not_none = next(i for i, j in enumerate(my_corr.content) if np.all(j))
-144 dummy_array = np.empty((my_corr.N, my_corr.N), dtype=object)
-145 dummy_array[:] = _nan_Obs_like(my_corr.content[first_not_none].ravel()[0])
-146 content = [o if o is not None else dummy_array for o in my_corr.content]
-147 dat = write_Array_to_dict(np.array(content, dtype=object))
-148 dat['type'] = 'Corr'
-149 corr_meta_data = str(my_corr.tag)
-150 if 'tag' in dat.keys():
-151 dat['tag'].append(corr_meta_data)
-152 else:
-153 dat['tag'] = [corr_meta_data]
-154 taglist = dat['tag']
-155 dat['tag'] = {} # tag is now a dictionary, that contains the previous taglist in the key "tag"
-156 dat['tag']['tag'] = taglist
-157 if my_corr.prange is not None:
-158 dat['tag']['prange'] = my_corr.prange
-159 return dat
-160
-161 if not isinstance(ol, list):
-162 ol = [ol]
-163
-164 d = {}
-165 d['program'] = 'pyerrors %s' % (pyerrorsversion.__version__)
-166 d['version'] = '1.1'
-167 d['who'] = getpass.getuser()
-168 d['date'] = datetime.datetime.now().astimezone().strftime('%Y-%m-%d %H:%M:%S %z')
-169 d['host'] = socket.gethostname() + ', ' + platform.platform()
-170
-171 if description:
-172 d['description'] = description
-173
-174 d['obsdata'] = []
-175 for io in ol:
-176 if isinstance(io, Obs):
-177 d['obsdata'].append(write_Obs_to_dict(io))
-178 elif isinstance(io, list):
-179 d['obsdata'].append(write_List_to_dict(io))
-180 elif isinstance(io, np.ndarray):
-181 d['obsdata'].append(write_Array_to_dict(io))
-182 elif isinstance(io, Corr):
-183 d['obsdata'].append(write_Corr_to_dict(io))
-184 else:
-185 raise Exception("Unkown datatype.")
-186
-187 def _jsonifier(o):
-188 if isinstance(o, np.int64):
-189 return int(o)
-190 raise TypeError('%r is not JSON serializable' % o)
+ 33
+ 34 Returns
+ 35 -------
+ 36 json_string : str
+ 37 String for export to .json(.gz) file
+ 38 """
+ 39
+ 40 def _gen_data_d_from_list(ol):
+ 41 dl = []
+ 42 No = len(ol)
+ 43 for name in ol[0].mc_names:
+ 44 ed = {}
+ 45 ed['id'] = name
+ 46 ed['replica'] = []
+ 47 for r_name in ol[0].e_content[name]:
+ 48 rd = {}
+ 49 rd['name'] = r_name
+ 50 rd['deltas'] = []
+ 51 offsets = [o.r_values[r_name] - o.value for o in ol]
+ 52 deltas = np.column_stack([ol[oi].deltas[r_name] + offsets[oi] for oi in range(No)])
+ 53 for i in range(len(ol[0].idl[r_name])):
+ 54 rd['deltas'].append([ol[0].idl[r_name][i]])
+ 55 rd['deltas'][-1] += deltas[i].tolist()
+ 56 ed['replica'].append(rd)
+ 57 dl.append(ed)
+ 58 return dl
+ 59
+ 60 def _gen_cdata_d_from_list(ol):
+ 61 dl = []
+ 62 for name in ol[0].cov_names:
+ 63 ed = {}
+ 64 ed['id'] = name
+ 65 ed['layout'] = str(ol[0].covobs[name].cov.shape).lstrip('(').rstrip(')').rstrip(',')
+ 66 ed['cov'] = list(np.ravel(ol[0].covobs[name].cov))
+ 67 ncov = ol[0].covobs[name].cov.shape[0]
+ 68 ed['grad'] = []
+ 69 for i in range(ncov):
+ 70 ed['grad'].append([])
+ 71 for o in ol:
+ 72 ed['grad'][-1].append(o.covobs[name].grad[i][0])
+ 73 dl.append(ed)
+ 74 return dl
+ 75
+ 76 def write_Obs_to_dict(o):
+ 77 d = {}
+ 78 d['type'] = 'Obs'
+ 79 d['layout'] = '1'
+ 80 if o.tag:
+ 81 d['tag'] = [o.tag]
+ 82 if o.reweighted:
+ 83 d['reweighted'] = o.reweighted
+ 84 d['value'] = [o.value]
+ 85 data = _gen_data_d_from_list([o])
+ 86 if len(data) > 0:
+ 87 d['data'] = data
+ 88 cdata = _gen_cdata_d_from_list([o])
+ 89 if len(cdata) > 0:
+ 90 d['cdata'] = cdata
+ 91 return d
+ 92
+ 93 def write_List_to_dict(ol):
+ 94 _assert_equal_properties(ol)
+ 95 d = {}
+ 96 d['type'] = 'List'
+ 97 d['layout'] = '%d' % len(ol)
+ 98 taglist = [o.tag for o in ol]
+ 99 if np.any([tag is not None for tag in taglist]):
+100 d['tag'] = taglist
+101 if ol[0].reweighted:
+102 d['reweighted'] = ol[0].reweighted
+103 d['value'] = [o.value for o in ol]
+104 data = _gen_data_d_from_list(ol)
+105 if len(data) > 0:
+106 d['data'] = data
+107 cdata = _gen_cdata_d_from_list(ol)
+108 if len(cdata) > 0:
+109 d['cdata'] = cdata
+110 return d
+111
+112 def write_Array_to_dict(oa):
+113 ol = np.ravel(oa)
+114 _assert_equal_properties(ol)
+115 d = {}
+116 d['type'] = 'Array'
+117 d['layout'] = str(oa.shape).lstrip('(').rstrip(')').rstrip(',')
+118 taglist = [o.tag for o in ol]
+119 if np.any([tag is not None for tag in taglist]):
+120 d['tag'] = taglist
+121 if ol[0].reweighted:
+122 d['reweighted'] = ol[0].reweighted
+123 d['value'] = [o.value for o in ol]
+124 data = _gen_data_d_from_list(ol)
+125 if len(data) > 0:
+126 d['data'] = data
+127 cdata = _gen_cdata_d_from_list(ol)
+128 if len(cdata) > 0:
+129 d['cdata'] = cdata
+130 return d
+131
+132 def _nan_Obs_like(obs):
+133 samples = []
+134 names = []
+135 idl = []
+136 for key, value in obs.idl.items():
+137 samples.append([np.nan] * len(value))
+138 names.append(key)
+139 idl.append(value)
+140 my_obs = Obs(samples, names, idl)
+141 my_obs._covobs = obs._covobs
+142 for name in obs._covobs:
+143 my_obs.names.append(name)
+144 my_obs.reweighted = obs.reweighted
+145 return my_obs
+146
+147 def write_Corr_to_dict(my_corr):
+148 first_not_none = next(i for i, j in enumerate(my_corr.content) if np.all(j))
+149 dummy_array = np.empty((my_corr.N, my_corr.N), dtype=object)
+150 dummy_array[:] = _nan_Obs_like(my_corr.content[first_not_none].ravel()[0])
+151 content = [o if o is not None else dummy_array for o in my_corr.content]
+152 dat = write_Array_to_dict(np.array(content, dtype=object))
+153 dat['type'] = 'Corr'
+154 corr_meta_data = str(my_corr.tag)
+155 if 'tag' in dat.keys():
+156 dat['tag'].append(corr_meta_data)
+157 else:
+158 dat['tag'] = [corr_meta_data]
+159 taglist = dat['tag']
+160 dat['tag'] = {} # tag is now a dictionary, that contains the previous taglist in the key "tag"
+161 dat['tag']['tag'] = taglist
+162 if my_corr.prange is not None:
+163 dat['tag']['prange'] = my_corr.prange
+164 return dat
+165
+166 if not isinstance(ol, list):
+167 ol = [ol]
+168
+169 d = {}
+170 d['program'] = 'pyerrors %s' % (pyerrorsversion.__version__)
+171 d['version'] = '1.1'
+172 d['who'] = getpass.getuser()
+173 d['date'] = datetime.datetime.now().astimezone().strftime('%Y-%m-%d %H:%M:%S %z')
+174 d['host'] = socket.gethostname() + ', ' + platform.platform()
+175
+176 if description:
+177 d['description'] = description
+178
+179 d['obsdata'] = []
+180 for io in ol:
+181 if isinstance(io, Obs):
+182 d['obsdata'].append(write_Obs_to_dict(io))
+183 elif isinstance(io, list):
+184 d['obsdata'].append(write_List_to_dict(io))
+185 elif isinstance(io, np.ndarray):
+186 d['obsdata'].append(write_Array_to_dict(io))
+187 elif isinstance(io, Corr):
+188 d['obsdata'].append(write_Corr_to_dict(io))
+189 else:
+190 raise Exception("Unkown datatype.")
191
-192 if indent:
-193 return json.dumps(d, indent=indent, ensure_ascii=False, default=_jsonifier, write_mode=json.WM_SINGLE_LINE_ARRAY)
-194 else:
-195 return json.dumps(d, indent=indent, ensure_ascii=False, default=_jsonifier, write_mode=json.WM_COMPACT)
+192 def _jsonifier(o):
+193 if isinstance(o, np.int64):
+194 return int(o)
+195 raise TypeError('%r is not JSON serializable' % o)
+196
+197 if indent:
+198 return json.dumps(d, indent=indent, ensure_ascii=False, default=_jsonifier, write_mode=json.WM_SINGLE_LINE_ARRAY)
+199 else:
+200 return json.dumps(d, indent=indent, ensure_ascii=False, default=_jsonifier, write_mode=json.WM_COMPACT)
@@ -1001,6 +1060,13 @@ Optional string that describes the contents of the json file.
Specify the indentation level of the json file. None or 0 is permissible and
saves disk space.
+
+Returns
+
+
+- json_string (str):
+String for export to .json(.gz) file
+
@@ -1016,41 +1082,45 @@ saves disk space.
- 198def dump_to_json(ol, fname, description='', indent=1, gz=True):
-199 """Export a list of Obs or structures containing Obs to a .json(.gz) file
-200
-201 Parameters
-202 ----------
-203 ol : list
-204 List of objects that will be exported. At the moment, these objects can be
-205 either of: Obs, list, numpy.ndarray, Corr.
-206 All Obs inside a structure have to be defined on the same set of configurations.
-207 fname : str
-208 Filename of the output file.
-209 description : str
-210 Optional string that describes the contents of the json file.
-211 indent : int
-212 Specify the indentation level of the json file. None or 0 is permissible and
-213 saves disk space.
-214 gz : bool
-215 If True, the output is a gzipped json. If False, the output is a json file.
-216 """
-217
-218 jsonstring = create_json_string(ol, description, indent)
-219
-220 if not fname.endswith('.json') and not fname.endswith('.gz'):
-221 fname += '.json'
-222
-223 if gz:
-224 if not fname.endswith('.gz'):
-225 fname += '.gz'
+ 203def dump_to_json(ol, fname, description='', indent=1, gz=True):
+204 """Export a list of Obs or structures containing Obs to a .json(.gz) file
+205
+206 Parameters
+207 ----------
+208 ol : list
+209 List of objects that will be exported. At the moment, these objects can be
+210 either of: Obs, list, numpy.ndarray, Corr.
+211 All Obs inside a structure have to be defined on the same set of configurations.
+212 fname : str
+213 Filename of the output file.
+214 description : str
+215 Optional string that describes the contents of the json file.
+216 indent : int
+217 Specify the indentation level of the json file. None or 0 is permissible and
+218 saves disk space.
+219 gz : bool
+220 If True, the output is a gzipped json. If False, the output is a json file.
+221
+222 Returns
+223 -------
+224 Null
+225 """
226
-227 fp = gzip.open(fname, 'wb')
-228 fp.write(jsonstring.encode('utf-8'))
-229 else:
-230 fp = open(fname, 'w', encoding='utf-8')
-231 fp.write(jsonstring)
-232 fp.close()
+227 jsonstring = create_json_string(ol, description, indent)
+228
+229 if not fname.endswith('.json') and not fname.endswith('.gz'):
+230 fname += '.json'
+231
+232 if gz:
+233 if not fname.endswith('.gz'):
+234 fname += '.gz'
+235
+236 fp = gzip.open(fname, 'wb')
+237 fp.write(jsonstring.encode('utf-8'))
+238 else:
+239 fp = open(fname, 'w', encoding='utf-8')
+240 fp.write(jsonstring)
+241 fp.close()
@@ -1073,6 +1143,12 @@ saves disk space.
gz (bool):
If True, the output is a gzipped json. If False, the output is a json file.
+
+Returns
+
+
@@ -1088,24 +1164,35 @@ If True, the output is a gzipped json. If False, the output is a json file.
- 436def import_json_string(json_string, verbose=True, full_output=False):
-437 """Reconstruct a list of Obs or structures containing Obs from a json string.
-438
-439 The following structures are supported: Obs, list, numpy.ndarray, Corr
-440 If the list contains only one element, it is unpacked from the list.
-441
-442 Parameters
-443 ----------
-444 json_string : str
-445 json string containing the data.
-446 verbose : bool
-447 Print additional information that was written to the file.
-448 full_output : bool
-449 If True, a dict containing auxiliary information and the data is returned.
-450 If False, only the data is returned.
-451 """
-452
-453 return _parse_json_dict(json.loads(json_string), verbose, full_output)
+ 456def import_json_string(json_string, verbose=True, full_output=False):
+457 """Reconstruct a list of Obs or structures containing Obs from a json string.
+458
+459 The following structures are supported: Obs, list, numpy.ndarray, Corr
+460 If the list contains only one element, it is unpacked from the list.
+461
+462 Parameters
+463 ----------
+464 json_string : str
+465 json string containing the data.
+466 verbose : bool
+467 Print additional information that was written to the file.
+468 full_output : bool
+469 If True, a dict containing auxiliary information and the data is returned.
+470 If False, only the data is returned.
+471
+472 Returns
+473 -------
+474 result : list[Obs]
+475 reconstructed list of observables from the json string
+476 or
+477 result : Obs
+478 only one observable if the list only has one entry
+479 or
+480 result : dict
+481 if full_output=True
+482 """
+483
+484 return _parse_json_dict(json.loads(json_string), verbose, full_output)
@@ -1125,6 +1212,19 @@ Print additional information that was written to the file.
If True, a dict containing auxiliary information and the data is returned.
If False, only the data is returned.
+
+Returns
+
+
+- result (list[Obs]):
+reconstructed list of observables from the json string
+- or
+- result (Obs):
+only one observable if the list only has one entry
+- or
+- result (dict):
+if full_output=True
+
@@ -1140,38 +1240,49 @@ If False, only the data is returned.
- 456def load_json(fname, verbose=True, gz=True, full_output=False):
-457 """Import a list of Obs or structures containing Obs from a .json(.gz) file.
-458
-459 The following structures are supported: Obs, list, numpy.ndarray, Corr
-460 If the list contains only one element, it is unpacked from the list.
-461
-462 Parameters
-463 ----------
-464 fname : str
-465 Filename of the input file.
-466 verbose : bool
-467 Print additional information that was written to the file.
-468 gz : bool
-469 If True, assumes that data is gzipped. If False, assumes JSON file.
-470 full_output : bool
-471 If True, a dict containing auxiliary information and the data is returned.
-472 If False, only the data is returned.
-473 """
-474 if not fname.endswith('.json') and not fname.endswith('.gz'):
-475 fname += '.json'
-476 if gz:
-477 if not fname.endswith('.gz'):
-478 fname += '.gz'
-479 with gzip.open(fname, 'r') as fin:
-480 d = json.load(fin)
-481 else:
-482 if fname.endswith('.gz'):
-483 warnings.warn("Trying to read from %s without unzipping!" % fname, UserWarning)
-484 with open(fname, 'r', encoding='utf-8') as fin:
-485 d = json.loads(fin.read())
-486
-487 return _parse_json_dict(d, verbose, full_output)
+ 487def load_json(fname, verbose=True, gz=True, full_output=False):
+488 """Import a list of Obs or structures containing Obs from a .json(.gz) file.
+489
+490 The following structures are supported: Obs, list, numpy.ndarray, Corr
+491 If the list contains only one element, it is unpacked from the list.
+492
+493 Parameters
+494 ----------
+495 fname : str
+496 Filename of the input file.
+497 verbose : bool
+498 Print additional information that was written to the file.
+499 gz : bool
+500 If True, assumes that data is gzipped. If False, assumes JSON file.
+501 full_output : bool
+502 If True, a dict containing auxiliary information and the data is returned.
+503 If False, only the data is returned.
+504
+505 Returns
+506 -------
+507 result : list[Obs]
+508 reconstructed list of observables from the json string
+509 or
+510 result : Obs
+511 only one observable if the list only has one entry
+512 or
+513 result : dict
+514 if full_output=True
+515 """
+516 if not fname.endswith('.json') and not fname.endswith('.gz'):
+517 fname += '.json'
+518 if gz:
+519 if not fname.endswith('.gz'):
+520 fname += '.gz'
+521 with gzip.open(fname, 'r') as fin:
+522 d = json.load(fin)
+523 else:
+524 if fname.endswith('.gz'):
+525 warnings.warn("Trying to read from %s without unzipping!" % fname, UserWarning)
+526 with open(fname, 'r', encoding='utf-8') as fin:
+527 d = json.loads(fin.read())
+528
+529 return _parse_json_dict(d, verbose, full_output)
@@ -1193,6 +1304,19 @@ If True, assumes that data is gzipped. If False, assumes JSON file.
If True, a dict containing auxiliary information and the data is returned.
If False, only the data is returned.
+
+Returns
+
+
+- result (list[Obs]):
+reconstructed list of observables from the json string
+- or
+- result (Obs):
+only one observable if the list only has one entry
+- or
+- result (dict):
+if full_output=True
+
@@ -1208,40 +1332,44 @@ If False, only the data is returned.
- 570def dump_dict_to_json(od, fname, description='', indent=1, reps='DICTOBS', gz=True):
-571 """Export a dict of Obs or structures containing Obs to a .json(.gz) file
-572
-573 Parameters
-574 ----------
-575 od : dict
-576 Dict of JSON valid structures and objects that will be exported.
-577 At the moment, these objects can be either of: Obs, list, numpy.ndarray, Corr.
-578 All Obs inside a structure have to be defined on the same set of configurations.
-579 fname : str
-580 Filename of the output file.
-581 description : str
-582 Optional string that describes the contents of the json file.
-583 indent : int
-584 Specify the indentation level of the json file. None or 0 is permissible and
-585 saves disk space.
-586 reps : str
-587 Specify the structure of the placeholder in exported dict to be reps[0-9]+.
-588 gz : bool
-589 If True, the output is a gzipped json. If False, the output is a json file.
-590 """
-591
-592 if not isinstance(od, dict):
-593 raise Exception('od has to be a dictionary. Did you want to use dump_to_json?')
-594
-595 infostring = ('This JSON file contains a python dictionary that has been parsed to a list of structures. '
-596 'OBSDICT contains the dictionary, where Obs or other structures have been replaced by '
-597 '' + reps + '[0-9]+. The field description contains the additional description of this JSON file. '
-598 'This file may be parsed to a dict with the pyerrors routine load_json_dict.')
-599
-600 desc_dict = {'INFO': infostring, 'OBSDICT': {}, 'description': description}
-601 ol, desc_dict['OBSDICT'] = _ol_from_dict(od, reps=reps)
-602
-603 dump_to_json(ol, fname, description=desc_dict, indent=indent, gz=gz)
+ 612def dump_dict_to_json(od, fname, description='', indent=1, reps='DICTOBS', gz=True):
+613 """Export a dict of Obs or structures containing Obs to a .json(.gz) file
+614
+615 Parameters
+616 ----------
+617 od : dict
+618 Dict of JSON valid structures and objects that will be exported.
+619 At the moment, these objects can be either of: Obs, list, numpy.ndarray, Corr.
+620 All Obs inside a structure have to be defined on the same set of configurations.
+621 fname : str
+622 Filename of the output file.
+623 description : str
+624 Optional string that describes the contents of the json file.
+625 indent : int
+626 Specify the indentation level of the json file. None or 0 is permissible and
+627 saves disk space.
+628 reps : str
+629 Specify the structure of the placeholder in exported dict to be reps[0-9]+.
+630 gz : bool
+631 If True, the output is a gzipped json. If False, the output is a json file.
+632
+633 Returns
+634 -------
+635 None
+636 """
+637
+638 if not isinstance(od, dict):
+639 raise Exception('od has to be a dictionary. Did you want to use dump_to_json?')
+640
+641 infostring = ('This JSON file contains a python dictionary that has been parsed to a list of structures. '
+642 'OBSDICT contains the dictionary, where Obs or other structures have been replaced by '
+643 '' + reps + '[0-9]+. The field description contains the additional description of this JSON file. '
+644 'This file may be parsed to a dict with the pyerrors routine load_json_dict.')
+645
+646 desc_dict = {'INFO': infostring, 'OBSDICT': {}, 'description': description}
+647 ol, desc_dict['OBSDICT'] = _ol_from_dict(od, reps=reps)
+648
+649 dump_to_json(ol, fname, description=desc_dict, indent=indent, gz=gz)
@@ -1266,6 +1394,12 @@ Specify the structure of the placeholder in exported dict to be reps[0-9]+.
gz (bool):
If True, the output is a gzipped json. If False, the output is a json file.
+
+Returns
+
+
@@ -1281,37 +1415,45 @@ If True, the output is a gzipped json. If False, the output is a json file.
- 669def load_json_dict(fname, verbose=True, gz=True, full_output=False, reps='DICTOBS'):
-670 """Import a dict of Obs or structures containing Obs from a .json(.gz) file.
-671
-672 The following structures are supported: Obs, list, numpy.ndarray, Corr
-673
-674 Parameters
-675 ----------
-676 fname : str
-677 Filename of the input file.
-678 verbose : bool
-679 Print additional information that was written to the file.
-680 gz : bool
-681 If True, assumes that data is gzipped. If False, assumes JSON file.
-682 full_output : bool
-683 If True, a dict containing auxiliary information and the data is returned.
-684 If False, only the data is returned.
-685 reps : str
-686 Specify the structure of the placeholder in imported dict to be reps[0-9]+.
-687 """
-688 indata = load_json(fname, verbose=verbose, gz=gz, full_output=True)
-689 description = indata['description']['description']
-690 indict = indata['description']['OBSDICT']
-691 ol = indata['obsdata']
-692 od = _od_from_list_and_dict(ol, indict, reps=reps)
-693
-694 if full_output:
-695 indata['description'] = description
-696 indata['obsdata'] = od
-697 return indata
-698 else:
-699 return od
+ 715def load_json_dict(fname, verbose=True, gz=True, full_output=False, reps='DICTOBS'):
+716 """Import a dict of Obs or structures containing Obs from a .json(.gz) file.
+717
+718 The following structures are supported: Obs, list, numpy.ndarray, Corr
+719
+720 Parameters
+721 ----------
+722 fname : str
+723 Filename of the input file.
+724 verbose : bool
+725 Print additional information that was written to the file.
+726 gz : bool
+727 If True, assumes that data is gzipped. If False, assumes JSON file.
+728 full_output : bool
+729 If True, a dict containing auxiliary information and the data is returned.
+730 If False, only the data is returned.
+731 reps : str
+732 Specify the structure of the placeholder in imported dict to be reps[0-9]+.
+733
+734 Returns
+735 -------
+736 data : Obs / list / Corr
+737 Read data
+738 or
+739 data : dict
+740 Read data and meta-data
+741 """
+742 indata = load_json(fname, verbose=verbose, gz=gz, full_output=True)
+743 description = indata['description']['description']
+744 indict = indata['description']['OBSDICT']
+745 ol = indata['obsdata']
+746 od = _od_from_list_and_dict(ol, indict, reps=reps)
+747
+748 if full_output:
+749 indata['description'] = description
+750 indata['obsdata'] = od
+751 return indata
+752 else:
+753 return od
@@ -1334,6 +1476,16 @@ If False, only the data is returned.
reps (str):
Specify the structure of the placeholder in imported dict to be reps[0-9]+.
+
+Returns
+
+
+- data (Obs / list / Corr):
+Read data
+- or
+- data (dict):
+Read data and meta-data
+
diff --git a/docs/pyerrors/input/misc.html b/docs/pyerrors/input/misc.html
index 307135a1..d7729156 100644
--- a/docs/pyerrors/input/misc.html
+++ b/docs/pyerrors/input/misc.html
@@ -85,7 +85,7 @@
7
8
9def read_pbp(path, prefix, **kwargs):
- 10 """Read pbp format from given folder structure. Returns a list of length nrw
+ 10 """Read pbp format from given folder structure.
11
12 Parameters
13 ----------
@@ -93,111 +93,116 @@
15 list which contains the first config to be read for each replicum
16 r_stop : list
17 list which contains the last config to be read for each replicum
- 18 """
- 19
- 20 ls = []
- 21 for (dirpath, dirnames, filenames) in os.walk(path):
- 22 ls.extend(filenames)
- 23 break
+ 18
+ 19 Returns
+ 20 -------
+ 21 result : list[Obs]
+ 22 list of observables read
+ 23 """
24
- 25 if not ls:
- 26 raise Exception('Error, directory not found')
- 27
- 28 # Exclude files with different names
- 29 for exc in ls:
- 30 if not fnmatch.fnmatch(exc, prefix + '*.dat'):
- 31 ls = list(set(ls) - set([exc]))
- 32 if len(ls) > 1:
- 33 ls.sort(key=lambda x: int(re.findall(r'\d+', x[len(prefix):])[0]))
- 34 replica = len(ls)
- 35
- 36 if 'r_start' in kwargs:
- 37 r_start = kwargs.get('r_start')
- 38 if len(r_start) != replica:
- 39 raise Exception('r_start does not match number of replicas')
- 40 # Adjust Configuration numbering to python index
- 41 r_start = [o - 1 if o else None for o in r_start]
- 42 else:
- 43 r_start = [None] * replica
- 44
- 45 if 'r_stop' in kwargs:
- 46 r_stop = kwargs.get('r_stop')
- 47 if len(r_stop) != replica:
- 48 raise Exception('r_stop does not match number of replicas')
- 49 else:
- 50 r_stop = [None] * replica
- 51
- 52 print(r'Read <bar{psi}\psi> from', prefix[:-1], ',', replica, 'replica', end='')
- 53
- 54 print_err = 0
- 55 if 'print_err' in kwargs:
- 56 print_err = 1
- 57 print()
+ 25 ls = []
+ 26 for (dirpath, dirnames, filenames) in os.walk(path):
+ 27 ls.extend(filenames)
+ 28 break
+ 29
+ 30 if not ls:
+ 31 raise Exception('Error, directory not found')
+ 32
+ 33 # Exclude files with different names
+ 34 for exc in ls:
+ 35 if not fnmatch.fnmatch(exc, prefix + '*.dat'):
+ 36 ls = list(set(ls) - set([exc]))
+ 37 if len(ls) > 1:
+ 38 ls.sort(key=lambda x: int(re.findall(r'\d+', x[len(prefix):])[0]))
+ 39 replica = len(ls)
+ 40
+ 41 if 'r_start' in kwargs:
+ 42 r_start = kwargs.get('r_start')
+ 43 if len(r_start) != replica:
+ 44 raise Exception('r_start does not match number of replicas')
+ 45 # Adjust Configuration numbering to python index
+ 46 r_start = [o - 1 if o else None for o in r_start]
+ 47 else:
+ 48 r_start = [None] * replica
+ 49
+ 50 if 'r_stop' in kwargs:
+ 51 r_stop = kwargs.get('r_stop')
+ 52 if len(r_stop) != replica:
+ 53 raise Exception('r_stop does not match number of replicas')
+ 54 else:
+ 55 r_stop = [None] * replica
+ 56
+ 57 print(r'Read <bar{psi}\psi> from', prefix[:-1], ',', replica, 'replica', end='')
58
- 59 deltas = []
- 60
- 61 for rep in range(replica):
- 62 tmp_array = []
- 63 with open(path + '/' + ls[rep], 'rb') as fp:
- 64
- 65 t = fp.read(4) # number of reweighting factors
- 66 if rep == 0:
- 67 nrw = struct.unpack('i', t)[0]
- 68 for k in range(nrw):
- 69 deltas.append([])
- 70 else:
- 71 if nrw != struct.unpack('i', t)[0]:
- 72 raise Exception('Error: different number of factors for replicum', rep)
- 73
- 74 for k in range(nrw):
- 75 tmp_array.append([])
- 76
- 77 # This block is necessary for openQCD1.6 ms1 files
- 78 nfct = []
- 79 for i in range(nrw):
- 80 t = fp.read(4)
- 81 nfct.append(struct.unpack('i', t)[0])
- 82 print('nfct: ', nfct) # Hasenbusch factor, 1 for rat reweighting
- 83
- 84 nsrc = []
- 85 for i in range(nrw):
- 86 t = fp.read(4)
- 87 nsrc.append(struct.unpack('i', t)[0])
+ 59 print_err = 0
+ 60 if 'print_err' in kwargs:
+ 61 print_err = 1
+ 62 print()
+ 63
+ 64 deltas = []
+ 65
+ 66 for rep in range(replica):
+ 67 tmp_array = []
+ 68 with open(path + '/' + ls[rep], 'rb') as fp:
+ 69
+ 70 t = fp.read(4) # number of reweighting factors
+ 71 if rep == 0:
+ 72 nrw = struct.unpack('i', t)[0]
+ 73 for k in range(nrw):
+ 74 deltas.append([])
+ 75 else:
+ 76 if nrw != struct.unpack('i', t)[0]:
+ 77 raise Exception('Error: different number of factors for replicum', rep)
+ 78
+ 79 for k in range(nrw):
+ 80 tmp_array.append([])
+ 81
+ 82 # This block is necessary for openQCD1.6 ms1 files
+ 83 nfct = []
+ 84 for i in range(nrw):
+ 85 t = fp.read(4)
+ 86 nfct.append(struct.unpack('i', t)[0])
+ 87 print('nfct: ', nfct) # Hasenbusch factor, 1 for rat reweighting
88
- 89 # body
- 90 while True:
+ 89 nsrc = []
+ 90 for i in range(nrw):
91 t = fp.read(4)
- 92 if len(t) < 4:
- 93 break
- 94 if print_err:
- 95 config_no = struct.unpack('i', t)
- 96 for i in range(nrw):
- 97 tmp_nfct = 1.0
- 98 for j in range(nfct[i]):
- 99 t = fp.read(8 * nsrc[i])
-100 t = fp.read(8 * nsrc[i])
-101 tmp_rw = struct.unpack('d' * nsrc[i], t)
-102 tmp_nfct *= np.mean(np.asarray(tmp_rw))
-103 if print_err:
-104 print(config_no, i, j, np.mean(np.asarray(tmp_rw)), np.std(np.asarray(tmp_rw)))
-105 print('Sources:', np.asarray(tmp_rw))
-106 print('Partial factor:', tmp_nfct)
-107 tmp_array[i].append(tmp_nfct)
-108
-109 for k in range(nrw):
-110 deltas[k].append(tmp_array[k][r_start[rep]:r_stop[rep]])
-111
-112 rep_names = []
-113 for entry in ls:
-114 truncated_entry = entry.split('.')[0]
-115 idx = truncated_entry.index('r')
-116 rep_names.append(truncated_entry[:idx] + '|' + truncated_entry[idx:])
-117 print(',', nrw, r'<bar{psi}\psi> with', nsrc, 'sources')
-118 result = []
-119 for t in range(nrw):
-120 result.append(Obs(deltas[t], rep_names))
-121
-122 return result
+ 92 nsrc.append(struct.unpack('i', t)[0])
+ 93
+ 94 # body
+ 95 while True:
+ 96 t = fp.read(4)
+ 97 if len(t) < 4:
+ 98 break
+ 99 if print_err:
+100 config_no = struct.unpack('i', t)
+101 for i in range(nrw):
+102 tmp_nfct = 1.0
+103 for j in range(nfct[i]):
+104 t = fp.read(8 * nsrc[i])
+105 t = fp.read(8 * nsrc[i])
+106 tmp_rw = struct.unpack('d' * nsrc[i], t)
+107 tmp_nfct *= np.mean(np.asarray(tmp_rw))
+108 if print_err:
+109 print(config_no, i, j, np.mean(np.asarray(tmp_rw)), np.std(np.asarray(tmp_rw)))
+110 print('Sources:', np.asarray(tmp_rw))
+111 print('Partial factor:', tmp_nfct)
+112 tmp_array[i].append(tmp_nfct)
+113
+114 for k in range(nrw):
+115 deltas[k].append(tmp_array[k][r_start[rep]:r_stop[rep]])
+116
+117 rep_names = []
+118 for entry in ls:
+119 truncated_entry = entry.split('.')[0]
+120 idx = truncated_entry.index('r')
+121 rep_names.append(truncated_entry[:idx] + '|' + truncated_entry[idx:])
+122 print(',', nrw, r'<bar{psi}\psi> with', nsrc, 'sources')
+123 result = []
+124 for t in range(nrw):
+125 result.append(Obs(deltas[t], rep_names))
+126
+127 return result
@@ -214,7 +219,7 @@
10def read_pbp(path, prefix, **kwargs):
- 11 """Read pbp format from given folder structure. Returns a list of length nrw
+ 11 """Read pbp format from given folder structure.
12
13 Parameters
14 ----------
@@ -222,115 +227,120 @@
16 list which contains the first config to be read for each replicum
17 r_stop : list
18 list which contains the last config to be read for each replicum
- 19 """
- 20
- 21 ls = []
- 22 for (dirpath, dirnames, filenames) in os.walk(path):
- 23 ls.extend(filenames)
- 24 break
+ 19
+ 20 Returns
+ 21 -------
+ 22 result : list[Obs]
+ 23 list of observables read
+ 24 """
25
- 26 if not ls:
- 27 raise Exception('Error, directory not found')
- 28
- 29 # Exclude files with different names
- 30 for exc in ls:
- 31 if not fnmatch.fnmatch(exc, prefix + '*.dat'):
- 32 ls = list(set(ls) - set([exc]))
- 33 if len(ls) > 1:
- 34 ls.sort(key=lambda x: int(re.findall(r'\d+', x[len(prefix):])[0]))
- 35 replica = len(ls)
- 36
- 37 if 'r_start' in kwargs:
- 38 r_start = kwargs.get('r_start')
- 39 if len(r_start) != replica:
- 40 raise Exception('r_start does not match number of replicas')
- 41 # Adjust Configuration numbering to python index
- 42 r_start = [o - 1 if o else None for o in r_start]
- 43 else:
- 44 r_start = [None] * replica
- 45
- 46 if 'r_stop' in kwargs:
- 47 r_stop = kwargs.get('r_stop')
- 48 if len(r_stop) != replica:
- 49 raise Exception('r_stop does not match number of replicas')
- 50 else:
- 51 r_stop = [None] * replica
- 52
- 53 print(r'Read <bar{psi}\psi> from', prefix[:-1], ',', replica, 'replica', end='')
- 54
- 55 print_err = 0
- 56 if 'print_err' in kwargs:
- 57 print_err = 1
- 58 print()
+ 26 ls = []
+ 27 for (dirpath, dirnames, filenames) in os.walk(path):
+ 28 ls.extend(filenames)
+ 29 break
+ 30
+ 31 if not ls:
+ 32 raise Exception('Error, directory not found')
+ 33
+ 34 # Exclude files with different names
+ 35 for exc in ls:
+ 36 if not fnmatch.fnmatch(exc, prefix + '*.dat'):
+ 37 ls = list(set(ls) - set([exc]))
+ 38 if len(ls) > 1:
+ 39 ls.sort(key=lambda x: int(re.findall(r'\d+', x[len(prefix):])[0]))
+ 40 replica = len(ls)
+ 41
+ 42 if 'r_start' in kwargs:
+ 43 r_start = kwargs.get('r_start')
+ 44 if len(r_start) != replica:
+ 45 raise Exception('r_start does not match number of replicas')
+ 46 # Adjust Configuration numbering to python index
+ 47 r_start = [o - 1 if o else None for o in r_start]
+ 48 else:
+ 49 r_start = [None] * replica
+ 50
+ 51 if 'r_stop' in kwargs:
+ 52 r_stop = kwargs.get('r_stop')
+ 53 if len(r_stop) != replica:
+ 54 raise Exception('r_stop does not match number of replicas')
+ 55 else:
+ 56 r_stop = [None] * replica
+ 57
+ 58 print(r'Read <bar{psi}\psi> from', prefix[:-1], ',', replica, 'replica', end='')
59
- 60 deltas = []
- 61
- 62 for rep in range(replica):
- 63 tmp_array = []
- 64 with open(path + '/' + ls[rep], 'rb') as fp:
- 65
- 66 t = fp.read(4) # number of reweighting factors
- 67 if rep == 0:
- 68 nrw = struct.unpack('i', t)[0]
- 69 for k in range(nrw):
- 70 deltas.append([])
- 71 else:
- 72 if nrw != struct.unpack('i', t)[0]:
- 73 raise Exception('Error: different number of factors for replicum', rep)
- 74
- 75 for k in range(nrw):
- 76 tmp_array.append([])
- 77
- 78 # This block is necessary for openQCD1.6 ms1 files
- 79 nfct = []
- 80 for i in range(nrw):
- 81 t = fp.read(4)
- 82 nfct.append(struct.unpack('i', t)[0])
- 83 print('nfct: ', nfct) # Hasenbusch factor, 1 for rat reweighting
- 84
- 85 nsrc = []
- 86 for i in range(nrw):
- 87 t = fp.read(4)
- 88 nsrc.append(struct.unpack('i', t)[0])
+ 60 print_err = 0
+ 61 if 'print_err' in kwargs:
+ 62 print_err = 1
+ 63 print()
+ 64
+ 65 deltas = []
+ 66
+ 67 for rep in range(replica):
+ 68 tmp_array = []
+ 69 with open(path + '/' + ls[rep], 'rb') as fp:
+ 70
+ 71 t = fp.read(4) # number of reweighting factors
+ 72 if rep == 0:
+ 73 nrw = struct.unpack('i', t)[0]
+ 74 for k in range(nrw):
+ 75 deltas.append([])
+ 76 else:
+ 77 if nrw != struct.unpack('i', t)[0]:
+ 78 raise Exception('Error: different number of factors for replicum', rep)
+ 79
+ 80 for k in range(nrw):
+ 81 tmp_array.append([])
+ 82
+ 83 # This block is necessary for openQCD1.6 ms1 files
+ 84 nfct = []
+ 85 for i in range(nrw):
+ 86 t = fp.read(4)
+ 87 nfct.append(struct.unpack('i', t)[0])
+ 88 print('nfct: ', nfct) # Hasenbusch factor, 1 for rat reweighting
89
- 90 # body
- 91 while True:
+ 90 nsrc = []
+ 91 for i in range(nrw):
92 t = fp.read(4)
- 93 if len(t) < 4:
- 94 break
- 95 if print_err:
- 96 config_no = struct.unpack('i', t)
- 97 for i in range(nrw):
- 98 tmp_nfct = 1.0
- 99 for j in range(nfct[i]):
-100 t = fp.read(8 * nsrc[i])
-101 t = fp.read(8 * nsrc[i])
-102 tmp_rw = struct.unpack('d' * nsrc[i], t)
-103 tmp_nfct *= np.mean(np.asarray(tmp_rw))
-104 if print_err:
-105 print(config_no, i, j, np.mean(np.asarray(tmp_rw)), np.std(np.asarray(tmp_rw)))
-106 print('Sources:', np.asarray(tmp_rw))
-107 print('Partial factor:', tmp_nfct)
-108 tmp_array[i].append(tmp_nfct)
-109
-110 for k in range(nrw):
-111 deltas[k].append(tmp_array[k][r_start[rep]:r_stop[rep]])
-112
-113 rep_names = []
-114 for entry in ls:
-115 truncated_entry = entry.split('.')[0]
-116 idx = truncated_entry.index('r')
-117 rep_names.append(truncated_entry[:idx] + '|' + truncated_entry[idx:])
-118 print(',', nrw, r'<bar{psi}\psi> with', nsrc, 'sources')
-119 result = []
-120 for t in range(nrw):
-121 result.append(Obs(deltas[t], rep_names))
-122
-123 return result
+ 93 nsrc.append(struct.unpack('i', t)[0])
+ 94
+ 95 # body
+ 96 while True:
+ 97 t = fp.read(4)
+ 98 if len(t) < 4:
+ 99 break
+100 if print_err:
+101 config_no = struct.unpack('i', t)
+102 for i in range(nrw):
+103 tmp_nfct = 1.0
+104 for j in range(nfct[i]):
+105 t = fp.read(8 * nsrc[i])
+106 t = fp.read(8 * nsrc[i])
+107 tmp_rw = struct.unpack('d' * nsrc[i], t)
+108 tmp_nfct *= np.mean(np.asarray(tmp_rw))
+109 if print_err:
+110 print(config_no, i, j, np.mean(np.asarray(tmp_rw)), np.std(np.asarray(tmp_rw)))
+111 print('Sources:', np.asarray(tmp_rw))
+112 print('Partial factor:', tmp_nfct)
+113 tmp_array[i].append(tmp_nfct)
+114
+115 for k in range(nrw):
+116 deltas[k].append(tmp_array[k][r_start[rep]:r_stop[rep]])
+117
+118 rep_names = []
+119 for entry in ls:
+120 truncated_entry = entry.split('.')[0]
+121 idx = truncated_entry.index('r')
+122 rep_names.append(truncated_entry[:idx] + '|' + truncated_entry[idx:])
+123 print(',', nrw, r'<bar{psi}\psi> with', nsrc, 'sources')
+124 result = []
+125 for t in range(nrw):
+126 result.append(Obs(deltas[t], rep_names))
+127
+128 return result
- Read pbp format from given folder structure. Returns a list of length nrw
+
Read pbp format from given folder structure.
Parameters
@@ -340,6 +350,13 @@ list which contains the first config to be read for each replicum
r_stop (list):
list which contains the last config to be read for each replicum
+
+
Returns
+
+
+- result (list[Obs]):
+list of observables read
+
diff --git a/docs/pyerrors/input/openQCD.html b/docs/pyerrors/input/openQCD.html
index e61d699f..9cef2ade 100644
--- a/docs/pyerrors/input/openQCD.html
+++ b/docs/pyerrors/input/openQCD.html
@@ -138,1094 +138,1124 @@
42 files performed if given.
43 print_err : bool
44 Print additional information that is useful for debugging.
- 45 """
- 46 known_oqcd_versions = ['1.4', '1.6', '2.0']
- 47 if not (version in known_oqcd_versions):
- 48 raise Exception('Unknown openQCD version defined!')
- 49 print("Working with openQCD version " + version)
- 50 if 'postfix' in kwargs:
- 51 postfix = kwargs.get('postfix')
- 52 else:
- 53 postfix = ''
- 54 ls = []
- 55 for (dirpath, dirnames, filenames) in os.walk(path):
- 56 ls.extend(filenames)
- 57 break
- 58
- 59 if not ls:
- 60 raise Exception(f"Error, directory '{path}' not found")
- 61 if 'files' in kwargs:
- 62 ls = kwargs.get('files')
- 63 else:
- 64 for exc in ls:
- 65 if not fnmatch.fnmatch(exc, prefix + '*' + postfix + '.dat'):
- 66 ls = list(set(ls) - set([exc]))
- 67 if len(ls) > 1:
- 68 ls.sort(key=lambda x: int(re.findall(r'\d+', x[len(prefix):])[0]))
- 69 replica = len(ls)
- 70
- 71 if 'r_start' in kwargs:
- 72 r_start = kwargs.get('r_start')
- 73 if len(r_start) != replica:
- 74 raise Exception('r_start does not match number of replicas')
- 75 r_start = [o if o else None for o in r_start]
- 76 else:
- 77 r_start = [None] * replica
- 78
- 79 if 'r_stop' in kwargs:
- 80 r_stop = kwargs.get('r_stop')
- 81 if len(r_stop) != replica:
- 82 raise Exception('r_stop does not match number of replicas')
- 83 else:
- 84 r_stop = [None] * replica
- 85
- 86 if 'r_step' in kwargs:
- 87 r_step = kwargs.get('r_step')
+ 45
+ 46 Returns
+ 47 -------
+ 48 rwms : Obs
+ 49 Reweighting factors read
+ 50 """
+ 51 known_oqcd_versions = ['1.4', '1.6', '2.0']
+ 52 if not (version in known_oqcd_versions):
+ 53 raise Exception('Unknown openQCD version defined!')
+ 54 print("Working with openQCD version " + version)
+ 55 if 'postfix' in kwargs:
+ 56 postfix = kwargs.get('postfix')
+ 57 else:
+ 58 postfix = ''
+ 59 ls = []
+ 60 for (dirpath, dirnames, filenames) in os.walk(path):
+ 61 ls.extend(filenames)
+ 62 break
+ 63
+ 64 if not ls:
+ 65 raise Exception(f"Error, directory '{path}' not found")
+ 66 if 'files' in kwargs:
+ 67 ls = kwargs.get('files')
+ 68 else:
+ 69 for exc in ls:
+ 70 if not fnmatch.fnmatch(exc, prefix + '*' + postfix + '.dat'):
+ 71 ls = list(set(ls) - set([exc]))
+ 72 if len(ls) > 1:
+ 73 ls.sort(key=lambda x: int(re.findall(r'\d+', x[len(prefix):])[0]))
+ 74 replica = len(ls)
+ 75
+ 76 if 'r_start' in kwargs:
+ 77 r_start = kwargs.get('r_start')
+ 78 if len(r_start) != replica:
+ 79 raise Exception('r_start does not match number of replicas')
+ 80 r_start = [o if o else None for o in r_start]
+ 81 else:
+ 82 r_start = [None] * replica
+ 83
+ 84 if 'r_stop' in kwargs:
+ 85 r_stop = kwargs.get('r_stop')
+ 86 if len(r_stop) != replica:
+ 87 raise Exception('r_stop does not match number of replicas')
88 else:
- 89 r_step = 1
+ 89 r_stop = [None] * replica
90
- 91 print('Read reweighting factors from', prefix[:-1], ',',
- 92 replica, 'replica', end='')
- 93
- 94 if names is None:
- 95 rep_names = []
- 96 for entry in ls:
- 97 truncated_entry = entry
- 98 suffixes = [".dat", ".rwms", ".ms1"]
- 99 for suffix in suffixes:
- 100 if truncated_entry.endswith(suffix):
- 101 truncated_entry = truncated_entry[0:-len(suffix)]
- 102 idx = truncated_entry.index('r')
- 103 rep_names.append(truncated_entry[:idx] + '|' + truncated_entry[idx:])
- 104 else:
- 105 rep_names = names
- 106
- 107 print_err = 0
- 108 if 'print_err' in kwargs:
- 109 print_err = 1
- 110 print()
+ 91 if 'r_step' in kwargs:
+ 92 r_step = kwargs.get('r_step')
+ 93 else:
+ 94 r_step = 1
+ 95
+ 96 print('Read reweighting factors from', prefix[:-1], ',',
+ 97 replica, 'replica', end='')
+ 98
+ 99 if names is None:
+ 100 rep_names = []
+ 101 for entry in ls:
+ 102 truncated_entry = entry
+ 103 suffixes = [".dat", ".rwms", ".ms1"]
+ 104 for suffix in suffixes:
+ 105 if truncated_entry.endswith(suffix):
+ 106 truncated_entry = truncated_entry[0:-len(suffix)]
+ 107 idx = truncated_entry.index('r')
+ 108 rep_names.append(truncated_entry[:idx] + '|' + truncated_entry[idx:])
+ 109 else:
+ 110 rep_names = names
111
- 112 deltas = []
- 113
- 114 configlist = []
- 115 r_start_index = []
- 116 r_stop_index = []
- 117
- 118 for rep in range(replica):
- 119 tmp_array = []
- 120 with open(path + '/' + ls[rep], 'rb') as fp:
- 121
- 122 t = fp.read(4) # number of reweighting factors
- 123 if rep == 0:
- 124 nrw = struct.unpack('i', t)[0]
- 125 if version == '2.0':
- 126 nrw = int(nrw / 2)
- 127 for k in range(nrw):
- 128 deltas.append([])
- 129 else:
- 130 if ((nrw != struct.unpack('i', t)[0] and (not version == '2.0')) or (nrw != struct.unpack('i', t)[0] / 2 and version == '2.0')):
- 131 raise Exception('Error: different number of reweighting factors for replicum', rep)
- 132
- 133 for k in range(nrw):
- 134 tmp_array.append([])
- 135
- 136 # This block is necessary for openQCD1.6 and openQCD2.0 ms1 files
- 137 nfct = []
- 138 if version in ['1.6', '2.0']:
- 139 for i in range(nrw):
- 140 t = fp.read(4)
- 141 nfct.append(struct.unpack('i', t)[0])
- 142 else:
- 143 for i in range(nrw):
- 144 nfct.append(1)
- 145
- 146 nsrc = []
- 147 for i in range(nrw):
- 148 t = fp.read(4)
- 149 nsrc.append(struct.unpack('i', t)[0])
- 150 if version == '2.0':
- 151 if not struct.unpack('i', fp.read(4))[0] == 0:
- 152 print('something is wrong!')
- 153
- 154 configlist.append([])
- 155 while True:
- 156 t = fp.read(4)
- 157 if len(t) < 4:
- 158 break
- 159 config_no = struct.unpack('i', t)[0]
- 160 configlist[-1].append(config_no)
- 161 for i in range(nrw):
- 162 if (version == '2.0'):
- 163 tmpd = _read_array_openQCD2(fp)
- 164 tmpd = _read_array_openQCD2(fp)
- 165 tmp_rw = tmpd['arr']
- 166 tmp_nfct = 1.0
- 167 for j in range(tmpd['n'][0]):
- 168 tmp_nfct *= np.mean(np.exp(-np.asarray(tmp_rw[j])))
- 169 if print_err:
- 170 print(config_no, i, j,
- 171 np.mean(np.exp(-np.asarray(tmp_rw[j]))),
- 172 np.std(np.exp(-np.asarray(tmp_rw[j]))))
- 173 print('Sources:',
- 174 np.exp(-np.asarray(tmp_rw[j])))
- 175 print('Partial factor:', tmp_nfct)
- 176 elif version == '1.6' or version == '1.4':
- 177 tmp_nfct = 1.0
- 178 for j in range(nfct[i]):
- 179 t = fp.read(8 * nsrc[i])
- 180 t = fp.read(8 * nsrc[i])
- 181 tmp_rw = struct.unpack('d' * nsrc[i], t)
- 182 tmp_nfct *= np.mean(np.exp(-np.asarray(tmp_rw)))
- 183 if print_err:
- 184 print(config_no, i, j,
- 185 np.mean(np.exp(-np.asarray(tmp_rw))),
- 186 np.std(np.exp(-np.asarray(tmp_rw))))
- 187 print('Sources:', np.exp(-np.asarray(tmp_rw)))
- 188 print('Partial factor:', tmp_nfct)
- 189 tmp_array[i].append(tmp_nfct)
- 190
- 191 diffmeas = configlist[-1][-1] - configlist[-1][-2]
- 192 configlist[-1] = [item // diffmeas for item in configlist[-1]]
- 193 if configlist[-1][0] > 1 and diffmeas > 1:
- 194 warnings.warn('Assume thermalization and that the first measurement belongs to the first config.')
- 195 offset = configlist[-1][0] - 1
- 196 configlist[-1] = [item - offset for item in configlist[-1]]
- 197
- 198 if r_start[rep] is None:
- 199 r_start_index.append(0)
- 200 else:
- 201 try:
- 202 r_start_index.append(configlist[-1].index(r_start[rep]))
- 203 except ValueError:
- 204 raise Exception('Config %d not in file with range [%d, %d]' % (
- 205 r_start[rep], configlist[-1][0], configlist[-1][-1])) from None
- 206
- 207 if r_stop[rep] is None:
- 208 r_stop_index.append(len(configlist[-1]) - 1)
- 209 else:
- 210 try:
- 211 r_stop_index.append(configlist[-1].index(r_stop[rep]))
- 212 except ValueError:
- 213 raise Exception('Config %d not in file with range [%d, %d]' % (
- 214 r_stop[rep], configlist[-1][0], configlist[-1][-1])) from None
- 215
- 216 for k in range(nrw):
- 217 deltas[k].append(tmp_array[k][r_start_index[rep]:r_stop_index[rep] + 1][::r_step])
- 218
- 219 if np.any([len(np.unique(np.diff(cl))) != 1 for cl in configlist]):
- 220 raise Exception('Irregular spaced data in input file!', [len(np.unique(np.diff(cl))) for cl in configlist])
- 221 stepsizes = [list(np.unique(np.diff(cl)))[0] for cl in configlist]
- 222 if np.any([step != 1 for step in stepsizes]):
- 223 warnings.warn('Stepsize between configurations is greater than one!' + str(stepsizes), RuntimeWarning)
- 224
- 225 print(',', nrw, 'reweighting factors with', nsrc, 'sources')
- 226 result = []
- 227 idl = [range(configlist[rep][r_start_index[rep]], configlist[rep][r_stop_index[rep]] + 1, r_step) for rep in range(replica)]
- 228
- 229 for t in range(nrw):
- 230 result.append(Obs(deltas[t], rep_names, idl=idl))
- 231 return result
- 232
+ 112 print_err = 0
+ 113 if 'print_err' in kwargs:
+ 114 print_err = 1
+ 115 print()
+ 116
+ 117 deltas = []
+ 118
+ 119 configlist = []
+ 120 r_start_index = []
+ 121 r_stop_index = []
+ 122
+ 123 for rep in range(replica):
+ 124 tmp_array = []
+ 125 with open(path + '/' + ls[rep], 'rb') as fp:
+ 126
+ 127 t = fp.read(4) # number of reweighting factors
+ 128 if rep == 0:
+ 129 nrw = struct.unpack('i', t)[0]
+ 130 if version == '2.0':
+ 131 nrw = int(nrw / 2)
+ 132 for k in range(nrw):
+ 133 deltas.append([])
+ 134 else:
+ 135 if ((nrw != struct.unpack('i', t)[0] and (not version == '2.0')) or (nrw != struct.unpack('i', t)[0] / 2 and version == '2.0')):
+ 136 raise Exception('Error: different number of reweighting factors for replicum', rep)
+ 137
+ 138 for k in range(nrw):
+ 139 tmp_array.append([])
+ 140
+ 141 # This block is necessary for openQCD1.6 and openQCD2.0 ms1 files
+ 142 nfct = []
+ 143 if version in ['1.6', '2.0']:
+ 144 for i in range(nrw):
+ 145 t = fp.read(4)
+ 146 nfct.append(struct.unpack('i', t)[0])
+ 147 else:
+ 148 for i in range(nrw):
+ 149 nfct.append(1)
+ 150
+ 151 nsrc = []
+ 152 for i in range(nrw):
+ 153 t = fp.read(4)
+ 154 nsrc.append(struct.unpack('i', t)[0])
+ 155 if version == '2.0':
+ 156 if not struct.unpack('i', fp.read(4))[0] == 0:
+ 157 print('something is wrong!')
+ 158
+ 159 configlist.append([])
+ 160 while True:
+ 161 t = fp.read(4)
+ 162 if len(t) < 4:
+ 163 break
+ 164 config_no = struct.unpack('i', t)[0]
+ 165 configlist[-1].append(config_no)
+ 166 for i in range(nrw):
+ 167 if (version == '2.0'):
+ 168 tmpd = _read_array_openQCD2(fp)
+ 169 tmpd = _read_array_openQCD2(fp)
+ 170 tmp_rw = tmpd['arr']
+ 171 tmp_nfct = 1.0
+ 172 for j in range(tmpd['n'][0]):
+ 173 tmp_nfct *= np.mean(np.exp(-np.asarray(tmp_rw[j])))
+ 174 if print_err:
+ 175 print(config_no, i, j,
+ 176 np.mean(np.exp(-np.asarray(tmp_rw[j]))),
+ 177 np.std(np.exp(-np.asarray(tmp_rw[j]))))
+ 178 print('Sources:',
+ 179 np.exp(-np.asarray(tmp_rw[j])))
+ 180 print('Partial factor:', tmp_nfct)
+ 181 elif version == '1.6' or version == '1.4':
+ 182 tmp_nfct = 1.0
+ 183 for j in range(nfct[i]):
+ 184 t = fp.read(8 * nsrc[i])
+ 185 t = fp.read(8 * nsrc[i])
+ 186 tmp_rw = struct.unpack('d' * nsrc[i], t)
+ 187 tmp_nfct *= np.mean(np.exp(-np.asarray(tmp_rw)))
+ 188 if print_err:
+ 189 print(config_no, i, j,
+ 190 np.mean(np.exp(-np.asarray(tmp_rw))),
+ 191 np.std(np.exp(-np.asarray(tmp_rw))))
+ 192 print('Sources:', np.exp(-np.asarray(tmp_rw)))
+ 193 print('Partial factor:', tmp_nfct)
+ 194 tmp_array[i].append(tmp_nfct)
+ 195
+ 196 diffmeas = configlist[-1][-1] - configlist[-1][-2]
+ 197 configlist[-1] = [item // diffmeas for item in configlist[-1]]
+ 198 if configlist[-1][0] > 1 and diffmeas > 1:
+ 199 warnings.warn('Assume thermalization and that the first measurement belongs to the first config.')
+ 200 offset = configlist[-1][0] - 1
+ 201 configlist[-1] = [item - offset for item in configlist[-1]]
+ 202
+ 203 if r_start[rep] is None:
+ 204 r_start_index.append(0)
+ 205 else:
+ 206 try:
+ 207 r_start_index.append(configlist[-1].index(r_start[rep]))
+ 208 except ValueError:
+ 209 raise Exception('Config %d not in file with range [%d, %d]' % (
+ 210 r_start[rep], configlist[-1][0], configlist[-1][-1])) from None
+ 211
+ 212 if r_stop[rep] is None:
+ 213 r_stop_index.append(len(configlist[-1]) - 1)
+ 214 else:
+ 215 try:
+ 216 r_stop_index.append(configlist[-1].index(r_stop[rep]))
+ 217 except ValueError:
+ 218 raise Exception('Config %d not in file with range [%d, %d]' % (
+ 219 r_stop[rep], configlist[-1][0], configlist[-1][-1])) from None
+ 220
+ 221 for k in range(nrw):
+ 222 deltas[k].append(tmp_array[k][r_start_index[rep]:r_stop_index[rep] + 1][::r_step])
+ 223
+ 224 if np.any([len(np.unique(np.diff(cl))) != 1 for cl in configlist]):
+ 225 raise Exception('Irregular spaced data in input file!', [len(np.unique(np.diff(cl))) for cl in configlist])
+ 226 stepsizes = [list(np.unique(np.diff(cl)))[0] for cl in configlist]
+ 227 if np.any([step != 1 for step in stepsizes]):
+ 228 warnings.warn('Stepsize between configurations is greater than one!' + str(stepsizes), RuntimeWarning)
+ 229
+ 230 print(',', nrw, 'reweighting factors with', nsrc, 'sources')
+ 231 result = []
+ 232 idl = [range(configlist[rep][r_start_index[rep]], configlist[rep][r_stop_index[rep]] + 1, r_step) for rep in range(replica)]
233
- 234def extract_t0(path, prefix, dtr_read, xmin, spatial_extent, fit_range=5, **kwargs):
- 235 """Extract t0 from given .ms.dat files. Returns t0 as Obs.
- 236
- 237 It is assumed that all boundary effects have
- 238 sufficiently decayed at x0=xmin.
- 239 The data around the zero crossing of t^2<E> - 0.3
- 240 is fitted with a linear function
- 241 from which the exact root is extracted.
- 242
- 243 It is assumed that one measurement is performed for each config.
- 244 If this is not the case, the resulting idl, as well as the handling
- 245 of r_start, r_stop and r_step is wrong and the user has to correct
- 246 this in the resulting observable.
+ 234 for t in range(nrw):
+ 235 result.append(Obs(deltas[t], rep_names, idl=idl))
+ 236 return result
+ 237
+ 238
+ 239def extract_t0(path, prefix, dtr_read, xmin, spatial_extent, fit_range=5, **kwargs):
+ 240 """Extract t0 from given .ms.dat files. Returns t0 as Obs.
+ 241
+ 242 It is assumed that all boundary effects have
+ 243 sufficiently decayed at x0=xmin.
+ 244 The data around the zero crossing of t^2<E> - 0.3
+ 245 is fitted with a linear function
+ 246 from which the exact root is extracted.
247
- 248 Parameters
- 249 ----------
- 250 path : str
- 251 Path to .ms.dat files
- 252 prefix : str
- 253 Ensemble prefix
- 254 dtr_read : int
- 255 Determines how many trajectories should be skipped
- 256 when reading the ms.dat files.
- 257 Corresponds to dtr_cnfg / dtr_ms in the openQCD input file.
- 258 xmin : int
- 259 First timeslice where the boundary
- 260 effects have sufficiently decayed.
- 261 spatial_extent : int
- 262 spatial extent of the lattice, required for normalization.
- 263 fit_range : int
- 264 Number of data points left and right of the zero
- 265 crossing to be included in the linear fit. (Default: 5)
- 266 r_start : list
- 267 list which contains the first config to be read for each replicum.
- 268 r_stop : list
- 269 list which contains the last config to be read for each replicum.
- 270 r_step : int
- 271 integer that defines a fixed step size between two measurements (in units of configs)
- 272 If not given, r_step=1 is assumed.
- 273 plaquette : bool
- 274 If true extract the plaquette estimate of t0 instead.
- 275 names : list
- 276 list of names that is assigned to the data according according
- 277 to the order in the file list. Use careful, if you do not provide file names!
- 278 files : list
- 279 list which contains the filenames to be read. No automatic detection of
- 280 files performed if given.
- 281 plot_fit : bool
- 282 If true, the fit for the extraction of t0 is shown together with the data.
- 283 assume_thermalization : bool
- 284 If True: If the first record divided by the distance between two measurements is larger than
- 285 1, it is assumed that this is due to thermalization and the first measurement belongs
- 286 to the first config (default).
- 287 If False: The config numbers are assumed to be traj_number // difference
- 288 """
- 289
- 290 ls = []
- 291 for (dirpath, dirnames, filenames) in os.walk(path):
- 292 ls.extend(filenames)
- 293 break
- 294
- 295 if not ls:
- 296 raise Exception('Error, directory not found')
- 297
- 298 if 'files' in kwargs:
- 299 ls = kwargs.get('files')
- 300 else:
- 301 for exc in ls:
- 302 if not fnmatch.fnmatch(exc, prefix + '*.ms.dat'):
- 303 ls = list(set(ls) - set([exc]))
- 304 if len(ls) > 1:
- 305 ls.sort(key=lambda x: int(re.findall(r'\d+', x[len(prefix):])[0]))
- 306 replica = len(ls)
+ 248 It is assumed that one measurement is performed for each config.
+ 249 If this is not the case, the resulting idl, as well as the handling
+ 250 of r_start, r_stop and r_step is wrong and the user has to correct
+ 251 this in the resulting observable.
+ 252
+ 253 Parameters
+ 254 ----------
+ 255 path : str
+ 256 Path to .ms.dat files
+ 257 prefix : str
+ 258 Ensemble prefix
+ 259 dtr_read : int
+ 260 Determines how many trajectories should be skipped
+ 261 when reading the ms.dat files.
+ 262 Corresponds to dtr_cnfg / dtr_ms in the openQCD input file.
+ 263 xmin : int
+ 264 First timeslice where the boundary
+ 265 effects have sufficiently decayed.
+ 266 spatial_extent : int
+ 267 spatial extent of the lattice, required for normalization.
+ 268 fit_range : int
+ 269 Number of data points left and right of the zero
+ 270 crossing to be included in the linear fit. (Default: 5)
+ 271 r_start : list
+ 272 list which contains the first config to be read for each replicum.
+ 273 r_stop : list
+ 274 list which contains the last config to be read for each replicum.
+ 275 r_step : int
+ 276 integer that defines a fixed step size between two measurements (in units of configs)
+ 277 If not given, r_step=1 is assumed.
+ 278 plaquette : bool
+ 279 If true extract the plaquette estimate of t0 instead.
+ 280 names : list
+ 281 list of names that is assigned to the data according according
+ 282 to the order in the file list. Use careful, if you do not provide file names!
+ 283 files : list
+ 284 list which contains the filenames to be read. No automatic detection of
+ 285 files performed if given.
+ 286 plot_fit : bool
+ 287 If true, the fit for the extraction of t0 is shown together with the data.
+ 288 assume_thermalization : bool
+ 289 If True: If the first record divided by the distance between two measurements is larger than
+ 290 1, it is assumed that this is due to thermalization and the first measurement belongs
+ 291 to the first config (default).
+ 292 If False: The config numbers are assumed to be traj_number // difference
+ 293
+ 294 Returns
+ 295 -------
+ 296 t0 : Obs
+ 297 Extracted t0
+ 298 """
+ 299
+ 300 ls = []
+ 301 for (dirpath, dirnames, filenames) in os.walk(path):
+ 302 ls.extend(filenames)
+ 303 break
+ 304
+ 305 if not ls:
+ 306 raise Exception('Error, directory not found')
307
- 308 if 'r_start' in kwargs:
- 309 r_start = kwargs.get('r_start')
- 310 if len(r_start) != replica:
- 311 raise Exception('r_start does not match number of replicas')
- 312 r_start = [o if o else None for o in r_start]
- 313 else:
- 314 r_start = [None] * replica
- 315
- 316 if 'r_stop' in kwargs:
- 317 r_stop = kwargs.get('r_stop')
- 318 if len(r_stop) != replica:
- 319 raise Exception('r_stop does not match number of replicas')
- 320 else:
- 321 r_stop = [None] * replica
- 322
- 323 if 'r_step' in kwargs:
- 324 r_step = kwargs.get('r_step')
- 325 else:
- 326 r_step = 1
- 327
- 328 print('Extract t0 from', prefix, ',', replica, 'replica')
- 329
- 330 if 'names' in kwargs:
- 331 rep_names = kwargs.get('names')
- 332 else:
- 333 rep_names = []
- 334 for entry in ls:
- 335 truncated_entry = entry.split('.')[0]
- 336 idx = truncated_entry.index('r')
- 337 rep_names.append(truncated_entry[:idx] + '|' + truncated_entry[idx:])
- 338
- 339 Ysum = []
- 340
- 341 configlist = []
- 342 r_start_index = []
- 343 r_stop_index = []
- 344
- 345 for rep in range(replica):
- 346
- 347 with open(path + '/' + ls[rep], 'rb') as fp:
- 348 t = fp.read(12)
- 349 header = struct.unpack('iii', t)
- 350 if rep == 0:
- 351 dn = header[0]
- 352 nn = header[1]
- 353 tmax = header[2]
- 354 elif dn != header[0] or nn != header[1] or tmax != header[2]:
- 355 raise Exception('Replica parameters do not match.')
+ 308 if 'files' in kwargs:
+ 309 ls = kwargs.get('files')
+ 310 else:
+ 311 for exc in ls:
+ 312 if not fnmatch.fnmatch(exc, prefix + '*.ms.dat'):
+ 313 ls = list(set(ls) - set([exc]))
+ 314 if len(ls) > 1:
+ 315 ls.sort(key=lambda x: int(re.findall(r'\d+', x[len(prefix):])[0]))
+ 316 replica = len(ls)
+ 317
+ 318 if 'r_start' in kwargs:
+ 319 r_start = kwargs.get('r_start')
+ 320 if len(r_start) != replica:
+ 321 raise Exception('r_start does not match number of replicas')
+ 322 r_start = [o if o else None for o in r_start]
+ 323 else:
+ 324 r_start = [None] * replica
+ 325
+ 326 if 'r_stop' in kwargs:
+ 327 r_stop = kwargs.get('r_stop')
+ 328 if len(r_stop) != replica:
+ 329 raise Exception('r_stop does not match number of replicas')
+ 330 else:
+ 331 r_stop = [None] * replica
+ 332
+ 333 if 'r_step' in kwargs:
+ 334 r_step = kwargs.get('r_step')
+ 335 else:
+ 336 r_step = 1
+ 337
+ 338 print('Extract t0 from', prefix, ',', replica, 'replica')
+ 339
+ 340 if 'names' in kwargs:
+ 341 rep_names = kwargs.get('names')
+ 342 else:
+ 343 rep_names = []
+ 344 for entry in ls:
+ 345 truncated_entry = entry.split('.')[0]
+ 346 idx = truncated_entry.index('r')
+ 347 rep_names.append(truncated_entry[:idx] + '|' + truncated_entry[idx:])
+ 348
+ 349 Ysum = []
+ 350
+ 351 configlist = []
+ 352 r_start_index = []
+ 353 r_stop_index = []
+ 354
+ 355 for rep in range(replica):
356
- 357 t = fp.read(8)
- 358 if rep == 0:
- 359 eps = struct.unpack('d', t)[0]
- 360 print('Step size:', eps, ', Maximal t value:', dn * (nn) * eps)
- 361 elif eps != struct.unpack('d', t)[0]:
- 362 raise Exception('Values for eps do not match among replica.')
- 363
- 364 Ysl = []
- 365
- 366 configlist.append([])
- 367 while True:
- 368 t = fp.read(4)
- 369 if (len(t) < 4):
- 370 break
- 371 nc = struct.unpack('i', t)[0]
- 372 configlist[-1].append(nc)
+ 357 with open(path + '/' + ls[rep], 'rb') as fp:
+ 358 t = fp.read(12)
+ 359 header = struct.unpack('iii', t)
+ 360 if rep == 0:
+ 361 dn = header[0]
+ 362 nn = header[1]
+ 363 tmax = header[2]
+ 364 elif dn != header[0] or nn != header[1] or tmax != header[2]:
+ 365 raise Exception('Replica parameters do not match.')
+ 366
+ 367 t = fp.read(8)
+ 368 if rep == 0:
+ 369 eps = struct.unpack('d', t)[0]
+ 370 print('Step size:', eps, ', Maximal t value:', dn * (nn) * eps)
+ 371 elif eps != struct.unpack('d', t)[0]:
+ 372 raise Exception('Values for eps do not match among replica.')
373
- 374 t = fp.read(8 * tmax * (nn + 1))
- 375 if kwargs.get('plaquette'):
- 376 if nc % dtr_read == 0:
- 377 Ysl.append(struct.unpack('d' * tmax * (nn + 1), t))
- 378 t = fp.read(8 * tmax * (nn + 1))
- 379 if not kwargs.get('plaquette'):
- 380 if nc % dtr_read == 0:
- 381 Ysl.append(struct.unpack('d' * tmax * (nn + 1), t))
- 382 t = fp.read(8 * tmax * (nn + 1))
+ 374 Ysl = []
+ 375
+ 376 configlist.append([])
+ 377 while True:
+ 378 t = fp.read(4)
+ 379 if (len(t) < 4):
+ 380 break
+ 381 nc = struct.unpack('i', t)[0]
+ 382 configlist[-1].append(nc)
383
- 384 Ysum.append([])
- 385 for i, item in enumerate(Ysl):
- 386 Ysum[-1].append([np.mean(item[current + xmin:
- 387 current + tmax - xmin])
- 388 for current in range(0, len(item), tmax)])
- 389
- 390 diffmeas = configlist[-1][-1] - configlist[-1][-2]
- 391 configlist[-1] = [item // diffmeas for item in configlist[-1]]
- 392 if kwargs.get('assume_thermalization', True) and configlist[-1][0] > 1:
- 393 warnings.warn('Assume thermalization and that the first measurement belongs to the first config.')
- 394 offset = configlist[-1][0] - 1
- 395 configlist[-1] = [item - offset for item in configlist[-1]]
- 396
- 397 if r_start[rep] is None:
- 398 r_start_index.append(0)
- 399 else:
- 400 try:
- 401 r_start_index.append(configlist[-1].index(r_start[rep]))
- 402 except ValueError:
- 403 raise Exception('Config %d not in file with range [%d, %d]' % (
- 404 r_start[rep], configlist[-1][0], configlist[-1][-1])) from None
- 405
- 406 if r_stop[rep] is None:
- 407 r_stop_index.append(len(configlist[-1]) - 1)
- 408 else:
- 409 try:
- 410 r_stop_index.append(configlist[-1].index(r_stop[rep]))
- 411 except ValueError:
- 412 raise Exception('Config %d not in file with range [%d, %d]' % (
- 413 r_stop[rep], configlist[-1][0], configlist[-1][-1])) from None
- 414
- 415 if np.any([len(np.unique(np.diff(cl))) != 1 for cl in configlist]):
- 416 raise Exception('Irregular spaced data in input file!', [len(np.unique(np.diff(cl))) for cl in configlist])
- 417 stepsizes = [list(np.unique(np.diff(cl)))[0] for cl in configlist]
- 418 if np.any([step != 1 for step in stepsizes]):
- 419 warnings.warn('Stepsize between configurations is greater than one!' + str(stepsizes), RuntimeWarning)
- 420
- 421 idl = [range(configlist[rep][r_start_index[rep]], configlist[rep][r_stop_index[rep]] + 1, r_step) for rep in range(replica)]
- 422 t2E_dict = {}
- 423 for n in range(nn + 1):
- 424 samples = []
- 425 for nrep, rep in enumerate(Ysum):
- 426 samples.append([])
- 427 for cnfg in rep:
- 428 samples[-1].append(cnfg[n])
- 429 samples[-1] = samples[-1][r_start_index[nrep]:r_stop_index[nrep] + 1][::r_step]
- 430 new_obs = Obs(samples, rep_names, idl=idl)
- 431 t2E_dict[n * dn * eps] = (n * dn * eps) ** 2 * new_obs / (spatial_extent ** 3) - 0.3
- 432
- 433 zero_crossing = np.argmax(np.array(
- 434 [o.value for o in t2E_dict.values()]) > 0.0)
- 435
- 436 x = list(t2E_dict.keys())[zero_crossing - fit_range:
- 437 zero_crossing + fit_range]
- 438 y = list(t2E_dict.values())[zero_crossing - fit_range:
- 439 zero_crossing + fit_range]
- 440 [o.gamma_method() for o in y]
- 441
- 442 fit_result = fit_lin(x, y)
- 443
- 444 if kwargs.get('plot_fit'):
- 445 plt.figure()
- 446 gs = gridspec.GridSpec(2, 1, height_ratios=[3, 1], wspace=0.0, hspace=0.0)
- 447 ax0 = plt.subplot(gs[0])
- 448 xmore = list(t2E_dict.keys())[zero_crossing - fit_range - 2: zero_crossing + fit_range + 2]
- 449 ymore = list(t2E_dict.values())[zero_crossing - fit_range - 2: zero_crossing + fit_range + 2]
- 450 [o.gamma_method() for o in ymore]
- 451 ax0.errorbar(xmore, [yi.value for yi in ymore], yerr=[yi.dvalue for yi in ymore], fmt='x')
- 452 xplot = np.linspace(np.min(x), np.max(x))
- 453 yplot = [fit_result[0] + fit_result[1] * xi for xi in xplot]
- 454 [yi.gamma_method() for yi in yplot]
- 455 ax0.fill_between(xplot, y1=[yi.value - yi.dvalue for yi in yplot], y2=[yi.value + yi.dvalue for yi in yplot])
- 456 retval = (-fit_result[0] / fit_result[1])
- 457 retval.gamma_method()
- 458 ylim = ax0.get_ylim()
- 459 ax0.fill_betweenx(ylim, x1=retval.value - retval.dvalue, x2=retval.value + retval.dvalue, color='gray', alpha=0.4)
- 460 ax0.set_ylim(ylim)
- 461 ax0.set_ylabel(r'$t^2 \langle E(t) \rangle - 0.3 $')
- 462 xlim = ax0.get_xlim()
- 463
- 464 fit_res = [fit_result[0] + fit_result[1] * xi for xi in x]
- 465 residuals = (np.asarray([o.value for o in y]) - [o.value for o in fit_res]) / np.asarray([o.dvalue for o in y])
- 466 ax1 = plt.subplot(gs[1])
- 467 ax1.plot(x, residuals, 'ko', ls='none', markersize=5)
- 468 ax1.tick_params(direction='out')
- 469 ax1.tick_params(axis="x", bottom=True, top=True, labelbottom=True)
- 470 ax1.axhline(y=0.0, ls='--', color='k')
- 471 ax1.fill_between(xlim, -1.0, 1.0, alpha=0.1, facecolor='k')
- 472 ax1.set_xlim(xlim)
- 473 ax1.set_ylabel('Residuals')
- 474 ax1.set_xlabel(r'$t/a^2$')
- 475
- 476 plt.draw()
- 477 return -fit_result[0] / fit_result[1]
- 478
- 479
- 480def _parse_array_openQCD2(d, n, size, wa, quadrupel=False):
- 481 arr = []
- 482 if d == 2:
- 483 for i in range(n[0]):
- 484 tmp = wa[i * n[1]:(i + 1) * n[1]]
- 485 if quadrupel:
- 486 tmp2 = []
- 487 for j in range(0, len(tmp), 2):
- 488 tmp2.append(tmp[j])
- 489 arr.append(tmp2)
- 490 else:
- 491 arr.append(np.asarray(tmp))
- 492
- 493 else:
- 494 raise Exception('Only two-dimensional arrays supported!')
- 495
- 496 return arr
- 497
- 498
- 499def _read_array_openQCD2(fp):
- 500 t = fp.read(4)
- 501 d = struct.unpack('i', t)[0]
- 502 t = fp.read(4 * d)
- 503 n = struct.unpack('%di' % (d), t)
- 504 t = fp.read(4)
- 505 size = struct.unpack('i', t)[0]
- 506 if size == 4:
- 507 types = 'i'
- 508 elif size == 8:
- 509 types = 'd'
- 510 elif size == 16:
- 511 types = 'dd'
- 512 else:
- 513 raise Exception("Type for size '" + str(size) + "' not known.")
- 514 m = n[0]
- 515 for i in range(1, d):
- 516 m *= n[i]
- 517
- 518 t = fp.read(m * size)
- 519 tmp = struct.unpack('%d%s' % (m, types), t)
- 520
- 521 arr = _parse_array_openQCD2(d, n, size, tmp, quadrupel=True)
- 522 return {'d': d, 'n': n, 'size': size, 'arr': arr}
- 523
- 524
- 525def read_qtop(path, prefix, c, dtr_cnfg=1, version="openQCD", **kwargs):
- 526 """Read the topologial charge based on openQCD gradient flow measurements.
+ 384 t = fp.read(8 * tmax * (nn + 1))
+ 385 if kwargs.get('plaquette'):
+ 386 if nc % dtr_read == 0:
+ 387 Ysl.append(struct.unpack('d' * tmax * (nn + 1), t))
+ 388 t = fp.read(8 * tmax * (nn + 1))
+ 389 if not kwargs.get('plaquette'):
+ 390 if nc % dtr_read == 0:
+ 391 Ysl.append(struct.unpack('d' * tmax * (nn + 1), t))
+ 392 t = fp.read(8 * tmax * (nn + 1))
+ 393
+ 394 Ysum.append([])
+ 395 for i, item in enumerate(Ysl):
+ 396 Ysum[-1].append([np.mean(item[current + xmin:
+ 397 current + tmax - xmin])
+ 398 for current in range(0, len(item), tmax)])
+ 399
+ 400 diffmeas = configlist[-1][-1] - configlist[-1][-2]
+ 401 configlist[-1] = [item // diffmeas for item in configlist[-1]]
+ 402 if kwargs.get('assume_thermalization', True) and configlist[-1][0] > 1:
+ 403 warnings.warn('Assume thermalization and that the first measurement belongs to the first config.')
+ 404 offset = configlist[-1][0] - 1
+ 405 configlist[-1] = [item - offset for item in configlist[-1]]
+ 406
+ 407 if r_start[rep] is None:
+ 408 r_start_index.append(0)
+ 409 else:
+ 410 try:
+ 411 r_start_index.append(configlist[-1].index(r_start[rep]))
+ 412 except ValueError:
+ 413 raise Exception('Config %d not in file with range [%d, %d]' % (
+ 414 r_start[rep], configlist[-1][0], configlist[-1][-1])) from None
+ 415
+ 416 if r_stop[rep] is None:
+ 417 r_stop_index.append(len(configlist[-1]) - 1)
+ 418 else:
+ 419 try:
+ 420 r_stop_index.append(configlist[-1].index(r_stop[rep]))
+ 421 except ValueError:
+ 422 raise Exception('Config %d not in file with range [%d, %d]' % (
+ 423 r_stop[rep], configlist[-1][0], configlist[-1][-1])) from None
+ 424
+ 425 if np.any([len(np.unique(np.diff(cl))) != 1 for cl in configlist]):
+ 426 raise Exception('Irregular spaced data in input file!', [len(np.unique(np.diff(cl))) for cl in configlist])
+ 427 stepsizes = [list(np.unique(np.diff(cl)))[0] for cl in configlist]
+ 428 if np.any([step != 1 for step in stepsizes]):
+ 429 warnings.warn('Stepsize between configurations is greater than one!' + str(stepsizes), RuntimeWarning)
+ 430
+ 431 idl = [range(configlist[rep][r_start_index[rep]], configlist[rep][r_stop_index[rep]] + 1, r_step) for rep in range(replica)]
+ 432 t2E_dict = {}
+ 433 for n in range(nn + 1):
+ 434 samples = []
+ 435 for nrep, rep in enumerate(Ysum):
+ 436 samples.append([])
+ 437 for cnfg in rep:
+ 438 samples[-1].append(cnfg[n])
+ 439 samples[-1] = samples[-1][r_start_index[nrep]:r_stop_index[nrep] + 1][::r_step]
+ 440 new_obs = Obs(samples, rep_names, idl=idl)
+ 441 t2E_dict[n * dn * eps] = (n * dn * eps) ** 2 * new_obs / (spatial_extent ** 3) - 0.3
+ 442
+ 443 zero_crossing = np.argmax(np.array(
+ 444 [o.value for o in t2E_dict.values()]) > 0.0)
+ 445
+ 446 x = list(t2E_dict.keys())[zero_crossing - fit_range:
+ 447 zero_crossing + fit_range]
+ 448 y = list(t2E_dict.values())[zero_crossing - fit_range:
+ 449 zero_crossing + fit_range]
+ 450 [o.gamma_method() for o in y]
+ 451
+ 452 fit_result = fit_lin(x, y)
+ 453
+ 454 if kwargs.get('plot_fit'):
+ 455 plt.figure()
+ 456 gs = gridspec.GridSpec(2, 1, height_ratios=[3, 1], wspace=0.0, hspace=0.0)
+ 457 ax0 = plt.subplot(gs[0])
+ 458 xmore = list(t2E_dict.keys())[zero_crossing - fit_range - 2: zero_crossing + fit_range + 2]
+ 459 ymore = list(t2E_dict.values())[zero_crossing - fit_range - 2: zero_crossing + fit_range + 2]
+ 460 [o.gamma_method() for o in ymore]
+ 461 ax0.errorbar(xmore, [yi.value for yi in ymore], yerr=[yi.dvalue for yi in ymore], fmt='x')
+ 462 xplot = np.linspace(np.min(x), np.max(x))
+ 463 yplot = [fit_result[0] + fit_result[1] * xi for xi in xplot]
+ 464 [yi.gamma_method() for yi in yplot]
+ 465 ax0.fill_between(xplot, y1=[yi.value - yi.dvalue for yi in yplot], y2=[yi.value + yi.dvalue for yi in yplot])
+ 466 retval = (-fit_result[0] / fit_result[1])
+ 467 retval.gamma_method()
+ 468 ylim = ax0.get_ylim()
+ 469 ax0.fill_betweenx(ylim, x1=retval.value - retval.dvalue, x2=retval.value + retval.dvalue, color='gray', alpha=0.4)
+ 470 ax0.set_ylim(ylim)
+ 471 ax0.set_ylabel(r'$t^2 \langle E(t) \rangle - 0.3 $')
+ 472 xlim = ax0.get_xlim()
+ 473
+ 474 fit_res = [fit_result[0] + fit_result[1] * xi for xi in x]
+ 475 residuals = (np.asarray([o.value for o in y]) - [o.value for o in fit_res]) / np.asarray([o.dvalue for o in y])
+ 476 ax1 = plt.subplot(gs[1])
+ 477 ax1.plot(x, residuals, 'ko', ls='none', markersize=5)
+ 478 ax1.tick_params(direction='out')
+ 479 ax1.tick_params(axis="x", bottom=True, top=True, labelbottom=True)
+ 480 ax1.axhline(y=0.0, ls='--', color='k')
+ 481 ax1.fill_between(xlim, -1.0, 1.0, alpha=0.1, facecolor='k')
+ 482 ax1.set_xlim(xlim)
+ 483 ax1.set_ylabel('Residuals')
+ 484 ax1.set_xlabel(r'$t/a^2$')
+ 485
+ 486 plt.draw()
+ 487 return -fit_result[0] / fit_result[1]
+ 488
+ 489
+ 490def _parse_array_openQCD2(d, n, size, wa, quadrupel=False):
+ 491 arr = []
+ 492 if d == 2:
+ 493 for i in range(n[0]):
+ 494 tmp = wa[i * n[1]:(i + 1) * n[1]]
+ 495 if quadrupel:
+ 496 tmp2 = []
+ 497 for j in range(0, len(tmp), 2):
+ 498 tmp2.append(tmp[j])
+ 499 arr.append(tmp2)
+ 500 else:
+ 501 arr.append(np.asarray(tmp))
+ 502
+ 503 else:
+ 504 raise Exception('Only two-dimensional arrays supported!')
+ 505
+ 506 return arr
+ 507
+ 508
+ 509def _read_array_openQCD2(fp):
+ 510 t = fp.read(4)
+ 511 d = struct.unpack('i', t)[0]
+ 512 t = fp.read(4 * d)
+ 513 n = struct.unpack('%di' % (d), t)
+ 514 t = fp.read(4)
+ 515 size = struct.unpack('i', t)[0]
+ 516 if size == 4:
+ 517 types = 'i'
+ 518 elif size == 8:
+ 519 types = 'd'
+ 520 elif size == 16:
+ 521 types = 'dd'
+ 522 else:
+ 523 raise Exception("Type for size '" + str(size) + "' not known.")
+ 524 m = n[0]
+ 525 for i in range(1, d):
+ 526 m *= n[i]
527
- 528 Parameters
- 529 ----------
- 530 path : str
- 531 path of the measurement files
- 532 prefix : str
- 533 prefix of the measurement files, e.g. <prefix>_id0_r0.ms.dat.
- 534 Ignored if file names are passed explicitly via keyword files.
- 535 c : double
- 536 Smearing radius in units of the lattice extent, c = sqrt(8 t0) / L.
- 537 dtr_cnfg : int
- 538 (optional) parameter that specifies the number of measurements
- 539 between two configs.
- 540 If it is not set, the distance between two measurements
- 541 in the file is assumed to be the distance between two configurations.
- 542 steps : int
- 543 (optional) Distance between two configurations in units of trajectories /
- 544 cycles. Assumed to be the distance between two measurements * dtr_cnfg if not given
- 545 version : str
- 546 Either openQCD or sfqcd, depending on the data.
- 547 L : int
- 548 spatial length of the lattice in L/a.
- 549 HAS to be set if version != sfqcd, since openQCD does not provide
- 550 this in the header
- 551 r_start : list
- 552 list which contains the first config to be read for each replicum.
- 553 r_stop : list
- 554 list which contains the last config to be read for each replicum.
- 555 files : list
- 556 specify the exact files that need to be read
- 557 from path, practical if e.g. only one replicum is needed
- 558 postfix : str
- 559 postfix of the file to read, e.g. '.gfms.dat' for openQCD-files
- 560 names : list
- 561 Alternative labeling for replicas/ensembles.
- 562 Has to have the appropriate length.
- 563 Zeuthen_flow : bool
- 564 (optional) If True, the Zeuthen flow is used for Qtop. Only possible
- 565 for version=='sfqcd' If False, the Wilson flow is used.
- 566 integer_charge : bool
- 567 If True, the charge is rounded towards the nearest integer on each config.
- 568 """
- 569
- 570 return _read_flow_obs(path, prefix, c, dtr_cnfg=dtr_cnfg, version=version, obspos=0, **kwargs)
- 571
- 572
- 573def read_gf_coupling(path, prefix, c, dtr_cnfg=1, Zeuthen_flow=True, **kwargs):
- 574 """Read the gradient flow coupling based on sfqcd gradient flow measurements. See 1607.06423 for details.
- 575
- 576 Note: The current implementation only works for c=0.3 and T=L. The definition of the coupling in 1607.06423 requires projection to topological charge zero which is not done within this function but has to be performed in a separate step.
- 577
- 578 Parameters
- 579 ----------
- 580 path : str
- 581 path of the measurement files
- 582 prefix : str
- 583 prefix of the measurement files, e.g. <prefix>_id0_r0.ms.dat.
- 584 Ignored if file names are passed explicitly via keyword files.
- 585 c : double
- 586 Smearing radius in units of the lattice extent, c = sqrt(8 t0) / L.
- 587 dtr_cnfg : int
- 588 (optional) parameter that specifies the number of measurements
- 589 between two configs.
- 590 If it is not set, the distance between two measurements
- 591 in the file is assumed to be the distance between two configurations.
- 592 steps : int
- 593 (optional) Distance between two configurations in units of trajectories /
- 594 cycles. Assumed to be the distance between two measurements * dtr_cnfg if not given
- 595 r_start : list
- 596 list which contains the first config to be read for each replicum.
- 597 r_stop : list
- 598 list which contains the last config to be read for each replicum.
- 599 files : list
- 600 specify the exact files that need to be read
- 601 from path, practical if e.g. only one replicum is needed
- 602 names : list
- 603 Alternative labeling for replicas/ensembles.
- 604 Has to have the appropriate length.
- 605 postfix : str
- 606 postfix of the file to read, e.g. '.gfms.dat' for openQCD-files
- 607 Zeuthen_flow : bool
- 608 (optional) If True, the Zeuthen flow is used for the coupling. If False, the Wilson flow is used.
- 609 """
- 610
- 611 if c != 0.3:
- 612 raise Exception("The required lattice norm is only implemented for c=0.3 at the moment.")
- 613
- 614 plaq = _read_flow_obs(path, prefix, c, dtr_cnfg=dtr_cnfg, version="sfqcd", obspos=6, sum_t=False, Zeuthen_flow=Zeuthen_flow, integer_charge=False, **kwargs)
- 615 C2x1 = _read_flow_obs(path, prefix, c, dtr_cnfg=dtr_cnfg, version="sfqcd", obspos=7, sum_t=False, Zeuthen_flow=Zeuthen_flow, integer_charge=False, **kwargs)
- 616 L = plaq.tag["L"]
- 617 T = plaq.tag["T"]
- 618
- 619 if T != L:
- 620 raise Exception("The required lattice norm is only implemented for T=L at the moment.")
- 621
- 622 if Zeuthen_flow is not True:
- 623 raise Exception("The required lattice norm is only implemented for the Zeuthen flow at the moment.")
- 624
- 625 t = (c * L) ** 2 / 8
- 626
- 627 normdict = {4: 0.012341170468270,
- 628 6: 0.010162691462430,
- 629 8: 0.009031614807931,
- 630 10: 0.008744966371393,
- 631 12: 0.008650917856809,
- 632 14: 8.611154391267955E-03,
- 633 16: 0.008591758449508,
- 634 20: 0.008575359627103,
- 635 24: 0.008569387847540,
- 636 28: 8.566803713382559E-03,
- 637 32: 0.008565541650006,
- 638 40: 8.564480684962046E-03,
- 639 48: 8.564098025073460E-03,
- 640 64: 8.563853943383087E-03}
+ 528 t = fp.read(m * size)
+ 529 tmp = struct.unpack('%d%s' % (m, types), t)
+ 530
+ 531 arr = _parse_array_openQCD2(d, n, size, tmp, quadrupel=True)
+ 532 return {'d': d, 'n': n, 'size': size, 'arr': arr}
+ 533
+ 534
+ 535def read_qtop(path, prefix, c, dtr_cnfg=1, version="openQCD", **kwargs):
+ 536 """Read the topologial charge based on openQCD gradient flow measurements.
+ 537
+ 538 Parameters
+ 539 ----------
+ 540 path : str
+ 541 path of the measurement files
+ 542 prefix : str
+ 543 prefix of the measurement files, e.g. <prefix>_id0_r0.ms.dat.
+ 544 Ignored if file names are passed explicitly via keyword files.
+ 545 c : double
+ 546 Smearing radius in units of the lattice extent, c = sqrt(8 t0) / L.
+ 547 dtr_cnfg : int
+ 548 (optional) parameter that specifies the number of measurements
+ 549 between two configs.
+ 550 If it is not set, the distance between two measurements
+ 551 in the file is assumed to be the distance between two configurations.
+ 552 steps : int
+ 553 (optional) Distance between two configurations in units of trajectories /
+ 554 cycles. Assumed to be the distance between two measurements * dtr_cnfg if not given
+ 555 version : str
+ 556 Either openQCD or sfqcd, depending on the data.
+ 557 L : int
+ 558 spatial length of the lattice in L/a.
+ 559 HAS to be set if version != sfqcd, since openQCD does not provide
+ 560 this in the header
+ 561 r_start : list
+ 562 list which contains the first config to be read for each replicum.
+ 563 r_stop : list
+ 564 list which contains the last config to be read for each replicum.
+ 565 files : list
+ 566 specify the exact files that need to be read
+ 567 from path, practical if e.g. only one replicum is needed
+ 568 postfix : str
+ 569 postfix of the file to read, e.g. '.gfms.dat' for openQCD-files
+ 570 names : list
+ 571 Alternative labeling for replicas/ensembles.
+ 572 Has to have the appropriate length.
+ 573 Zeuthen_flow : bool
+ 574 (optional) If True, the Zeuthen flow is used for Qtop. Only possible
+ 575 for version=='sfqcd' If False, the Wilson flow is used.
+ 576 integer_charge : bool
+ 577 If True, the charge is rounded towards the nearest integer on each config.
+ 578
+ 579 Returns
+ 580 -------
+ 581 result : Obs
+ 582 Read topological charge
+ 583 """
+ 584
+ 585 return _read_flow_obs(path, prefix, c, dtr_cnfg=dtr_cnfg, version=version, obspos=0, **kwargs)
+ 586
+ 587
+ 588def read_gf_coupling(path, prefix, c, dtr_cnfg=1, Zeuthen_flow=True, **kwargs):
+ 589 """Read the gradient flow coupling based on sfqcd gradient flow measurements. See 1607.06423 for details.
+ 590
+ 591 Note: The current implementation only works for c=0.3 and T=L. The definition of the coupling in 1607.06423 requires projection to topological charge zero which is not done within this function but has to be performed in a separate step.
+ 592
+ 593 Parameters
+ 594 ----------
+ 595 path : str
+ 596 path of the measurement files
+ 597 prefix : str
+ 598 prefix of the measurement files, e.g. <prefix>_id0_r0.ms.dat.
+ 599 Ignored if file names are passed explicitly via keyword files.
+ 600 c : double
+ 601 Smearing radius in units of the lattice extent, c = sqrt(8 t0) / L.
+ 602 dtr_cnfg : int
+ 603 (optional) parameter that specifies the number of measurements
+ 604 between two configs.
+ 605 If it is not set, the distance between two measurements
+ 606 in the file is assumed to be the distance between two configurations.
+ 607 steps : int
+ 608 (optional) Distance between two configurations in units of trajectories /
+ 609 cycles. Assumed to be the distance between two measurements * dtr_cnfg if not given
+ 610 r_start : list
+ 611 list which contains the first config to be read for each replicum.
+ 612 r_stop : list
+ 613 list which contains the last config to be read for each replicum.
+ 614 files : list
+ 615 specify the exact files that need to be read
+ 616 from path, practical if e.g. only one replicum is needed
+ 617 names : list
+ 618 Alternative labeling for replicas/ensembles.
+ 619 Has to have the appropriate length.
+ 620 postfix : str
+ 621 postfix of the file to read, e.g. '.gfms.dat' for openQCD-files
+ 622 Zeuthen_flow : bool
+ 623 (optional) If True, the Zeuthen flow is used for the coupling. If False, the Wilson flow is used.
+ 624 """
+ 625
+ 626 if c != 0.3:
+ 627 raise Exception("The required lattice norm is only implemented for c=0.3 at the moment.")
+ 628
+ 629 plaq = _read_flow_obs(path, prefix, c, dtr_cnfg=dtr_cnfg, version="sfqcd", obspos=6, sum_t=False, Zeuthen_flow=Zeuthen_flow, integer_charge=False, **kwargs)
+ 630 C2x1 = _read_flow_obs(path, prefix, c, dtr_cnfg=dtr_cnfg, version="sfqcd", obspos=7, sum_t=False, Zeuthen_flow=Zeuthen_flow, integer_charge=False, **kwargs)
+ 631 L = plaq.tag["L"]
+ 632 T = plaq.tag["T"]
+ 633
+ 634 if T != L:
+ 635 raise Exception("The required lattice norm is only implemented for T=L at the moment.")
+ 636
+ 637 if Zeuthen_flow is not True:
+ 638 raise Exception("The required lattice norm is only implemented for the Zeuthen flow at the moment.")
+ 639
+ 640 t = (c * L) ** 2 / 8
641
- 642 return t * t * (5 / 3 * plaq - 1 / 12 * C2x1) / normdict[L]
- 643
- 644
- 645def _read_flow_obs(path, prefix, c, dtr_cnfg=1, version="openQCD", obspos=0, sum_t=True, **kwargs):
- 646 """Read a flow observable based on openQCD gradient flow measurements.
- 647
- 648 Parameters
- 649 ----------
- 650 path : str
- 651 path of the measurement files
- 652 prefix : str
- 653 prefix of the measurement files, e.g. <prefix>_id0_r0.ms.dat.
- 654 Ignored if file names are passed explicitly via keyword files.
- 655 c : double
- 656 Smearing radius in units of the lattice extent, c = sqrt(8 t0) / L.
- 657 dtr_cnfg : int
- 658 (optional) parameter that specifies the number of measurements
- 659 between two configs.
- 660 If it is not set, the distance between two measurements
- 661 in the file is assumed to be the distance between two configurations.
- 662 steps : int
- 663 (optional) Distance between two configurations in units of trajectories /
- 664 cycles. Assumed to be the distance between two measurements * dtr_cnfg if not given
- 665 version : str
- 666 Either openQCD or sfqcd, depending on the data.
- 667 obspos : int
- 668 position of the obeservable in the measurement file. Only relevant for sfqcd files.
- 669 sum_t : bool
- 670 If true sum over all timeslices, if false only take the value at T/2.
- 671 L : int
- 672 spatial length of the lattice in L/a.
- 673 HAS to be set if version != sfqcd, since openQCD does not provide
- 674 this in the header
- 675 r_start : list
- 676 list which contains the first config to be read for each replicum.
- 677 r_stop : list
- 678 list which contains the last config to be read for each replicum.
- 679 files : list
- 680 specify the exact files that need to be read
- 681 from path, practical if e.g. only one replicum is needed
- 682 names : list
- 683 Alternative labeling for replicas/ensembles.
- 684 Has to have the appropriate length.
- 685 postfix : str
- 686 postfix of the file to read, e.g. '.gfms.dat' for openQCD-files
- 687 Zeuthen_flow : bool
- 688 (optional) If True, the Zeuthen flow is used for Qtop. Only possible
- 689 for version=='sfqcd' If False, the Wilson flow is used.
- 690 integer_charge : bool
- 691 If True, the charge is rounded towards the nearest integer on each config.
- 692 """
- 693 known_versions = ["openQCD", "sfqcd"]
- 694
- 695 if version not in known_versions:
- 696 raise Exception("Unknown openQCD version.")
- 697 if "steps" in kwargs:
- 698 steps = kwargs.get("steps")
- 699 if version == "sfqcd":
- 700 if "L" in kwargs:
- 701 supposed_L = kwargs.get("L")
- 702 else:
- 703 supposed_L = None
- 704 postfix = ".gfms.dat"
- 705 else:
- 706 if "L" not in kwargs:
- 707 raise Exception("This version of openQCD needs you to provide the spatial length of the lattice as parameter 'L'.")
- 708 else:
- 709 L = kwargs.get("L")
- 710 postfix = ".ms.dat"
- 711
- 712 if "postfix" in kwargs:
- 713 postfix = kwargs.get("postfix")
+ 642 normdict = {4: 0.012341170468270,
+ 643 6: 0.010162691462430,
+ 644 8: 0.009031614807931,
+ 645 10: 0.008744966371393,
+ 646 12: 0.008650917856809,
+ 647 14: 8.611154391267955E-03,
+ 648 16: 0.008591758449508,
+ 649 20: 0.008575359627103,
+ 650 24: 0.008569387847540,
+ 651 28: 8.566803713382559E-03,
+ 652 32: 0.008565541650006,
+ 653 40: 8.564480684962046E-03,
+ 654 48: 8.564098025073460E-03,
+ 655 64: 8.563853943383087E-03}
+ 656
+ 657 return t * t * (5 / 3 * plaq - 1 / 12 * C2x1) / normdict[L]
+ 658
+ 659
+ 660def _read_flow_obs(path, prefix, c, dtr_cnfg=1, version="openQCD", obspos=0, sum_t=True, **kwargs):
+ 661 """Read a flow observable based on openQCD gradient flow measurements.
+ 662
+ 663 Parameters
+ 664 ----------
+ 665 path : str
+ 666 path of the measurement files
+ 667 prefix : str
+ 668 prefix of the measurement files, e.g. <prefix>_id0_r0.ms.dat.
+ 669 Ignored if file names are passed explicitly via keyword files.
+ 670 c : double
+ 671 Smearing radius in units of the lattice extent, c = sqrt(8 t0) / L.
+ 672 dtr_cnfg : int
+ 673 (optional) parameter that specifies the number of measurements
+ 674 between two configs.
+ 675 If it is not set, the distance between two measurements
+ 676 in the file is assumed to be the distance between two configurations.
+ 677 steps : int
+ 678 (optional) Distance between two configurations in units of trajectories /
+ 679 cycles. Assumed to be the distance between two measurements * dtr_cnfg if not given
+ 680 version : str
+ 681 Either openQCD or sfqcd, depending on the data.
+ 682 obspos : int
+ 683 position of the obeservable in the measurement file. Only relevant for sfqcd files.
+ 684 sum_t : bool
+ 685 If true sum over all timeslices, if false only take the value at T/2.
+ 686 L : int
+ 687 spatial length of the lattice in L/a.
+ 688 HAS to be set if version != sfqcd, since openQCD does not provide
+ 689 this in the header
+ 690 r_start : list
+ 691 list which contains the first config to be read for each replicum.
+ 692 r_stop : list
+ 693 list which contains the last config to be read for each replicum.
+ 694 files : list
+ 695 specify the exact files that need to be read
+ 696 from path, practical if e.g. only one replicum is needed
+ 697 names : list
+ 698 Alternative labeling for replicas/ensembles.
+ 699 Has to have the appropriate length.
+ 700 postfix : str
+ 701 postfix of the file to read, e.g. '.gfms.dat' for openQCD-files
+ 702 Zeuthen_flow : bool
+ 703 (optional) If True, the Zeuthen flow is used for Qtop. Only possible
+ 704 for version=='sfqcd' If False, the Wilson flow is used.
+ 705 integer_charge : bool
+ 706 If True, the charge is rounded towards the nearest integer on each config.
+ 707
+ 708 Returns
+ 709 -------
+ 710 result : Obs
+ 711 flow observable specified
+ 712 """
+ 713 known_versions = ["openQCD", "sfqcd"]
714
- 715 if "files" in kwargs:
- 716 files = kwargs.get("files")
- 717 postfix = ''
- 718 else:
- 719 found = []
- 720 files = []
- 721 for (dirpath, dirnames, filenames) in os.walk(path + "/"):
- 722 found.extend(filenames)
- 723 break
- 724 for f in found:
- 725 if fnmatch.fnmatch(f, prefix + "*" + postfix):
- 726 files.append(f)
- 727
- 728 files = sorted(files)
- 729
- 730 if 'r_start' in kwargs:
- 731 r_start = kwargs.get('r_start')
- 732 if len(r_start) != len(files):
- 733 raise Exception('r_start does not match number of replicas')
- 734 r_start = [o if o else None for o in r_start]
- 735 else:
- 736 r_start = [None] * len(files)
- 737
- 738 if 'r_stop' in kwargs:
- 739 r_stop = kwargs.get('r_stop')
- 740 if len(r_stop) != len(files):
- 741 raise Exception('r_stop does not match number of replicas')
- 742 else:
- 743 r_stop = [None] * len(files)
- 744 rep_names = []
- 745
- 746 zeuthen = kwargs.get('Zeuthen_flow', False)
- 747 if zeuthen and version not in ['sfqcd']:
- 748 raise Exception('Zeuthen flow can only be used for version==sfqcd')
+ 715 if version not in known_versions:
+ 716 raise Exception("Unknown openQCD version.")
+ 717 if "steps" in kwargs:
+ 718 steps = kwargs.get("steps")
+ 719 if version == "sfqcd":
+ 720 if "L" in kwargs:
+ 721 supposed_L = kwargs.get("L")
+ 722 else:
+ 723 supposed_L = None
+ 724 postfix = ".gfms.dat"
+ 725 else:
+ 726 if "L" not in kwargs:
+ 727 raise Exception("This version of openQCD needs you to provide the spatial length of the lattice as parameter 'L'.")
+ 728 else:
+ 729 L = kwargs.get("L")
+ 730 postfix = ".ms.dat"
+ 731
+ 732 if "postfix" in kwargs:
+ 733 postfix = kwargs.get("postfix")
+ 734
+ 735 if "files" in kwargs:
+ 736 files = kwargs.get("files")
+ 737 postfix = ''
+ 738 else:
+ 739 found = []
+ 740 files = []
+ 741 for (dirpath, dirnames, filenames) in os.walk(path + "/"):
+ 742 found.extend(filenames)
+ 743 break
+ 744 for f in found:
+ 745 if fnmatch.fnmatch(f, prefix + "*" + postfix):
+ 746 files.append(f)
+ 747
+ 748 files = sorted(files)
749
- 750 r_start_index = []
- 751 r_stop_index = []
- 752 deltas = []
- 753 configlist = []
- 754 if not zeuthen:
- 755 obspos += 8
- 756 for rep, file in enumerate(files):
- 757 with open(path + "/" + file, "rb") as fp:
- 758
- 759 Q = []
- 760 traj_list = []
- 761 if version in ['sfqcd']:
- 762 t = fp.read(12)
- 763 header = struct.unpack('<iii', t)
- 764 zthfl = header[0] # Zeuthen flow -> if it's equal to 2 it means that the Zeuthen flow is also 'measured' (apart from the Wilson flow)
- 765 ncs = header[1] # number of different values for c in t_flow=1/8 c² L² -> measurements done for ncs c's
- 766 tmax = header[2] # lattice T/a
- 767
- 768 t = fp.read(12)
- 769 Ls = struct.unpack('<iii', t)
- 770 if (Ls[0] == Ls[1] and Ls[1] == Ls[2]):
- 771 L = Ls[0]
- 772 if not (supposed_L == L) and supposed_L:
- 773 raise Exception("It seems the length given in the header and by you contradict each other")
- 774 else:
- 775 raise Exception("Found more than one spatial length in header!")
- 776
- 777 t = fp.read(16)
- 778 header2 = struct.unpack('<dd', t)
- 779 tol = header2[0]
- 780 cmax = header2[1] # highest value of c used
- 781
- 782 if c > cmax:
- 783 raise Exception('Flow has been determined between c=0 and c=%lf with tolerance %lf' % (cmax, tol))
- 784
- 785 if (zthfl == 2):
- 786 nfl = 2 # number of flows
- 787 else:
- 788 nfl = 1
- 789 iobs = 8 * nfl # number of flow observables calculated
- 790
- 791 while True:
- 792 t = fp.read(4)
- 793 if (len(t) < 4):
- 794 break
- 795 traj_list.append(struct.unpack('i', t)[0]) # trajectory number when measurement was done
+ 750 if 'r_start' in kwargs:
+ 751 r_start = kwargs.get('r_start')
+ 752 if len(r_start) != len(files):
+ 753 raise Exception('r_start does not match number of replicas')
+ 754 r_start = [o if o else None for o in r_start]
+ 755 else:
+ 756 r_start = [None] * len(files)
+ 757
+ 758 if 'r_stop' in kwargs:
+ 759 r_stop = kwargs.get('r_stop')
+ 760 if len(r_stop) != len(files):
+ 761 raise Exception('r_stop does not match number of replicas')
+ 762 else:
+ 763 r_stop = [None] * len(files)
+ 764 rep_names = []
+ 765
+ 766 zeuthen = kwargs.get('Zeuthen_flow', False)
+ 767 if zeuthen and version not in ['sfqcd']:
+ 768 raise Exception('Zeuthen flow can only be used for version==sfqcd')
+ 769
+ 770 r_start_index = []
+ 771 r_stop_index = []
+ 772 deltas = []
+ 773 configlist = []
+ 774 if not zeuthen:
+ 775 obspos += 8
+ 776 for rep, file in enumerate(files):
+ 777 with open(path + "/" + file, "rb") as fp:
+ 778
+ 779 Q = []
+ 780 traj_list = []
+ 781 if version in ['sfqcd']:
+ 782 t = fp.read(12)
+ 783 header = struct.unpack('<iii', t)
+ 784 zthfl = header[0] # Zeuthen flow -> if it's equal to 2 it means that the Zeuthen flow is also 'measured' (apart from the Wilson flow)
+ 785 ncs = header[1] # number of different values for c in t_flow=1/8 c² L² -> measurements done for ncs c's
+ 786 tmax = header[2] # lattice T/a
+ 787
+ 788 t = fp.read(12)
+ 789 Ls = struct.unpack('<iii', t)
+ 790 if (Ls[0] == Ls[1] and Ls[1] == Ls[2]):
+ 791 L = Ls[0]
+ 792 if not (supposed_L == L) and supposed_L:
+ 793 raise Exception("It seems the length given in the header and by you contradict each other")
+ 794 else:
+ 795 raise Exception("Found more than one spatial length in header!")
796
- 797 for j in range(ncs + 1):
- 798 for i in range(iobs):
- 799 t = fp.read(8 * tmax)
- 800 if (i == obspos): # determines the flow observable -> i=0 <-> Zeuthen flow
- 801 Q.append(struct.unpack('d' * tmax, t))
- 802
- 803 else:
- 804 t = fp.read(12)
- 805 header = struct.unpack('<iii', t)
- 806 # step size in integration steps "dnms"
- 807 dn = header[0]
- 808 # number of measurements, so "ntot"/dn
- 809 nn = header[1]
- 810 # lattice T/a
- 811 tmax = header[2]
- 812
- 813 t = fp.read(8)
- 814 eps = struct.unpack('d', t)[0]
- 815
- 816 while True:
- 817 t = fp.read(4)
- 818 if (len(t) < 4):
- 819 break
- 820 traj_list.append(struct.unpack('i', t)[0])
- 821 # Wsl
- 822 t = fp.read(8 * tmax * (nn + 1))
- 823 # Ysl
- 824 t = fp.read(8 * tmax * (nn + 1))
- 825 # Qsl, which is asked for in this method
- 826 t = fp.read(8 * tmax * (nn + 1))
- 827 # unpack the array of Qtops,
- 828 # on each timeslice t=0,...,tmax-1 and the
- 829 # measurement number in = 0...nn (see README.qcd1)
- 830 tmpd = struct.unpack('d' * tmax * (nn + 1), t)
- 831 Q.append(tmpd)
+ 797 t = fp.read(16)
+ 798 header2 = struct.unpack('<dd', t)
+ 799 tol = header2[0]
+ 800 cmax = header2[1] # highest value of c used
+ 801
+ 802 if c > cmax:
+ 803 raise Exception('Flow has been determined between c=0 and c=%lf with tolerance %lf' % (cmax, tol))
+ 804
+ 805 if (zthfl == 2):
+ 806 nfl = 2 # number of flows
+ 807 else:
+ 808 nfl = 1
+ 809 iobs = 8 * nfl # number of flow observables calculated
+ 810
+ 811 while True:
+ 812 t = fp.read(4)
+ 813 if (len(t) < 4):
+ 814 break
+ 815 traj_list.append(struct.unpack('i', t)[0]) # trajectory number when measurement was done
+ 816
+ 817 for j in range(ncs + 1):
+ 818 for i in range(iobs):
+ 819 t = fp.read(8 * tmax)
+ 820 if (i == obspos): # determines the flow observable -> i=0 <-> Zeuthen flow
+ 821 Q.append(struct.unpack('d' * tmax, t))
+ 822
+ 823 else:
+ 824 t = fp.read(12)
+ 825 header = struct.unpack('<iii', t)
+ 826 # step size in integration steps "dnms"
+ 827 dn = header[0]
+ 828 # number of measurements, so "ntot"/dn
+ 829 nn = header[1]
+ 830 # lattice T/a
+ 831 tmax = header[2]
832
- 833 if len(np.unique(np.diff(traj_list))) != 1:
- 834 raise Exception("Irregularities in stepsize found")
- 835 else:
- 836 if 'steps' in kwargs:
- 837 if steps != traj_list[1] - traj_list[0]:
- 838 raise Exception("steps and the found stepsize are not the same")
- 839 else:
- 840 steps = traj_list[1] - traj_list[0]
- 841
- 842 configlist.append([tr // steps // dtr_cnfg for tr in traj_list])
- 843 if configlist[-1][0] > 1:
- 844 offset = configlist[-1][0] - 1
- 845 warnings.warn('Assume thermalization and that the first measurement belongs to the first config. Offset = %d configs (%d trajectories / cycles)' % (
- 846 offset, offset * steps))
- 847 configlist[-1] = [item - offset for item in configlist[-1]]
- 848
- 849 if r_start[rep] is None:
- 850 r_start_index.append(0)
- 851 else:
- 852 try:
- 853 r_start_index.append(configlist[-1].index(r_start[rep]))
- 854 except ValueError:
- 855 raise Exception('Config %d not in file with range [%d, %d]' % (
- 856 r_start[rep], configlist[-1][0], configlist[-1][-1])) from None
- 857
- 858 if r_stop[rep] is None:
- 859 r_stop_index.append(len(configlist[-1]) - 1)
- 860 else:
- 861 try:
- 862 r_stop_index.append(configlist[-1].index(r_stop[rep]))
- 863 except ValueError:
- 864 raise Exception('Config %d not in file with range [%d, %d]' % (
- 865 r_stop[rep], configlist[-1][0], configlist[-1][-1])) from None
- 866
- 867 if version in ['sfqcd']:
- 868 cstepsize = cmax / ncs
- 869 index_aim = round(c / cstepsize)
- 870 else:
- 871 t_aim = (c * L) ** 2 / 8
- 872 index_aim = round(t_aim / eps / dn)
- 873
- 874 Q_sum = []
- 875 for i, item in enumerate(Q):
- 876 if sum_t is True:
- 877 Q_sum.append([sum(item[current:current + tmax])
- 878 for current in range(0, len(item), tmax)])
- 879 else:
- 880 Q_sum.append([item[int(tmax / 2)]])
- 881 Q_top = []
- 882 if version in ['sfqcd']:
- 883 for i in range(len(Q_sum) // (ncs + 1)):
- 884 Q_top.append(Q_sum[i * (ncs + 1) + index_aim][0])
- 885 else:
- 886 for i in range(len(Q) // dtr_cnfg):
- 887 Q_top.append(Q_sum[dtr_cnfg * i][index_aim])
- 888 if len(Q_top) != len(traj_list) // dtr_cnfg:
- 889 raise Exception("qtops and traj_list dont have the same length")
- 890
- 891 if kwargs.get('integer_charge', False):
- 892 Q_top = [round(q) for q in Q_top]
+ 833 t = fp.read(8)
+ 834 eps = struct.unpack('d', t)[0]
+ 835
+ 836 while True:
+ 837 t = fp.read(4)
+ 838 if (len(t) < 4):
+ 839 break
+ 840 traj_list.append(struct.unpack('i', t)[0])
+ 841 # Wsl
+ 842 t = fp.read(8 * tmax * (nn + 1))
+ 843 # Ysl
+ 844 t = fp.read(8 * tmax * (nn + 1))
+ 845 # Qsl, which is asked for in this method
+ 846 t = fp.read(8 * tmax * (nn + 1))
+ 847 # unpack the array of Qtops,
+ 848 # on each timeslice t=0,...,tmax-1 and the
+ 849 # measurement number in = 0...nn (see README.qcd1)
+ 850 tmpd = struct.unpack('d' * tmax * (nn + 1), t)
+ 851 Q.append(tmpd)
+ 852
+ 853 if len(np.unique(np.diff(traj_list))) != 1:
+ 854 raise Exception("Irregularities in stepsize found")
+ 855 else:
+ 856 if 'steps' in kwargs:
+ 857 if steps != traj_list[1] - traj_list[0]:
+ 858 raise Exception("steps and the found stepsize are not the same")
+ 859 else:
+ 860 steps = traj_list[1] - traj_list[0]
+ 861
+ 862 configlist.append([tr // steps // dtr_cnfg for tr in traj_list])
+ 863 if configlist[-1][0] > 1:
+ 864 offset = configlist[-1][0] - 1
+ 865 warnings.warn('Assume thermalization and that the first measurement belongs to the first config. Offset = %d configs (%d trajectories / cycles)' % (
+ 866 offset, offset * steps))
+ 867 configlist[-1] = [item - offset for item in configlist[-1]]
+ 868
+ 869 if r_start[rep] is None:
+ 870 r_start_index.append(0)
+ 871 else:
+ 872 try:
+ 873 r_start_index.append(configlist[-1].index(r_start[rep]))
+ 874 except ValueError:
+ 875 raise Exception('Config %d not in file with range [%d, %d]' % (
+ 876 r_start[rep], configlist[-1][0], configlist[-1][-1])) from None
+ 877
+ 878 if r_stop[rep] is None:
+ 879 r_stop_index.append(len(configlist[-1]) - 1)
+ 880 else:
+ 881 try:
+ 882 r_stop_index.append(configlist[-1].index(r_stop[rep]))
+ 883 except ValueError:
+ 884 raise Exception('Config %d not in file with range [%d, %d]' % (
+ 885 r_stop[rep], configlist[-1][0], configlist[-1][-1])) from None
+ 886
+ 887 if version in ['sfqcd']:
+ 888 cstepsize = cmax / ncs
+ 889 index_aim = round(c / cstepsize)
+ 890 else:
+ 891 t_aim = (c * L) ** 2 / 8
+ 892 index_aim = round(t_aim / eps / dn)
893
- 894 truncated_file = file[:-len(postfix)]
- 895
- 896 if "names" not in kwargs:
- 897 try:
- 898 idx = truncated_file.index('r')
- 899 except Exception:
- 900 if "names" not in kwargs:
- 901 raise Exception("Automatic recognition of replicum failed, please enter the key word 'names'.")
- 902 ens_name = truncated_file[:idx]
- 903 rep_names.append(ens_name + '|' + truncated_file[idx:])
- 904 else:
- 905 names = kwargs.get("names")
- 906 rep_names = names
- 907 deltas.append(Q_top)
- 908
- 909 idl = [range(int(configlist[rep][r_start_index[rep]]), int(configlist[rep][r_stop_index[rep]]) + 1, 1) for rep in range(len(deltas))]
- 910 deltas = [deltas[nrep][r_start_index[nrep]:r_stop_index[nrep] + 1] for nrep in range(len(deltas))]
- 911 result = Obs(deltas, rep_names, idl=idl)
- 912 result.tag = {"T": tmax - 1,
- 913 "L": L}
- 914 return result
+ 894 Q_sum = []
+ 895 for i, item in enumerate(Q):
+ 896 if sum_t is True:
+ 897 Q_sum.append([sum(item[current:current + tmax])
+ 898 for current in range(0, len(item), tmax)])
+ 899 else:
+ 900 Q_sum.append([item[int(tmax / 2)]])
+ 901 Q_top = []
+ 902 if version in ['sfqcd']:
+ 903 for i in range(len(Q_sum) // (ncs + 1)):
+ 904 Q_top.append(Q_sum[i * (ncs + 1) + index_aim][0])
+ 905 else:
+ 906 for i in range(len(Q) // dtr_cnfg):
+ 907 Q_top.append(Q_sum[dtr_cnfg * i][index_aim])
+ 908 if len(Q_top) != len(traj_list) // dtr_cnfg:
+ 909 raise Exception("qtops and traj_list dont have the same length")
+ 910
+ 911 if kwargs.get('integer_charge', False):
+ 912 Q_top = [round(q) for q in Q_top]
+ 913
+ 914 truncated_file = file[:-len(postfix)]
915
- 916
- 917def qtop_projection(qtop, target=0):
- 918 """Returns the projection to the topological charge sector defined by target.
- 919
- 920 Parameters
- 921 ----------
- 922 path : Obs
- 923 Topological charge.
- 924 target : int
- 925 Specifies the topological sector to be reweighted to (default 0)
- 926 """
- 927 if qtop.reweighted:
- 928 raise Exception('You can not use a reweighted observable for reweighting!')
- 929
- 930 proj_qtop = []
- 931 for n in qtop.deltas:
- 932 proj_qtop.append(np.array([1 if round(qtop.r_values[n] + q) == target else 0 for q in qtop.deltas[n]]))
- 933
- 934 reto = Obs(proj_qtop, qtop.names, idl=[qtop.idl[name] for name in qtop.names])
- 935 return reto
+ 916 if "names" not in kwargs:
+ 917 try:
+ 918 idx = truncated_file.index('r')
+ 919 except Exception:
+ 920 if "names" not in kwargs:
+ 921 raise Exception("Automatic recognition of replicum failed, please enter the key word 'names'.")
+ 922 ens_name = truncated_file[:idx]
+ 923 rep_names.append(ens_name + '|' + truncated_file[idx:])
+ 924 else:
+ 925 names = kwargs.get("names")
+ 926 rep_names = names
+ 927 deltas.append(Q_top)
+ 928
+ 929 idl = [range(int(configlist[rep][r_start_index[rep]]), int(configlist[rep][r_stop_index[rep]]) + 1, 1) for rep in range(len(deltas))]
+ 930 deltas = [deltas[nrep][r_start_index[nrep]:r_stop_index[nrep] + 1] for nrep in range(len(deltas))]
+ 931 result = Obs(deltas, rep_names, idl=idl)
+ 932 result.tag = {"T": tmax - 1,
+ 933 "L": L}
+ 934 return result
+ 935
936
- 937
- 938def read_qtop_sector(path, prefix, c, target=0, **kwargs):
- 939 """Constructs reweighting factors to a specified topological sector.
- 940
- 941 Parameters
- 942 ----------
- 943 path : str
- 944 path of the measurement files
- 945 prefix : str
- 946 prefix of the measurement files, e.g. <prefix>_id0_r0.ms.dat
- 947 c : double
- 948 Smearing radius in units of the lattice extent, c = sqrt(8 t0) / L
- 949 target : int
- 950 Specifies the topological sector to be reweighted to (default 0)
- 951 dtr_cnfg : int
- 952 (optional) parameter that specifies the number of trajectories
- 953 between two configs.
- 954 if it is not set, the distance between two measurements
- 955 in the file is assumed to be the distance between two configurations.
- 956 steps : int
- 957 (optional) Distance between two configurations in units of trajectories /
- 958 cycles. Assumed to be the distance between two measurements * dtr_cnfg if not given
- 959 version : str
- 960 version string of the openQCD (sfqcd) version used to create
- 961 the ensemble. Default is 2.0. May also be set to sfqcd.
- 962 L : int
- 963 spatial length of the lattice in L/a.
- 964 HAS to be set if version != sfqcd, since openQCD does not provide
- 965 this in the header
- 966 r_start : list
- 967 offset of the first ensemble, making it easier to match
- 968 later on with other Obs
- 969 r_stop : list
- 970 last configurations that need to be read (per replicum)
- 971 files : list
- 972 specify the exact files that need to be read
- 973 from path, practical if e.g. only one replicum is needed
- 974 names : list
- 975 Alternative labeling for replicas/ensembles.
- 976 Has to have the appropriate length
- 977 Zeuthen_flow : bool
- 978 (optional) If True, the Zeuthen flow is used for Qtop. Only possible
- 979 for version=='sfqcd' If False, the Wilson flow is used.
- 980 """
- 981
- 982 if not isinstance(target, int):
- 983 raise Exception("'target' has to be an integer.")
- 984
- 985 kwargs['integer_charge'] = True
- 986 qtop = read_qtop(path, prefix, c, **kwargs)
- 987
- 988 return qtop_projection(qtop, target=target)
- 989
- 990
- 991def read_ms5_xsf(path, prefix, qc, corr, sep="r", **kwargs):
- 992 """
- 993 Read data from files in the specified directory with the specified prefix and quark combination extension, and return a `Corr` object containing the data.
- 994
- 995 Parameters
- 996 ----------
- 997 path : str
- 998 The directory to search for the files in.
- 999 prefix : str
-1000 The prefix to match the files against.
-1001 qc : str
-1002 The quark combination extension to match the files against.
-1003 corr : str
-1004 The correlator to extract data for.
-1005 sep : str, optional
-1006 The separator to use when parsing the replika names.
-1007 **kwargs
-1008 Additional keyword arguments. The following keyword arguments are recognized:
-1009
-1010 - names (List[str]): A list of names to use for the replicas.
+ 937def qtop_projection(qtop, target=0):
+ 938 """Returns the projection to the topological charge sector defined by target.
+ 939
+ 940 Parameters
+ 941 ----------
+ 942 path : Obs
+ 943 Topological charge.
+ 944 target : int
+ 945 Specifies the topological sector to be reweighted to (default 0)
+ 946
+ 947 Returns
+ 948 -------
+ 949 reto : Obs
+ 950 projection to the topological charge sector defined by target
+ 951 """
+ 952 if qtop.reweighted:
+ 953 raise Exception('You can not use a reweighted observable for reweighting!')
+ 954
+ 955 proj_qtop = []
+ 956 for n in qtop.deltas:
+ 957 proj_qtop.append(np.array([1 if round(qtop.r_values[n] + q) == target else 0 for q in qtop.deltas[n]]))
+ 958
+ 959 reto = Obs(proj_qtop, qtop.names, idl=[qtop.idl[name] for name in qtop.names])
+ 960 return reto
+ 961
+ 962
+ 963def read_qtop_sector(path, prefix, c, target=0, **kwargs):
+ 964 """Constructs reweighting factors to a specified topological sector.
+ 965
+ 966 Parameters
+ 967 ----------
+ 968 path : str
+ 969 path of the measurement files
+ 970 prefix : str
+ 971 prefix of the measurement files, e.g. <prefix>_id0_r0.ms.dat
+ 972 c : double
+ 973 Smearing radius in units of the lattice extent, c = sqrt(8 t0) / L
+ 974 target : int
+ 975 Specifies the topological sector to be reweighted to (default 0)
+ 976 dtr_cnfg : int
+ 977 (optional) parameter that specifies the number of trajectories
+ 978 between two configs.
+ 979 if it is not set, the distance between two measurements
+ 980 in the file is assumed to be the distance between two configurations.
+ 981 steps : int
+ 982 (optional) Distance between two configurations in units of trajectories /
+ 983 cycles. Assumed to be the distance between two measurements * dtr_cnfg if not given
+ 984 version : str
+ 985 version string of the openQCD (sfqcd) version used to create
+ 986 the ensemble. Default is 2.0. May also be set to sfqcd.
+ 987 L : int
+ 988 spatial length of the lattice in L/a.
+ 989 HAS to be set if version != sfqcd, since openQCD does not provide
+ 990 this in the header
+ 991 r_start : list
+ 992 offset of the first ensemble, making it easier to match
+ 993 later on with other Obs
+ 994 r_stop : list
+ 995 last configurations that need to be read (per replicum)
+ 996 files : list
+ 997 specify the exact files that need to be read
+ 998 from path, practical if e.g. only one replicum is needed
+ 999 names : list
+1000 Alternative labeling for replicas/ensembles.
+1001 Has to have the appropriate length
+1002 Zeuthen_flow : bool
+1003 (optional) If True, the Zeuthen flow is used for Qtop. Only possible
+1004 for version=='sfqcd' If False, the Wilson flow is used.
+1005
+1006 Returns
+1007 -------
+1008 reto : Obs
+1009 projection to the topological charge sector defined by target
+1010 """
1011
-1012 Returns
-1013 -------
-1014 Corr
-1015 A complex valued `Corr` object containing the data read from the files. In case of boudary to bulk correlators.
-1016 or
-1017 CObs
-1018 A complex valued `CObs` object containing the data read from the files. In case of boudary to boundary correlators.
+1012 if not isinstance(target, int):
+1013 raise Exception("'target' has to be an integer.")
+1014
+1015 kwargs['integer_charge'] = True
+1016 qtop = read_qtop(path, prefix, c, **kwargs)
+1017
+1018 return qtop_projection(qtop, target=target)
1019
1020
-1021 Raises
-1022 ------
-1023 FileNotFoundError
-1024 If no files matching the specified prefix and quark combination extension are found in the specified directory.
-1025 IOError
-1026 If there is an error reading a file.
-1027 struct.error
-1028 If there is an error unpacking binary data.
-1029 """
-1030
-1031 found = []
-1032 files = []
-1033 names = []
-1034 for (dirpath, dirnames, filenames) in os.walk(path + "/"):
-1035 found.extend(filenames)
-1036 break
-1037
-1038 for f in found:
-1039 if fnmatch.fnmatch(f, prefix + "*.ms5_xsf_" + qc + ".dat"):
-1040 files.append(f)
-1041 if not sep == "":
-1042 names.append(prefix + "|r" + f.split(".")[0].split(sep)[1])
-1043 else:
-1044 names.append(prefix)
-1045 files = sorted(files)
-1046
-1047 if "names" in kwargs:
-1048 names = kwargs.get("names")
-1049 else:
-1050 names = sorted(names)
-1051
-1052 cnfgs = []
-1053 realsamples = []
-1054 imagsamples = []
-1055 repnum = 0
-1056 for file in files:
-1057 with open(path + "/" + file, "rb") as fp:
-1058
-1059 t = fp.read(8)
-1060 kappa = struct.unpack('d', t)[0]
-1061 t = fp.read(8)
-1062 csw = struct.unpack('d', t)[0]
-1063 t = fp.read(8)
-1064 dF = struct.unpack('d', t)[0]
-1065 t = fp.read(8)
-1066 zF = struct.unpack('d', t)[0]
+1021def read_ms5_xsf(path, prefix, qc, corr, sep="r", **kwargs):
+1022 """
+1023 Read data from files in the specified directory with the specified prefix and quark combination extension, and return a `Corr` object containing the data.
+1024
+1025 Parameters
+1026 ----------
+1027 path : str
+1028 The directory to search for the files in.
+1029 prefix : str
+1030 The prefix to match the files against.
+1031 qc : str
+1032 The quark combination extension to match the files against.
+1033 corr : str
+1034 The correlator to extract data for.
+1035 sep : str, optional
+1036 The separator to use when parsing the replika names.
+1037 **kwargs
+1038 Additional keyword arguments. The following keyword arguments are recognized:
+1039
+1040 - names (List[str]): A list of names to use for the replicas.
+1041
+1042 Returns
+1043 -------
+1044 Corr
+1045 A complex valued `Corr` object containing the data read from the files. In case of boudary to bulk correlators.
+1046 or
+1047 CObs
+1048 A complex valued `CObs` object containing the data read from the files. In case of boudary to boundary correlators.
+1049
+1050
+1051 Raises
+1052 ------
+1053 FileNotFoundError
+1054 If no files matching the specified prefix and quark combination extension are found in the specified directory.
+1055 IOError
+1056 If there is an error reading a file.
+1057 struct.error
+1058 If there is an error unpacking binary data.
+1059 """
+1060
+1061 found = []
+1062 files = []
+1063 names = []
+1064 for (dirpath, dirnames, filenames) in os.walk(path + "/"):
+1065 found.extend(filenames)
+1066 break
1067
-1068 t = fp.read(4)
-1069 tmax = struct.unpack('i', t)[0]
-1070 t = fp.read(4)
-1071 bnd = struct.unpack('i', t)[0]
-1072
-1073 placesBI = ["gS", "gP",
-1074 "gA", "gV",
-1075 "gVt", "lA",
-1076 "lV", "lVt",
-1077 "lT", "lTt"]
-1078 placesBB = ["g1", "l1"]
-1079
-1080 # the chunks have the following structure:
-1081 # confignumber, 10x timedependent complex correlators as doubles, 2x timeindependent complex correlators as doubles
-1082
-1083 chunksize = 4 + (8 * 2 * tmax * 10) + (8 * 2 * 2)
-1084 packstr = '=i' + ('d' * 2 * tmax * 10) + ('d' * 2 * 2)
-1085 cnfgs.append([])
-1086 realsamples.append([])
-1087 imagsamples.append([])
-1088 for t in range(tmax):
-1089 realsamples[repnum].append([])
-1090 imagsamples[repnum].append([])
-1091
-1092 while True:
-1093 cnfgt = fp.read(chunksize)
-1094 if not cnfgt:
-1095 break
-1096 asascii = struct.unpack(packstr, cnfgt)
-1097 cnfg = asascii[0]
-1098 cnfgs[repnum].append(cnfg)
-1099
-1100 if corr not in placesBB:
-1101 tmpcorr = asascii[1 + 2 * tmax * placesBI.index(corr):1 + 2 * tmax * placesBI.index(corr) + 2 * tmax]
-1102 else:
-1103 tmpcorr = asascii[1 + 2 * tmax * len(placesBI) + 2 * placesBB.index(corr):1 + 2 * tmax * len(placesBI) + 2 * placesBB.index(corr) + 2]
-1104
-1105 corrres = [[], []]
-1106 for i in range(len(tmpcorr)):
-1107 corrres[i % 2].append(tmpcorr[i])
-1108 for t in range(int(len(tmpcorr) / 2)):
-1109 realsamples[repnum][t].append(corrres[0][t])
-1110 for t in range(int(len(tmpcorr) / 2)):
-1111 imagsamples[repnum][t].append(corrres[1][t])
-1112 repnum += 1
-1113
-1114 s = "Read correlator " + corr + " from " + str(repnum) + " replika with " + str(len(realsamples[0][t]))
-1115 for rep in range(1, repnum):
-1116 s += ", " + str(len(realsamples[rep][t]))
-1117 s += " samples"
-1118 print(s)
-1119 print("Asserted run parameters:\n T:", tmax, "kappa:", kappa, "csw:", csw, "dF:", dF, "zF:", zF, "bnd:", bnd)
-1120
-1121 # we have the data now... but we need to re format the whole thing and put it into Corr objects.
-1122
-1123 compObs = []
-1124
-1125 for t in range(int(len(tmpcorr) / 2)):
-1126 compObs.append(CObs(Obs([realsamples[rep][t] for rep in range(repnum)], names=names, idl=cnfgs),
-1127 Obs([imagsamples[rep][t] for rep in range(repnum)], names=names, idl=cnfgs)))
-1128
-1129 if len(compObs) == 1:
-1130 return compObs[0]
-1131 else:
-1132 return Corr(compObs)
+1068 for f in found:
+1069 if fnmatch.fnmatch(f, prefix + "*.ms5_xsf_" + qc + ".dat"):
+1070 files.append(f)
+1071 if not sep == "":
+1072 names.append(prefix + "|r" + f.split(".")[0].split(sep)[1])
+1073 else:
+1074 names.append(prefix)
+1075 files = sorted(files)
+1076
+1077 if "names" in kwargs:
+1078 names = kwargs.get("names")
+1079 else:
+1080 names = sorted(names)
+1081
+1082 cnfgs = []
+1083 realsamples = []
+1084 imagsamples = []
+1085 repnum = 0
+1086 for file in files:
+1087 with open(path + "/" + file, "rb") as fp:
+1088
+1089 t = fp.read(8)
+1090 kappa = struct.unpack('d', t)[0]
+1091 t = fp.read(8)
+1092 csw = struct.unpack('d', t)[0]
+1093 t = fp.read(8)
+1094 dF = struct.unpack('d', t)[0]
+1095 t = fp.read(8)
+1096 zF = struct.unpack('d', t)[0]
+1097
+1098 t = fp.read(4)
+1099 tmax = struct.unpack('i', t)[0]
+1100 t = fp.read(4)
+1101 bnd = struct.unpack('i', t)[0]
+1102
+1103 placesBI = ["gS", "gP",
+1104 "gA", "gV",
+1105 "gVt", "lA",
+1106 "lV", "lVt",
+1107 "lT", "lTt"]
+1108 placesBB = ["g1", "l1"]
+1109
+1110 # the chunks have the following structure:
+1111 # confignumber, 10x timedependent complex correlators as doubles, 2x timeindependent complex correlators as doubles
+1112
+1113 chunksize = 4 + (8 * 2 * tmax * 10) + (8 * 2 * 2)
+1114 packstr = '=i' + ('d' * 2 * tmax * 10) + ('d' * 2 * 2)
+1115 cnfgs.append([])
+1116 realsamples.append([])
+1117 imagsamples.append([])
+1118 for t in range(tmax):
+1119 realsamples[repnum].append([])
+1120 imagsamples[repnum].append([])
+1121
+1122 while True:
+1123 cnfgt = fp.read(chunksize)
+1124 if not cnfgt:
+1125 break
+1126 asascii = struct.unpack(packstr, cnfgt)
+1127 cnfg = asascii[0]
+1128 cnfgs[repnum].append(cnfg)
+1129
+1130 if corr not in placesBB:
+1131 tmpcorr = asascii[1 + 2 * tmax * placesBI.index(corr):1 + 2 * tmax * placesBI.index(corr) + 2 * tmax]
+1132 else:
+1133 tmpcorr = asascii[1 + 2 * tmax * len(placesBI) + 2 * placesBB.index(corr):1 + 2 * tmax * len(placesBI) + 2 * placesBB.index(corr) + 2]
+1134
+1135 corrres = [[], []]
+1136 for i in range(len(tmpcorr)):
+1137 corrres[i % 2].append(tmpcorr[i])
+1138 for t in range(int(len(tmpcorr) / 2)):
+1139 realsamples[repnum][t].append(corrres[0][t])
+1140 for t in range(int(len(tmpcorr) / 2)):
+1141 imagsamples[repnum][t].append(corrres[1][t])
+1142 repnum += 1
+1143
+1144 s = "Read correlator " + corr + " from " + str(repnum) + " replika with " + str(len(realsamples[0][t]))
+1145 for rep in range(1, repnum):
+1146 s += ", " + str(len(realsamples[rep][t]))
+1147 s += " samples"
+1148 print(s)
+1149 print("Asserted run parameters:\n T:", tmax, "kappa:", kappa, "csw:", csw, "dF:", dF, "zF:", zF, "bnd:", bnd)
+1150
+1151 # we have the data now... but we need to re format the whole thing and put it into Corr objects.
+1152
+1153 compObs = []
+1154
+1155 for t in range(int(len(tmpcorr) / 2)):
+1156 compObs.append(CObs(Obs([realsamples[rep][t] for rep in range(repnum)], names=names, idl=cnfgs),
+1157 Obs([imagsamples[rep][t] for rep in range(repnum)], names=names, idl=cnfgs)))
+1158
+1159 if len(compObs) == 1:
+1160 return compObs[0]
+1161 else:
+1162 return Corr(compObs)
@@ -1271,193 +1301,198 @@
43 files performed if given.
44 print_err : bool
45 Print additional information that is useful for debugging.
- 46 """
- 47 known_oqcd_versions = ['1.4', '1.6', '2.0']
- 48 if not (version in known_oqcd_versions):
- 49 raise Exception('Unknown openQCD version defined!')
- 50 print("Working with openQCD version " + version)
- 51 if 'postfix' in kwargs:
- 52 postfix = kwargs.get('postfix')
- 53 else:
- 54 postfix = ''
- 55 ls = []
- 56 for (dirpath, dirnames, filenames) in os.walk(path):
- 57 ls.extend(filenames)
- 58 break
- 59
- 60 if not ls:
- 61 raise Exception(f"Error, directory '{path}' not found")
- 62 if 'files' in kwargs:
- 63 ls = kwargs.get('files')
- 64 else:
- 65 for exc in ls:
- 66 if not fnmatch.fnmatch(exc, prefix + '*' + postfix + '.dat'):
- 67 ls = list(set(ls) - set([exc]))
- 68 if len(ls) > 1:
- 69 ls.sort(key=lambda x: int(re.findall(r'\d+', x[len(prefix):])[0]))
- 70 replica = len(ls)
- 71
- 72 if 'r_start' in kwargs:
- 73 r_start = kwargs.get('r_start')
- 74 if len(r_start) != replica:
- 75 raise Exception('r_start does not match number of replicas')
- 76 r_start = [o if o else None for o in r_start]
- 77 else:
- 78 r_start = [None] * replica
- 79
- 80 if 'r_stop' in kwargs:
- 81 r_stop = kwargs.get('r_stop')
- 82 if len(r_stop) != replica:
- 83 raise Exception('r_stop does not match number of replicas')
- 84 else:
- 85 r_stop = [None] * replica
- 86
- 87 if 'r_step' in kwargs:
- 88 r_step = kwargs.get('r_step')
+ 46
+ 47 Returns
+ 48 -------
+ 49 rwms : Obs
+ 50 Reweighting factors read
+ 51 """
+ 52 known_oqcd_versions = ['1.4', '1.6', '2.0']
+ 53 if not (version in known_oqcd_versions):
+ 54 raise Exception('Unknown openQCD version defined!')
+ 55 print("Working with openQCD version " + version)
+ 56 if 'postfix' in kwargs:
+ 57 postfix = kwargs.get('postfix')
+ 58 else:
+ 59 postfix = ''
+ 60 ls = []
+ 61 for (dirpath, dirnames, filenames) in os.walk(path):
+ 62 ls.extend(filenames)
+ 63 break
+ 64
+ 65 if not ls:
+ 66 raise Exception(f"Error, directory '{path}' not found")
+ 67 if 'files' in kwargs:
+ 68 ls = kwargs.get('files')
+ 69 else:
+ 70 for exc in ls:
+ 71 if not fnmatch.fnmatch(exc, prefix + '*' + postfix + '.dat'):
+ 72 ls = list(set(ls) - set([exc]))
+ 73 if len(ls) > 1:
+ 74 ls.sort(key=lambda x: int(re.findall(r'\d+', x[len(prefix):])[0]))
+ 75 replica = len(ls)
+ 76
+ 77 if 'r_start' in kwargs:
+ 78 r_start = kwargs.get('r_start')
+ 79 if len(r_start) != replica:
+ 80 raise Exception('r_start does not match number of replicas')
+ 81 r_start = [o if o else None for o in r_start]
+ 82 else:
+ 83 r_start = [None] * replica
+ 84
+ 85 if 'r_stop' in kwargs:
+ 86 r_stop = kwargs.get('r_stop')
+ 87 if len(r_stop) != replica:
+ 88 raise Exception('r_stop does not match number of replicas')
89 else:
- 90 r_step = 1
+ 90 r_stop = [None] * replica
91
- 92 print('Read reweighting factors from', prefix[:-1], ',',
- 93 replica, 'replica', end='')
- 94
- 95 if names is None:
- 96 rep_names = []
- 97 for entry in ls:
- 98 truncated_entry = entry
- 99 suffixes = [".dat", ".rwms", ".ms1"]
-100 for suffix in suffixes:
-101 if truncated_entry.endswith(suffix):
-102 truncated_entry = truncated_entry[0:-len(suffix)]
-103 idx = truncated_entry.index('r')
-104 rep_names.append(truncated_entry[:idx] + '|' + truncated_entry[idx:])
-105 else:
-106 rep_names = names
-107
-108 print_err = 0
-109 if 'print_err' in kwargs:
-110 print_err = 1
-111 print()
+ 92 if 'r_step' in kwargs:
+ 93 r_step = kwargs.get('r_step')
+ 94 else:
+ 95 r_step = 1
+ 96
+ 97 print('Read reweighting factors from', prefix[:-1], ',',
+ 98 replica, 'replica', end='')
+ 99
+100 if names is None:
+101 rep_names = []
+102 for entry in ls:
+103 truncated_entry = entry
+104 suffixes = [".dat", ".rwms", ".ms1"]
+105 for suffix in suffixes:
+106 if truncated_entry.endswith(suffix):
+107 truncated_entry = truncated_entry[0:-len(suffix)]
+108 idx = truncated_entry.index('r')
+109 rep_names.append(truncated_entry[:idx] + '|' + truncated_entry[idx:])
+110 else:
+111 rep_names = names
112
-113 deltas = []
-114
-115 configlist = []
-116 r_start_index = []
-117 r_stop_index = []
-118
-119 for rep in range(replica):
-120 tmp_array = []
-121 with open(path + '/' + ls[rep], 'rb') as fp:
-122
-123 t = fp.read(4) # number of reweighting factors
-124 if rep == 0:
-125 nrw = struct.unpack('i', t)[0]
-126 if version == '2.0':
-127 nrw = int(nrw / 2)
-128 for k in range(nrw):
-129 deltas.append([])
-130 else:
-131 if ((nrw != struct.unpack('i', t)[0] and (not version == '2.0')) or (nrw != struct.unpack('i', t)[0] / 2 and version == '2.0')):
-132 raise Exception('Error: different number of reweighting factors for replicum', rep)
-133
-134 for k in range(nrw):
-135 tmp_array.append([])
-136
-137 # This block is necessary for openQCD1.6 and openQCD2.0 ms1 files
-138 nfct = []
-139 if version in ['1.6', '2.0']:
-140 for i in range(nrw):
-141 t = fp.read(4)
-142 nfct.append(struct.unpack('i', t)[0])
-143 else:
-144 for i in range(nrw):
-145 nfct.append(1)
-146
-147 nsrc = []
-148 for i in range(nrw):
-149 t = fp.read(4)
-150 nsrc.append(struct.unpack('i', t)[0])
-151 if version == '2.0':
-152 if not struct.unpack('i', fp.read(4))[0] == 0:
-153 print('something is wrong!')
-154
-155 configlist.append([])
-156 while True:
-157 t = fp.read(4)
-158 if len(t) < 4:
-159 break
-160 config_no = struct.unpack('i', t)[0]
-161 configlist[-1].append(config_no)
-162 for i in range(nrw):
-163 if (version == '2.0'):
-164 tmpd = _read_array_openQCD2(fp)
-165 tmpd = _read_array_openQCD2(fp)
-166 tmp_rw = tmpd['arr']
-167 tmp_nfct = 1.0
-168 for j in range(tmpd['n'][0]):
-169 tmp_nfct *= np.mean(np.exp(-np.asarray(tmp_rw[j])))
-170 if print_err:
-171 print(config_no, i, j,
-172 np.mean(np.exp(-np.asarray(tmp_rw[j]))),
-173 np.std(np.exp(-np.asarray(tmp_rw[j]))))
-174 print('Sources:',
-175 np.exp(-np.asarray(tmp_rw[j])))
-176 print('Partial factor:', tmp_nfct)
-177 elif version == '1.6' or version == '1.4':
-178 tmp_nfct = 1.0
-179 for j in range(nfct[i]):
-180 t = fp.read(8 * nsrc[i])
-181 t = fp.read(8 * nsrc[i])
-182 tmp_rw = struct.unpack('d' * nsrc[i], t)
-183 tmp_nfct *= np.mean(np.exp(-np.asarray(tmp_rw)))
-184 if print_err:
-185 print(config_no, i, j,
-186 np.mean(np.exp(-np.asarray(tmp_rw))),
-187 np.std(np.exp(-np.asarray(tmp_rw))))
-188 print('Sources:', np.exp(-np.asarray(tmp_rw)))
-189 print('Partial factor:', tmp_nfct)
-190 tmp_array[i].append(tmp_nfct)
-191
-192 diffmeas = configlist[-1][-1] - configlist[-1][-2]
-193 configlist[-1] = [item // diffmeas for item in configlist[-1]]
-194 if configlist[-1][0] > 1 and diffmeas > 1:
-195 warnings.warn('Assume thermalization and that the first measurement belongs to the first config.')
-196 offset = configlist[-1][0] - 1
-197 configlist[-1] = [item - offset for item in configlist[-1]]
-198
-199 if r_start[rep] is None:
-200 r_start_index.append(0)
-201 else:
-202 try:
-203 r_start_index.append(configlist[-1].index(r_start[rep]))
-204 except ValueError:
-205 raise Exception('Config %d not in file with range [%d, %d]' % (
-206 r_start[rep], configlist[-1][0], configlist[-1][-1])) from None
-207
-208 if r_stop[rep] is None:
-209 r_stop_index.append(len(configlist[-1]) - 1)
-210 else:
-211 try:
-212 r_stop_index.append(configlist[-1].index(r_stop[rep]))
-213 except ValueError:
-214 raise Exception('Config %d not in file with range [%d, %d]' % (
-215 r_stop[rep], configlist[-1][0], configlist[-1][-1])) from None
-216
-217 for k in range(nrw):
-218 deltas[k].append(tmp_array[k][r_start_index[rep]:r_stop_index[rep] + 1][::r_step])
-219
-220 if np.any([len(np.unique(np.diff(cl))) != 1 for cl in configlist]):
-221 raise Exception('Irregular spaced data in input file!', [len(np.unique(np.diff(cl))) for cl in configlist])
-222 stepsizes = [list(np.unique(np.diff(cl)))[0] for cl in configlist]
-223 if np.any([step != 1 for step in stepsizes]):
-224 warnings.warn('Stepsize between configurations is greater than one!' + str(stepsizes), RuntimeWarning)
-225
-226 print(',', nrw, 'reweighting factors with', nsrc, 'sources')
-227 result = []
-228 idl = [range(configlist[rep][r_start_index[rep]], configlist[rep][r_stop_index[rep]] + 1, r_step) for rep in range(replica)]
-229
-230 for t in range(nrw):
-231 result.append(Obs(deltas[t], rep_names, idl=idl))
-232 return result
+113 print_err = 0
+114 if 'print_err' in kwargs:
+115 print_err = 1
+116 print()
+117
+118 deltas = []
+119
+120 configlist = []
+121 r_start_index = []
+122 r_stop_index = []
+123
+124 for rep in range(replica):
+125 tmp_array = []
+126 with open(path + '/' + ls[rep], 'rb') as fp:
+127
+128 t = fp.read(4) # number of reweighting factors
+129 if rep == 0:
+130 nrw = struct.unpack('i', t)[0]
+131 if version == '2.0':
+132 nrw = int(nrw / 2)
+133 for k in range(nrw):
+134 deltas.append([])
+135 else:
+136 if ((nrw != struct.unpack('i', t)[0] and (not version == '2.0')) or (nrw != struct.unpack('i', t)[0] / 2 and version == '2.0')):
+137 raise Exception('Error: different number of reweighting factors for replicum', rep)
+138
+139 for k in range(nrw):
+140 tmp_array.append([])
+141
+142 # This block is necessary for openQCD1.6 and openQCD2.0 ms1 files
+143 nfct = []
+144 if version in ['1.6', '2.0']:
+145 for i in range(nrw):
+146 t = fp.read(4)
+147 nfct.append(struct.unpack('i', t)[0])
+148 else:
+149 for i in range(nrw):
+150 nfct.append(1)
+151
+152 nsrc = []
+153 for i in range(nrw):
+154 t = fp.read(4)
+155 nsrc.append(struct.unpack('i', t)[0])
+156 if version == '2.0':
+157 if not struct.unpack('i', fp.read(4))[0] == 0:
+158 print('something is wrong!')
+159
+160 configlist.append([])
+161 while True:
+162 t = fp.read(4)
+163 if len(t) < 4:
+164 break
+165 config_no = struct.unpack('i', t)[0]
+166 configlist[-1].append(config_no)
+167 for i in range(nrw):
+168 if (version == '2.0'):
+169 tmpd = _read_array_openQCD2(fp)
+170 tmpd = _read_array_openQCD2(fp)
+171 tmp_rw = tmpd['arr']
+172 tmp_nfct = 1.0
+173 for j in range(tmpd['n'][0]):
+174 tmp_nfct *= np.mean(np.exp(-np.asarray(tmp_rw[j])))
+175 if print_err:
+176 print(config_no, i, j,
+177 np.mean(np.exp(-np.asarray(tmp_rw[j]))),
+178 np.std(np.exp(-np.asarray(tmp_rw[j]))))
+179 print('Sources:',
+180 np.exp(-np.asarray(tmp_rw[j])))
+181 print('Partial factor:', tmp_nfct)
+182 elif version == '1.6' or version == '1.4':
+183 tmp_nfct = 1.0
+184 for j in range(nfct[i]):
+185 t = fp.read(8 * nsrc[i])
+186 t = fp.read(8 * nsrc[i])
+187 tmp_rw = struct.unpack('d' * nsrc[i], t)
+188 tmp_nfct *= np.mean(np.exp(-np.asarray(tmp_rw)))
+189 if print_err:
+190 print(config_no, i, j,
+191 np.mean(np.exp(-np.asarray(tmp_rw))),
+192 np.std(np.exp(-np.asarray(tmp_rw))))
+193 print('Sources:', np.exp(-np.asarray(tmp_rw)))
+194 print('Partial factor:', tmp_nfct)
+195 tmp_array[i].append(tmp_nfct)
+196
+197 diffmeas = configlist[-1][-1] - configlist[-1][-2]
+198 configlist[-1] = [item // diffmeas for item in configlist[-1]]
+199 if configlist[-1][0] > 1 and diffmeas > 1:
+200 warnings.warn('Assume thermalization and that the first measurement belongs to the first config.')
+201 offset = configlist[-1][0] - 1
+202 configlist[-1] = [item - offset for item in configlist[-1]]
+203
+204 if r_start[rep] is None:
+205 r_start_index.append(0)
+206 else:
+207 try:
+208 r_start_index.append(configlist[-1].index(r_start[rep]))
+209 except ValueError:
+210 raise Exception('Config %d not in file with range [%d, %d]' % (
+211 r_start[rep], configlist[-1][0], configlist[-1][-1])) from None
+212
+213 if r_stop[rep] is None:
+214 r_stop_index.append(len(configlist[-1]) - 1)
+215 else:
+216 try:
+217 r_stop_index.append(configlist[-1].index(r_stop[rep]))
+218 except ValueError:
+219 raise Exception('Config %d not in file with range [%d, %d]' % (
+220 r_stop[rep], configlist[-1][0], configlist[-1][-1])) from None
+221
+222 for k in range(nrw):
+223 deltas[k].append(tmp_array[k][r_start_index[rep]:r_stop_index[rep] + 1][::r_step])
+224
+225 if np.any([len(np.unique(np.diff(cl))) != 1 for cl in configlist]):
+226 raise Exception('Irregular spaced data in input file!', [len(np.unique(np.diff(cl))) for cl in configlist])
+227 stepsizes = [list(np.unique(np.diff(cl)))[0] for cl in configlist]
+228 if np.any([step != 1 for step in stepsizes]):
+229 warnings.warn('Stepsize between configurations is greater than one!' + str(stepsizes), RuntimeWarning)
+230
+231 print(',', nrw, 'reweighting factors with', nsrc, 'sources')
+232 result = []
+233 idl = [range(configlist[rep][r_start_index[rep]], configlist[rep][r_stop_index[rep]] + 1, r_step) for rep in range(replica)]
+234
+235 for t in range(nrw):
+236 result.append(Obs(deltas[t], rep_names, idl=idl))
+237 return result
@@ -1492,6 +1527,13 @@ files performed if given.
print_err (bool):
Print additional information that is useful for debugging.
+
+Returns
+
+
+- rwms (Obs):
+Reweighting factors read
+
@@ -1507,250 +1549,255 @@ Print additional information that is useful for debugging.
-
@@ -1824,52 +1878,57 @@ If False: The config numbers are assumed to be traj_number // difference
- 526def read_qtop(path, prefix, c, dtr_cnfg=1, version="openQCD", **kwargs):
-527 """Read the topologial charge based on openQCD gradient flow measurements.
-528
-529 Parameters
-530 ----------
-531 path : str
-532 path of the measurement files
-533 prefix : str
-534 prefix of the measurement files, e.g. <prefix>_id0_r0.ms.dat.
-535 Ignored if file names are passed explicitly via keyword files.
-536 c : double
-537 Smearing radius in units of the lattice extent, c = sqrt(8 t0) / L.
-538 dtr_cnfg : int
-539 (optional) parameter that specifies the number of measurements
-540 between two configs.
-541 If it is not set, the distance between two measurements
-542 in the file is assumed to be the distance between two configurations.
-543 steps : int
-544 (optional) Distance between two configurations in units of trajectories /
-545 cycles. Assumed to be the distance between two measurements * dtr_cnfg if not given
-546 version : str
-547 Either openQCD or sfqcd, depending on the data.
-548 L : int
-549 spatial length of the lattice in L/a.
-550 HAS to be set if version != sfqcd, since openQCD does not provide
-551 this in the header
-552 r_start : list
-553 list which contains the first config to be read for each replicum.
-554 r_stop : list
-555 list which contains the last config to be read for each replicum.
-556 files : list
-557 specify the exact files that need to be read
-558 from path, practical if e.g. only one replicum is needed
-559 postfix : str
-560 postfix of the file to read, e.g. '.gfms.dat' for openQCD-files
-561 names : list
-562 Alternative labeling for replicas/ensembles.
-563 Has to have the appropriate length.
-564 Zeuthen_flow : bool
-565 (optional) If True, the Zeuthen flow is used for Qtop. Only possible
-566 for version=='sfqcd' If False, the Wilson flow is used.
-567 integer_charge : bool
-568 If True, the charge is rounded towards the nearest integer on each config.
-569 """
-570
-571 return _read_flow_obs(path, prefix, c, dtr_cnfg=dtr_cnfg, version=version, obspos=0, **kwargs)
+ 536def read_qtop(path, prefix, c, dtr_cnfg=1, version="openQCD", **kwargs):
+537 """Read the topologial charge based on openQCD gradient flow measurements.
+538
+539 Parameters
+540 ----------
+541 path : str
+542 path of the measurement files
+543 prefix : str
+544 prefix of the measurement files, e.g. <prefix>_id0_r0.ms.dat.
+545 Ignored if file names are passed explicitly via keyword files.
+546 c : double
+547 Smearing radius in units of the lattice extent, c = sqrt(8 t0) / L.
+548 dtr_cnfg : int
+549 (optional) parameter that specifies the number of measurements
+550 between two configs.
+551 If it is not set, the distance between two measurements
+552 in the file is assumed to be the distance between two configurations.
+553 steps : int
+554 (optional) Distance between two configurations in units of trajectories /
+555 cycles. Assumed to be the distance between two measurements * dtr_cnfg if not given
+556 version : str
+557 Either openQCD or sfqcd, depending on the data.
+558 L : int
+559 spatial length of the lattice in L/a.
+560 HAS to be set if version != sfqcd, since openQCD does not provide
+561 this in the header
+562 r_start : list
+563 list which contains the first config to be read for each replicum.
+564 r_stop : list
+565 list which contains the last config to be read for each replicum.
+566 files : list
+567 specify the exact files that need to be read
+568 from path, practical if e.g. only one replicum is needed
+569 postfix : str
+570 postfix of the file to read, e.g. '.gfms.dat' for openQCD-files
+571 names : list
+572 Alternative labeling for replicas/ensembles.
+573 Has to have the appropriate length.
+574 Zeuthen_flow : bool
+575 (optional) If True, the Zeuthen flow is used for Qtop. Only possible
+576 for version=='sfqcd' If False, the Wilson flow is used.
+577 integer_charge : bool
+578 If True, the charge is rounded towards the nearest integer on each config.
+579
+580 Returns
+581 -------
+582 result : Obs
+583 Read topological charge
+584 """
+585
+586 return _read_flow_obs(path, prefix, c, dtr_cnfg=dtr_cnfg, version=version, obspos=0, **kwargs)
@@ -1917,6 +1976,13 @@ for version=='sfqcd' If False, the Wilson flow is used.
integer_charge (bool):
If True, the charge is rounded towards the nearest integer on each config.
+
+Returns
+
+
+- result (Obs):
+Read topological charge
+
@@ -1932,76 +1998,76 @@ If True, the charge is rounded towards the nearest integer on each config.
- 574def read_gf_coupling(path, prefix, c, dtr_cnfg=1, Zeuthen_flow=True, **kwargs):
-575 """Read the gradient flow coupling based on sfqcd gradient flow measurements. See 1607.06423 for details.
-576
-577 Note: The current implementation only works for c=0.3 and T=L. The definition of the coupling in 1607.06423 requires projection to topological charge zero which is not done within this function but has to be performed in a separate step.
-578
-579 Parameters
-580 ----------
-581 path : str
-582 path of the measurement files
-583 prefix : str
-584 prefix of the measurement files, e.g. <prefix>_id0_r0.ms.dat.
-585 Ignored if file names are passed explicitly via keyword files.
-586 c : double
-587 Smearing radius in units of the lattice extent, c = sqrt(8 t0) / L.
-588 dtr_cnfg : int
-589 (optional) parameter that specifies the number of measurements
-590 between two configs.
-591 If it is not set, the distance between two measurements
-592 in the file is assumed to be the distance between two configurations.
-593 steps : int
-594 (optional) Distance between two configurations in units of trajectories /
-595 cycles. Assumed to be the distance between two measurements * dtr_cnfg if not given
-596 r_start : list
-597 list which contains the first config to be read for each replicum.
-598 r_stop : list
-599 list which contains the last config to be read for each replicum.
-600 files : list
-601 specify the exact files that need to be read
-602 from path, practical if e.g. only one replicum is needed
-603 names : list
-604 Alternative labeling for replicas/ensembles.
-605 Has to have the appropriate length.
-606 postfix : str
-607 postfix of the file to read, e.g. '.gfms.dat' for openQCD-files
-608 Zeuthen_flow : bool
-609 (optional) If True, the Zeuthen flow is used for the coupling. If False, the Wilson flow is used.
-610 """
-611
-612 if c != 0.3:
-613 raise Exception("The required lattice norm is only implemented for c=0.3 at the moment.")
-614
-615 plaq = _read_flow_obs(path, prefix, c, dtr_cnfg=dtr_cnfg, version="sfqcd", obspos=6, sum_t=False, Zeuthen_flow=Zeuthen_flow, integer_charge=False, **kwargs)
-616 C2x1 = _read_flow_obs(path, prefix, c, dtr_cnfg=dtr_cnfg, version="sfqcd", obspos=7, sum_t=False, Zeuthen_flow=Zeuthen_flow, integer_charge=False, **kwargs)
-617 L = plaq.tag["L"]
-618 T = plaq.tag["T"]
-619
-620 if T != L:
-621 raise Exception("The required lattice norm is only implemented for T=L at the moment.")
-622
-623 if Zeuthen_flow is not True:
-624 raise Exception("The required lattice norm is only implemented for the Zeuthen flow at the moment.")
-625
-626 t = (c * L) ** 2 / 8
-627
-628 normdict = {4: 0.012341170468270,
-629 6: 0.010162691462430,
-630 8: 0.009031614807931,
-631 10: 0.008744966371393,
-632 12: 0.008650917856809,
-633 14: 8.611154391267955E-03,
-634 16: 0.008591758449508,
-635 20: 0.008575359627103,
-636 24: 0.008569387847540,
-637 28: 8.566803713382559E-03,
-638 32: 0.008565541650006,
-639 40: 8.564480684962046E-03,
-640 48: 8.564098025073460E-03,
-641 64: 8.563853943383087E-03}
+ 589def read_gf_coupling(path, prefix, c, dtr_cnfg=1, Zeuthen_flow=True, **kwargs):
+590 """Read the gradient flow coupling based on sfqcd gradient flow measurements. See 1607.06423 for details.
+591
+592 Note: The current implementation only works for c=0.3 and T=L. The definition of the coupling in 1607.06423 requires projection to topological charge zero which is not done within this function but has to be performed in a separate step.
+593
+594 Parameters
+595 ----------
+596 path : str
+597 path of the measurement files
+598 prefix : str
+599 prefix of the measurement files, e.g. <prefix>_id0_r0.ms.dat.
+600 Ignored if file names are passed explicitly via keyword files.
+601 c : double
+602 Smearing radius in units of the lattice extent, c = sqrt(8 t0) / L.
+603 dtr_cnfg : int
+604 (optional) parameter that specifies the number of measurements
+605 between two configs.
+606 If it is not set, the distance between two measurements
+607 in the file is assumed to be the distance between two configurations.
+608 steps : int
+609 (optional) Distance between two configurations in units of trajectories /
+610 cycles. Assumed to be the distance between two measurements * dtr_cnfg if not given
+611 r_start : list
+612 list which contains the first config to be read for each replicum.
+613 r_stop : list
+614 list which contains the last config to be read for each replicum.
+615 files : list
+616 specify the exact files that need to be read
+617 from path, practical if e.g. only one replicum is needed
+618 names : list
+619 Alternative labeling for replicas/ensembles.
+620 Has to have the appropriate length.
+621 postfix : str
+622 postfix of the file to read, e.g. '.gfms.dat' for openQCD-files
+623 Zeuthen_flow : bool
+624 (optional) If True, the Zeuthen flow is used for the coupling. If False, the Wilson flow is used.
+625 """
+626
+627 if c != 0.3:
+628 raise Exception("The required lattice norm is only implemented for c=0.3 at the moment.")
+629
+630 plaq = _read_flow_obs(path, prefix, c, dtr_cnfg=dtr_cnfg, version="sfqcd", obspos=6, sum_t=False, Zeuthen_flow=Zeuthen_flow, integer_charge=False, **kwargs)
+631 C2x1 = _read_flow_obs(path, prefix, c, dtr_cnfg=dtr_cnfg, version="sfqcd", obspos=7, sum_t=False, Zeuthen_flow=Zeuthen_flow, integer_charge=False, **kwargs)
+632 L = plaq.tag["L"]
+633 T = plaq.tag["T"]
+634
+635 if T != L:
+636 raise Exception("The required lattice norm is only implemented for T=L at the moment.")
+637
+638 if Zeuthen_flow is not True:
+639 raise Exception("The required lattice norm is only implemented for the Zeuthen flow at the moment.")
+640
+641 t = (c * L) ** 2 / 8
642
-643 return t * t * (5 / 3 * plaq - 1 / 12 * C2x1) / normdict[L]
+643 normdict = {4: 0.012341170468270,
+644 6: 0.010162691462430,
+645 8: 0.009031614807931,
+646 10: 0.008744966371393,
+647 12: 0.008650917856809,
+648 14: 8.611154391267955E-03,
+649 16: 0.008591758449508,
+650 20: 0.008575359627103,
+651 24: 0.008569387847540,
+652 28: 8.566803713382559E-03,
+653 32: 0.008565541650006,
+654 40: 8.564480684962046E-03,
+655 48: 8.564098025073460E-03,
+656 64: 8.563853943383087E-03}
+657
+658 return t * t * (5 / 3 * plaq - 1 / 12 * C2x1) / normdict[L]
@@ -2057,25 +2123,30 @@ postfix of the file to read, e.g. '.gfms.dat' for openQCD-files
- 918def qtop_projection(qtop, target=0):
-919 """Returns the projection to the topological charge sector defined by target.
-920
-921 Parameters
-922 ----------
-923 path : Obs
-924 Topological charge.
-925 target : int
-926 Specifies the topological sector to be reweighted to (default 0)
-927 """
-928 if qtop.reweighted:
-929 raise Exception('You can not use a reweighted observable for reweighting!')
-930
-931 proj_qtop = []
-932 for n in qtop.deltas:
-933 proj_qtop.append(np.array([1 if round(qtop.r_values[n] + q) == target else 0 for q in qtop.deltas[n]]))
-934
-935 reto = Obs(proj_qtop, qtop.names, idl=[qtop.idl[name] for name in qtop.names])
-936 return reto
+ 938def qtop_projection(qtop, target=0):
+939 """Returns the projection to the topological charge sector defined by target.
+940
+941 Parameters
+942 ----------
+943 path : Obs
+944 Topological charge.
+945 target : int
+946 Specifies the topological sector to be reweighted to (default 0)
+947
+948 Returns
+949 -------
+950 reto : Obs
+951 projection to the topological charge sector defined by target
+952 """
+953 if qtop.reweighted:
+954 raise Exception('You can not use a reweighted observable for reweighting!')
+955
+956 proj_qtop = []
+957 for n in qtop.deltas:
+958 proj_qtop.append(np.array([1 if round(qtop.r_values[n] + q) == target else 0 for q in qtop.deltas[n]]))
+959
+960 reto = Obs(proj_qtop, qtop.names, idl=[qtop.idl[name] for name in qtop.names])
+961 return reto
@@ -2089,6 +2160,13 @@ Topological charge.
target (int):
Specifies the topological sector to be reweighted to (default 0)
+
+Returns
+
+
+- reto (Obs):
+projection to the topological charge sector defined by target
+
@@ -2104,57 +2182,62 @@ Specifies the topological sector to be reweighted to (default 0)
- 939def read_qtop_sector(path, prefix, c, target=0, **kwargs):
-940 """Constructs reweighting factors to a specified topological sector.
-941
-942 Parameters
-943 ----------
-944 path : str
-945 path of the measurement files
-946 prefix : str
-947 prefix of the measurement files, e.g. <prefix>_id0_r0.ms.dat
-948 c : double
-949 Smearing radius in units of the lattice extent, c = sqrt(8 t0) / L
-950 target : int
-951 Specifies the topological sector to be reweighted to (default 0)
-952 dtr_cnfg : int
-953 (optional) parameter that specifies the number of trajectories
-954 between two configs.
-955 if it is not set, the distance between two measurements
-956 in the file is assumed to be the distance between two configurations.
-957 steps : int
-958 (optional) Distance between two configurations in units of trajectories /
-959 cycles. Assumed to be the distance between two measurements * dtr_cnfg if not given
-960 version : str
-961 version string of the openQCD (sfqcd) version used to create
-962 the ensemble. Default is 2.0. May also be set to sfqcd.
-963 L : int
-964 spatial length of the lattice in L/a.
-965 HAS to be set if version != sfqcd, since openQCD does not provide
-966 this in the header
-967 r_start : list
-968 offset of the first ensemble, making it easier to match
-969 later on with other Obs
-970 r_stop : list
-971 last configurations that need to be read (per replicum)
-972 files : list
-973 specify the exact files that need to be read
-974 from path, practical if e.g. only one replicum is needed
-975 names : list
-976 Alternative labeling for replicas/ensembles.
-977 Has to have the appropriate length
-978 Zeuthen_flow : bool
-979 (optional) If True, the Zeuthen flow is used for Qtop. Only possible
-980 for version=='sfqcd' If False, the Wilson flow is used.
-981 """
-982
-983 if not isinstance(target, int):
-984 raise Exception("'target' has to be an integer.")
-985
-986 kwargs['integer_charge'] = True
-987 qtop = read_qtop(path, prefix, c, **kwargs)
-988
-989 return qtop_projection(qtop, target=target)
+ 964def read_qtop_sector(path, prefix, c, target=0, **kwargs):
+ 965 """Constructs reweighting factors to a specified topological sector.
+ 966
+ 967 Parameters
+ 968 ----------
+ 969 path : str
+ 970 path of the measurement files
+ 971 prefix : str
+ 972 prefix of the measurement files, e.g. <prefix>_id0_r0.ms.dat
+ 973 c : double
+ 974 Smearing radius in units of the lattice extent, c = sqrt(8 t0) / L
+ 975 target : int
+ 976 Specifies the topological sector to be reweighted to (default 0)
+ 977 dtr_cnfg : int
+ 978 (optional) parameter that specifies the number of trajectories
+ 979 between two configs.
+ 980 if it is not set, the distance between two measurements
+ 981 in the file is assumed to be the distance between two configurations.
+ 982 steps : int
+ 983 (optional) Distance between two configurations in units of trajectories /
+ 984 cycles. Assumed to be the distance between two measurements * dtr_cnfg if not given
+ 985 version : str
+ 986 version string of the openQCD (sfqcd) version used to create
+ 987 the ensemble. Default is 2.0. May also be set to sfqcd.
+ 988 L : int
+ 989 spatial length of the lattice in L/a.
+ 990 HAS to be set if version != sfqcd, since openQCD does not provide
+ 991 this in the header
+ 992 r_start : list
+ 993 offset of the first ensemble, making it easier to match
+ 994 later on with other Obs
+ 995 r_stop : list
+ 996 last configurations that need to be read (per replicum)
+ 997 files : list
+ 998 specify the exact files that need to be read
+ 999 from path, practical if e.g. only one replicum is needed
+1000 names : list
+1001 Alternative labeling for replicas/ensembles.
+1002 Has to have the appropriate length
+1003 Zeuthen_flow : bool
+1004 (optional) If True, the Zeuthen flow is used for Qtop. Only possible
+1005 for version=='sfqcd' If False, the Wilson flow is used.
+1006
+1007 Returns
+1008 -------
+1009 reto : Obs
+1010 projection to the topological charge sector defined by target
+1011 """
+1012
+1013 if not isinstance(target, int):
+1014 raise Exception("'target' has to be an integer.")
+1015
+1016 kwargs['integer_charge'] = True
+1017 qtop = read_qtop(path, prefix, c, **kwargs)
+1018
+1019 return qtop_projection(qtop, target=target)
@@ -2201,6 +2284,13 @@ Has to have the appropriate length
(optional) If True, the Zeuthen flow is used for Qtop. Only possible
for version=='sfqcd' If False, the Wilson flow is used.
+
+Returns
+
+
+- reto (Obs):
+projection to the topological charge sector defined by target
+
@@ -2216,148 +2306,148 @@ for version=='sfqcd' If False, the Wilson flow is used.
- 992def read_ms5_xsf(path, prefix, qc, corr, sep="r", **kwargs):
- 993 """
- 994 Read data from files in the specified directory with the specified prefix and quark combination extension, and return a `Corr` object containing the data.
- 995
- 996 Parameters
- 997 ----------
- 998 path : str
- 999 The directory to search for the files in.
-1000 prefix : str
-1001 The prefix to match the files against.
-1002 qc : str
-1003 The quark combination extension to match the files against.
-1004 corr : str
-1005 The correlator to extract data for.
-1006 sep : str, optional
-1007 The separator to use when parsing the replika names.
-1008 **kwargs
-1009 Additional keyword arguments. The following keyword arguments are recognized:
-1010
-1011 - names (List[str]): A list of names to use for the replicas.
-1012
-1013 Returns
-1014 -------
-1015 Corr
-1016 A complex valued `Corr` object containing the data read from the files. In case of boudary to bulk correlators.
-1017 or
-1018 CObs
-1019 A complex valued `CObs` object containing the data read from the files. In case of boudary to boundary correlators.
-1020
-1021
-1022 Raises
-1023 ------
-1024 FileNotFoundError
-1025 If no files matching the specified prefix and quark combination extension are found in the specified directory.
-1026 IOError
-1027 If there is an error reading a file.
-1028 struct.error
-1029 If there is an error unpacking binary data.
-1030 """
-1031
-1032 found = []
-1033 files = []
-1034 names = []
-1035 for (dirpath, dirnames, filenames) in os.walk(path + "/"):
-1036 found.extend(filenames)
-1037 break
-1038
-1039 for f in found:
-1040 if fnmatch.fnmatch(f, prefix + "*.ms5_xsf_" + qc + ".dat"):
-1041 files.append(f)
-1042 if not sep == "":
-1043 names.append(prefix + "|r" + f.split(".")[0].split(sep)[1])
-1044 else:
-1045 names.append(prefix)
-1046 files = sorted(files)
-1047
-1048 if "names" in kwargs:
-1049 names = kwargs.get("names")
-1050 else:
-1051 names = sorted(names)
-1052
-1053 cnfgs = []
-1054 realsamples = []
-1055 imagsamples = []
-1056 repnum = 0
-1057 for file in files:
-1058 with open(path + "/" + file, "rb") as fp:
-1059
-1060 t = fp.read(8)
-1061 kappa = struct.unpack('d', t)[0]
-1062 t = fp.read(8)
-1063 csw = struct.unpack('d', t)[0]
-1064 t = fp.read(8)
-1065 dF = struct.unpack('d', t)[0]
-1066 t = fp.read(8)
-1067 zF = struct.unpack('d', t)[0]
+ 1022def read_ms5_xsf(path, prefix, qc, corr, sep="r", **kwargs):
+1023 """
+1024 Read data from files in the specified directory with the specified prefix and quark combination extension, and return a `Corr` object containing the data.
+1025
+1026 Parameters
+1027 ----------
+1028 path : str
+1029 The directory to search for the files in.
+1030 prefix : str
+1031 The prefix to match the files against.
+1032 qc : str
+1033 The quark combination extension to match the files against.
+1034 corr : str
+1035 The correlator to extract data for.
+1036 sep : str, optional
+1037 The separator to use when parsing the replika names.
+1038 **kwargs
+1039 Additional keyword arguments. The following keyword arguments are recognized:
+1040
+1041 - names (List[str]): A list of names to use for the replicas.
+1042
+1043 Returns
+1044 -------
+1045 Corr
+1046 A complex valued `Corr` object containing the data read from the files. In case of boudary to bulk correlators.
+1047 or
+1048 CObs
+1049 A complex valued `CObs` object containing the data read from the files. In case of boudary to boundary correlators.
+1050
+1051
+1052 Raises
+1053 ------
+1054 FileNotFoundError
+1055 If no files matching the specified prefix and quark combination extension are found in the specified directory.
+1056 IOError
+1057 If there is an error reading a file.
+1058 struct.error
+1059 If there is an error unpacking binary data.
+1060 """
+1061
+1062 found = []
+1063 files = []
+1064 names = []
+1065 for (dirpath, dirnames, filenames) in os.walk(path + "/"):
+1066 found.extend(filenames)
+1067 break
1068
-1069 t = fp.read(4)
-1070 tmax = struct.unpack('i', t)[0]
-1071 t = fp.read(4)
-1072 bnd = struct.unpack('i', t)[0]
-1073
-1074 placesBI = ["gS", "gP",
-1075 "gA", "gV",
-1076 "gVt", "lA",
-1077 "lV", "lVt",
-1078 "lT", "lTt"]
-1079 placesBB = ["g1", "l1"]
-1080
-1081 # the chunks have the following structure:
-1082 # confignumber, 10x timedependent complex correlators as doubles, 2x timeindependent complex correlators as doubles
-1083
-1084 chunksize = 4 + (8 * 2 * tmax * 10) + (8 * 2 * 2)
-1085 packstr = '=i' + ('d' * 2 * tmax * 10) + ('d' * 2 * 2)
-1086 cnfgs.append([])
-1087 realsamples.append([])
-1088 imagsamples.append([])
-1089 for t in range(tmax):
-1090 realsamples[repnum].append([])
-1091 imagsamples[repnum].append([])
-1092
-1093 while True:
-1094 cnfgt = fp.read(chunksize)
-1095 if not cnfgt:
-1096 break
-1097 asascii = struct.unpack(packstr, cnfgt)
-1098 cnfg = asascii[0]
-1099 cnfgs[repnum].append(cnfg)
-1100
-1101 if corr not in placesBB:
-1102 tmpcorr = asascii[1 + 2 * tmax * placesBI.index(corr):1 + 2 * tmax * placesBI.index(corr) + 2 * tmax]
-1103 else:
-1104 tmpcorr = asascii[1 + 2 * tmax * len(placesBI) + 2 * placesBB.index(corr):1 + 2 * tmax * len(placesBI) + 2 * placesBB.index(corr) + 2]
-1105
-1106 corrres = [[], []]
-1107 for i in range(len(tmpcorr)):
-1108 corrres[i % 2].append(tmpcorr[i])
-1109 for t in range(int(len(tmpcorr) / 2)):
-1110 realsamples[repnum][t].append(corrres[0][t])
-1111 for t in range(int(len(tmpcorr) / 2)):
-1112 imagsamples[repnum][t].append(corrres[1][t])
-1113 repnum += 1
-1114
-1115 s = "Read correlator " + corr + " from " + str(repnum) + " replika with " + str(len(realsamples[0][t]))
-1116 for rep in range(1, repnum):
-1117 s += ", " + str(len(realsamples[rep][t]))
-1118 s += " samples"
-1119 print(s)
-1120 print("Asserted run parameters:\n T:", tmax, "kappa:", kappa, "csw:", csw, "dF:", dF, "zF:", zF, "bnd:", bnd)
-1121
-1122 # we have the data now... but we need to re format the whole thing and put it into Corr objects.
-1123
-1124 compObs = []
-1125
-1126 for t in range(int(len(tmpcorr) / 2)):
-1127 compObs.append(CObs(Obs([realsamples[rep][t] for rep in range(repnum)], names=names, idl=cnfgs),
-1128 Obs([imagsamples[rep][t] for rep in range(repnum)], names=names, idl=cnfgs)))
-1129
-1130 if len(compObs) == 1:
-1131 return compObs[0]
-1132 else:
-1133 return Corr(compObs)
+1069 for f in found:
+1070 if fnmatch.fnmatch(f, prefix + "*.ms5_xsf_" + qc + ".dat"):
+1071 files.append(f)
+1072 if not sep == "":
+1073 names.append(prefix + "|r" + f.split(".")[0].split(sep)[1])
+1074 else:
+1075 names.append(prefix)
+1076 files = sorted(files)
+1077
+1078 if "names" in kwargs:
+1079 names = kwargs.get("names")
+1080 else:
+1081 names = sorted(names)
+1082
+1083 cnfgs = []
+1084 realsamples = []
+1085 imagsamples = []
+1086 repnum = 0
+1087 for file in files:
+1088 with open(path + "/" + file, "rb") as fp:
+1089
+1090 t = fp.read(8)
+1091 kappa = struct.unpack('d', t)[0]
+1092 t = fp.read(8)
+1093 csw = struct.unpack('d', t)[0]
+1094 t = fp.read(8)
+1095 dF = struct.unpack('d', t)[0]
+1096 t = fp.read(8)
+1097 zF = struct.unpack('d', t)[0]
+1098
+1099 t = fp.read(4)
+1100 tmax = struct.unpack('i', t)[0]
+1101 t = fp.read(4)
+1102 bnd = struct.unpack('i', t)[0]
+1103
+1104 placesBI = ["gS", "gP",
+1105 "gA", "gV",
+1106 "gVt", "lA",
+1107 "lV", "lVt",
+1108 "lT", "lTt"]
+1109 placesBB = ["g1", "l1"]
+1110
+1111 # the chunks have the following structure:
+1112 # confignumber, 10x timedependent complex correlators as doubles, 2x timeindependent complex correlators as doubles
+1113
+1114 chunksize = 4 + (8 * 2 * tmax * 10) + (8 * 2 * 2)
+1115 packstr = '=i' + ('d' * 2 * tmax * 10) + ('d' * 2 * 2)
+1116 cnfgs.append([])
+1117 realsamples.append([])
+1118 imagsamples.append([])
+1119 for t in range(tmax):
+1120 realsamples[repnum].append([])
+1121 imagsamples[repnum].append([])
+1122
+1123 while True:
+1124 cnfgt = fp.read(chunksize)
+1125 if not cnfgt:
+1126 break
+1127 asascii = struct.unpack(packstr, cnfgt)
+1128 cnfg = asascii[0]
+1129 cnfgs[repnum].append(cnfg)
+1130
+1131 if corr not in placesBB:
+1132 tmpcorr = asascii[1 + 2 * tmax * placesBI.index(corr):1 + 2 * tmax * placesBI.index(corr) + 2 * tmax]
+1133 else:
+1134 tmpcorr = asascii[1 + 2 * tmax * len(placesBI) + 2 * placesBB.index(corr):1 + 2 * tmax * len(placesBI) + 2 * placesBB.index(corr) + 2]
+1135
+1136 corrres = [[], []]
+1137 for i in range(len(tmpcorr)):
+1138 corrres[i % 2].append(tmpcorr[i])
+1139 for t in range(int(len(tmpcorr) / 2)):
+1140 realsamples[repnum][t].append(corrres[0][t])
+1141 for t in range(int(len(tmpcorr) / 2)):
+1142 imagsamples[repnum][t].append(corrres[1][t])
+1143 repnum += 1
+1144
+1145 s = "Read correlator " + corr + " from " + str(repnum) + " replika with " + str(len(realsamples[0][t]))
+1146 for rep in range(1, repnum):
+1147 s += ", " + str(len(realsamples[rep][t]))
+1148 s += " samples"
+1149 print(s)
+1150 print("Asserted run parameters:\n T:", tmax, "kappa:", kappa, "csw:", csw, "dF:", dF, "zF:", zF, "bnd:", bnd)
+1151
+1152 # we have the data now... but we need to re format the whole thing and put it into Corr objects.
+1153
+1154 compObs = []
+1155
+1156 for t in range(int(len(tmpcorr) / 2)):
+1157 compObs.append(CObs(Obs([realsamples[rep][t] for rep in range(repnum)], names=names, idl=cnfgs),
+1158 Obs([imagsamples[rep][t] for rep in range(repnum)], names=names, idl=cnfgs)))
+1159
+1160 if len(compObs) == 1:
+1161 return compObs[0]
+1162 else:
+1163 return Corr(compObs)
diff --git a/docs/pyerrors/input/pandas.html b/docs/pyerrors/input/pandas.html
index af4857f6..7c1526c8 100644
--- a/docs/pyerrors/input/pandas.html
+++ b/docs/pyerrors/input/pandas.html
@@ -109,133 +109,151 @@
22 How to behave if table already exists. Options 'fail', 'replace', 'append'.
23 gz : bool
24 If True the json strings are gzipped.
- 25 """
- 26 se_df = _serialize_df(df, gz=gz)
- 27 con = sqlite3.connect(db)
- 28 se_df.to_sql(table_name, con, if_exists=if_exists, index=False, **kwargs)
- 29 con.close()
- 30
- 31
- 32def read_sql(sql, db, auto_gamma=False, **kwargs):
- 33 """Execute SQL query on sqlite database and obtain DataFrame including Obs or Corr valued columns.
+ 25
+ 26 Returns
+ 27 -------
+ 28 None
+ 29 """
+ 30 se_df = _serialize_df(df, gz=gz)
+ 31 con = sqlite3.connect(db)
+ 32 se_df.to_sql(table_name, con, if_exists=if_exists, index=False, **kwargs)
+ 33 con.close()
34
- 35 Parameters
- 36 ----------
- 37 sql : str
- 38 SQL query to be executed.
- 39 db : str
- 40 Path to the sqlite database.
- 41 auto_gamma : bool
- 42 If True applies the gamma_method to all imported Obs objects with the default parameters for
- 43 the error analysis. Default False.
- 44 """
- 45 con = sqlite3.connect(db)
- 46 extract_df = pd.read_sql(sql, con, **kwargs)
- 47 con.close()
- 48 return _deserialize_df(extract_df, auto_gamma=auto_gamma)
- 49
- 50
- 51def dump_df(df, fname, gz=True):
- 52 """Exports a pandas DataFrame containing Obs valued columns to a (gzipped) csv file.
- 53
- 54 Before making use of pandas to_csv functionality Obs objects are serialized via the standardized
- 55 json format of pyerrors.
- 56
- 57 Parameters
- 58 ----------
- 59 df : pandas.DataFrame
- 60 Dataframe to be dumped to a file.
- 61 fname : str
- 62 Filename of the output file.
- 63 gz : bool
- 64 If True, the output is a gzipped csv file. If False, the output is a csv file.
- 65 """
- 66 out = _serialize_df(df, gz=False)
- 67
- 68 if not fname.endswith('.csv'):
- 69 fname += '.csv'
- 70
- 71 if gz is True:
- 72 if not fname.endswith('.gz'):
- 73 fname += '.gz'
- 74 out.to_csv(fname, index=False, compression='gzip')
- 75 else:
- 76 out.to_csv(fname, index=False)
- 77
- 78
- 79def load_df(fname, auto_gamma=False, gz=True):
- 80 """Imports a pandas DataFrame from a csv.(gz) file in which Obs objects are serialized as json strings.
- 81
- 82 Parameters
- 83 ----------
- 84 fname : str
- 85 Filename of the input file.
- 86 auto_gamma : bool
- 87 If True applies the gamma_method to all imported Obs objects with the default parameters for
- 88 the error analysis. Default False.
- 89 gz : bool
- 90 If True, assumes that data is gzipped. If False, assumes JSON file.
- 91 """
- 92 if not fname.endswith('.csv') and not fname.endswith('.gz'):
- 93 fname += '.csv'
+ 35
+ 36def read_sql(sql, db, auto_gamma=False, **kwargs):
+ 37 """Execute SQL query on sqlite database and obtain DataFrame including Obs or Corr valued columns.
+ 38
+ 39 Parameters
+ 40 ----------
+ 41 sql : str
+ 42 SQL query to be executed.
+ 43 db : str
+ 44 Path to the sqlite database.
+ 45 auto_gamma : bool
+ 46 If True applies the gamma_method to all imported Obs objects with the default parameters for
+ 47 the error analysis. Default False.
+ 48
+ 49 Returns
+ 50 -------
+ 51 data : pandas.DataFrame
+ 52 Dataframe with the content of the sqlite database.
+ 53 """
+ 54 con = sqlite3.connect(db)
+ 55 extract_df = pd.read_sql(sql, con, **kwargs)
+ 56 con.close()
+ 57 return _deserialize_df(extract_df, auto_gamma=auto_gamma)
+ 58
+ 59
+ 60def dump_df(df, fname, gz=True):
+ 61 """Exports a pandas DataFrame containing Obs valued columns to a (gzipped) csv file.
+ 62
+ 63 Before making use of pandas to_csv functionality Obs objects are serialized via the standardized
+ 64 json format of pyerrors.
+ 65
+ 66 Parameters
+ 67 ----------
+ 68 df : pandas.DataFrame
+ 69 Dataframe to be dumped to a file.
+ 70 fname : str
+ 71 Filename of the output file.
+ 72 gz : bool
+ 73 If True, the output is a gzipped csv file. If False, the output is a csv file.
+ 74
+ 75 Returns
+ 76 -------
+ 77 None
+ 78 """
+ 79 out = _serialize_df(df, gz=False)
+ 80
+ 81 if not fname.endswith('.csv'):
+ 82 fname += '.csv'
+ 83
+ 84 if gz is True:
+ 85 if not fname.endswith('.gz'):
+ 86 fname += '.gz'
+ 87 out.to_csv(fname, index=False, compression='gzip')
+ 88 else:
+ 89 out.to_csv(fname, index=False)
+ 90
+ 91
+ 92def load_df(fname, auto_gamma=False, gz=True):
+ 93 """Imports a pandas DataFrame from a csv.(gz) file in which Obs objects are serialized as json strings.
94
- 95 if gz is True:
- 96 if not fname.endswith('.gz'):
- 97 fname += '.gz'
- 98 with gzip.open(fname) as f:
- 99 re_import = pd.read_csv(f)
-100 else:
-101 if fname.endswith('.gz'):
-102 warnings.warn("Trying to read from %s without unzipping!" % fname, UserWarning)
-103 re_import = pd.read_csv(fname)
+ 95 Parameters
+ 96 ----------
+ 97 fname : str
+ 98 Filename of the input file.
+ 99 auto_gamma : bool
+100 If True applies the gamma_method to all imported Obs objects with the default parameters for
+101 the error analysis. Default False.
+102 gz : bool
+103 If True, assumes that data is gzipped. If False, assumes JSON file.
104
-105 return _deserialize_df(re_import, auto_gamma=auto_gamma)
-106
-107
-108def _serialize_df(df, gz=False):
-109 """Serializes all Obs or Corr valued columns into json strings according to the pyerrors json specification.
-110
-111 Parameters
-112 ----------
-113 df : pandas.DataFrame
-114 DataFrame to be serilized.
-115 gz: bool
-116 gzip the json string representation. Default False.
-117 """
-118 out = df.copy()
-119 for column in out:
-120 if isinstance(out[column][0], (Obs, Corr)):
-121 out[column] = out[column].transform(lambda x: create_json_string(x, indent=0))
-122 if gz is True:
-123 out[column] = out[column].transform(lambda x: gzip.compress(x.encode('utf-8')))
-124 return out
+105 Returns
+106 -------
+107 data : pandas.DataFrame
+108 Dataframe with the content of the sqlite database.
+109 """
+110 if not fname.endswith('.csv') and not fname.endswith('.gz'):
+111 fname += '.csv'
+112
+113 if gz is True:
+114 if not fname.endswith('.gz'):
+115 fname += '.gz'
+116 with gzip.open(fname) as f:
+117 re_import = pd.read_csv(f)
+118 else:
+119 if fname.endswith('.gz'):
+120 warnings.warn("Trying to read from %s without unzipping!" % fname, UserWarning)
+121 re_import = pd.read_csv(fname)
+122
+123 return _deserialize_df(re_import, auto_gamma=auto_gamma)
+124
125
-126
-127def _deserialize_df(df, auto_gamma=False):
-128 """Deserializes all pyerrors json strings into Obs or Corr objects according to the pyerrors json specification.
-129
-130 Parameters
-131 ----------
-132 df : pandas.DataFrame
-133 DataFrame to be deserilized.
-134 auto_gamma : bool
-135 If True applies the gamma_method to all imported Obs objects with the default parameters for
-136 the error analysis. Default False.
-137
-138 Notes:
-139 ------
-140 In case any column of the DataFrame is gzipped it is gunzipped in the process.
-141 """
-142 for column in df.select_dtypes(include="object"):
-143 if isinstance(df[column][0], bytes):
-144 if df[column][0].startswith(b"\x1f\x8b\x08\x00"):
-145 df[column] = df[column].transform(lambda x: gzip.decompress(x).decode('utf-8'))
-146 if isinstance(df[column][0], str):
-147 if '"program":' in df[column][0][:20]:
-148 df[column] = df[column].transform(lambda x: import_json_string(x, verbose=False))
-149 if auto_gamma is True:
-150 df[column].apply(lambda x: x.gamma_method())
-151 return df
+126def _serialize_df(df, gz=False):
+127 """Serializes all Obs or Corr valued columns into json strings according to the pyerrors json specification.
+128
+129 Parameters
+130 ----------
+131 df : pandas.DataFrame
+132 DataFrame to be serilized.
+133 gz: bool
+134 gzip the json string representation. Default False.
+135 """
+136 out = df.copy()
+137 for column in out:
+138 if isinstance(out[column][0], (Obs, Corr)):
+139 out[column] = out[column].transform(lambda x: create_json_string(x, indent=0))
+140 if gz is True:
+141 out[column] = out[column].transform(lambda x: gzip.compress(x.encode('utf-8')))
+142 return out
+143
+144
+145def _deserialize_df(df, auto_gamma=False):
+146 """Deserializes all pyerrors json strings into Obs or Corr objects according to the pyerrors json specification.
+147
+148 Parameters
+149 ----------
+150 df : pandas.DataFrame
+151 DataFrame to be deserilized.
+152 auto_gamma : bool
+153 If True applies the gamma_method to all imported Obs objects with the default parameters for
+154 the error analysis. Default False.
+155
+156 Notes:
+157 ------
+158 In case any column of the DataFrame is gzipped it is gunzipped in the process.
+159 """
+160 for column in df.select_dtypes(include="object"):
+161 if isinstance(df[column][0], bytes):
+162 if df[column][0].startswith(b"\x1f\x8b\x08\x00"):
+163 df[column] = df[column].transform(lambda x: gzip.decompress(x).decode('utf-8'))
+164 if isinstance(df[column][0], str):
+165 if '"program":' in df[column][0][:20]:
+166 df[column] = df[column].transform(lambda x: import_json_string(x, verbose=False))
+167 if auto_gamma is True:
+168 df[column].apply(lambda x: x.gamma_method())
+169 return df
@@ -266,11 +284,15 @@
23 How to behave if table already exists. Options 'fail', 'replace', 'append'.
24 gz : bool
25 If True the json strings are gzipped.
-26 """
-27 se_df = _serialize_df(df, gz=gz)
-28 con = sqlite3.connect(db)
-29 se_df.to_sql(table_name, con, if_exists=if_exists, index=False, **kwargs)
-30 con.close()
+26
+27 Returns
+28 -------
+29 None
+30 """
+31 se_df = _serialize_df(df, gz=gz)
+32 con = sqlite3.connect(db)
+33 se_df.to_sql(table_name, con, if_exists=if_exists, index=False, **kwargs)
+34 con.close()
@@ -290,6 +312,12 @@ How to behave if table already exists. Options 'fail', 'replace', 'append'.
gz (bool):
If True the json strings are gzipped.
+
+Returns
+
+
@@ -305,23 +333,28 @@ If True the json strings are gzipped.
- 33def read_sql(sql, db, auto_gamma=False, **kwargs):
-34 """Execute SQL query on sqlite database and obtain DataFrame including Obs or Corr valued columns.
-35
-36 Parameters
-37 ----------
-38 sql : str
-39 SQL query to be executed.
-40 db : str
-41 Path to the sqlite database.
-42 auto_gamma : bool
-43 If True applies the gamma_method to all imported Obs objects with the default parameters for
-44 the error analysis. Default False.
-45 """
-46 con = sqlite3.connect(db)
-47 extract_df = pd.read_sql(sql, con, **kwargs)
-48 con.close()
-49 return _deserialize_df(extract_df, auto_gamma=auto_gamma)
+ 37def read_sql(sql, db, auto_gamma=False, **kwargs):
+38 """Execute SQL query on sqlite database and obtain DataFrame including Obs or Corr valued columns.
+39
+40 Parameters
+41 ----------
+42 sql : str
+43 SQL query to be executed.
+44 db : str
+45 Path to the sqlite database.
+46 auto_gamma : bool
+47 If True applies the gamma_method to all imported Obs objects with the default parameters for
+48 the error analysis. Default False.
+49
+50 Returns
+51 -------
+52 data : pandas.DataFrame
+53 Dataframe with the content of the sqlite database.
+54 """
+55 con = sqlite3.connect(db)
+56 extract_df = pd.read_sql(sql, con, **kwargs)
+57 con.close()
+58 return _deserialize_df(extract_df, auto_gamma=auto_gamma)
@@ -338,6 +371,13 @@ Path to the sqlite database.
If True applies the gamma_method to all imported Obs objects with the default parameters for
the error analysis. Default False.
+
+Returns
+
+
+- data (pandas.DataFrame):
+Dataframe with the content of the sqlite database.
+
@@ -353,32 +393,36 @@ the error analysis. Default False.
- 52def dump_df(df, fname, gz=True):
-53 """Exports a pandas DataFrame containing Obs valued columns to a (gzipped) csv file.
-54
-55 Before making use of pandas to_csv functionality Obs objects are serialized via the standardized
-56 json format of pyerrors.
-57
-58 Parameters
-59 ----------
-60 df : pandas.DataFrame
-61 Dataframe to be dumped to a file.
-62 fname : str
-63 Filename of the output file.
-64 gz : bool
-65 If True, the output is a gzipped csv file. If False, the output is a csv file.
-66 """
-67 out = _serialize_df(df, gz=False)
-68
-69 if not fname.endswith('.csv'):
-70 fname += '.csv'
-71
-72 if gz is True:
-73 if not fname.endswith('.gz'):
-74 fname += '.gz'
-75 out.to_csv(fname, index=False, compression='gzip')
-76 else:
-77 out.to_csv(fname, index=False)
+ 61def dump_df(df, fname, gz=True):
+62 """Exports a pandas DataFrame containing Obs valued columns to a (gzipped) csv file.
+63
+64 Before making use of pandas to_csv functionality Obs objects are serialized via the standardized
+65 json format of pyerrors.
+66
+67 Parameters
+68 ----------
+69 df : pandas.DataFrame
+70 Dataframe to be dumped to a file.
+71 fname : str
+72 Filename of the output file.
+73 gz : bool
+74 If True, the output is a gzipped csv file. If False, the output is a csv file.
+75
+76 Returns
+77 -------
+78 None
+79 """
+80 out = _serialize_df(df, gz=False)
+81
+82 if not fname.endswith('.csv'):
+83 fname += '.csv'
+84
+85 if gz is True:
+86 if not fname.endswith('.gz'):
+87 fname += '.gz'
+88 out.to_csv(fname, index=False, compression='gzip')
+89 else:
+90 out.to_csv(fname, index=False)
@@ -397,6 +441,12 @@ Filename of the output file.
gz (bool):
If True, the output is a gzipped csv file. If False, the output is a csv file.
+
+Returns
+
+
@@ -412,33 +462,38 @@ If True, the output is a gzipped csv file. If False, the output is a csv file.
- 80def load_df(fname, auto_gamma=False, gz=True):
- 81 """Imports a pandas DataFrame from a csv.(gz) file in which Obs objects are serialized as json strings.
- 82
- 83 Parameters
- 84 ----------
- 85 fname : str
- 86 Filename of the input file.
- 87 auto_gamma : bool
- 88 If True applies the gamma_method to all imported Obs objects with the default parameters for
- 89 the error analysis. Default False.
- 90 gz : bool
- 91 If True, assumes that data is gzipped. If False, assumes JSON file.
- 92 """
- 93 if not fname.endswith('.csv') and not fname.endswith('.gz'):
- 94 fname += '.csv'
+ 93def load_df(fname, auto_gamma=False, gz=True):
+ 94 """Imports a pandas DataFrame from a csv.(gz) file in which Obs objects are serialized as json strings.
95
- 96 if gz is True:
- 97 if not fname.endswith('.gz'):
- 98 fname += '.gz'
- 99 with gzip.open(fname) as f:
-100 re_import = pd.read_csv(f)
-101 else:
-102 if fname.endswith('.gz'):
-103 warnings.warn("Trying to read from %s without unzipping!" % fname, UserWarning)
-104 re_import = pd.read_csv(fname)
+ 96 Parameters
+ 97 ----------
+ 98 fname : str
+ 99 Filename of the input file.
+100 auto_gamma : bool
+101 If True applies the gamma_method to all imported Obs objects with the default parameters for
+102 the error analysis. Default False.
+103 gz : bool
+104 If True, assumes that data is gzipped. If False, assumes JSON file.
105
-106 return _deserialize_df(re_import, auto_gamma=auto_gamma)
+106 Returns
+107 -------
+108 data : pandas.DataFrame
+109 Dataframe with the content of the sqlite database.
+110 """
+111 if not fname.endswith('.csv') and not fname.endswith('.gz'):
+112 fname += '.csv'
+113
+114 if gz is True:
+115 if not fname.endswith('.gz'):
+116 fname += '.gz'
+117 with gzip.open(fname) as f:
+118 re_import = pd.read_csv(f)
+119 else:
+120 if fname.endswith('.gz'):
+121 warnings.warn("Trying to read from %s without unzipping!" % fname, UserWarning)
+122 re_import = pd.read_csv(fname)
+123
+124 return _deserialize_df(re_import, auto_gamma=auto_gamma)
@@ -455,6 +510,13 @@ the error analysis. Default False.
gz (bool):
If True, assumes that data is gzipped. If False, assumes JSON file.
+
+Returns
+
+
+- data (pandas.DataFrame):
+Dataframe with the content of the sqlite database.
+
diff --git a/docs/pyerrors/input/sfcf.html b/docs/pyerrors/input/sfcf.html
index 14a86390..16205ce2 100644
--- a/docs/pyerrors/input/sfcf.html
+++ b/docs/pyerrors/input/sfcf.html
@@ -133,314 +133,320 @@
55 files: list
56 list of files to be read per replica, default is all.
57 for non-compact output format, hand the folders to be read here.
- 58 check_configs:
+ 58 check_configs: list[list[int]]
59 list of list of supposed configs, eg. [range(1,1000)]
60 for one replicum with 1000 configs
- 61 """
- 62 if kwargs.get('im'):
- 63 im = 1
- 64 part = 'imaginary'
- 65 else:
- 66 im = 0
- 67 part = 'real'
- 68 if "replica" in kwargs:
- 69 reps = kwargs.get("replica")
- 70 if corr_type == 'bb':
- 71 b2b = True
- 72 single = True
- 73 elif corr_type == 'bib':
- 74 b2b = True
- 75 single = False
- 76 else:
- 77 b2b = False
- 78 single = False
- 79 compact = True
- 80 appended = False
- 81 known_versions = ["0.0", "1.0", "2.0", "1.0c", "2.0c", "1.0a", "2.0a"]
- 82
- 83 if version not in known_versions:
- 84 raise Exception("This version is not known!")
- 85 if (version[-1] == "c"):
- 86 appended = False
- 87 compact = True
- 88 version = version[:-1]
- 89 elif (version[-1] == "a"):
- 90 appended = True
- 91 compact = False
- 92 version = version[:-1]
- 93 else:
- 94 compact = False
- 95 appended = False
- 96 read = 0
- 97 T = 0
- 98 start = 0
- 99 ls = []
-100 if "replica" in kwargs:
-101 ls = reps
-102 else:
-103 for (dirpath, dirnames, filenames) in os.walk(path):
-104 if not appended:
-105 ls.extend(dirnames)
-106 else:
-107 ls.extend(filenames)
-108 break
-109 if not ls:
-110 raise Exception('Error, directory not found')
-111 # Exclude folders with different names
-112 for exc in ls:
-113 if not fnmatch.fnmatch(exc, prefix + '*'):
-114 ls = list(set(ls) - set([exc]))
-115
-116 if not appended:
-117 if len(ls) > 1:
-118 # New version, to cope with ids, etc.
-119 ls.sort(key=lambda x: int(re.findall(r'\d+', x[len(prefix):])[0]))
-120 replica = len(ls)
-121 else:
-122 replica = len([file.split(".")[-1] for file in ls]) // len(set([file.split(".")[-1] for file in ls]))
-123 print('Read', part, 'part of', name, 'from', prefix[:-1],
-124 ',', replica, 'replica')
-125 if 'names' in kwargs:
-126 new_names = kwargs.get('names')
-127 if len(new_names) != len(set(new_names)):
-128 raise Exception("names are not unique!")
-129 if len(new_names) != replica:
-130 raise Exception('Names does not have the required length', replica)
-131 else:
-132 new_names = []
-133 if not appended:
-134 for entry in ls:
-135 try:
-136 idx = entry.index('r')
-137 except Exception:
-138 raise Exception("Automatic recognition of replicum failed, please enter the key word 'names'.")
-139
-140 if 'ens_name' in kwargs:
-141 new_names.append(kwargs.get('ens_name') + '|' + entry[idx:])
-142 else:
-143 new_names.append(entry[:idx] + '|' + entry[idx:])
-144 else:
+ 61
+ 62 Returns
+ 63 -------
+ 64 result: list[Obs]
+ 65 list of Observables with length T, observable per timeslice.
+ 66 bb-type correlators have length 1.
+ 67 """
+ 68 if kwargs.get('im'):
+ 69 im = 1
+ 70 part = 'imaginary'
+ 71 else:
+ 72 im = 0
+ 73 part = 'real'
+ 74 if "replica" in kwargs:
+ 75 reps = kwargs.get("replica")
+ 76 if corr_type == 'bb':
+ 77 b2b = True
+ 78 single = True
+ 79 elif corr_type == 'bib':
+ 80 b2b = True
+ 81 single = False
+ 82 else:
+ 83 b2b = False
+ 84 single = False
+ 85 compact = True
+ 86 appended = False
+ 87 known_versions = ["0.0", "1.0", "2.0", "1.0c", "2.0c", "1.0a", "2.0a"]
+ 88
+ 89 if version not in known_versions:
+ 90 raise Exception("This version is not known!")
+ 91 if (version[-1] == "c"):
+ 92 appended = False
+ 93 compact = True
+ 94 version = version[:-1]
+ 95 elif (version[-1] == "a"):
+ 96 appended = True
+ 97 compact = False
+ 98 version = version[:-1]
+ 99 else:
+100 compact = False
+101 appended = False
+102 read = 0
+103 T = 0
+104 start = 0
+105 ls = []
+106 if "replica" in kwargs:
+107 ls = reps
+108 else:
+109 for (dirpath, dirnames, filenames) in os.walk(path):
+110 if not appended:
+111 ls.extend(dirnames)
+112 else:
+113 ls.extend(filenames)
+114 break
+115 if not ls:
+116 raise Exception('Error, directory not found')
+117 # Exclude folders with different names
+118 for exc in ls:
+119 if not fnmatch.fnmatch(exc, prefix + '*'):
+120 ls = list(set(ls) - set([exc]))
+121
+122 if not appended:
+123 if len(ls) > 1:
+124 # New version, to cope with ids, etc.
+125 ls.sort(key=lambda x: int(re.findall(r'\d+', x[len(prefix):])[0]))
+126 replica = len(ls)
+127 else:
+128 replica = len([file.split(".")[-1] for file in ls]) // len(set([file.split(".")[-1] for file in ls]))
+129 print('Read', part, 'part of', name, 'from', prefix[:-1],
+130 ',', replica, 'replica')
+131 if 'names' in kwargs:
+132 new_names = kwargs.get('names')
+133 if len(new_names) != len(set(new_names)):
+134 raise Exception("names are not unique!")
+135 if len(new_names) != replica:
+136 raise Exception('Names does not have the required length', replica)
+137 else:
+138 new_names = []
+139 if not appended:
+140 for entry in ls:
+141 try:
+142 idx = entry.index('r')
+143 except Exception:
+144 raise Exception("Automatic recognition of replicum failed, please enter the key word 'names'.")
145
-146 for exc in ls:
-147 if not fnmatch.fnmatch(exc, prefix + '*.' + name):
-148 ls = list(set(ls) - set([exc]))
-149 ls.sort(key=lambda x: int(re.findall(r'\d+', x)[-1]))
-150 for entry in ls:
-151 myentry = entry[:-len(name) - 1]
-152 try:
-153 idx = myentry.index('r')
-154 except Exception:
-155 raise Exception("Automatic recognition of replicum failed, please enter the key word 'names'.")
-156
-157 if 'ens_name' in kwargs:
-158 new_names.append(kwargs.get('ens_name') + '|' + myentry[idx:])
-159 else:
-160 new_names.append(myentry[:idx] + '|' + myentry[idx:])
-161 idl = []
-162 if not appended:
-163 for i, item in enumerate(ls):
-164 sub_ls = []
-165 if "files" in kwargs:
-166 sub_ls = kwargs.get("files")
-167 sub_ls.sort(key=lambda x: int(re.findall(r'\d+', x)[-1]))
-168 else:
-169 for (dirpath, dirnames, filenames) in os.walk(path + '/' + item):
-170 if compact:
-171 sub_ls.extend(filenames)
-172 else:
-173 sub_ls.extend(dirnames)
-174 break
-175 if compact:
-176 for exc in sub_ls:
-177 if not fnmatch.fnmatch(exc, prefix + '*'):
-178 sub_ls = list(set(sub_ls) - set([exc]))
-179 sub_ls.sort(key=lambda x: int(re.findall(r'\d+', x)[-1]))
-180 else:
-181 for exc in sub_ls:
-182 if not fnmatch.fnmatch(exc, 'cfg*'):
-183 sub_ls = list(set(sub_ls) - set([exc]))
-184 sub_ls.sort(key=lambda x: int(x[3:]))
-185 rep_idl = []
-186 no_cfg = len(sub_ls)
-187 for cfg in sub_ls:
-188 try:
-189 if compact:
-190 rep_idl.append(int(cfg.split(cfg_separator)[-1]))
-191 else:
-192 rep_idl.append(int(cfg[3:]))
-193 except Exception:
-194 raise Exception("Couldn't parse idl from directroy, problem with file " + cfg)
-195 rep_idl.sort()
-196 # maybe there is a better way to print the idls
-197 print(item, ':', no_cfg, ' configurations')
-198 idl.append(rep_idl)
-199 # here we have found all the files we need to look into.
-200 if i == 0:
-201 # here, we want to find the place within the file,
-202 # where the correlator we need is stored.
-203 # to do so, the pattern needed is put together
-204 # from the input values
-205 if version == "0.0":
-206 pattern = "# " + name + " : offset " + str(noffset) + ", wf " + str(wf)
-207 # if b2b, a second wf is needed
-208 if b2b:
-209 pattern += ", wf_2 " + str(wf2)
-210 qs = quarks.split(" ")
-211 pattern += " : " + qs[0] + " - " + qs[1]
-212 file = open(path + '/' + item + '/' + sub_ls[0] + '/' + name, "r")
-213 for k, line in enumerate(file):
-214 if read == 1 and not line.strip() and k > start + 1:
-215 break
-216 if read == 1 and k >= start:
-217 T += 1
-218 if pattern in line:
-219 read = 1
-220 start = k + 1
-221 print(str(T) + " entries found.")
-222 file.close()
-223 else:
-224 pattern = 'name ' + name + '\nquarks ' + quarks + '\noffset ' + str(noffset) + '\nwf ' + str(wf)
-225 if b2b:
-226 pattern += '\nwf_2 ' + str(wf2)
-227 # and the file is parsed through to find the pattern
-228 if compact:
-229 file = open(path + '/' + item + '/' + sub_ls[0], "r")
-230 else:
-231 # for non-compactified versions of the files
-232 file = open(path + '/' + item + '/' + sub_ls[0] + '/' + name, "r")
-233
-234 content = file.read()
-235 match = re.search(pattern, content)
-236 if match:
-237 start_read = content.count('\n', 0, match.start()) + 5 + b2b
-238 end_match = re.search(r'\n\s*\n', content[match.start():])
-239 T = content[match.start():].count('\n', 0, end_match.start()) - 4 - b2b
-240 assert T > 0
-241 print(T, 'entries, starting to read in line', start_read)
-242 file.close()
-243 else:
-244 file.close()
-245 raise Exception('Correlator with pattern\n' + pattern + '\nnot found.')
-246
-247 # we found where the correlator
-248 # that is to be read is in the files
-249 # after preparing the datastructure
-250 # the correlators get parsed into...
-251 deltas = []
-252 for j in range(T):
-253 deltas.append([])
-254
-255 for t in range(T):
-256 deltas[t].append(np.zeros(no_cfg))
-257 if compact:
-258 for cfg in range(no_cfg):
-259 with open(path + '/' + item + '/' + sub_ls[cfg]) as fp:
-260 lines = fp.readlines()
-261 # check, if the correlator is in fact
-262 # printed completely
-263 if (start_read + T > len(lines)):
-264 raise Exception("EOF before end of correlator data! Maybe " + path + '/' + item + '/' + sub_ls[cfg] + " is corrupted?")
-265 # and start to read the correlator.
-266 # the range here is chosen like this,
-267 # since this allows for implementing
-268 # a security check for every read correlator later...
-269 for k in range(start_read - 6, start_read + T):
-270 if k == start_read - 5 - b2b:
-271 if lines[k].strip() != 'name ' + name:
-272 raise Exception('Wrong format', sub_ls[cfg])
-273 if (k >= start_read and k < start_read + T):
-274 floats = list(map(float, lines[k].split()))
-275 deltas[k - start_read][i][cfg] = floats[-2:][im]
-276 else:
-277 for cnfg, subitem in enumerate(sub_ls):
-278 with open(path + '/' + item + '/' + subitem + '/' + name) as fp:
-279 # since the non-compatified files
-280 # are typically not so long,
-281 # we can iterate over the whole file.
-282 # here one can also implement the chekc from above.
-283 for k, line in enumerate(fp):
-284 if (k >= start_read and k < start_read + T):
-285 floats = list(map(float, line.split()))
-286 if version == "0.0":
-287 deltas[k - start][i][cnfg] = floats[im - single]
-288 else:
-289 deltas[k - start_read][i][cnfg] = floats[1 + im - single]
-290
-291 else:
-292 if "files" in kwargs:
-293 ls = kwargs.get("files")
-294 else:
-295 for exc in ls:
-296 if not fnmatch.fnmatch(exc, prefix + '*.' + name):
-297 ls = list(set(ls) - set([exc]))
-298 ls.sort(key=lambda x: int(re.findall(r'\d+', x)[-1]))
-299 pattern = 'name ' + name + '\nquarks ' + quarks + '\noffset ' + str(noffset) + '\nwf ' + str(wf)
-300 if b2b:
-301 pattern += '\nwf_2 ' + str(wf2)
-302 for rep, file in enumerate(ls):
-303 rep_idl = []
-304 with open(path + '/' + file, 'r') as fp:
-305 content = fp.readlines()
-306 data_starts = []
-307 for linenumber, line in enumerate(content):
-308 if "[run]" in line:
-309 data_starts.append(linenumber)
-310 if len(set([data_starts[i] - data_starts[i - 1] for i in range(1, len(data_starts))])) > 1:
-311 raise Exception("Irregularities in file structure found, not all runs have the same output length")
-312 chunk = content[:data_starts[1]]
-313 for linenumber, line in enumerate(chunk):
-314 if line.startswith("gauge_name"):
-315 gauge_line = linenumber
-316 elif line.startswith("[correlator]"):
-317 corr_line = linenumber
-318 found_pat = ""
-319 for li in chunk[corr_line + 1: corr_line + 6 + b2b]:
-320 found_pat += li
-321 if re.search(pattern, found_pat):
-322 start_read = corr_line + 7 + b2b
-323 break
-324 endline = corr_line + 6 + b2b
-325 while not chunk[endline] == "\n":
-326 endline += 1
-327 T = endline - start_read
-328 if rep == 0:
-329 deltas = []
-330 for t in range(T):
-331 deltas.append([])
-332 for t in range(T):
-333 deltas[t].append(np.zeros(len(data_starts)))
-334 # all other chunks should follow the same structure
-335 for cnfg in range(len(data_starts)):
-336 start = data_starts[cnfg]
-337 stop = start + data_starts[1]
-338 chunk = content[start:stop]
-339 try:
-340 rep_idl.append(int(chunk[gauge_line].split(cfg_separator)[-1]))
-341 except Exception:
-342 raise Exception("Couldn't parse idl from directory, problem with chunk around line ", gauge_line)
-343
-344 found_pat = ""
-345 for li in chunk[corr_line + 1:corr_line + 6 + b2b]:
-346 found_pat += li
-347 if re.search(pattern, found_pat):
-348 for t, line in enumerate(chunk[start_read:start_read + T]):
-349 floats = list(map(float, line.split()))
-350 deltas[t][rep][cnfg] = floats[im + 1 - single]
-351 idl.append(rep_idl)
-352
-353 if "check_configs" in kwargs:
-354 print("Checking for missing configs...")
-355 che = kwargs.get("check_configs")
-356 if not (len(che) == len(idl)):
-357 raise Exception("check_configs has to be the same length as replica!")
-358 for r in range(len(idl)):
-359 print("checking " + new_names[r])
-360 utils.check_idl(idl[r], che[r])
-361 print("Done")
-362 result = []
-363 for t in range(T):
-364 result.append(Obs(deltas[t], new_names, idl=idl))
-365 return result
+146 if 'ens_name' in kwargs:
+147 new_names.append(kwargs.get('ens_name') + '|' + entry[idx:])
+148 else:
+149 new_names.append(entry[:idx] + '|' + entry[idx:])
+150 else:
+151
+152 for exc in ls:
+153 if not fnmatch.fnmatch(exc, prefix + '*.' + name):
+154 ls = list(set(ls) - set([exc]))
+155 ls.sort(key=lambda x: int(re.findall(r'\d+', x)[-1]))
+156 for entry in ls:
+157 myentry = entry[:-len(name) - 1]
+158 try:
+159 idx = myentry.index('r')
+160 except Exception:
+161 raise Exception("Automatic recognition of replicum failed, please enter the key word 'names'.")
+162
+163 if 'ens_name' in kwargs:
+164 new_names.append(kwargs.get('ens_name') + '|' + myentry[idx:])
+165 else:
+166 new_names.append(myentry[:idx] + '|' + myentry[idx:])
+167 idl = []
+168 if not appended:
+169 for i, item in enumerate(ls):
+170 sub_ls = []
+171 if "files" in kwargs:
+172 sub_ls = kwargs.get("files")
+173 sub_ls.sort(key=lambda x: int(re.findall(r'\d+', x)[-1]))
+174 else:
+175 for (dirpath, dirnames, filenames) in os.walk(path + '/' + item):
+176 if compact:
+177 sub_ls.extend(filenames)
+178 else:
+179 sub_ls.extend(dirnames)
+180 break
+181 if compact:
+182 for exc in sub_ls:
+183 if not fnmatch.fnmatch(exc, prefix + '*'):
+184 sub_ls = list(set(sub_ls) - set([exc]))
+185 sub_ls.sort(key=lambda x: int(re.findall(r'\d+', x)[-1]))
+186 else:
+187 for exc in sub_ls:
+188 if not fnmatch.fnmatch(exc, 'cfg*'):
+189 sub_ls = list(set(sub_ls) - set([exc]))
+190 sub_ls.sort(key=lambda x: int(x[3:]))
+191 rep_idl = []
+192 no_cfg = len(sub_ls)
+193 for cfg in sub_ls:
+194 try:
+195 if compact:
+196 rep_idl.append(int(cfg.split(cfg_separator)[-1]))
+197 else:
+198 rep_idl.append(int(cfg[3:]))
+199 except Exception:
+200 raise Exception("Couldn't parse idl from directroy, problem with file " + cfg)
+201 rep_idl.sort()
+202 # maybe there is a better way to print the idls
+203 print(item, ':', no_cfg, ' configurations')
+204 idl.append(rep_idl)
+205 # here we have found all the files we need to look into.
+206 if i == 0:
+207 # here, we want to find the place within the file,
+208 # where the correlator we need is stored.
+209 # to do so, the pattern needed is put together
+210 # from the input values
+211 if version == "0.0":
+212 pattern = "# " + name + " : offset " + str(noffset) + ", wf " + str(wf)
+213 # if b2b, a second wf is needed
+214 if b2b:
+215 pattern += ", wf_2 " + str(wf2)
+216 qs = quarks.split(" ")
+217 pattern += " : " + qs[0] + " - " + qs[1]
+218 file = open(path + '/' + item + '/' + sub_ls[0] + '/' + name, "r")
+219 for k, line in enumerate(file):
+220 if read == 1 and not line.strip() and k > start + 1:
+221 break
+222 if read == 1 and k >= start:
+223 T += 1
+224 if pattern in line:
+225 read = 1
+226 start = k + 1
+227 print(str(T) + " entries found.")
+228 file.close()
+229 else:
+230 pattern = 'name ' + name + '\nquarks ' + quarks + '\noffset ' + str(noffset) + '\nwf ' + str(wf)
+231 if b2b:
+232 pattern += '\nwf_2 ' + str(wf2)
+233 # and the file is parsed through to find the pattern
+234 if compact:
+235 file = open(path + '/' + item + '/' + sub_ls[0], "r")
+236 else:
+237 # for non-compactified versions of the files
+238 file = open(path + '/' + item + '/' + sub_ls[0] + '/' + name, "r")
+239
+240 content = file.read()
+241 match = re.search(pattern, content)
+242 if match:
+243 start_read = content.count('\n', 0, match.start()) + 5 + b2b
+244 end_match = re.search(r'\n\s*\n', content[match.start():])
+245 T = content[match.start():].count('\n', 0, end_match.start()) - 4 - b2b
+246 assert T > 0
+247 print(T, 'entries, starting to read in line', start_read)
+248 file.close()
+249 else:
+250 file.close()
+251 raise Exception('Correlator with pattern\n' + pattern + '\nnot found.')
+252
+253 # we found where the correlator
+254 # that is to be read is in the files
+255 # after preparing the datastructure
+256 # the correlators get parsed into...
+257 deltas = []
+258 for j in range(T):
+259 deltas.append([])
+260
+261 for t in range(T):
+262 deltas[t].append(np.zeros(no_cfg))
+263 if compact:
+264 for cfg in range(no_cfg):
+265 with open(path + '/' + item + '/' + sub_ls[cfg]) as fp:
+266 lines = fp.readlines()
+267 # check, if the correlator is in fact
+268 # printed completely
+269 if (start_read + T > len(lines)):
+270 raise Exception("EOF before end of correlator data! Maybe " + path + '/' + item + '/' + sub_ls[cfg] + " is corrupted?")
+271 # and start to read the correlator.
+272 # the range here is chosen like this,
+273 # since this allows for implementing
+274 # a security check for every read correlator later...
+275 for k in range(start_read - 6, start_read + T):
+276 if k == start_read - 5 - b2b:
+277 if lines[k].strip() != 'name ' + name:
+278 raise Exception('Wrong format', sub_ls[cfg])
+279 if (k >= start_read and k < start_read + T):
+280 floats = list(map(float, lines[k].split()))
+281 deltas[k - start_read][i][cfg] = floats[-2:][im]
+282 else:
+283 for cnfg, subitem in enumerate(sub_ls):
+284 with open(path + '/' + item + '/' + subitem + '/' + name) as fp:
+285 # since the non-compatified files
+286 # are typically not so long,
+287 # we can iterate over the whole file.
+288 # here one can also implement the chekc from above.
+289 for k, line in enumerate(fp):
+290 if (k >= start_read and k < start_read + T):
+291 floats = list(map(float, line.split()))
+292 if version == "0.0":
+293 deltas[k - start][i][cnfg] = floats[im - single]
+294 else:
+295 deltas[k - start_read][i][cnfg] = floats[1 + im - single]
+296
+297 else:
+298 if "files" in kwargs:
+299 ls = kwargs.get("files")
+300 else:
+301 for exc in ls:
+302 if not fnmatch.fnmatch(exc, prefix + '*.' + name):
+303 ls = list(set(ls) - set([exc]))
+304 ls.sort(key=lambda x: int(re.findall(r'\d+', x)[-1]))
+305 pattern = 'name ' + name + '\nquarks ' + quarks + '\noffset ' + str(noffset) + '\nwf ' + str(wf)
+306 if b2b:
+307 pattern += '\nwf_2 ' + str(wf2)
+308 for rep, file in enumerate(ls):
+309 rep_idl = []
+310 with open(path + '/' + file, 'r') as fp:
+311 content = fp.readlines()
+312 data_starts = []
+313 for linenumber, line in enumerate(content):
+314 if "[run]" in line:
+315 data_starts.append(linenumber)
+316 if len(set([data_starts[i] - data_starts[i - 1] for i in range(1, len(data_starts))])) > 1:
+317 raise Exception("Irregularities in file structure found, not all runs have the same output length")
+318 chunk = content[:data_starts[1]]
+319 for linenumber, line in enumerate(chunk):
+320 if line.startswith("gauge_name"):
+321 gauge_line = linenumber
+322 elif line.startswith("[correlator]"):
+323 corr_line = linenumber
+324 found_pat = ""
+325 for li in chunk[corr_line + 1: corr_line + 6 + b2b]:
+326 found_pat += li
+327 if re.search(pattern, found_pat):
+328 start_read = corr_line + 7 + b2b
+329 break
+330 endline = corr_line + 6 + b2b
+331 while not chunk[endline] == "\n":
+332 endline += 1
+333 T = endline - start_read
+334 if rep == 0:
+335 deltas = []
+336 for t in range(T):
+337 deltas.append([])
+338 for t in range(T):
+339 deltas[t].append(np.zeros(len(data_starts)))
+340 # all other chunks should follow the same structure
+341 for cnfg in range(len(data_starts)):
+342 start = data_starts[cnfg]
+343 stop = start + data_starts[1]
+344 chunk = content[start:stop]
+345 try:
+346 rep_idl.append(int(chunk[gauge_line].split(cfg_separator)[-1]))
+347 except Exception:
+348 raise Exception("Couldn't parse idl from directory, problem with chunk around line ", gauge_line)
+349
+350 found_pat = ""
+351 for li in chunk[corr_line + 1:corr_line + 6 + b2b]:
+352 found_pat += li
+353 if re.search(pattern, found_pat):
+354 for t, line in enumerate(chunk[start_read:start_read + T]):
+355 floats = list(map(float, line.split()))
+356 deltas[t][rep][cnfg] = floats[im + 1 - single]
+357 idl.append(rep_idl)
+358
+359 if "check_configs" in kwargs:
+360 print("Checking for missing configs...")
+361 che = kwargs.get("check_configs")
+362 if not (len(che) == len(idl)):
+363 raise Exception("check_configs has to be the same length as replica!")
+364 for r in range(len(idl)):
+365 print("checking " + new_names[r])
+366 utils.check_idl(idl[r], che[r])
+367 print("Done")
+368 result = []
+369 for t in range(T):
+370 result.append(Obs(deltas[t], new_names, idl=idl))
+371 return result
@@ -505,314 +511,320 @@
56 files: list
57 list of files to be read per replica, default is all.
58 for non-compact output format, hand the folders to be read here.
- 59 check_configs:
+ 59 check_configs: list[list[int]]
60 list of list of supposed configs, eg. [range(1,1000)]
61 for one replicum with 1000 configs
- 62 """
- 63 if kwargs.get('im'):
- 64 im = 1
- 65 part = 'imaginary'
- 66 else:
- 67 im = 0
- 68 part = 'real'
- 69 if "replica" in kwargs:
- 70 reps = kwargs.get("replica")
- 71 if corr_type == 'bb':
- 72 b2b = True
- 73 single = True
- 74 elif corr_type == 'bib':
- 75 b2b = True
- 76 single = False
- 77 else:
- 78 b2b = False
- 79 single = False
- 80 compact = True
- 81 appended = False
- 82 known_versions = ["0.0", "1.0", "2.0", "1.0c", "2.0c", "1.0a", "2.0a"]
- 83
- 84 if version not in known_versions:
- 85 raise Exception("This version is not known!")
- 86 if (version[-1] == "c"):
- 87 appended = False
- 88 compact = True
- 89 version = version[:-1]
- 90 elif (version[-1] == "a"):
- 91 appended = True
- 92 compact = False
- 93 version = version[:-1]
- 94 else:
- 95 compact = False
- 96 appended = False
- 97 read = 0
- 98 T = 0
- 99 start = 0
-100 ls = []
-101 if "replica" in kwargs:
-102 ls = reps
-103 else:
-104 for (dirpath, dirnames, filenames) in os.walk(path):
-105 if not appended:
-106 ls.extend(dirnames)
-107 else:
-108 ls.extend(filenames)
-109 break
-110 if not ls:
-111 raise Exception('Error, directory not found')
-112 # Exclude folders with different names
-113 for exc in ls:
-114 if not fnmatch.fnmatch(exc, prefix + '*'):
-115 ls = list(set(ls) - set([exc]))
-116
-117 if not appended:
-118 if len(ls) > 1:
-119 # New version, to cope with ids, etc.
-120 ls.sort(key=lambda x: int(re.findall(r'\d+', x[len(prefix):])[0]))
-121 replica = len(ls)
-122 else:
-123 replica = len([file.split(".")[-1] for file in ls]) // len(set([file.split(".")[-1] for file in ls]))
-124 print('Read', part, 'part of', name, 'from', prefix[:-1],
-125 ',', replica, 'replica')
-126 if 'names' in kwargs:
-127 new_names = kwargs.get('names')
-128 if len(new_names) != len(set(new_names)):
-129 raise Exception("names are not unique!")
-130 if len(new_names) != replica:
-131 raise Exception('Names does not have the required length', replica)
-132 else:
-133 new_names = []
-134 if not appended:
-135 for entry in ls:
-136 try:
-137 idx = entry.index('r')
-138 except Exception:
-139 raise Exception("Automatic recognition of replicum failed, please enter the key word 'names'.")
-140
-141 if 'ens_name' in kwargs:
-142 new_names.append(kwargs.get('ens_name') + '|' + entry[idx:])
-143 else:
-144 new_names.append(entry[:idx] + '|' + entry[idx:])
-145 else:
+ 62
+ 63 Returns
+ 64 -------
+ 65 result: list[Obs]
+ 66 list of Observables with length T, observable per timeslice.
+ 67 bb-type correlators have length 1.
+ 68 """
+ 69 if kwargs.get('im'):
+ 70 im = 1
+ 71 part = 'imaginary'
+ 72 else:
+ 73 im = 0
+ 74 part = 'real'
+ 75 if "replica" in kwargs:
+ 76 reps = kwargs.get("replica")
+ 77 if corr_type == 'bb':
+ 78 b2b = True
+ 79 single = True
+ 80 elif corr_type == 'bib':
+ 81 b2b = True
+ 82 single = False
+ 83 else:
+ 84 b2b = False
+ 85 single = False
+ 86 compact = True
+ 87 appended = False
+ 88 known_versions = ["0.0", "1.0", "2.0", "1.0c", "2.0c", "1.0a", "2.0a"]
+ 89
+ 90 if version not in known_versions:
+ 91 raise Exception("This version is not known!")
+ 92 if (version[-1] == "c"):
+ 93 appended = False
+ 94 compact = True
+ 95 version = version[:-1]
+ 96 elif (version[-1] == "a"):
+ 97 appended = True
+ 98 compact = False
+ 99 version = version[:-1]
+100 else:
+101 compact = False
+102 appended = False
+103 read = 0
+104 T = 0
+105 start = 0
+106 ls = []
+107 if "replica" in kwargs:
+108 ls = reps
+109 else:
+110 for (dirpath, dirnames, filenames) in os.walk(path):
+111 if not appended:
+112 ls.extend(dirnames)
+113 else:
+114 ls.extend(filenames)
+115 break
+116 if not ls:
+117 raise Exception('Error, directory not found')
+118 # Exclude folders with different names
+119 for exc in ls:
+120 if not fnmatch.fnmatch(exc, prefix + '*'):
+121 ls = list(set(ls) - set([exc]))
+122
+123 if not appended:
+124 if len(ls) > 1:
+125 # New version, to cope with ids, etc.
+126 ls.sort(key=lambda x: int(re.findall(r'\d+', x[len(prefix):])[0]))
+127 replica = len(ls)
+128 else:
+129 replica = len([file.split(".")[-1] for file in ls]) // len(set([file.split(".")[-1] for file in ls]))
+130 print('Read', part, 'part of', name, 'from', prefix[:-1],
+131 ',', replica, 'replica')
+132 if 'names' in kwargs:
+133 new_names = kwargs.get('names')
+134 if len(new_names) != len(set(new_names)):
+135 raise Exception("names are not unique!")
+136 if len(new_names) != replica:
+137 raise Exception('Names does not have the required length', replica)
+138 else:
+139 new_names = []
+140 if not appended:
+141 for entry in ls:
+142 try:
+143 idx = entry.index('r')
+144 except Exception:
+145 raise Exception("Automatic recognition of replicum failed, please enter the key word 'names'.")
146
-147 for exc in ls:
-148 if not fnmatch.fnmatch(exc, prefix + '*.' + name):
-149 ls = list(set(ls) - set([exc]))
-150 ls.sort(key=lambda x: int(re.findall(r'\d+', x)[-1]))
-151 for entry in ls:
-152 myentry = entry[:-len(name) - 1]
-153 try:
-154 idx = myentry.index('r')
-155 except Exception:
-156 raise Exception("Automatic recognition of replicum failed, please enter the key word 'names'.")
-157
-158 if 'ens_name' in kwargs:
-159 new_names.append(kwargs.get('ens_name') + '|' + myentry[idx:])
-160 else:
-161 new_names.append(myentry[:idx] + '|' + myentry[idx:])
-162 idl = []
-163 if not appended:
-164 for i, item in enumerate(ls):
-165 sub_ls = []
-166 if "files" in kwargs:
-167 sub_ls = kwargs.get("files")
-168 sub_ls.sort(key=lambda x: int(re.findall(r'\d+', x)[-1]))
-169 else:
-170 for (dirpath, dirnames, filenames) in os.walk(path + '/' + item):
-171 if compact:
-172 sub_ls.extend(filenames)
-173 else:
-174 sub_ls.extend(dirnames)
-175 break
-176 if compact:
-177 for exc in sub_ls:
-178 if not fnmatch.fnmatch(exc, prefix + '*'):
-179 sub_ls = list(set(sub_ls) - set([exc]))
-180 sub_ls.sort(key=lambda x: int(re.findall(r'\d+', x)[-1]))
-181 else:
-182 for exc in sub_ls:
-183 if not fnmatch.fnmatch(exc, 'cfg*'):
-184 sub_ls = list(set(sub_ls) - set([exc]))
-185 sub_ls.sort(key=lambda x: int(x[3:]))
-186 rep_idl = []
-187 no_cfg = len(sub_ls)
-188 for cfg in sub_ls:
-189 try:
-190 if compact:
-191 rep_idl.append(int(cfg.split(cfg_separator)[-1]))
-192 else:
-193 rep_idl.append(int(cfg[3:]))
-194 except Exception:
-195 raise Exception("Couldn't parse idl from directroy, problem with file " + cfg)
-196 rep_idl.sort()
-197 # maybe there is a better way to print the idls
-198 print(item, ':', no_cfg, ' configurations')
-199 idl.append(rep_idl)
-200 # here we have found all the files we need to look into.
-201 if i == 0:
-202 # here, we want to find the place within the file,
-203 # where the correlator we need is stored.
-204 # to do so, the pattern needed is put together
-205 # from the input values
-206 if version == "0.0":
-207 pattern = "# " + name + " : offset " + str(noffset) + ", wf " + str(wf)
-208 # if b2b, a second wf is needed
-209 if b2b:
-210 pattern += ", wf_2 " + str(wf2)
-211 qs = quarks.split(" ")
-212 pattern += " : " + qs[0] + " - " + qs[1]
-213 file = open(path + '/' + item + '/' + sub_ls[0] + '/' + name, "r")
-214 for k, line in enumerate(file):
-215 if read == 1 and not line.strip() and k > start + 1:
-216 break
-217 if read == 1 and k >= start:
-218 T += 1
-219 if pattern in line:
-220 read = 1
-221 start = k + 1
-222 print(str(T) + " entries found.")
-223 file.close()
-224 else:
-225 pattern = 'name ' + name + '\nquarks ' + quarks + '\noffset ' + str(noffset) + '\nwf ' + str(wf)
-226 if b2b:
-227 pattern += '\nwf_2 ' + str(wf2)
-228 # and the file is parsed through to find the pattern
-229 if compact:
-230 file = open(path + '/' + item + '/' + sub_ls[0], "r")
-231 else:
-232 # for non-compactified versions of the files
-233 file = open(path + '/' + item + '/' + sub_ls[0] + '/' + name, "r")
-234
-235 content = file.read()
-236 match = re.search(pattern, content)
-237 if match:
-238 start_read = content.count('\n', 0, match.start()) + 5 + b2b
-239 end_match = re.search(r'\n\s*\n', content[match.start():])
-240 T = content[match.start():].count('\n', 0, end_match.start()) - 4 - b2b
-241 assert T > 0
-242 print(T, 'entries, starting to read in line', start_read)
-243 file.close()
-244 else:
-245 file.close()
-246 raise Exception('Correlator with pattern\n' + pattern + '\nnot found.')
-247
-248 # we found where the correlator
-249 # that is to be read is in the files
-250 # after preparing the datastructure
-251 # the correlators get parsed into...
-252 deltas = []
-253 for j in range(T):
-254 deltas.append([])
-255
-256 for t in range(T):
-257 deltas[t].append(np.zeros(no_cfg))
-258 if compact:
-259 for cfg in range(no_cfg):
-260 with open(path + '/' + item + '/' + sub_ls[cfg]) as fp:
-261 lines = fp.readlines()
-262 # check, if the correlator is in fact
-263 # printed completely
-264 if (start_read + T > len(lines)):
-265 raise Exception("EOF before end of correlator data! Maybe " + path + '/' + item + '/' + sub_ls[cfg] + " is corrupted?")
-266 # and start to read the correlator.
-267 # the range here is chosen like this,
-268 # since this allows for implementing
-269 # a security check for every read correlator later...
-270 for k in range(start_read - 6, start_read + T):
-271 if k == start_read - 5 - b2b:
-272 if lines[k].strip() != 'name ' + name:
-273 raise Exception('Wrong format', sub_ls[cfg])
-274 if (k >= start_read and k < start_read + T):
-275 floats = list(map(float, lines[k].split()))
-276 deltas[k - start_read][i][cfg] = floats[-2:][im]
-277 else:
-278 for cnfg, subitem in enumerate(sub_ls):
-279 with open(path + '/' + item + '/' + subitem + '/' + name) as fp:
-280 # since the non-compatified files
-281 # are typically not so long,
-282 # we can iterate over the whole file.
-283 # here one can also implement the chekc from above.
-284 for k, line in enumerate(fp):
-285 if (k >= start_read and k < start_read + T):
-286 floats = list(map(float, line.split()))
-287 if version == "0.0":
-288 deltas[k - start][i][cnfg] = floats[im - single]
-289 else:
-290 deltas[k - start_read][i][cnfg] = floats[1 + im - single]
-291
-292 else:
-293 if "files" in kwargs:
-294 ls = kwargs.get("files")
-295 else:
-296 for exc in ls:
-297 if not fnmatch.fnmatch(exc, prefix + '*.' + name):
-298 ls = list(set(ls) - set([exc]))
-299 ls.sort(key=lambda x: int(re.findall(r'\d+', x)[-1]))
-300 pattern = 'name ' + name + '\nquarks ' + quarks + '\noffset ' + str(noffset) + '\nwf ' + str(wf)
-301 if b2b:
-302 pattern += '\nwf_2 ' + str(wf2)
-303 for rep, file in enumerate(ls):
-304 rep_idl = []
-305 with open(path + '/' + file, 'r') as fp:
-306 content = fp.readlines()
-307 data_starts = []
-308 for linenumber, line in enumerate(content):
-309 if "[run]" in line:
-310 data_starts.append(linenumber)
-311 if len(set([data_starts[i] - data_starts[i - 1] for i in range(1, len(data_starts))])) > 1:
-312 raise Exception("Irregularities in file structure found, not all runs have the same output length")
-313 chunk = content[:data_starts[1]]
-314 for linenumber, line in enumerate(chunk):
-315 if line.startswith("gauge_name"):
-316 gauge_line = linenumber
-317 elif line.startswith("[correlator]"):
-318 corr_line = linenumber
-319 found_pat = ""
-320 for li in chunk[corr_line + 1: corr_line + 6 + b2b]:
-321 found_pat += li
-322 if re.search(pattern, found_pat):
-323 start_read = corr_line + 7 + b2b
-324 break
-325 endline = corr_line + 6 + b2b
-326 while not chunk[endline] == "\n":
-327 endline += 1
-328 T = endline - start_read
-329 if rep == 0:
-330 deltas = []
-331 for t in range(T):
-332 deltas.append([])
-333 for t in range(T):
-334 deltas[t].append(np.zeros(len(data_starts)))
-335 # all other chunks should follow the same structure
-336 for cnfg in range(len(data_starts)):
-337 start = data_starts[cnfg]
-338 stop = start + data_starts[1]
-339 chunk = content[start:stop]
-340 try:
-341 rep_idl.append(int(chunk[gauge_line].split(cfg_separator)[-1]))
-342 except Exception:
-343 raise Exception("Couldn't parse idl from directory, problem with chunk around line ", gauge_line)
-344
-345 found_pat = ""
-346 for li in chunk[corr_line + 1:corr_line + 6 + b2b]:
-347 found_pat += li
-348 if re.search(pattern, found_pat):
-349 for t, line in enumerate(chunk[start_read:start_read + T]):
-350 floats = list(map(float, line.split()))
-351 deltas[t][rep][cnfg] = floats[im + 1 - single]
-352 idl.append(rep_idl)
-353
-354 if "check_configs" in kwargs:
-355 print("Checking for missing configs...")
-356 che = kwargs.get("check_configs")
-357 if not (len(che) == len(idl)):
-358 raise Exception("check_configs has to be the same length as replica!")
-359 for r in range(len(idl)):
-360 print("checking " + new_names[r])
-361 utils.check_idl(idl[r], che[r])
-362 print("Done")
-363 result = []
-364 for t in range(T):
-365 result.append(Obs(deltas[t], new_names, idl=idl))
-366 return result
+147 if 'ens_name' in kwargs:
+148 new_names.append(kwargs.get('ens_name') + '|' + entry[idx:])
+149 else:
+150 new_names.append(entry[:idx] + '|' + entry[idx:])
+151 else:
+152
+153 for exc in ls:
+154 if not fnmatch.fnmatch(exc, prefix + '*.' + name):
+155 ls = list(set(ls) - set([exc]))
+156 ls.sort(key=lambda x: int(re.findall(r'\d+', x)[-1]))
+157 for entry in ls:
+158 myentry = entry[:-len(name) - 1]
+159 try:
+160 idx = myentry.index('r')
+161 except Exception:
+162 raise Exception("Automatic recognition of replicum failed, please enter the key word 'names'.")
+163
+164 if 'ens_name' in kwargs:
+165 new_names.append(kwargs.get('ens_name') + '|' + myentry[idx:])
+166 else:
+167 new_names.append(myentry[:idx] + '|' + myentry[idx:])
+168 idl = []
+169 if not appended:
+170 for i, item in enumerate(ls):
+171 sub_ls = []
+172 if "files" in kwargs:
+173 sub_ls = kwargs.get("files")
+174 sub_ls.sort(key=lambda x: int(re.findall(r'\d+', x)[-1]))
+175 else:
+176 for (dirpath, dirnames, filenames) in os.walk(path + '/' + item):
+177 if compact:
+178 sub_ls.extend(filenames)
+179 else:
+180 sub_ls.extend(dirnames)
+181 break
+182 if compact:
+183 for exc in sub_ls:
+184 if not fnmatch.fnmatch(exc, prefix + '*'):
+185 sub_ls = list(set(sub_ls) - set([exc]))
+186 sub_ls.sort(key=lambda x: int(re.findall(r'\d+', x)[-1]))
+187 else:
+188 for exc in sub_ls:
+189 if not fnmatch.fnmatch(exc, 'cfg*'):
+190 sub_ls = list(set(sub_ls) - set([exc]))
+191 sub_ls.sort(key=lambda x: int(x[3:]))
+192 rep_idl = []
+193 no_cfg = len(sub_ls)
+194 for cfg in sub_ls:
+195 try:
+196 if compact:
+197 rep_idl.append(int(cfg.split(cfg_separator)[-1]))
+198 else:
+199 rep_idl.append(int(cfg[3:]))
+200 except Exception:
+201 raise Exception("Couldn't parse idl from directroy, problem with file " + cfg)
+202 rep_idl.sort()
+203 # maybe there is a better way to print the idls
+204 print(item, ':', no_cfg, ' configurations')
+205 idl.append(rep_idl)
+206 # here we have found all the files we need to look into.
+207 if i == 0:
+208 # here, we want to find the place within the file,
+209 # where the correlator we need is stored.
+210 # to do so, the pattern needed is put together
+211 # from the input values
+212 if version == "0.0":
+213 pattern = "# " + name + " : offset " + str(noffset) + ", wf " + str(wf)
+214 # if b2b, a second wf is needed
+215 if b2b:
+216 pattern += ", wf_2 " + str(wf2)
+217 qs = quarks.split(" ")
+218 pattern += " : " + qs[0] + " - " + qs[1]
+219 file = open(path + '/' + item + '/' + sub_ls[0] + '/' + name, "r")
+220 for k, line in enumerate(file):
+221 if read == 1 and not line.strip() and k > start + 1:
+222 break
+223 if read == 1 and k >= start:
+224 T += 1
+225 if pattern in line:
+226 read = 1
+227 start = k + 1
+228 print(str(T) + " entries found.")
+229 file.close()
+230 else:
+231 pattern = 'name ' + name + '\nquarks ' + quarks + '\noffset ' + str(noffset) + '\nwf ' + str(wf)
+232 if b2b:
+233 pattern += '\nwf_2 ' + str(wf2)
+234 # and the file is parsed through to find the pattern
+235 if compact:
+236 file = open(path + '/' + item + '/' + sub_ls[0], "r")
+237 else:
+238 # for non-compactified versions of the files
+239 file = open(path + '/' + item + '/' + sub_ls[0] + '/' + name, "r")
+240
+241 content = file.read()
+242 match = re.search(pattern, content)
+243 if match:
+244 start_read = content.count('\n', 0, match.start()) + 5 + b2b
+245 end_match = re.search(r'\n\s*\n', content[match.start():])
+246 T = content[match.start():].count('\n', 0, end_match.start()) - 4 - b2b
+247 assert T > 0
+248 print(T, 'entries, starting to read in line', start_read)
+249 file.close()
+250 else:
+251 file.close()
+252 raise Exception('Correlator with pattern\n' + pattern + '\nnot found.')
+253
+254 # we found where the correlator
+255 # that is to be read is in the files
+256 # after preparing the datastructure
+257 # the correlators get parsed into...
+258 deltas = []
+259 for j in range(T):
+260 deltas.append([])
+261
+262 for t in range(T):
+263 deltas[t].append(np.zeros(no_cfg))
+264 if compact:
+265 for cfg in range(no_cfg):
+266 with open(path + '/' + item + '/' + sub_ls[cfg]) as fp:
+267 lines = fp.readlines()
+268 # check, if the correlator is in fact
+269 # printed completely
+270 if (start_read + T > len(lines)):
+271 raise Exception("EOF before end of correlator data! Maybe " + path + '/' + item + '/' + sub_ls[cfg] + " is corrupted?")
+272 # and start to read the correlator.
+273 # the range here is chosen like this,
+274 # since this allows for implementing
+275 # a security check for every read correlator later...
+276 for k in range(start_read - 6, start_read + T):
+277 if k == start_read - 5 - b2b:
+278 if lines[k].strip() != 'name ' + name:
+279 raise Exception('Wrong format', sub_ls[cfg])
+280 if (k >= start_read and k < start_read + T):
+281 floats = list(map(float, lines[k].split()))
+282 deltas[k - start_read][i][cfg] = floats[-2:][im]
+283 else:
+284 for cnfg, subitem in enumerate(sub_ls):
+285 with open(path + '/' + item + '/' + subitem + '/' + name) as fp:
+286 # since the non-compatified files
+287 # are typically not so long,
+288 # we can iterate over the whole file.
+289 # here one can also implement the chekc from above.
+290 for k, line in enumerate(fp):
+291 if (k >= start_read and k < start_read + T):
+292 floats = list(map(float, line.split()))
+293 if version == "0.0":
+294 deltas[k - start][i][cnfg] = floats[im - single]
+295 else:
+296 deltas[k - start_read][i][cnfg] = floats[1 + im - single]
+297
+298 else:
+299 if "files" in kwargs:
+300 ls = kwargs.get("files")
+301 else:
+302 for exc in ls:
+303 if not fnmatch.fnmatch(exc, prefix + '*.' + name):
+304 ls = list(set(ls) - set([exc]))
+305 ls.sort(key=lambda x: int(re.findall(r'\d+', x)[-1]))
+306 pattern = 'name ' + name + '\nquarks ' + quarks + '\noffset ' + str(noffset) + '\nwf ' + str(wf)
+307 if b2b:
+308 pattern += '\nwf_2 ' + str(wf2)
+309 for rep, file in enumerate(ls):
+310 rep_idl = []
+311 with open(path + '/' + file, 'r') as fp:
+312 content = fp.readlines()
+313 data_starts = []
+314 for linenumber, line in enumerate(content):
+315 if "[run]" in line:
+316 data_starts.append(linenumber)
+317 if len(set([data_starts[i] - data_starts[i - 1] for i in range(1, len(data_starts))])) > 1:
+318 raise Exception("Irregularities in file structure found, not all runs have the same output length")
+319 chunk = content[:data_starts[1]]
+320 for linenumber, line in enumerate(chunk):
+321 if line.startswith("gauge_name"):
+322 gauge_line = linenumber
+323 elif line.startswith("[correlator]"):
+324 corr_line = linenumber
+325 found_pat = ""
+326 for li in chunk[corr_line + 1: corr_line + 6 + b2b]:
+327 found_pat += li
+328 if re.search(pattern, found_pat):
+329 start_read = corr_line + 7 + b2b
+330 break
+331 endline = corr_line + 6 + b2b
+332 while not chunk[endline] == "\n":
+333 endline += 1
+334 T = endline - start_read
+335 if rep == 0:
+336 deltas = []
+337 for t in range(T):
+338 deltas.append([])
+339 for t in range(T):
+340 deltas[t].append(np.zeros(len(data_starts)))
+341 # all other chunks should follow the same structure
+342 for cnfg in range(len(data_starts)):
+343 start = data_starts[cnfg]
+344 stop = start + data_starts[1]
+345 chunk = content[start:stop]
+346 try:
+347 rep_idl.append(int(chunk[gauge_line].split(cfg_separator)[-1]))
+348 except Exception:
+349 raise Exception("Couldn't parse idl from directory, problem with chunk around line ", gauge_line)
+350
+351 found_pat = ""
+352 for li in chunk[corr_line + 1:corr_line + 6 + b2b]:
+353 found_pat += li
+354 if re.search(pattern, found_pat):
+355 for t, line in enumerate(chunk[start_read:start_read + T]):
+356 floats = list(map(float, line.split()))
+357 deltas[t][rep][cnfg] = floats[im + 1 - single]
+358 idl.append(rep_idl)
+359
+360 if "check_configs" in kwargs:
+361 print("Checking for missing configs...")
+362 che = kwargs.get("check_configs")
+363 if not (len(che) == len(idl)):
+364 raise Exception("check_configs has to be the same length as replica!")
+365 for r in range(len(idl)):
+366 print("checking " + new_names[r])
+367 utils.check_idl(idl[r], che[r])
+368 print("Done")
+369 result = []
+370 for t in range(T):
+371 result.append(Obs(deltas[t], new_names, idl=idl))
+372 return result
@@ -867,9 +879,18 @@ list of replica to be read, default is all
files (list):
list of files to be read per replica, default is all.
for non-compact output format, hand the folders to be read here.
-check_configs:: list of list of supposed configs, eg. [range(1,1000)]
+check_configs (list[list[int]]):
+list of list of supposed configs, eg. [range(1,1000)]
for one replicum with 1000 configs
+
+Returns
+
+
+- result (list[Obs]):
+list of Observables with length T, observable per timeslice.
+bb-type correlators have length 1.
+
diff --git a/docs/pyerrors/input/utils.html b/docs/pyerrors/input/utils.html
index a6b742d0..18e4bca3 100644
--- a/docs/pyerrors/input/utils.html
+++ b/docs/pyerrors/input/utils.html
@@ -90,18 +90,24 @@
10 idl of the current replicum
11 che : list
12 list of configurations to be checked against
-13 """
-14 missing = []
-15 for c in che:
-16 if c not in idl:
-17 missing.append(c)
-18 # print missing configurations such that it can directly be parsed to slurm terminal
-19 if not (len(missing) == 0):
-20 print(len(missing), "configs missing")
-21 miss_str = str(missing[0])
-22 for i in missing[1:]:
-23 miss_str += "," + str(i)
-24 print(miss_str)
+13
+14 Returns
+15 -------
+16 miss_str : str
+17 string with integers of which idls are missing
+18 """
+19 missing = []
+20 for c in che:
+21 if c not in idl:
+22 missing.append(c)
+23 # print missing configurations such that it can directly be parsed to slurm terminal
+24 if not (len(missing) == 0):
+25 print(len(missing), "configs missing")
+26 miss_str = str(missing[0])
+27 for i in missing[1:]:
+28 miss_str += "," + str(i)
+29 print(miss_str)
+30 return miss_str
@@ -126,18 +132,24 @@
11 idl of the current replicum
12 che : list
13 list of configurations to be checked against
-14 """
-15 missing = []
-16 for c in che:
-17 if c not in idl:
-18 missing.append(c)
-19 # print missing configurations such that it can directly be parsed to slurm terminal
-20 if not (len(missing) == 0):
-21 print(len(missing), "configs missing")
-22 miss_str = str(missing[0])
-23 for i in missing[1:]:
-24 miss_str += "," + str(i)
-25 print(miss_str)
+14
+15 Returns
+16 -------
+17 miss_str : str
+18 string with integers of which idls are missing
+19 """
+20 missing = []
+21 for c in che:
+22 if c not in idl:
+23 missing.append(c)
+24 # print missing configurations such that it can directly be parsed to slurm terminal
+25 if not (len(missing) == 0):
+26 print(len(missing), "configs missing")
+27 miss_str = str(missing[0])
+28 for i in missing[1:]:
+29 miss_str += "," + str(i)
+30 print(miss_str)
+31 return miss_str
@@ -151,6 +163,13 @@ idl of the current replicum
che (list):
list of configurations to be checked against
+
+Returns
+
+
+- miss_str (str):
+string with integers of which idls are missing
+
diff --git a/docs/pyerrors/misc.html b/docs/pyerrors/misc.html
index 444e13fc..0cab6773 100644
--- a/docs/pyerrors/misc.html
+++ b/docs/pyerrors/misc.html
@@ -101,105 +101,124 @@
14 name of the file
15 path : str
16 specifies a custom path for the file (default '.')
- 17 """
- 18 if 'path' in kwargs:
- 19 file_name = kwargs.get('path') + '/' + name + '.p'
- 20 else:
- 21 file_name = name + '.p'
- 22 with open(file_name, 'wb') as fb:
- 23 pickle.dump(obj, fb)
- 24
- 25
- 26def load_object(path):
- 27 """Load object from pickle file.
+ 17
+ 18 Returns
+ 19 -------
+ 20 None
+ 21 """
+ 22 if 'path' in kwargs:
+ 23 file_name = kwargs.get('path') + '/' + name + '.p'
+ 24 else:
+ 25 file_name = name + '.p'
+ 26 with open(file_name, 'wb') as fb:
+ 27 pickle.dump(obj, fb)
28
- 29 Parameters
- 30 ----------
- 31 path : str
- 32 path to the file
- 33 """
- 34 with open(path, 'rb') as file:
- 35 return pickle.load(file)
- 36
+ 29
+ 30def load_object(path):
+ 31 """Load object from pickle file.
+ 32
+ 33 Parameters
+ 34 ----------
+ 35 path : str
+ 36 path to the file
37
- 38def pseudo_Obs(value, dvalue, name, samples=1000):
- 39 """Generate an Obs object with given value, dvalue and name for test purposes
- 40
- 41 Parameters
- 42 ----------
- 43 value : float
- 44 central value of the Obs to be generated.
- 45 dvalue : float
- 46 error of the Obs to be generated.
- 47 name : str
- 48 name of the ensemble for which the Obs is to be generated.
- 49 samples: int
- 50 number of samples for the Obs (default 1000).
- 51 """
- 52 if dvalue <= 0.0:
- 53 return Obs([np.zeros(samples) + value], [name])
- 54 else:
- 55 for _ in range(100):
- 56 deltas = [np.random.normal(0.0, dvalue * np.sqrt(samples), samples)]
- 57 deltas -= np.mean(deltas)
- 58 deltas *= dvalue / np.sqrt((np.var(deltas) / samples)) / np.sqrt(1 + 3 / samples)
- 59 deltas += value
- 60 res = Obs(deltas, [name])
- 61 res.gamma_method(S=2, tau_exp=0)
- 62 if abs(res.dvalue - dvalue) < 1e-10 * dvalue:
- 63 break
- 64
- 65 res._value = float(value)
- 66
- 67 return res
- 68
- 69
- 70def gen_correlated_data(means, cov, name, tau=0.5, samples=1000):
- 71 """ Generate observables with given covariance and autocorrelation times.
- 72
- 73 Parameters
- 74 ----------
- 75 means : list
- 76 list containing the mean value of each observable.
- 77 cov : numpy.ndarray
- 78 covariance matrix for the data to be generated.
- 79 name : str
- 80 ensemble name for the data to be geneated.
- 81 tau : float or list
- 82 can either be a real number or a list with an entry for
- 83 every dataset.
- 84 samples : int
- 85 number of samples to be generated for each observable.
- 86 """
- 87
- 88 assert len(means) == cov.shape[-1]
- 89 tau = np.asarray(tau)
- 90 if np.min(tau) < 0.5:
- 91 raise Exception('All integrated autocorrelations have to be >= 0.5.')
- 92
- 93 a = (2 * tau - 1) / (2 * tau + 1)
- 94 rand = np.random.multivariate_normal(np.zeros_like(means), cov * samples, samples)
- 95
- 96 # Normalize samples such that sample variance matches input
- 97 norm = np.array([np.var(o, ddof=1) / samples for o in rand.T])
- 98 rand = rand @ np.diag(np.sqrt(np.diag(cov))) @ np.diag(1 / np.sqrt(norm))
- 99
-100 data = [rand[0]]
-101 for i in range(1, samples):
-102 data.append(np.sqrt(1 - a ** 2) * rand[i] + a * data[-1])
-103 corr_data = np.array(data) - np.mean(data, axis=0) + means
-104 return [Obs([dat], [name]) for dat in corr_data.T]
-105
+ 38 Returns
+ 39 -------
+ 40 object : Obs
+ 41 Loaded Object
+ 42 """
+ 43 with open(path, 'rb') as file:
+ 44 return pickle.load(file)
+ 45
+ 46
+ 47def pseudo_Obs(value, dvalue, name, samples=1000):
+ 48 """Generate an Obs object with given value, dvalue and name for test purposes
+ 49
+ 50 Parameters
+ 51 ----------
+ 52 value : float
+ 53 central value of the Obs to be generated.
+ 54 dvalue : float
+ 55 error of the Obs to be generated.
+ 56 name : str
+ 57 name of the ensemble for which the Obs is to be generated.
+ 58 samples: int
+ 59 number of samples for the Obs (default 1000).
+ 60
+ 61 Returns
+ 62 -------
+ 63 res : Obs
+ 64 Generated Observable
+ 65 """
+ 66 if dvalue <= 0.0:
+ 67 return Obs([np.zeros(samples) + value], [name])
+ 68 else:
+ 69 for _ in range(100):
+ 70 deltas = [np.random.normal(0.0, dvalue * np.sqrt(samples), samples)]
+ 71 deltas -= np.mean(deltas)
+ 72 deltas *= dvalue / np.sqrt((np.var(deltas) / samples)) / np.sqrt(1 + 3 / samples)
+ 73 deltas += value
+ 74 res = Obs(deltas, [name])
+ 75 res.gamma_method(S=2, tau_exp=0)
+ 76 if abs(res.dvalue - dvalue) < 1e-10 * dvalue:
+ 77 break
+ 78
+ 79 res._value = float(value)
+ 80
+ 81 return res
+ 82
+ 83
+ 84def gen_correlated_data(means, cov, name, tau=0.5, samples=1000):
+ 85 """ Generate observables with given covariance and autocorrelation times.
+ 86
+ 87 Parameters
+ 88 ----------
+ 89 means : list
+ 90 list containing the mean value of each observable.
+ 91 cov : numpy.ndarray
+ 92 covariance matrix for the data to be generated.
+ 93 name : str
+ 94 ensemble name for the data to be geneated.
+ 95 tau : float or list
+ 96 can either be a real number or a list with an entry for
+ 97 every dataset.
+ 98 samples : int
+ 99 number of samples to be generated for each observable.
+100
+101 Returns
+102 -------
+103 corr_obs : list[Obs]
+104 Generated observable list
+105 """
106
-107def _assert_equal_properties(ol, otype=Obs):
-108 otype = type(ol[0])
-109 for o in ol[1:]:
-110 if not isinstance(o, otype):
-111 raise Exception("Wrong data type in list.")
-112 for attr in ["reweighted", "e_content", "idl"]:
-113 if hasattr(ol[0], attr):
-114 if not getattr(ol[0], attr) == getattr(o, attr):
-115 raise Exception(f"All Obs in list have to have the same state '{attr}'.")
+107 assert len(means) == cov.shape[-1]
+108 tau = np.asarray(tau)
+109 if np.min(tau) < 0.5:
+110 raise Exception('All integrated autocorrelations have to be >= 0.5.')
+111
+112 a = (2 * tau - 1) / (2 * tau + 1)
+113 rand = np.random.multivariate_normal(np.zeros_like(means), cov * samples, samples)
+114
+115 # Normalize samples such that sample variance matches input
+116 norm = np.array([np.var(o, ddof=1) / samples for o in rand.T])
+117 rand = rand @ np.diag(np.sqrt(np.diag(cov))) @ np.diag(1 / np.sqrt(norm))
+118
+119 data = [rand[0]]
+120 for i in range(1, samples):
+121 data.append(np.sqrt(1 - a ** 2) * rand[i] + a * data[-1])
+122 corr_data = np.array(data) - np.mean(data, axis=0) + means
+123 return [Obs([dat], [name]) for dat in corr_data.T]
+124
+125
+126def _assert_equal_properties(ol, otype=Obs):
+127 otype = type(ol[0])
+128 for o in ol[1:]:
+129 if not isinstance(o, otype):
+130 raise Exception("Wrong data type in list.")
+131 for attr in ["reweighted", "e_content", "idl"]:
+132 if hasattr(ol[0], attr):
+133 if not getattr(ol[0], attr) == getattr(o, attr):
+134 raise Exception(f"All Obs in list have to have the same state '{attr}'.")
@@ -226,13 +245,17 @@
15 name of the file
16 path : str
17 specifies a custom path for the file (default '.')
-18 """
-19 if 'path' in kwargs:
-20 file_name = kwargs.get('path') + '/' + name + '.p'
-21 else:
-22 file_name = name + '.p'
-23 with open(file_name, 'wb') as fb:
-24 pickle.dump(obj, fb)
+18
+19 Returns
+20 -------
+21 None
+22 """
+23 if 'path' in kwargs:
+24 file_name = kwargs.get('path') + '/' + name + '.p'
+25 else:
+26 file_name = name + '.p'
+27 with open(file_name, 'wb') as fb:
+28 pickle.dump(obj, fb)
@@ -248,6 +271,12 @@ name of the file
path (str):
specifies a custom path for the file (default '.')
+
+Returns
+
+
@@ -263,16 +292,21 @@ specifies a custom path for the file (default '.')
- 27def load_object(path):
-28 """Load object from pickle file.
-29
-30 Parameters
-31 ----------
-32 path : str
-33 path to the file
-34 """
-35 with open(path, 'rb') as file:
-36 return pickle.load(file)
+ 31def load_object(path):
+32 """Load object from pickle file.
+33
+34 Parameters
+35 ----------
+36 path : str
+37 path to the file
+38
+39 Returns
+40 -------
+41 object : Obs
+42 Loaded Object
+43 """
+44 with open(path, 'rb') as file:
+45 return pickle.load(file)
@@ -284,6 +318,13 @@ specifies a custom path for the file (default '.')
path (str):
path to the file
+
+Returns
+
+
+- object (Obs):
+Loaded Object
+
@@ -299,36 +340,41 @@ path to the file
- 39def pseudo_Obs(value, dvalue, name, samples=1000):
-40 """Generate an Obs object with given value, dvalue and name for test purposes
-41
-42 Parameters
-43 ----------
-44 value : float
-45 central value of the Obs to be generated.
-46 dvalue : float
-47 error of the Obs to be generated.
-48 name : str
-49 name of the ensemble for which the Obs is to be generated.
-50 samples: int
-51 number of samples for the Obs (default 1000).
-52 """
-53 if dvalue <= 0.0:
-54 return Obs([np.zeros(samples) + value], [name])
-55 else:
-56 for _ in range(100):
-57 deltas = [np.random.normal(0.0, dvalue * np.sqrt(samples), samples)]
-58 deltas -= np.mean(deltas)
-59 deltas *= dvalue / np.sqrt((np.var(deltas) / samples)) / np.sqrt(1 + 3 / samples)
-60 deltas += value
-61 res = Obs(deltas, [name])
-62 res.gamma_method(S=2, tau_exp=0)
-63 if abs(res.dvalue - dvalue) < 1e-10 * dvalue:
-64 break
-65
-66 res._value = float(value)
-67
-68 return res
+ 48def pseudo_Obs(value, dvalue, name, samples=1000):
+49 """Generate an Obs object with given value, dvalue and name for test purposes
+50
+51 Parameters
+52 ----------
+53 value : float
+54 central value of the Obs to be generated.
+55 dvalue : float
+56 error of the Obs to be generated.
+57 name : str
+58 name of the ensemble for which the Obs is to be generated.
+59 samples: int
+60 number of samples for the Obs (default 1000).
+61
+62 Returns
+63 -------
+64 res : Obs
+65 Generated Observable
+66 """
+67 if dvalue <= 0.0:
+68 return Obs([np.zeros(samples) + value], [name])
+69 else:
+70 for _ in range(100):
+71 deltas = [np.random.normal(0.0, dvalue * np.sqrt(samples), samples)]
+72 deltas -= np.mean(deltas)
+73 deltas *= dvalue / np.sqrt((np.var(deltas) / samples)) / np.sqrt(1 + 3 / samples)
+74 deltas += value
+75 res = Obs(deltas, [name])
+76 res.gamma_method(S=2, tau_exp=0)
+77 if abs(res.dvalue - dvalue) < 1e-10 * dvalue:
+78 break
+79
+80 res._value = float(value)
+81
+82 return res
@@ -346,6 +392,13 @@ name of the ensemble for which the Obs is to be generated.
samples (int):
number of samples for the Obs (default 1000).
+
+Returns
+
+
+- res (Obs):
+Generated Observable
+
@@ -361,41 +414,46 @@ number of samples for the Obs (default 1000).
- 71def gen_correlated_data(means, cov, name, tau=0.5, samples=1000):
- 72 """ Generate observables with given covariance and autocorrelation times.
- 73
- 74 Parameters
- 75 ----------
- 76 means : list
- 77 list containing the mean value of each observable.
- 78 cov : numpy.ndarray
- 79 covariance matrix for the data to be generated.
- 80 name : str
- 81 ensemble name for the data to be geneated.
- 82 tau : float or list
- 83 can either be a real number or a list with an entry for
- 84 every dataset.
- 85 samples : int
- 86 number of samples to be generated for each observable.
- 87 """
- 88
- 89 assert len(means) == cov.shape[-1]
- 90 tau = np.asarray(tau)
- 91 if np.min(tau) < 0.5:
- 92 raise Exception('All integrated autocorrelations have to be >= 0.5.')
- 93
- 94 a = (2 * tau - 1) / (2 * tau + 1)
- 95 rand = np.random.multivariate_normal(np.zeros_like(means), cov * samples, samples)
- 96
- 97 # Normalize samples such that sample variance matches input
- 98 norm = np.array([np.var(o, ddof=1) / samples for o in rand.T])
- 99 rand = rand @ np.diag(np.sqrt(np.diag(cov))) @ np.diag(1 / np.sqrt(norm))
-100
-101 data = [rand[0]]
-102 for i in range(1, samples):
-103 data.append(np.sqrt(1 - a ** 2) * rand[i] + a * data[-1])
-104 corr_data = np.array(data) - np.mean(data, axis=0) + means
-105 return [Obs([dat], [name]) for dat in corr_data.T]
+ 85def gen_correlated_data(means, cov, name, tau=0.5, samples=1000):
+ 86 """ Generate observables with given covariance and autocorrelation times.
+ 87
+ 88 Parameters
+ 89 ----------
+ 90 means : list
+ 91 list containing the mean value of each observable.
+ 92 cov : numpy.ndarray
+ 93 covariance matrix for the data to be generated.
+ 94 name : str
+ 95 ensemble name for the data to be geneated.
+ 96 tau : float or list
+ 97 can either be a real number or a list with an entry for
+ 98 every dataset.
+ 99 samples : int
+100 number of samples to be generated for each observable.
+101
+102 Returns
+103 -------
+104 corr_obs : list[Obs]
+105 Generated observable list
+106 """
+107
+108 assert len(means) == cov.shape[-1]
+109 tau = np.asarray(tau)
+110 if np.min(tau) < 0.5:
+111 raise Exception('All integrated autocorrelations have to be >= 0.5.')
+112
+113 a = (2 * tau - 1) / (2 * tau + 1)
+114 rand = np.random.multivariate_normal(np.zeros_like(means), cov * samples, samples)
+115
+116 # Normalize samples such that sample variance matches input
+117 norm = np.array([np.var(o, ddof=1) / samples for o in rand.T])
+118 rand = rand @ np.diag(np.sqrt(np.diag(cov))) @ np.diag(1 / np.sqrt(norm))
+119
+120 data = [rand[0]]
+121 for i in range(1, samples):
+122 data.append(np.sqrt(1 - a ** 2) * rand[i] + a * data[-1])
+123 corr_data = np.array(data) - np.mean(data, axis=0) + means
+124 return [Obs([dat], [name]) for dat in corr_data.T]
@@ -416,6 +474,13 @@ every dataset.
samples (int):
number of samples to be generated for each observable.
+
+Returns
+
+
+- corr_obs (list[Obs]):
+Generated observable list
+
diff --git a/docs/pyerrors/mpm.html b/docs/pyerrors/mpm.html
index 4063718f..43f00ae5 100644
--- a/docs/pyerrors/mpm.html
+++ b/docs/pyerrors/mpm.html
@@ -99,41 +99,46 @@
21 matrix pencil parameter which filters noise. The optimal value is expected between
22 len(data)/3 and 2*len(data)/3. The computation is more expensive the closer p is
23 to len(data)/2 but could possibly suppress more noise (default len(data)//2).
-24 """
-25 if isinstance(corrs[0], Obs):
-26 data = [corrs]
-27 else:
-28 data = corrs
-29
-30 lengths = [len(d) for d in data]
-31 if lengths.count(lengths[0]) != len(lengths):
-32 raise Exception('All datasets have to have the same length.')
-33
-34 data_sets = len(data)
-35 n_data = len(data[0])
-36
-37 if p is None:
-38 p = max(n_data // 2, k)
-39 if n_data <= p:
-40 raise Exception('The pencil p has to be smaller than the number of data samples.')
-41 if p < k or n_data - p < k:
-42 raise Exception('Cannot extract', k, 'energy levels with p=', p, 'and N-p=', n_data - p)
-43
-44 # Construct the hankel matrices
-45 matrix = []
-46 for n in range(data_sets):
-47 matrix.append(scipy.linalg.hankel(data[n][:n_data - p], data[n][n_data - p - 1:]))
-48 matrix = np.array(matrix)
-49 # Construct y1 and y2
-50 y1 = np.concatenate(matrix[:, :, :p])
-51 y2 = np.concatenate(matrix[:, :, 1:])
-52 # Apply SVD to y2
-53 u, s, vh = svd(y2, **kwargs)
-54 # Construct z from y1 and SVD of y2, setting all singular values beyond the kth to zero
-55 z = np.diag(1. / s[:k]) @ u[:, :k].T @ y1 @ vh.T[:, :k]
-56 # Return the sorted logarithms of the real eigenvalues as Obs
-57 energy_levels = np.log(np.abs(eig(z, **kwargs)))
-58 return sorted(energy_levels, key=lambda x: abs(x.value))
+24
+25 Returns
+26 -------
+27 energy_levels : list[Obs]
+28 Extracted energy levels
+29 """
+30 if isinstance(corrs[0], Obs):
+31 data = [corrs]
+32 else:
+33 data = corrs
+34
+35 lengths = [len(d) for d in data]
+36 if lengths.count(lengths[0]) != len(lengths):
+37 raise Exception('All datasets have to have the same length.')
+38
+39 data_sets = len(data)
+40 n_data = len(data[0])
+41
+42 if p is None:
+43 p = max(n_data // 2, k)
+44 if n_data <= p:
+45 raise Exception('The pencil p has to be smaller than the number of data samples.')
+46 if p < k or n_data - p < k:
+47 raise Exception('Cannot extract', k, 'energy levels with p=', p, 'and N-p=', n_data - p)
+48
+49 # Construct the hankel matrices
+50 matrix = []
+51 for n in range(data_sets):
+52 matrix.append(scipy.linalg.hankel(data[n][:n_data - p], data[n][n_data - p - 1:]))
+53 matrix = np.array(matrix)
+54 # Construct y1 and y2
+55 y1 = np.concatenate(matrix[:, :, :p])
+56 y2 = np.concatenate(matrix[:, :, 1:])
+57 # Apply SVD to y2
+58 u, s, vh = svd(y2, **kwargs)
+59 # Construct z from y1 and SVD of y2, setting all singular values beyond the kth to zero
+60 z = np.diag(1. / s[:k]) @ u[:, :k].T @ y1 @ vh.T[:, :k]
+61 # Return the sorted logarithms of the real eigenvalues as Obs
+62 energy_levels = np.log(np.abs(eig(z, **kwargs)))
+63 return sorted(energy_levels, key=lambda x: abs(x.value))
@@ -166,41 +171,46 @@
22 matrix pencil parameter which filters noise. The optimal value is expected between
23 len(data)/3 and 2*len(data)/3. The computation is more expensive the closer p is
24 to len(data)/2 but could possibly suppress more noise (default len(data)//2).
-25 """
-26 if isinstance(corrs[0], Obs):
-27 data = [corrs]
-28 else:
-29 data = corrs
-30
-31 lengths = [len(d) for d in data]
-32 if lengths.count(lengths[0]) != len(lengths):
-33 raise Exception('All datasets have to have the same length.')
-34
-35 data_sets = len(data)
-36 n_data = len(data[0])
-37
-38 if p is None:
-39 p = max(n_data // 2, k)
-40 if n_data <= p:
-41 raise Exception('The pencil p has to be smaller than the number of data samples.')
-42 if p < k or n_data - p < k:
-43 raise Exception('Cannot extract', k, 'energy levels with p=', p, 'and N-p=', n_data - p)
-44
-45 # Construct the hankel matrices
-46 matrix = []
-47 for n in range(data_sets):
-48 matrix.append(scipy.linalg.hankel(data[n][:n_data - p], data[n][n_data - p - 1:]))
-49 matrix = np.array(matrix)
-50 # Construct y1 and y2
-51 y1 = np.concatenate(matrix[:, :, :p])
-52 y2 = np.concatenate(matrix[:, :, 1:])
-53 # Apply SVD to y2
-54 u, s, vh = svd(y2, **kwargs)
-55 # Construct z from y1 and SVD of y2, setting all singular values beyond the kth to zero
-56 z = np.diag(1. / s[:k]) @ u[:, :k].T @ y1 @ vh.T[:, :k]
-57 # Return the sorted logarithms of the real eigenvalues as Obs
-58 energy_levels = np.log(np.abs(eig(z, **kwargs)))
-59 return sorted(energy_levels, key=lambda x: abs(x.value))
+25
+26 Returns
+27 -------
+28 energy_levels : list[Obs]
+29 Extracted energy levels
+30 """
+31 if isinstance(corrs[0], Obs):
+32 data = [corrs]
+33 else:
+34 data = corrs
+35
+36 lengths = [len(d) for d in data]
+37 if lengths.count(lengths[0]) != len(lengths):
+38 raise Exception('All datasets have to have the same length.')
+39
+40 data_sets = len(data)
+41 n_data = len(data[0])
+42
+43 if p is None:
+44 p = max(n_data // 2, k)
+45 if n_data <= p:
+46 raise Exception('The pencil p has to be smaller than the number of data samples.')
+47 if p < k or n_data - p < k:
+48 raise Exception('Cannot extract', k, 'energy levels with p=', p, 'and N-p=', n_data - p)
+49
+50 # Construct the hankel matrices
+51 matrix = []
+52 for n in range(data_sets):
+53 matrix.append(scipy.linalg.hankel(data[n][:n_data - p], data[n][n_data - p - 1:]))
+54 matrix = np.array(matrix)
+55 # Construct y1 and y2
+56 y1 = np.concatenate(matrix[:, :, :p])
+57 y2 = np.concatenate(matrix[:, :, 1:])
+58 # Apply SVD to y2
+59 u, s, vh = svd(y2, **kwargs)
+60 # Construct z from y1 and SVD of y2, setting all singular values beyond the kth to zero
+61 z = np.diag(1. / s[:k]) @ u[:, :k].T @ y1 @ vh.T[:, :k]
+62 # Return the sorted logarithms of the real eigenvalues as Obs
+63 energy_levels = np.log(np.abs(eig(z, **kwargs)))
+64 return sorted(energy_levels, key=lambda x: abs(x.value))
@@ -222,6 +232,13 @@ matrix pencil parameter which filters noise. The optimal value is expected betwe
len(data)/3 and 2*len(data)/3. The computation is more expensive the closer p is
to len(data)/2 but could possibly suppress more noise (default len(data)//2).
+
+Returns
+
+
+- energy_levels (list[Obs]):
+Extracted energy levels
+
diff --git a/docs/pyerrors/roots.html b/docs/pyerrors/roots.html
index fdfde017..96acde28 100644
--- a/docs/pyerrors/roots.html
+++ b/docs/pyerrors/roots.html
@@ -102,7 +102,7 @@
24
25 Returns
26 -------
-27 Obs
+27 res : Obs
28 `Obs` valued root of the function.
29 '''
30 d_val = np.vectorize(lambda x: x.value)(np.array(d))
@@ -154,7 +154,7 @@
25
26 Returns
27 -------
-28 Obs
+28 res : Obs
29 `Obs` valued root of the function.
30 '''
31 d_val = np.vectorize(lambda x: x.value)(np.array(d))
@@ -198,7 +198,8 @@ Initial guess for the minimization.
Returns
-- Obs:
Obs
valued root of the function.
+- res (Obs):
+
Obs
valued root of the function.
diff --git a/docs/search.js b/docs/search.js
index 38f25af0..9e6ea864 100644
--- a/docs/search.js
+++ b/docs/search.js
@@ -1,6 +1,6 @@
window.pdocSearch = (function(){
/** elasticlunr - http://weixsong.github.io * Copyright (C) 2017 Oliver Nightingale * Copyright (C) 2017 Wei Song * MIT Licensed */!function(){function e(e){if(null===e||"object"!=typeof e)return e;var t=e.constructor();for(var n in e)e.hasOwnProperty(n)&&(t[n]=e[n]);return t}var t=function(e){var n=new t.Index;return n.pipeline.add(t.trimmer,t.stopWordFilter,t.stemmer),e&&e.call(n,n),n};t.version="0.9.5",lunr=t,t.utils={},t.utils.warn=function(e){return function(t){e.console&&console.warn&&console.warn(t)}}(this),t.utils.toString=function(e){return void 0===e||null===e?"":e.toString()},t.EventEmitter=function(){this.events={}},t.EventEmitter.prototype.addListener=function(){var e=Array.prototype.slice.call(arguments),t=e.pop(),n=e;if("function"!=typeof t)throw new TypeError("last argument must be a function");n.forEach(function(e){this.hasHandler(e)||(this.events[e]=[]),this.events[e].push(t)},this)},t.EventEmitter.prototype.removeListener=function(e,t){if(this.hasHandler(e)){var n=this.events[e].indexOf(t);-1!==n&&(this.events[e].splice(n,1),0==this.events[e].length&&delete this.events[e])}},t.EventEmitter.prototype.emit=function(e){if(this.hasHandler(e)){var t=Array.prototype.slice.call(arguments,1);this.events[e].forEach(function(e){e.apply(void 0,t)},this)}},t.EventEmitter.prototype.hasHandler=function(e){return e in this.events},t.tokenizer=function(e){if(!arguments.length||null===e||void 0===e)return[];if(Array.isArray(e)){var n=e.filter(function(e){return null===e||void 0===e?!1:!0});n=n.map(function(e){return t.utils.toString(e).toLowerCase()});var i=[];return n.forEach(function(e){var n=e.split(t.tokenizer.seperator);i=i.concat(n)},this),i}return e.toString().trim().toLowerCase().split(t.tokenizer.seperator)},t.tokenizer.defaultSeperator=/[\s\-]+/,t.tokenizer.seperator=t.tokenizer.defaultSeperator,t.tokenizer.setSeperator=function(e){null!==e&&void 0!==e&&"object"==typeof e&&(t.tokenizer.seperator=e)},t.tokenizer.resetSeperator=function(){t.tokenizer.seperator=t.tokenizer.defaultSeperator},t.tokenizer.getSeperator=function(){return t.tokenizer.seperator},t.Pipeline=function(){this._queue=[]},t.Pipeline.registeredFunctions={},t.Pipeline.registerFunction=function(e,n){n in t.Pipeline.registeredFunctions&&t.utils.warn("Overwriting existing registered function: "+n),e.label=n,t.Pipeline.registeredFunctions[n]=e},t.Pipeline.getRegisteredFunction=function(e){return e in t.Pipeline.registeredFunctions!=!0?null:t.Pipeline.registeredFunctions[e]},t.Pipeline.warnIfFunctionNotRegistered=function(e){var n=e.label&&e.label in this.registeredFunctions;n||t.utils.warn("Function is not registered with pipeline. This may cause problems when serialising the index.\n",e)},t.Pipeline.load=function(e){var n=new t.Pipeline;return e.forEach(function(e){var i=t.Pipeline.getRegisteredFunction(e);if(!i)throw new Error("Cannot load un-registered function: "+e);n.add(i)}),n},t.Pipeline.prototype.add=function(){var e=Array.prototype.slice.call(arguments);e.forEach(function(e){t.Pipeline.warnIfFunctionNotRegistered(e),this._queue.push(e)},this)},t.Pipeline.prototype.after=function(e,n){t.Pipeline.warnIfFunctionNotRegistered(n);var i=this._queue.indexOf(e);if(-1===i)throw new Error("Cannot find existingFn");this._queue.splice(i+1,0,n)},t.Pipeline.prototype.before=function(e,n){t.Pipeline.warnIfFunctionNotRegistered(n);var i=this._queue.indexOf(e);if(-1===i)throw new Error("Cannot find existingFn");this._queue.splice(i,0,n)},t.Pipeline.prototype.remove=function(e){var t=this._queue.indexOf(e);-1!==t&&this._queue.splice(t,1)},t.Pipeline.prototype.run=function(e){for(var t=[],n=e.length,i=this._queue.length,o=0;n>o;o++){for(var r=e[o],s=0;i>s&&(r=this._queue[s](r,o,e),void 0!==r&&null!==r);s++);void 0!==r&&null!==r&&t.push(r)}return t},t.Pipeline.prototype.reset=function(){this._queue=[]},t.Pipeline.prototype.get=function(){return this._queue},t.Pipeline.prototype.toJSON=function(){return this._queue.map(function(e){return t.Pipeline.warnIfFunctionNotRegistered(e),e.label})},t.Index=function(){this._fields=[],this._ref="id",this.pipeline=new t.Pipeline,this.documentStore=new t.DocumentStore,this.index={},this.eventEmitter=new t.EventEmitter,this._idfCache={},this.on("add","remove","update",function(){this._idfCache={}}.bind(this))},t.Index.prototype.on=function(){var e=Array.prototype.slice.call(arguments);return this.eventEmitter.addListener.apply(this.eventEmitter,e)},t.Index.prototype.off=function(e,t){return this.eventEmitter.removeListener(e,t)},t.Index.load=function(e){e.version!==t.version&&t.utils.warn("version mismatch: current "+t.version+" importing "+e.version);var n=new this;n._fields=e.fields,n._ref=e.ref,n.documentStore=t.DocumentStore.load(e.documentStore),n.pipeline=t.Pipeline.load(e.pipeline),n.index={};for(var i in e.index)n.index[i]=t.InvertedIndex.load(e.index[i]);return n},t.Index.prototype.addField=function(e){return this._fields.push(e),this.index[e]=new t.InvertedIndex,this},t.Index.prototype.setRef=function(e){return this._ref=e,this},t.Index.prototype.saveDocument=function(e){return this.documentStore=new t.DocumentStore(e),this},t.Index.prototype.addDoc=function(e,n){if(e){var n=void 0===n?!0:n,i=e[this._ref];this.documentStore.addDoc(i,e),this._fields.forEach(function(n){var o=this.pipeline.run(t.tokenizer(e[n]));this.documentStore.addFieldLength(i,n,o.length);var r={};o.forEach(function(e){e in r?r[e]+=1:r[e]=1},this);for(var s in r){var u=r[s];u=Math.sqrt(u),this.index[n].addToken(s,{ref:i,tf:u})}},this),n&&this.eventEmitter.emit("add",e,this)}},t.Index.prototype.removeDocByRef=function(e){if(e&&this.documentStore.isDocStored()!==!1&&this.documentStore.hasDoc(e)){var t=this.documentStore.getDoc(e);this.removeDoc(t,!1)}},t.Index.prototype.removeDoc=function(e,n){if(e){var n=void 0===n?!0:n,i=e[this._ref];this.documentStore.hasDoc(i)&&(this.documentStore.removeDoc(i),this._fields.forEach(function(n){var o=this.pipeline.run(t.tokenizer(e[n]));o.forEach(function(e){this.index[n].removeToken(e,i)},this)},this),n&&this.eventEmitter.emit("remove",e,this))}},t.Index.prototype.updateDoc=function(e,t){var t=void 0===t?!0:t;this.removeDocByRef(e[this._ref],!1),this.addDoc(e,!1),t&&this.eventEmitter.emit("update",e,this)},t.Index.prototype.idf=function(e,t){var n="@"+t+"/"+e;if(Object.prototype.hasOwnProperty.call(this._idfCache,n))return this._idfCache[n];var i=this.index[t].getDocFreq(e),o=1+Math.log(this.documentStore.length/(i+1));return this._idfCache[n]=o,o},t.Index.prototype.getFields=function(){return this._fields.slice()},t.Index.prototype.search=function(e,n){if(!e)return[];e="string"==typeof e?{any:e}:JSON.parse(JSON.stringify(e));var i=null;null!=n&&(i=JSON.stringify(n));for(var o=new t.Configuration(i,this.getFields()).get(),r={},s=Object.keys(e),u=0;u0&&t.push(e);for(var i in n)"docs"!==i&&"df"!==i&&this.expandToken(e+i,t,n[i]);return t},t.InvertedIndex.prototype.toJSON=function(){return{root:this.root}},t.Configuration=function(e,n){var e=e||"";if(void 0==n||null==n)throw new Error("fields should not be null");this.config={};var i;try{i=JSON.parse(e),this.buildUserConfig(i,n)}catch(o){t.utils.warn("user configuration parse failed, will use default configuration"),this.buildDefaultConfig(n)}},t.Configuration.prototype.buildDefaultConfig=function(e){this.reset(),e.forEach(function(e){this.config[e]={boost:1,bool:"OR",expand:!1}},this)},t.Configuration.prototype.buildUserConfig=function(e,n){var i="OR",o=!1;if(this.reset(),"bool"in e&&(i=e.bool||i),"expand"in e&&(o=e.expand||o),"fields"in e)for(var r in e.fields)if(n.indexOf(r)>-1){var s=e.fields[r],u=o;void 0!=s.expand&&(u=s.expand),this.config[r]={boost:s.boost||0===s.boost?s.boost:1,bool:s.bool||i,expand:u}}else t.utils.warn("field name in user configuration not found in index instance fields");else this.addAllFields2UserConfig(i,o,n)},t.Configuration.prototype.addAllFields2UserConfig=function(e,t,n){n.forEach(function(n){this.config[n]={boost:1,bool:e,expand:t}},this)},t.Configuration.prototype.get=function(){return this.config},t.Configuration.prototype.reset=function(){this.config={}},lunr.SortedSet=function(){this.length=0,this.elements=[]},lunr.SortedSet.load=function(e){var t=new this;return t.elements=e,t.length=e.length,t},lunr.SortedSet.prototype.add=function(){var e,t;for(e=0;e1;){if(r===e)return o;e>r&&(t=o),r>e&&(n=o),i=n-t,o=t+Math.floor(i/2),r=this.elements[o]}return r===e?o:-1},lunr.SortedSet.prototype.locationFor=function(e){for(var t=0,n=this.elements.length,i=n-t,o=t+Math.floor(i/2),r=this.elements[o];i>1;)e>r&&(t=o),r>e&&(n=o),i=n-t,o=t+Math.floor(i/2),r=this.elements[o];return r>e?o:e>r?o+1:void 0},lunr.SortedSet.prototype.intersect=function(e){for(var t=new lunr.SortedSet,n=0,i=0,o=this.length,r=e.length,s=this.elements,u=e.elements;;){if(n>o-1||i>r-1)break;s[n]!==u[i]?s[n]u[i]&&i++:(t.add(s[n]),n++,i++)}return t},lunr.SortedSet.prototype.clone=function(){var e=new lunr.SortedSet;return e.elements=this.toArray(),e.length=e.elements.length,e},lunr.SortedSet.prototype.union=function(e){var t,n,i;this.length>=e.length?(t=this,n=e):(t=e,n=this),i=t.clone();for(var o=0,r=n.toArray();oWhat is pyerrors?\n\npyerrors
is a python package for error computation and propagation of Markov chain Monte Carlo data.\nIt is based on the gamma method arXiv:hep-lat/0306017. Some of its features are:
\n\n\n- automatic differentiation for exact linear error propagation as suggested in arXiv:1809.01289 (partly based on the autograd package).
\n- treatment of slow modes in the simulation as suggested in arXiv:1009.5228.
\n- coherent error propagation for data from different Markov chains.
\n- non-linear fits with x- and y-errors and exact linear error propagation based on automatic differentiation as introduced in arXiv:1809.01289.
\n- real and complex matrix operations and their error propagation based on automatic differentiation (Matrix inverse, Cholesky decomposition, calculation of eigenvalues and eigenvectors, singular value decomposition...).
\n
\n\nMore detailed examples can found in the GitHub repository
.
\n\nIf you use pyerrors
for research that leads to a publication please consider citing:
\n\n\n- Fabian Joswig, Simon Kuberski, Justus T. Kuhlmann, Jan Neuendorf, pyerrors: a python framework for error analysis of Monte Carlo data. [arXiv:2209.14371 [hep-lat]].
\n- Ulli Wolff, Monte Carlo errors with less errors. Comput.Phys.Commun. 156 (2004) 143-153, Comput.Phys.Commun. 176 (2007) 383 (erratum).
\n- Alberto Ramos, Automatic differentiation for error analysis of Monte Carlo data. Comput.Phys.Commun. 238 (2019) 19-35.
\n
\n\nand
\n\n\n- Stefan Schaefer, Rainer Sommer, Francesco Virotta, Critical slowing down and error analysis in lattice QCD simulations. Nucl.Phys.B 845 (2011) 93-119.
\n
\n\nwhere applicable.
\n\nThere exist similar publicly available implementations of gamma method error analysis suites in Fortran, Julia and Python.
\n\nBasic example
\n\n\n
import numpy as np\nimport pyerrors as pe\n\nmy_obs = pe.Obs([samples], ['ensemble_name']) # Initialize an Obs object\nmy_new_obs = 2 * np.log(my_obs) / my_obs ** 2 # Construct derived Obs object\nmy_new_obs.gamma_method() # Estimate the statistical error\nprint(my_new_obs) # Print the result to stdout\n> 0.31498(72)\n
\n
\n\nThe Obs
class
\n\npyerrors
introduces a new datatype, Obs
, which simplifies error propagation and estimation for auto- and cross-correlated data.\nAn Obs
object can be initialized with two arguments, the first is a list containing the samples for an observable from a Monte Carlo chain.\nThe samples can either be provided as python list or as numpy array.\nThe second argument is a list containing the names of the respective Monte Carlo chains as strings. These strings uniquely identify a Monte Carlo chain/ensemble.
\n\n\n
import pyerrors as pe\n\nmy_obs = pe.Obs([samples], ['ensemble_name'])\n
\n
\n\nError propagation
\n\nWhen performing mathematical operations on Obs
objects the correct error propagation is intrinsically taken care of using a first order Taylor expansion\n$$\\delta_f^i=\\sum_\\alpha \\bar{f}_\\alpha \\delta_\\alpha^i\\,,\\quad \\delta_\\alpha^i=a_\\alpha^i-\\bar{a}_\\alpha\\,,$$\nas introduced in arXiv:hep-lat/0306017.\nThe required derivatives $\\bar{f}_\\alpha$ are evaluated up to machine precision via automatic differentiation as suggested in arXiv:1809.01289.
\n\nThe Obs
class is designed such that mathematical numpy functions can be used on Obs
just as for regular floats.
\n\n\n
import numpy as np\nimport pyerrors as pe\n\nmy_obs1 = pe.Obs([samples1], ['ensemble_name'])\nmy_obs2 = pe.Obs([samples2], ['ensemble_name'])\n\nmy_sum = my_obs1 + my_obs2\n\nmy_m_eff = np.log(my_obs1 / my_obs2)\n\niamzero = my_m_eff - my_m_eff\n# Check that value and fluctuations are zero within machine precision\nprint(iamzero == 0.0)\n> True\n
\n
\n\nError estimation
\n\nThe error estimation within pyerrors
is based on the gamma method introduced in arXiv:hep-lat/0306017.\nAfter having arrived at the derived quantity of interest the gamma_method
can be called as detailed in the following example.
\n\n\n
my_sum.gamma_method()\nprint(my_sum)\n> 1.70(57)\nmy_sum.details()\n> Result 1.70000000e+00 +/- 5.72046658e-01 +/- 7.56746598e-02 (33.650%)\n> t_int 2.71422900e+00 +/- 6.40320983e-01 S = 2.00\n> 1000 samples in 1 ensemble:\n> \u00b7 Ensemble 'ensemble_name' : 1000 configurations (from 1 to 1000)\n
\n
\n\nWe use the following definition of the integrated autocorrelation time established in Madras & Sokal 1988\n$$\\tau_\\mathrm{int}=\\frac{1}{2}+\\sum_{t=1}^{W}\\rho(t)\\geq \\frac{1}{2}\\,.$$\nThe window $W$ is determined via the automatic windowing procedure described in arXiv:hep-lat/0306017.\nThe standard value for the parameter $S$ of this automatic windowing procedure is $S=2$. Other values for $S$ can be passed to the gamma_method
as parameter.
\n\n\n
my_sum.gamma_method(S=3.0)\nmy_sum.details()\n> Result 1.70000000e+00 +/- 6.30675201e-01 +/- 1.04585650e-01 (37.099%)\n> t_int 3.29909703e+00 +/- 9.77310102e-01 S = 3.00\n> 1000 samples in 1 ensemble:\n> \u00b7 Ensemble 'ensemble_name' : 1000 configurations (from 1 to 1000)\n
\n
\n\nThe integrated autocorrelation time $\\tau_\\mathrm{int}$ and the autocorrelation function $\\rho(W)$ can be monitored via the methods pyerrors.obs.Obs.plot_tauint
and pyerrors.obs.Obs.plot_rho
.
\n\nIf the parameter $S$ is set to zero it is assumed that the dataset does not exhibit any autocorrelation and the window size is chosen to be zero.\nIn this case the error estimate is identical to the sample standard error.
\n\nExponential tails
\n\nSlow modes in the Monte Carlo history can be accounted for by attaching an exponential tail to the autocorrelation function $\\rho$ as suggested in arXiv:1009.5228. The longest autocorrelation time in the history, $\\tau_\\mathrm{exp}$, can be passed to the gamma_method
as parameter. In this case the automatic windowing procedure is vacated and the parameter $S$ does not affect the error estimate.
\n\n\n
my_sum.gamma_method(tau_exp=7.2)\nmy_sum.details()\n> Result 1.70000000e+00 +/- 6.28097762e-01 +/- 5.79077524e-02 (36.947%)\n> t_int 3.27218667e+00 +/- 7.99583654e-01 tau_exp = 7.20, N_sigma = 1\n> 1000 samples in 1 ensemble:\n> \u00b7 Ensemble 'ensemble_name' : 1000 configurations (from 1 to 1000)\n
\n
\n\nFor the full API see pyerrors.obs.Obs.gamma_method
.
\n\nMultiple ensembles/replica
\n\nError propagation for multiple ensembles (Markov chains with different simulation parameters) is handled automatically. Ensembles are uniquely identified by their name
.
\n\n\n
obs1 = pe.Obs([samples1], ['ensemble1'])\nobs2 = pe.Obs([samples2], ['ensemble2'])\n\nmy_sum = obs1 + obs2\nmy_sum.details()\n> Result 2.00697958e+00\n> 1500 samples in 2 ensembles:\n> \u00b7 Ensemble 'ensemble1' : 1000 configurations (from 1 to 1000)\n> \u00b7 Ensemble 'ensemble2' : 500 configurations (from 1 to 500)\n
\n
\n\npyerrors
identifies multiple replica (independent Markov chains with identical simulation parameters) by the vertical bar |
in the name of the data set.
\n\n\n
obs1 = pe.Obs([samples1], ['ensemble1|r01'])\nobs2 = pe.Obs([samples2], ['ensemble1|r02'])\n\n> my_sum = obs1 + obs2\n> my_sum.details()\n> Result 2.00697958e+00\n> 1500 samples in 1 ensemble:\n> \u00b7 Ensemble 'ensemble1'\n> \u00b7 Replicum 'r01' : 1000 configurations (from 1 to 1000)\n> \u00b7 Replicum 'r02' : 500 configurations (from 1 to 500)\n
\n
\n\nError estimation for multiple ensembles
\n\nIn order to keep track of different error analysis parameters for different ensembles one can make use of global dictionaries as detailed in the following example.
\n\n\n
pe.Obs.S_dict['ensemble1'] = 2.5\npe.Obs.tau_exp_dict['ensemble2'] = 8.0\npe.Obs.tau_exp_dict['ensemble3'] = 2.0\n
\n
\n\nIn case the gamma_method
is called without any parameters it will use the values specified in the dictionaries for the respective ensembles.\nPassing arguments to the gamma_method
still dominates over the dictionaries.
\n\nIrregular Monte Carlo chains
\n\nObs
objects defined on irregular Monte Carlo chains can be initialized with the parameter idl
.
\n\n\n
# Observable defined on configurations 20 to 519\nobs1 = pe.Obs([samples1], ['ensemble1'], idl=[range(20, 520)])\nobs1.details()\n> Result 9.98319881e-01\n> 500 samples in 1 ensemble:\n> \u00b7 Ensemble 'ensemble1' : 500 configurations (from 20 to 519)\n\n# Observable defined on every second configuration between 5 and 1003\nobs2 = pe.Obs([samples2], ['ensemble1'], idl=[range(5, 1005, 2)])\nobs2.details()\n> Result 9.99100712e-01\n> 500 samples in 1 ensemble:\n> \u00b7 Ensemble 'ensemble1' : 500 configurations (from 5 to 1003 in steps of 2)\n\n# Observable defined on configurations 2, 9, 28, 29 and 501\nobs3 = pe.Obs([samples3], ['ensemble1'], idl=[[2, 9, 28, 29, 501]])\nobs3.details()\n> Result 1.01718064e+00\n> 5 samples in 1 ensemble:\n> \u00b7 Ensemble 'ensemble1' : 5 configurations (irregular range)\n
\n
\n\nObs
objects defined on regular and irregular histories of the same ensemble can be combined with each other and the correct error propagation and estimation is automatically taken care of.
\n\nWarning: Irregular Monte Carlo chains can result in odd patterns in the autocorrelation functions.\nMake sure to check the autocorrelation time with e.g. pyerrors.obs.Obs.plot_rho
or pyerrors.obs.Obs.plot_tauint
.
\n\nFor the full API see pyerrors.obs.Obs
.
\n\nCorrelators
\n\nWhen one is not interested in single observables but correlation functions, pyerrors
offers the Corr
class which simplifies the corresponding error propagation and provides the user with a set of standard methods. In order to initialize a Corr
objects one needs to arrange the data as a list of Obs
\n\n\n
my_corr = pe.Corr([obs_0, obs_1, obs_2, obs_3])\nprint(my_corr)\n> x0/a Corr(x0/a)\n> ------------------\n> 0 0.7957(80)\n> 1 0.5156(51)\n> 2 0.3227(33)\n> 3 0.2041(21)\n
\n
\n\nIn case the correlation functions are not defined on the outermost timeslices, for example because of fixed boundary conditions, a padding can be introduced.
\n\n\n
my_corr = pe.Corr([obs_0, obs_1, obs_2, obs_3], padding=[1, 1])\nprint(my_corr)\n> x0/a Corr(x0/a)\n> ------------------\n> 0\n> 1 0.7957(80)\n> 2 0.5156(51)\n> 3 0.3227(33)\n> 4 0.2041(21)\n> 5\n
\n
\n\nThe individual entries of a correlator can be accessed via slicing
\n\n\n
print(my_corr[3])\n> 0.3227(33)\n
\n
\n\nError propagation with the Corr
class works very similar to Obs
objects. Mathematical operations are overloaded and Corr
objects can be computed together with other Corr
objects, Obs
objects or real numbers and integers.
\n\n\n
my_new_corr = 0.3 * my_corr[2] * my_corr * my_corr + 12 / my_corr\n
\n
\n\npyerrors
provides the user with a set of regularly used methods for the manipulation of correlator objects:
\n\n\nCorr.gamma_method
applies the gamma method to all entries of the correlator. \nCorr.m_eff
to construct effective masses. Various variants for periodic and fixed temporal boundary conditions are available. \nCorr.deriv
returns the first derivative of the correlator as Corr
. Different discretizations of the numerical derivative are available. \nCorr.second_deriv
returns the second derivative of the correlator as Corr
. Different discretizations of the numerical derivative are available. \nCorr.symmetric
symmetrizes parity even correlations functions, assuming periodic boundary conditions. \nCorr.anti_symmetric
anti-symmetrizes parity odd correlations functions, assuming periodic boundary conditions. \nCorr.T_symmetry
averages a correlator with its time symmetry partner, assuming fixed boundary conditions. \nCorr.plateau
extracts a plateau value from the correlator in a given range. \nCorr.roll
periodically shifts the correlator. \nCorr.reverse
reverses the time ordering of the correlator. \nCorr.correlate
constructs a disconnected correlation function from the correlator and another Corr
or Obs
object. \nCorr.reweight
reweights the correlator. \n
\n\npyerrors
can also handle matrices of correlation functions and extract energy states from these matrices via a generalized eigenvalue problem (see pyerrors.correlators.Corr.GEVP
).
\n\nFor the full API see pyerrors.correlators.Corr
.
\n\nComplex valued observables
\n\npyerrors
can handle complex valued observables via the class pyerrors.obs.CObs
.\nCObs
are initialized with a real and an imaginary part which both can be Obs
valued.
\n\n\n
my_real_part = pe.Obs([samples1], ['ensemble1'])\nmy_imag_part = pe.Obs([samples2], ['ensemble1'])\n\nmy_cobs = pe.CObs(my_real_part, my_imag_part)\nmy_cobs.gamma_method()\nprint(my_cobs)\n> (0.9959(91)+0.659(28)j)\n
\n
\n\nElementary mathematical operations are overloaded and samples are properly propagated as for the Obs
class.
\n\n\n
my_derived_cobs = (my_cobs + my_cobs.conjugate()) / np.abs(my_cobs)\nmy_derived_cobs.gamma_method()\nprint(my_derived_cobs)\n> (1.668(23)+0.0j)\n
\n
\n\nThe Covobs
class
\n\nIn many projects, auxiliary data that is not based on Monte Carlo chains enters. Examples are experimentally determined mesons masses which are used to set the scale or renormalization constants. These numbers come with an error that has to be propagated through the analysis. The Covobs
class allows to define such quantities in pyerrors
. Furthermore, external input might consist of correlated quantities. An example are the parameters of an interpolation formula, which are defined via mean values and a covariance matrix between all parameters. The contribution of the interpolation formula to the error of a derived quantity therefore might depend on the complete covariance matrix.
\n\nThis concept is built into the definition of Covobs
. In pyerrors
, external input is defined by $M$ mean values, a $M\\times M$ covariance matrix, where $M=1$ is permissible, and a name that uniquely identifies the covariance matrix. Below, we define the pion mass, based on its mean value and error, 134.9768(5). Note, that the square of the error enters cov_Obs
, since the second argument of this function is the covariance matrix of the Covobs
.
\n\n\n
import pyerrors.obs as pe\n\nmpi = pe.cov_Obs(134.9768, 0.0005**2, 'pi^0 mass')\nmpi.gamma_method()\nmpi.details()\n> Result 1.34976800e+02 +/- 5.00000000e-04 +/- 0.00000000e+00 (0.000%)\n> pi^0 mass 5.00000000e-04\n> 0 samples in 1 ensemble:\n> \u00b7 Covobs 'pi^0 mass'\n
\n
\n\nThe resulting object mpi
is an Obs
that contains a Covobs
. In the following, it may be handled as any other Obs
. The contribution of the covariance matrix to the error of an Obs
is determined from the $M \\times M$ covariance matrix $\\Sigma$ and the gradient of the Obs
with respect to the external quantities, which is the $1\\times M$ Jacobian matrix $J$, via\n$$s = \\sqrt{J^T \\Sigma J}\\,,$$\nwhere the Jacobian is computed for each derived quantity via automatic differentiation.
\n\nCorrelated auxiliary data is defined similarly to above, e.g., via
\n\n\n
RAP = pe.cov_Obs([16.7457, -19.0475], [[3.49591, -6.07560], [-6.07560, 10.5834]], 'R_AP, 1906.03445, (5.3a)')\nprint(RAP)\n> [Obs[16.7(1.9)], Obs[-19.0(3.3)]]\n
\n
\n\nwhere RAP
now is a list of two Obs
that contains the two correlated parameters.
\n\nSince the gradient of a derived observable with respect to an external covariance matrix is propagated through the entire analysis, the Covobs
class allows to quote the derivative of a result with respect to the external quantities. If these derivatives are published together with the result, small shifts in the definition of external quantities, e.g., the definition of the physical point, can be performed a posteriori based on the published information. This may help to compare results of different groups. The gradient of an Obs
o
with respect to a covariance matrix with the identifying string k
may be accessed via
\n\n\n\nError propagation in iterative algorithms
\n\npyerrors
supports exact linear error propagation for iterative algorithms like various variants of non-linear least squares fits or root finding. The derivatives required for the error propagation are calculated as described in arXiv:1809.01289.
\n\nLeast squares fits
\n\nStandard non-linear least square fits with errors on the dependent but not the independent variables can be performed with pyerrors.fits.least_squares
. As default solver the Levenberg-Marquardt algorithm implemented in scipy is used.
\n\nFit functions have to be of the following form
\n\n\n
import autograd.numpy as anp\n\ndef func(a, x):\n return a[1] * anp.exp(-a[0] * x)\n
\n
\n\nIt is important that numerical functions refer to autograd.numpy
instead of numpy
for the automatic differentiation in iterative algorithms to work properly.
\n\nFits can then be performed via
\n\n\n
fit_result = pe.fits.least_squares(x, y, func)\nprint("\\n", fit_result)\n> Fit with 2 parameters\n> Method: Levenberg-Marquardt\n> `ftol` termination condition is satisfied.\n> chisquare/d.o.f.: 0.9593035785160936\n\n> Goodness of fit:\n> \u03c7\u00b2/d.o.f. = 0.959304\n> p-value = 0.5673\n> Fit parameters:\n> 0 0.0548(28)\n> 1 1.933(64)\n
\n
\n\nwhere x is a list
or numpy.array
of floats
and y is a list
or numpy.array
of Obs
.
\n\nData stored in Corr
objects can be fitted directly using the Corr.fit
method.
\n\n\n
my_corr = pe.Corr(y)\nfit_result = my_corr.fit(func, fitrange=[12, 25])\n
\n
\n\nthis can simplify working with absolute fit ranges and takes care of gaps in the data automatically.
\n\nFor fit functions with multiple independent variables the fit function can be of the form
\n\n\n
def func(a, x):\n (x1, x2) = x\n return a[0] * x1 ** 2 + a[1] * x2\n
\n
\n\npyerrors
also supports correlated fits which can be triggered via the parameter correlated_fit=True
.\nDetails about how the required covariance matrix is estimated can be found in pyerrors.obs.covariance
.
\n\nDirect visualizations of the performed fits can be triggered via resplot=True
or qqplot=True
. For all available options see pyerrors.fits.least_squares
.
\n\nTotal least squares fits
\n\npyerrors
can also fit data with errors on both the dependent and independent variables using the total least squares method also referred to orthogonal distance regression as implemented in scipy, see pyerrors.fits.least_squares
. The syntax is identical to the standard least squares case, the only difference being that x
also has to be a list
or numpy.array
of Obs
.
\n\nFor the full API see pyerrors.fits
for fits and pyerrors.roots
for finding roots of functions.
\n\nMatrix operations
\n\npyerrors
provides wrappers for Obs
- and CObs
-valued matrix operations based on numpy.linalg
. The supported functions include:
\n\n\ninv
for the matrix inverse. \ncholseky
for the Cholesky decomposition. \ndet
for the matrix determinant. \neigh
for eigenvalues and eigenvectors of hermitean matrices. \neig
for eigenvalues of general matrices. \npinv
for the Moore-Penrose pseudoinverse. \nsvd
for the singular-value-decomposition. \n
\n\nFor the full API see pyerrors.linalg
.
\n\nExport data
\n\nThe preferred exported file format within pyerrors
is json.gz. Files written to this format are valid JSON files that have been compressed using gzip. The structure of the content is inspired by the dobs format of the ALPHA collaboration. The aim of the format is to facilitate the storage of data in a self-contained way such that, even years after the creation of the file, it is possible to extract all necessary information:
\n\n\n- What observables are stored? Possibly: How exactly are they defined.
\n- How does each single ensemble or external quantity contribute to the error of the observable?
\n- Who did write the file when and on which machine?
\n
\n\nThis can be achieved by storing all information in one single file. The export routines of pyerrors
are written such that as much information as possible is written automatically as described in the following example
\n\n\n
my_obs = pe.Obs([samples], ["test_ensemble"])\nmy_obs.tag = "My observable"\n\npe.input.json.dump_to_json(my_obs, "test_output_file", description="This file contains a test observable")\n# For a single observable one can equivalently use the class method dump\nmy_obs.dump("test_output_file", description="This file contains a test observable")\n\ncheck = pe.input.json.load_json("test_output_file")\n\nprint(my_obs == check)\n> True\n
\n
\n\nThe format also allows to directly write out the content of Corr
objects or lists and arrays of Obs
objects by passing the desired data to pyerrors.input.json.dump_to_json
.
\n\n\n\nThe first entries of the file provide optional auxiliary information:
\n\n\nprogram
is a string that indicates which program was used to write the file. \nversion
is a string that specifies the version of the format. \nwho
is a string that specifies the user name of the creator of the file. \ndate
is a string and contains the creation date of the file. \nhost
is a string and contains the hostname of the machine where the file has been written. \ndescription
contains information on the content of the file. This field is not filled automatically in pyerrors
. The user is advised to provide as detailed information as possible in this field. Examples are: Input files of measurements or simulations, LaTeX formulae or references to publications to specify how the observables have been computed, details on the analysis strategy, ... This field may be any valid JSON type. Strings, arrays or objects (equivalent to dicts in python) are well suited to provide information. \n
\n\nThe only necessary entry of the file is the field\n-obsdata
, an array that contains the actual data.
\n\nEach entry of the array belongs to a single structure of observables. Currently, these structures can be either of Obs
, list
, numpy.ndarray
, Corr
. All Obs
inside a structure (with dimension > 0) have to be defined on the same set of configurations. Different structures, that are represented by entries of the array obsdata
, are treated independently. Each entry of the array obsdata
has the following required entries:
\n\n\ntype
is a string that specifies the type of the structure. This allows to parse the content to the correct form after reading the file. It is always possible to interpret the content as list of Obs. \nvalue
is an array that contains the mean values of the Obs inside the structure.\nThe following entries are optional: \nlayout
is a string that specifies the layout of multi-dimensional structures. Examples are \"2, 2\" for a 2x2 dimensional matrix or \"64, 4, 4\" for a Corr with $T=64$ and 4x4 matrices on each time slices. \"1\" denotes a single Obs. Multi-dimensional structures are stored in row-major format (see below). \ntag
is any JSON type. It contains additional information concerning the structure. The tag
of an Obs
in pyerrors
is written here. \nreweighted
is a Bool that may be used to specify, whether the Obs
in the structure have been reweighted. \ndata
is an array that contains the data from MC chains. We will define it below. \ncdata
is an array that contains the data from external quantities with an error (Covobs
in pyerrors
). We will define it below. \n
\n\nThe array data
contains the data from MC chains. Each entry of the array corresponds to one ensemble and contains:
\n\n\nid
, a string that contains the name of the ensemble \nreplica
, an array that contains an entry per replica of the ensemble. \n
\n\nEach entry of replica
contains\nname
, a string that contains the name of the replica\ndeltas
, an array that contains the actual data.
\n\nEach entry in deltas
corresponds to one configuration of the replica and has $1+N$ many entries. The first entry is an integer that specifies the configuration number that, together with ensemble and replica name, may be used to uniquely identify the configuration on which the data has been obtained. The following N entries specify the deltas, i.e., the deviation of the observable from the mean value on this configuration, of each Obs
inside the structure. Multi-dimensional structures are stored in a row-major format. For primary observables, such as correlation functions, $value + delta_i$ matches the primary data obtained on the configuration.
\n\nThe array cdata
contains information about the contribution of auxiliary observables, represented by Covobs
in pyerrors
, to the total error of the observables. Each entry of the array belongs to one auxiliary covariance matrix and contains:
\n\n\nid
, a string that identifies the covariance matrix \nlayout
, a string that defines the dimensions of the $M\\times M$ covariance matrix (has to be \"M, M\" or \"1\"). \ncov
, an array that contains the $M\\times M$ many entries of the covariance matrix, stored in row-major format. \ngrad
, an array that contains N entries, one for each Obs
inside the structure. Each entry itself is an array, that contains the M gradients of the Nth observable with respect to the quantity that corresponds to the Mth diagonal entry of the covariance matrix. \n
\n\nA JSON schema that may be used to verify the correctness of a file with respect to the format definition is stored in ./examples/json_schema.json. The schema is a self-descriptive format definition and contains an exemplary file.
\n\nJulia I/O routines for the json.gz format, compatible with ADerrors.jl, can be found here.
\n"}, "pyerrors.correlators": {"fullname": "pyerrors.correlators", "modulename": "pyerrors.correlators", "kind": "module", "doc": "\n"}, "pyerrors.correlators.Corr": {"fullname": "pyerrors.correlators.Corr", "modulename": "pyerrors.correlators", "qualname": "Corr", "kind": "class", "doc": "The class for a correlator (time dependent sequence of pe.Obs).
\n\nEverything, this class does, can be achieved using lists or arrays of Obs.\nBut it is simply more convenient to have a dedicated object for correlators.\nOne often wants to add or multiply correlators of the same length at every timeslice and it is inconvenient\nto iterate over all timeslices for every operation. This is especially true, when dealing with matrices.
\n\nThe correlator can have two types of content: An Obs at every timeslice OR a GEVP\nmatrix at every timeslice. Other dependency (eg. spatial) are not supported.
\n"}, "pyerrors.correlators.Corr.__init__": {"fullname": "pyerrors.correlators.Corr.__init__", "modulename": "pyerrors.correlators", "qualname": "Corr.__init__", "kind": "function", "doc": "Initialize a Corr object.
\n\nParameters
\n\n\n- data_input (list or array):\nlist of Obs or list of arrays of Obs or array of Corrs
\n- padding (list, optional):\nList with two entries where the first labels the padding\nat the front of the correlator and the second the padding\nat the back.
\n- prange (list, optional):\nList containing the first and last timeslice of the plateau\nregion indentified for this correlator.
\n
\n", "signature": "(data_input, padding=[0, 0], prange=None)"}, "pyerrors.correlators.Corr.gamma_method": {"fullname": "pyerrors.correlators.Corr.gamma_method", "modulename": "pyerrors.correlators", "qualname": "Corr.gamma_method", "kind": "function", "doc": "Apply the gamma method to the content of the Corr.
\n", "signature": "(self, **kwargs):", "funcdef": "def"}, "pyerrors.correlators.Corr.gm": {"fullname": "pyerrors.correlators.Corr.gm", "modulename": "pyerrors.correlators", "qualname": "Corr.gm", "kind": "function", "doc": "Apply the gamma method to the content of the Corr.
\n", "signature": "(self, **kwargs):", "funcdef": "def"}, "pyerrors.correlators.Corr.projected": {"fullname": "pyerrors.correlators.Corr.projected", "modulename": "pyerrors.correlators", "qualname": "Corr.projected", "kind": "function", "doc": "We need to project the Correlator with a Vector to get a single value at each timeslice.
\n\nThe method can use one or two vectors.\nIf two are specified it returns v1@G@v2 (the order might be very important.)\nBy default it will return the lowest source, which usually means unsmeared-unsmeared (0,0), but it does not have to
\n", "signature": "(self, vector_l=None, vector_r=None, normalize=False):", "funcdef": "def"}, "pyerrors.correlators.Corr.item": {"fullname": "pyerrors.correlators.Corr.item", "modulename": "pyerrors.correlators", "qualname": "Corr.item", "kind": "function", "doc": "Picks the element [i,j] from every matrix and returns a correlator containing one Obs per timeslice.
\n\nParameters
\n\n\n- i (int):\nFirst index to be picked.
\n- j (int):\nSecond index to be picked.
\n
\n", "signature": "(self, i, j):", "funcdef": "def"}, "pyerrors.correlators.Corr.plottable": {"fullname": "pyerrors.correlators.Corr.plottable", "modulename": "pyerrors.correlators", "qualname": "Corr.plottable", "kind": "function", "doc": "Outputs the correlator in a plotable format.
\n\nOutputs three lists containing the timeslice index, the value on each\ntimeslice and the error on each timeslice.
\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.correlators.Corr.symmetric": {"fullname": "pyerrors.correlators.Corr.symmetric", "modulename": "pyerrors.correlators", "qualname": "Corr.symmetric", "kind": "function", "doc": "Symmetrize the correlator around x0=0.
\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.correlators.Corr.anti_symmetric": {"fullname": "pyerrors.correlators.Corr.anti_symmetric", "modulename": "pyerrors.correlators", "qualname": "Corr.anti_symmetric", "kind": "function", "doc": "Anti-symmetrize the correlator around x0=0.
\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.correlators.Corr.is_matrix_symmetric": {"fullname": "pyerrors.correlators.Corr.is_matrix_symmetric", "modulename": "pyerrors.correlators", "qualname": "Corr.is_matrix_symmetric", "kind": "function", "doc": "Checks whether a correlator matrices is symmetric on every timeslice.
\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.correlators.Corr.matrix_symmetric": {"fullname": "pyerrors.correlators.Corr.matrix_symmetric", "modulename": "pyerrors.correlators", "qualname": "Corr.matrix_symmetric", "kind": "function", "doc": "Symmetrizes the correlator matrices on every timeslice.
\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.correlators.Corr.GEVP": {"fullname": "pyerrors.correlators.Corr.GEVP", "modulename": "pyerrors.correlators", "qualname": "Corr.GEVP", "kind": "function", "doc": "Solve the generalized eigenvalue problem on the correlator matrix and returns the corresponding eigenvectors.
\n\nThe eigenvectors are sorted according to the descending eigenvalues, the zeroth eigenvector(s) correspond to the\nlargest eigenvalue(s). The eigenvector(s) for the individual states can be accessed via slicing
\n\n\n
C.GEVP(t0=2)[0] # Ground state vector(s)\nC.GEVP(t0=2)[:3] # Vectors for the lowest three states\n
\n
\n\nParameters
\n\n\n- t0 (int):\nThe time t0 for the right hand side of the GEVP according to $G(t)v_i=\\lambda_i G(t_0)v_i$
\n- ts (int):\nfixed time $G(t_s)v_i=\\lambda_i G(t_0)v_i$ if sort=None.\nIf sort=\"Eigenvector\" it gives a reference point for the sorting method.
\n- sort (string):\nIf this argument is set, a list of self.T vectors per state is returned. If it is set to None, only one vector is returned.\n
\n- \"Eigenvalue\": The eigenvector is chosen according to which eigenvalue it belongs individually on every timeslice.
\n- \"Eigenvector\": Use the method described in arXiv:2004.10472 to find the set of v(t) belonging to the state.\nThe reference state is identified by its eigenvalue at $t=t_s$.
\n
\n
\n\nOther Parameters
\n\n\n- state (int):\nReturns only the vector(s) for a specified state. The lowest state is zero.
\n
\n", "signature": "(self, t0, ts=None, sort='Eigenvalue', **kwargs):", "funcdef": "def"}, "pyerrors.correlators.Corr.Eigenvalue": {"fullname": "pyerrors.correlators.Corr.Eigenvalue", "modulename": "pyerrors.correlators", "qualname": "Corr.Eigenvalue", "kind": "function", "doc": "Determines the eigenvalue of the GEVP by solving and projecting the correlator
\n\nParameters
\n\n\n- state (int):\nThe state one is interested in ordered by energy. The lowest state is zero.
\n- All other parameters are identical to the ones of Corr.GEVP.
\n
\n", "signature": "(self, t0, ts=None, state=0, sort='Eigenvalue'):", "funcdef": "def"}, "pyerrors.correlators.Corr.Hankel": {"fullname": "pyerrors.correlators.Corr.Hankel", "modulename": "pyerrors.correlators", "qualname": "Corr.Hankel", "kind": "function", "doc": "Constructs an NxN Hankel matrix
\n\nC(t) c(t+1) ... c(t+n-1)\nC(t+1) c(t+2) ... c(t+n)\n.................\nC(t+(n-1)) c(t+n) ... c(t+2(n-1))
\n\nParameters
\n\n\n- N (int):\nDimension of the Hankel matrix
\n- periodic (bool, optional):\ndetermines whether the matrix is extended periodically
\n
\n", "signature": "(self, N, periodic=False):", "funcdef": "def"}, "pyerrors.correlators.Corr.roll": {"fullname": "pyerrors.correlators.Corr.roll", "modulename": "pyerrors.correlators", "qualname": "Corr.roll", "kind": "function", "doc": "Periodically shift the correlator by dt timeslices
\n\nParameters
\n\n\n- dt (int):\nnumber of timeslices
\n
\n", "signature": "(self, dt):", "funcdef": "def"}, "pyerrors.correlators.Corr.reverse": {"fullname": "pyerrors.correlators.Corr.reverse", "modulename": "pyerrors.correlators", "qualname": "Corr.reverse", "kind": "function", "doc": "Reverse the time ordering of the Corr
\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.correlators.Corr.thin": {"fullname": "pyerrors.correlators.Corr.thin", "modulename": "pyerrors.correlators", "qualname": "Corr.thin", "kind": "function", "doc": "Thin out a correlator to suppress correlations
\n\nParameters
\n\n\n- spacing (int):\nKeep only every 'spacing'th entry of the correlator
\n- offset (int):\nOffset the equal spacing
\n
\n", "signature": "(self, spacing=2, offset=0):", "funcdef": "def"}, "pyerrors.correlators.Corr.correlate": {"fullname": "pyerrors.correlators.Corr.correlate", "modulename": "pyerrors.correlators", "qualname": "Corr.correlate", "kind": "function", "doc": "Correlate the correlator with another correlator or Obs
\n\nParameters
\n\n\n- partner (Obs or Corr):\npartner to correlate the correlator with.\nCan either be an Obs which is correlated with all entries of the\ncorrelator or a Corr of same length.
\n
\n", "signature": "(self, partner):", "funcdef": "def"}, "pyerrors.correlators.Corr.reweight": {"fullname": "pyerrors.correlators.Corr.reweight", "modulename": "pyerrors.correlators", "qualname": "Corr.reweight", "kind": "function", "doc": "Reweight the correlator.
\n\nParameters
\n\n\n- weight (Obs):\nReweighting factor. An Observable that has to be defined on a superset of the\nconfigurations in obs[i].idl for all i.
\n- all_configs (bool):\nif True, the reweighted observables are normalized by the average of\nthe reweighting factor on all configurations in weight.idl and not\non the configurations in obs[i].idl.
\n
\n", "signature": "(self, weight, **kwargs):", "funcdef": "def"}, "pyerrors.correlators.Corr.T_symmetry": {"fullname": "pyerrors.correlators.Corr.T_symmetry", "modulename": "pyerrors.correlators", "qualname": "Corr.T_symmetry", "kind": "function", "doc": "Return the time symmetry average of the correlator and its partner
\n\nParameters
\n\n\n- partner (Corr):\nTime symmetry partner of the Corr
\n- partity (int):\nParity quantum number of the correlator, can be +1 or -1
\n
\n", "signature": "(self, partner, parity=1):", "funcdef": "def"}, "pyerrors.correlators.Corr.deriv": {"fullname": "pyerrors.correlators.Corr.deriv", "modulename": "pyerrors.correlators", "qualname": "Corr.deriv", "kind": "function", "doc": "Return the first derivative of the correlator with respect to x0.
\n\nParameters
\n\n\n- variant (str):\ndecides which definition of the finite differences derivative is used.\nAvailable choice: symmetric, forward, backward, improved, log, default: symmetric
\n
\n", "signature": "(self, variant='symmetric'):", "funcdef": "def"}, "pyerrors.correlators.Corr.second_deriv": {"fullname": "pyerrors.correlators.Corr.second_deriv", "modulename": "pyerrors.correlators", "qualname": "Corr.second_deriv", "kind": "function", "doc": "Return the second derivative of the correlator with respect to x0.
\n\nParameters
\n\n\n- variant (str):\ndecides which definition of the finite differences derivative is used.\nAvailable choice: symmetric, improved, log, default: symmetric
\n
\n", "signature": "(self, variant='symmetric'):", "funcdef": "def"}, "pyerrors.correlators.Corr.m_eff": {"fullname": "pyerrors.correlators.Corr.m_eff", "modulename": "pyerrors.correlators", "qualname": "Corr.m_eff", "kind": "function", "doc": "Returns the effective mass of the correlator as correlator object
\n\nParameters
\n\n\n- variant (str):\nlog : uses the standard effective mass log(C(t) / C(t+1))\ncosh, periodic : Use periodicitiy of the correlator by solving C(t) / C(t+1) = cosh(m * (t - T/2)) / cosh(m * (t + 1 - T/2)) for m.\nsinh : Use anti-periodicitiy of the correlator by solving C(t) / C(t+1) = sinh(m * (t - T/2)) / sinh(m * (t + 1 - T/2)) for m.\nSee, e.g., arXiv:1205.5380\narccosh : Uses the explicit form of the symmetrized correlator (not recommended)\nlogsym: uses the symmetric effective mass log(C(t-1) / C(t+1))/2
\n- guess (float):\nguess for the root finder, only relevant for the root variant
\n
\n", "signature": "(self, variant='log', guess=1.0):", "funcdef": "def"}, "pyerrors.correlators.Corr.fit": {"fullname": "pyerrors.correlators.Corr.fit", "modulename": "pyerrors.correlators", "qualname": "Corr.fit", "kind": "function", "doc": "Fits function to the data
\n\nParameters
\n\n\n- function (obj):\nfunction to fit to the data. See fits.least_squares for details.
\n- fitrange (list):\nTwo element list containing the timeslices on which the fit is supposed to start and stop.\nCaution: This range is inclusive as opposed to standard python indexing.\n
fitrange=[4, 6]
corresponds to the three entries 4, 5 and 6.\nIf not specified, self.prange or all timeslices are used. \n- silent (bool):\nDecides whether output is printed to the standard output.
\n
\n", "signature": "(self, function, fitrange=None, silent=False, **kwargs):", "funcdef": "def"}, "pyerrors.correlators.Corr.plateau": {"fullname": "pyerrors.correlators.Corr.plateau", "modulename": "pyerrors.correlators", "qualname": "Corr.plateau", "kind": "function", "doc": "Extract a plateau value from a Corr object
\n\nParameters
\n\n\n- plateau_range (list):\nlist with two entries, indicating the first and the last timeslice\nof the plateau region.
\n- method (str):\nmethod to extract the plateau.\n 'fit' fits a constant to the plateau region\n 'avg', 'average' or 'mean' just average over the given timeslices.
\n- auto_gamma (bool):\napply gamma_method with default parameters to the Corr. Defaults to None
\n
\n", "signature": "(self, plateau_range=None, method='fit', auto_gamma=False):", "funcdef": "def"}, "pyerrors.correlators.Corr.set_prange": {"fullname": "pyerrors.correlators.Corr.set_prange", "modulename": "pyerrors.correlators", "qualname": "Corr.set_prange", "kind": "function", "doc": "Sets the attribute prange of the Corr object.
\n", "signature": "(self, prange):", "funcdef": "def"}, "pyerrors.correlators.Corr.show": {"fullname": "pyerrors.correlators.Corr.show", "modulename": "pyerrors.correlators", "qualname": "Corr.show", "kind": "function", "doc": "Plots the correlator using the tag of the correlator as label if available.
\n\nParameters
\n\n\n- x_range (list):\nlist of two values, determining the range of the x-axis e.g. [4, 8].
\n- comp (Corr or list of Corr):\nCorrelator or list of correlators which are plotted for comparison.\nThe tags of these correlators are used as labels if available.
\n- logscale (bool):\nSets y-axis to logscale.
\n- plateau (Obs):\nPlateau value to be visualized in the figure.
\n- fit_res (Fit_result):\nFit_result object to be visualized.
\n- ylabel (str):\nLabel for the y-axis.
\n- save (str):\npath to file in which the figure should be saved.
\n- auto_gamma (bool):\nApply the gamma method with standard parameters to all correlators and plateau values before plotting.
\n- hide_sigma (float):\nHides data points from the first value on which is consistent with zero within 'hide_sigma' standard errors.
\n- references (list):\nList of floating point values that are displayed as horizontal lines for reference.
\n- title (string):\nOptional title of the figure.
\n
\n", "signature": "(\tself,\tx_range=None,\tcomp=None,\ty_range=None,\tlogscale=False,\tplateau=None,\tfit_res=None,\tylabel=None,\tsave=None,\tauto_gamma=False,\thide_sigma=None,\treferences=None,\ttitle=None):", "funcdef": "def"}, "pyerrors.correlators.Corr.spaghetti_plot": {"fullname": "pyerrors.correlators.Corr.spaghetti_plot", "modulename": "pyerrors.correlators", "qualname": "Corr.spaghetti_plot", "kind": "function", "doc": "Produces a spaghetti plot of the correlator suited to monitor exceptional configurations.
\n\nParameters
\n\n\n- logscale (bool):\nDetermines whether the scale of the y-axis is logarithmic or standard.
\n
\n", "signature": "(self, logscale=True):", "funcdef": "def"}, "pyerrors.correlators.Corr.dump": {"fullname": "pyerrors.correlators.Corr.dump", "modulename": "pyerrors.correlators", "qualname": "Corr.dump", "kind": "function", "doc": "Dumps the Corr into a file of chosen type
\n\nParameters
\n\n\n- filename (str):\nName of the file to be saved.
\n- datatype (str):\nFormat of the exported file. Supported formats include\n\"json.gz\" and \"pickle\"
\n- path (str):\nspecifies a custom path for the file (default '.')
\n
\n", "signature": "(self, filename, datatype='json.gz', **kwargs):", "funcdef": "def"}, "pyerrors.correlators.Corr.print": {"fullname": "pyerrors.correlators.Corr.print", "modulename": "pyerrors.correlators", "qualname": "Corr.print", "kind": "function", "doc": "\n", "signature": "(self, print_range=None):", "funcdef": "def"}, "pyerrors.correlators.Corr.sqrt": {"fullname": "pyerrors.correlators.Corr.sqrt", "modulename": "pyerrors.correlators", "qualname": "Corr.sqrt", "kind": "function", "doc": "\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.correlators.Corr.log": {"fullname": "pyerrors.correlators.Corr.log", "modulename": "pyerrors.correlators", "qualname": "Corr.log", "kind": "function", "doc": "\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.correlators.Corr.exp": {"fullname": "pyerrors.correlators.Corr.exp", "modulename": "pyerrors.correlators", "qualname": "Corr.exp", "kind": "function", "doc": "\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.correlators.Corr.sin": {"fullname": "pyerrors.correlators.Corr.sin", "modulename": "pyerrors.correlators", "qualname": "Corr.sin", "kind": "function", "doc": "\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.correlators.Corr.cos": {"fullname": "pyerrors.correlators.Corr.cos", "modulename": "pyerrors.correlators", "qualname": "Corr.cos", "kind": "function", "doc": "\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.correlators.Corr.tan": {"fullname": "pyerrors.correlators.Corr.tan", "modulename": "pyerrors.correlators", "qualname": "Corr.tan", "kind": "function", "doc": "\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.correlators.Corr.sinh": {"fullname": "pyerrors.correlators.Corr.sinh", "modulename": "pyerrors.correlators", "qualname": "Corr.sinh", "kind": "function", "doc": "\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.correlators.Corr.cosh": {"fullname": "pyerrors.correlators.Corr.cosh", "modulename": "pyerrors.correlators", "qualname": "Corr.cosh", "kind": "function", "doc": "\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.correlators.Corr.tanh": {"fullname": "pyerrors.correlators.Corr.tanh", "modulename": "pyerrors.correlators", "qualname": "Corr.tanh", "kind": "function", "doc": "\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.correlators.Corr.arcsin": {"fullname": "pyerrors.correlators.Corr.arcsin", "modulename": "pyerrors.correlators", "qualname": "Corr.arcsin", "kind": "function", "doc": "\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.correlators.Corr.arccos": {"fullname": "pyerrors.correlators.Corr.arccos", "modulename": "pyerrors.correlators", "qualname": "Corr.arccos", "kind": "function", "doc": "\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.correlators.Corr.arctan": {"fullname": "pyerrors.correlators.Corr.arctan", "modulename": "pyerrors.correlators", "qualname": "Corr.arctan", "kind": "function", "doc": "\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.correlators.Corr.arcsinh": {"fullname": "pyerrors.correlators.Corr.arcsinh", "modulename": "pyerrors.correlators", "qualname": "Corr.arcsinh", "kind": "function", "doc": "\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.correlators.Corr.arccosh": {"fullname": "pyerrors.correlators.Corr.arccosh", "modulename": "pyerrors.correlators", "qualname": "Corr.arccosh", "kind": "function", "doc": "\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.correlators.Corr.arctanh": {"fullname": "pyerrors.correlators.Corr.arctanh", "modulename": "pyerrors.correlators", "qualname": "Corr.arctanh", "kind": "function", "doc": "\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.correlators.Corr.prune": {"fullname": "pyerrors.correlators.Corr.prune", "modulename": "pyerrors.correlators", "qualname": "Corr.prune", "kind": "function", "doc": "Project large correlation matrix to lowest states
\n\nThis method can be used to reduce the size of an (N x N) correlation matrix\nto (Ntrunc x Ntrunc) by solving a GEVP at very early times where the noise\nis still small.
\n\nParameters
\n\n\n- Ntrunc (int):\nRank of the target matrix.
\n- tproj (int):\nTime where the eigenvectors are evaluated, corresponds to ts in the GEVP method.\nThe default value is 3.
\n- t0proj (int):\nTime where the correlation matrix is inverted. Choosing t0proj=1 is strongly\ndiscouraged for O(a) improved theories, since the correctness of the procedure\ncannot be granted in this case. The default value is 2.
\n- basematrix (Corr):\nCorrelation matrix that is used to determine the eigenvectors of the\nlowest states based on a GEVP. basematrix is taken to be the Corr itself if\nis is not specified.
\n
\n\nNotes
\n\nWe have the basematrix $C(t)$ and the target matrix $G(t)$. We start by solving\nthe GEVP $$C(t) v_n(t, t_0) = \\lambda_n(t, t_0) C(t_0) v_n(t, t_0)$$ where $t \\equiv t_\\mathrm{proj}$\nand $t_0 \\equiv t_{0, \\mathrm{proj}}$. The target matrix is projected onto the subspace of the\nresulting eigenvectors $v_n, n=1,\\dots,N_\\mathrm{trunc}$ via\n$$G^\\prime_{i, j}(t) = (v_i, G(t) v_j)$$. This allows to reduce the size of a large\ncorrelation matrix and to remove some noise that is added by irrelevant operators.\nThis may allow to use the GEVP on $G(t)$ at late times such that the theoretically motivated\nbound $t_0 \\leq t/2$ holds, since the condition number of $G(t)$ is decreased, compared to $C(t)$.
\n", "signature": "(self, Ntrunc, tproj=3, t0proj=2, basematrix=None):", "funcdef": "def"}, "pyerrors.covobs": {"fullname": "pyerrors.covobs", "modulename": "pyerrors.covobs", "kind": "module", "doc": "\n"}, "pyerrors.covobs.Covobs": {"fullname": "pyerrors.covobs.Covobs", "modulename": "pyerrors.covobs", "qualname": "Covobs", "kind": "class", "doc": "\n"}, "pyerrors.covobs.Covobs.__init__": {"fullname": "pyerrors.covobs.Covobs.__init__", "modulename": "pyerrors.covobs", "qualname": "Covobs.__init__", "kind": "function", "doc": "Initialize Covobs object.
\n\nParameters
\n\n\n- mean (float):\nMean value of the new Obs
\n- cov (list or array):\n2d Covariance matrix or 1d diagonal entries
\n- name (str):\nidentifier for the covariance matrix
\n- pos (int):\nPosition of the variance belonging to mean in cov.\nIs taken to be 1 if cov is 0-dimensional
\n- grad (list or array):\nGradient of the Covobs wrt. the means belonging to cov.
\n
\n", "signature": "(mean, cov, name, pos=None, grad=None)"}, "pyerrors.covobs.Covobs.errsq": {"fullname": "pyerrors.covobs.Covobs.errsq", "modulename": "pyerrors.covobs", "qualname": "Covobs.errsq", "kind": "function", "doc": "Return the variance (= square of the error) of the Covobs
\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.dirac": {"fullname": "pyerrors.dirac", "modulename": "pyerrors.dirac", "kind": "module", "doc": "\n"}, "pyerrors.dirac.epsilon_tensor": {"fullname": "pyerrors.dirac.epsilon_tensor", "modulename": "pyerrors.dirac", "qualname": "epsilon_tensor", "kind": "function", "doc": "Rank-3 epsilon tensor
\n\nBased on https://codegolf.stackexchange.com/a/160375
\n", "signature": "(i, j, k):", "funcdef": "def"}, "pyerrors.dirac.epsilon_tensor_rank4": {"fullname": "pyerrors.dirac.epsilon_tensor_rank4", "modulename": "pyerrors.dirac", "qualname": "epsilon_tensor_rank4", "kind": "function", "doc": "Rank-4 epsilon tensor
\n\nExtension of https://codegolf.stackexchange.com/a/160375
\n", "signature": "(i, j, k, o):", "funcdef": "def"}, "pyerrors.dirac.Grid_gamma": {"fullname": "pyerrors.dirac.Grid_gamma", "modulename": "pyerrors.dirac", "qualname": "Grid_gamma", "kind": "function", "doc": "Returns gamma matrix in Grid labeling.
\n", "signature": "(gamma_tag):", "funcdef": "def"}, "pyerrors.fits": {"fullname": "pyerrors.fits", "modulename": "pyerrors.fits", "kind": "module", "doc": "\n"}, "pyerrors.fits.Fit_result": {"fullname": "pyerrors.fits.Fit_result", "modulename": "pyerrors.fits", "qualname": "Fit_result", "kind": "class", "doc": "Represents fit results.
\n\nAttributes
\n\n\n- fit_parameters (list):\nresults for the individual fit parameters,\nalso accessible via indices.
\n- chisquare_by_dof (float):\nreduced chisquare.
\n- p_value (float):\np-value of the fit
\n- t2_p_value (float):\nHotelling t-squared p-value for correlated fits.
\n
\n", "bases": "collections.abc.Sequence"}, "pyerrors.fits.Fit_result.__init__": {"fullname": "pyerrors.fits.Fit_result.__init__", "modulename": "pyerrors.fits", "qualname": "Fit_result.__init__", "kind": "function", "doc": "\n", "signature": "()"}, "pyerrors.fits.Fit_result.gamma_method": {"fullname": "pyerrors.fits.Fit_result.gamma_method", "modulename": "pyerrors.fits", "qualname": "Fit_result.gamma_method", "kind": "function", "doc": "Apply the gamma method to all fit parameters
\n", "signature": "(self, **kwargs):", "funcdef": "def"}, "pyerrors.fits.Fit_result.gm": {"fullname": "pyerrors.fits.Fit_result.gm", "modulename": "pyerrors.fits", "qualname": "Fit_result.gm", "kind": "function", "doc": "Apply the gamma method to all fit parameters
\n", "signature": "(self, **kwargs):", "funcdef": "def"}, "pyerrors.fits.least_squares": {"fullname": "pyerrors.fits.least_squares", "modulename": "pyerrors.fits", "qualname": "least_squares", "kind": "function", "doc": "Performs a non-linear fit to y = func(x).
\n\nParameters
\n\n\n- x (list):\nlist of floats.
\n- y (list):\nlist of Obs.
\nfunc (object):\nfit function, has to be of the form
\n\n\n
import autograd.numpy as anp\n\ndef func(a, x):\n return a[0] + a[1] * x + a[2] * anp.sinh(x)\n
\n
\n\nFor multiple x values func can be of the form
\n\n\n
def func(a, x):\n (x1, x2) = x\n return a[0] * x1 ** 2 + a[1] * x2\n
\n
\n\nIt is important that all numpy functions refer to autograd.numpy, otherwise the differentiation\nwill not work.
\n- priors (list, optional):\npriors has to be a list with an entry for every parameter in the fit. The entries can either be\nObs (e.g. results from a previous fit) or strings containing a value and an error formatted like\n0.548(23), 500(40) or 0.5(0.4)
\n- silent (bool, optional):\nIf true all output to the console is omitted (default False).
\n- initial_guess (list):\ncan provide an initial guess for the input parameters. Relevant for\nnon-linear fits with many parameters. In case of correlated fits the guess is used to perform\nan uncorrelated fit which then serves as guess for the correlated fit.
\n- method (str, optional):\ncan be used to choose an alternative method for the minimization of chisquare.\nThe possible methods are the ones which can be used for scipy.optimize.minimize and\nmigrad of iminuit. If no method is specified, Levenberg-Marquard is used.\nReliable alternatives are migrad, Powell and Nelder-Mead.
\n- correlated_fit (bool):\nIf True, use the full inverse covariance matrix in the definition of the chisquare cost function.\nFor details about how the covariance matrix is estimated see
pyerrors.obs.covariance
.\nIn practice the correlation matrix is Cholesky decomposed and inverted (instead of the covariance matrix).\nThis procedure should be numerically more stable as the correlation matrix is typically better conditioned (Jacobi preconditioning).\nAt the moment this option only works for prior==None
and when no method
is given. \n- expected_chisquare (bool):\nIf True estimates the expected chisquare which is\ncorrected by effects caused by correlated input data (default False).
\n- resplot (bool):\nIf True, a plot which displays fit, data and residuals is generated (default False).
\n- qqplot (bool):\nIf True, a quantile-quantile plot of the fit result is generated (default False).
\n- num_grad (bool):\nUse numerical differentation instead of automatic differentiation to perform the error propagation (default False).
\n
\n", "signature": "(x, y, func, priors=None, silent=False, **kwargs):", "funcdef": "def"}, "pyerrors.fits.total_least_squares": {"fullname": "pyerrors.fits.total_least_squares", "modulename": "pyerrors.fits", "qualname": "total_least_squares", "kind": "function", "doc": "Performs a non-linear fit to y = func(x) and returns a list of Obs corresponding to the fit parameters.
\n\nParameters
\n\n\n- x (list):\nlist of Obs, or a tuple of lists of Obs
\n- y (list):\nlist of Obs. The dvalues of the Obs are used as x- and yerror for the fit.
\nfunc (object):\nfunc has to be of the form
\n\n\n
import autograd.numpy as anp\n\ndef func(a, x):\n return a[0] + a[1] * x + a[2] * anp.sinh(x)\n
\n
\n\nFor multiple x values func can be of the form
\n\n\n
def func(a, x):\n (x1, x2) = x\n return a[0] * x1 ** 2 + a[1] * x2\n
\n
\n\nIt is important that all numpy functions refer to autograd.numpy, otherwise the differentiation\nwill not work.
\n- silent (bool, optional):\nIf true all output to the console is omitted (default False).
\n- initial_guess (list):\ncan provide an initial guess for the input parameters. Relevant for non-linear\nfits with many parameters.
\n- expected_chisquare (bool):\nIf true prints the expected chisquare which is\ncorrected by effects caused by correlated input data.\nThis can take a while as the full correlation matrix\nhas to be calculated (default False).
\n- num_grad (bool):\nUse numerical differentation instead of automatic differentiation to perform the error propagation (default False).
\n
\n\nNotes
\n\nBased on the orthogonal distance regression module of scipy
\n", "signature": "(x, y, func, silent=False, **kwargs):", "funcdef": "def"}, "pyerrors.fits.fit_lin": {"fullname": "pyerrors.fits.fit_lin", "modulename": "pyerrors.fits", "qualname": "fit_lin", "kind": "function", "doc": "Performs a linear fit to y = n + m * x and returns two Obs n, m.
\n\nParameters
\n\n\n- x (list):\nCan either be a list of floats in which case no xerror is assumed, or\na list of Obs, where the dvalues of the Obs are used as xerror for the fit.
\n- y (list):\nList of Obs, the dvalues of the Obs are used as yerror for the fit.
\n
\n", "signature": "(x, y, **kwargs):", "funcdef": "def"}, "pyerrors.fits.qqplot": {"fullname": "pyerrors.fits.qqplot", "modulename": "pyerrors.fits", "qualname": "qqplot", "kind": "function", "doc": "Generates a quantile-quantile plot of the fit result which can be used to\ncheck if the residuals of the fit are gaussian distributed.
\n", "signature": "(x, o_y, func, p):", "funcdef": "def"}, "pyerrors.fits.residual_plot": {"fullname": "pyerrors.fits.residual_plot", "modulename": "pyerrors.fits", "qualname": "residual_plot", "kind": "function", "doc": "Generates a plot which compares the fit to the data and displays the corresponding residuals
\n", "signature": "(x, y, func, fit_res):", "funcdef": "def"}, "pyerrors.fits.error_band": {"fullname": "pyerrors.fits.error_band", "modulename": "pyerrors.fits", "qualname": "error_band", "kind": "function", "doc": "Returns the error band for an array of sample values x, for given fit function func with optimized parameters beta.
\n", "signature": "(x, func, beta):", "funcdef": "def"}, "pyerrors.fits.ks_test": {"fullname": "pyerrors.fits.ks_test", "modulename": "pyerrors.fits", "qualname": "ks_test", "kind": "function", "doc": "Performs a Kolmogorov\u2013Smirnov test for the p-values of all fit object.
\n\nParameters
\n\n\n- objects (list):\nList of fit results to include in the analysis (optional).
\n
\n", "signature": "(objects=None):", "funcdef": "def"}, "pyerrors.input": {"fullname": "pyerrors.input", "modulename": "pyerrors.input", "kind": "module", "doc": "pyerrors
includes an input
submodule in which input routines and parsers for the output of various numerical programs are contained.
\n\nJackknife samples
\n\nFor comparison with other analysis workflows pyerrors
can also generate jackknife samples from an Obs
object or import jackknife samples into an Obs
object.\nSee pyerrors.obs.Obs.export_jackknife
and pyerrors.obs.import_jackknife
for details.
\n"}, "pyerrors.input.bdio": {"fullname": "pyerrors.input.bdio", "modulename": "pyerrors.input.bdio", "kind": "module", "doc": "\n"}, "pyerrors.input.bdio.read_ADerrors": {"fullname": "pyerrors.input.bdio.read_ADerrors", "modulename": "pyerrors.input.bdio", "qualname": "read_ADerrors", "kind": "function", "doc": "Extract generic MCMC data from a bdio file
\n\nread_ADerrors requires bdio to be compiled into a shared library. This can be achieved by\nadding the flag -fPIC to CC and changing the all target to
\n\nall: bdio.o $(LIBDIR)\n gcc -shared -Wl,-soname,libbdio.so -o $(BUILDDIR)/libbdio.so $(BUILDDIR)/bdio.o\n cp $(BUILDDIR)/libbdio.so $(LIBDIR)/
\n\nParameters
\n\n\n- file_path -- path to the bdio file
\n- bdio_path -- path to the shared bdio library libbdio.so (default ./libbdio.so)
\n
\n", "signature": "(file_path, bdio_path='./libbdio.so', **kwargs):", "funcdef": "def"}, "pyerrors.input.bdio.write_ADerrors": {"fullname": "pyerrors.input.bdio.write_ADerrors", "modulename": "pyerrors.input.bdio", "qualname": "write_ADerrors", "kind": "function", "doc": "Write Obs to a bdio file according to ADerrors conventions
\n\nread_mesons requires bdio to be compiled into a shared library. This can be achieved by\nadding the flag -fPIC to CC and changing the all target to
\n\nall: bdio.o $(LIBDIR)\n gcc -shared -Wl,-soname,libbdio.so -o $(BUILDDIR)/libbdio.so $(BUILDDIR)/bdio.o\n cp $(BUILDDIR)/libbdio.so $(LIBDIR)/
\n\nParameters
\n\n\n- file_path -- path to the bdio file
\n- bdio_path -- path to the shared bdio library libbdio.so (default ./libbdio.so)
\n
\n", "signature": "(obs_list, file_path, bdio_path='./libbdio.so', **kwargs):", "funcdef": "def"}, "pyerrors.input.bdio.read_mesons": {"fullname": "pyerrors.input.bdio.read_mesons", "modulename": "pyerrors.input.bdio", "qualname": "read_mesons", "kind": "function", "doc": "Extract mesons data from a bdio file and return it as a dictionary
\n\nThe dictionary can be accessed with a tuple consisting of (type, source_position, kappa1, kappa2)
\n\nread_mesons requires bdio to be compiled into a shared library. This can be achieved by\nadding the flag -fPIC to CC and changing the all target to
\n\nall: bdio.o $(LIBDIR)\n gcc -shared -Wl,-soname,libbdio.so -o $(BUILDDIR)/libbdio.so $(BUILDDIR)/bdio.o\n cp $(BUILDDIR)/libbdio.so $(LIBDIR)/
\n\nParameters
\n\n\n- file_path (str):\npath to the bdio file
\n- bdio_path (str):\npath to the shared bdio library libbdio.so (default ./libbdio.so)
\n- start (int):\nThe first configuration to be read (default 1)
\n- stop (int):\nThe last configuration to be read (default None)
\n- step (int):\nFixed step size between two measurements (default 1)
\n- alternative_ensemble_name (str):\nManually overwrite ensemble name
\n
\n", "signature": "(file_path, bdio_path='./libbdio.so', **kwargs):", "funcdef": "def"}, "pyerrors.input.bdio.read_dSdm": {"fullname": "pyerrors.input.bdio.read_dSdm", "modulename": "pyerrors.input.bdio", "qualname": "read_dSdm", "kind": "function", "doc": "Extract dSdm data from a bdio file and return it as a dictionary
\n\nThe dictionary can be accessed with a tuple consisting of (type, kappa)
\n\nread_dSdm requires bdio to be compiled into a shared library. This can be achieved by\nadding the flag -fPIC to CC and changing the all target to
\n\nall: bdio.o $(LIBDIR)\n gcc -shared -Wl,-soname,libbdio.so -o $(BUILDDIR)/libbdio.so $(BUILDDIR)/bdio.o\n cp $(BUILDDIR)/libbdio.so $(LIBDIR)/
\n\nParameters
\n\n\n- file_path (str):\npath to the bdio file
\n- bdio_path (str):\npath to the shared bdio library libbdio.so (default ./libbdio.so)
\n- start (int):\nThe first configuration to be read (default 1)
\n- stop (int):\nThe last configuration to be read (default None)
\n- step (int):\nFixed step size between two measurements (default 1)
\n- alternative_ensemble_name (str):\nManually overwrite ensemble name
\n
\n", "signature": "(file_path, bdio_path='./libbdio.so', **kwargs):", "funcdef": "def"}, "pyerrors.input.dobs": {"fullname": "pyerrors.input.dobs", "modulename": "pyerrors.input.dobs", "kind": "module", "doc": "\n"}, "pyerrors.input.dobs.create_pobs_string": {"fullname": "pyerrors.input.dobs.create_pobs_string", "modulename": "pyerrors.input.dobs", "qualname": "create_pobs_string", "kind": "function", "doc": "Export a list of Obs or structures containing Obs to an xml string\naccording to the Zeuthen pobs format.
\n\nTags are not written or recovered automatically. The separator | is removed from the replica names.
\n\nParameters
\n\n\n- obsl (list):\nList of Obs that will be exported.\nThe Obs inside a structure have to be defined on the same ensemble.
\n- name (str):\nThe name of the observable.
\n- spec (str):\nOptional string that describes the contents of the file.
\n- origin (str):\nSpecify where the data has its origin.
\n- symbol (list):\nA list of symbols that describe the observables to be written. May be empty.
\n- enstag (str):\nEnstag that is written to pobs. If None, the ensemble name is used.
\n
\n", "signature": "(obsl, name, spec='', origin='', symbol=[], enstag=None):", "funcdef": "def"}, "pyerrors.input.dobs.write_pobs": {"fullname": "pyerrors.input.dobs.write_pobs", "modulename": "pyerrors.input.dobs", "qualname": "write_pobs", "kind": "function", "doc": "Export a list of Obs or structures containing Obs to a .xml.gz file\naccording to the Zeuthen pobs format.
\n\nTags are not written or recovered automatically. The separator | is removed from the replica names.
\n\nParameters
\n\n\n- obsl (list):\nList of Obs that will be exported.\nThe Obs inside a structure have to be defined on the same ensemble.
\n- fname (str):\nFilename of the output file.
\n- name (str):\nThe name of the observable.
\n- spec (str):\nOptional string that describes the contents of the file.
\n- origin (str):\nSpecify where the data has its origin.
\n- symbol (list):\nA list of symbols that describe the observables to be written. May be empty.
\n- enstag (str):\nEnstag that is written to pobs. If None, the ensemble name is used.
\n- gz (bool):\nIf True, the output is a gzipped xml. If False, the output is an xml file.
\n
\n", "signature": "(\tobsl,\tfname,\tname,\tspec='',\torigin='',\tsymbol=[],\tenstag=None,\tgz=True):", "funcdef": "def"}, "pyerrors.input.dobs.read_pobs": {"fullname": "pyerrors.input.dobs.read_pobs", "modulename": "pyerrors.input.dobs", "qualname": "read_pobs", "kind": "function", "doc": "Import a list of Obs from an xml.gz file in the Zeuthen pobs format.
\n\nTags are not written or recovered automatically.
\n\nParameters
\n\n\n- fname (str):\nFilename of the input file.
\n- full_output (bool):\nIf True, a dict containing auxiliary information and the data is returned.\nIf False, only the data is returned as list.
\n- separatior_insertion (str or int):\nstr: replace all occurences of \"separator_insertion\" within the replica names\nby \"|%s\" % (separator_insertion) when constructing the names of the replica.\nint: Insert the separator \"|\" at the position given by separator_insertion.\nNone (default): Replica names remain unchanged.
\n
\n", "signature": "(fname, full_output=False, gz=True, separator_insertion=None):", "funcdef": "def"}, "pyerrors.input.dobs.import_dobs_string": {"fullname": "pyerrors.input.dobs.import_dobs_string", "modulename": "pyerrors.input.dobs", "qualname": "import_dobs_string", "kind": "function", "doc": "Import a list of Obs from a string in the Zeuthen dobs format.
\n\nTags are not written or recovered automatically.
\n\nParameters
\n\n\n- content (str):\nXML string containing the data
\n- noemtpy (bool):\nIf True, ensembles with no contribution to the Obs are not included.\nIf False, ensembles are included as written in the file, possibly with vanishing entries.
\n- full_output (bool):\nIf True, a dict containing auxiliary information and the data is returned.\nIf False, only the data is returned as list.
\n- separatior_insertion (str, int or bool):\nstr: replace all occurences of \"separator_insertion\" within the replica names\nby \"|%s\" % (separator_insertion) when constructing the names of the replica.\nint: Insert the separator \"|\" at the position given by separator_insertion.\nTrue (default): separator \"|\" is inserted after len(ensname), assuming that the\nensemble name is a prefix to the replica name.\nNone or False: No separator is inserted.
\n
\n", "signature": "(content, noempty=False, full_output=False, separator_insertion=True):", "funcdef": "def"}, "pyerrors.input.dobs.read_dobs": {"fullname": "pyerrors.input.dobs.read_dobs", "modulename": "pyerrors.input.dobs", "qualname": "read_dobs", "kind": "function", "doc": "Import a list of Obs from an xml.gz file in the Zeuthen dobs format.
\n\nTags are not written or recovered automatically.
\n\nParameters
\n\n\n- fname (str):\nFilename of the input file.
\n- noemtpy (bool):\nIf True, ensembles with no contribution to the Obs are not included.\nIf False, ensembles are included as written in the file.
\n- full_output (bool):\nIf True, a dict containing auxiliary information and the data is returned.\nIf False, only the data is returned as list.
\n- gz (bool):\nIf True, assumes that data is gzipped. If False, assumes XML file.
\n- separatior_insertion (str, int or bool):\nstr: replace all occurences of \"separator_insertion\" within the replica names\nby \"|%s\" % (separator_insertion) when constructing the names of the replica.\nint: Insert the separator \"|\" at the position given by separator_insertion.\nTrue (default): separator \"|\" is inserted after len(ensname), assuming that the\nensemble name is a prefix to the replica name.\nNone or False: No separator is inserted.
\n
\n", "signature": "(\tfname,\tnoempty=False,\tfull_output=False,\tgz=True,\tseparator_insertion=True):", "funcdef": "def"}, "pyerrors.input.dobs.create_dobs_string": {"fullname": "pyerrors.input.dobs.create_dobs_string", "modulename": "pyerrors.input.dobs", "qualname": "create_dobs_string", "kind": "function", "doc": "Generate the string for the export of a list of Obs or structures containing Obs\nto a .xml.gz file according to the Zeuthen dobs format.
\n\nTags are not written or recovered automatically. The separator |is removed from the replica names.
\n\nParameters
\n\n\n- obsl (list):\nList of Obs that will be exported.\nThe Obs inside a structure do not have to be defined on the same set of configurations,\nbut the storage requirement is increased, if this is not the case.
\n- name (str):\nThe name of the observable.
\n- spec (str):\nOptional string that describes the contents of the file.
\n- origin (str):\nSpecify where the data has its origin.
\n- symbol (list):\nA list of symbols that describe the observables to be written. May be empty.
\n- who (str):\nProvide the name of the person that exports the data.
\n- enstags (dict):\nProvide alternative enstag for ensembles in the form enstags = {ename: enstag}\nOtherwise, the ensemble name is used.
\n
\n", "signature": "(\tobsl,\tname,\tspec='dobs v1.0',\torigin='',\tsymbol=[],\twho=None,\tenstags=None):", "funcdef": "def"}, "pyerrors.input.dobs.write_dobs": {"fullname": "pyerrors.input.dobs.write_dobs", "modulename": "pyerrors.input.dobs", "qualname": "write_dobs", "kind": "function", "doc": "Export a list of Obs or structures containing Obs to a .xml.gz file\naccording to the Zeuthen dobs format.
\n\nTags are not written or recovered automatically. The separator | is removed from the replica names.
\n\nParameters
\n\n\n- obsl (list):\nList of Obs that will be exported.\nThe Obs inside a structure do not have to be defined on the same set of configurations,\nbut the storage requirement is increased, if this is not the case.
\n- fname (str):\nFilename of the output file.
\n- name (str):\nThe name of the observable.
\n- spec (str):\nOptional string that describes the contents of the file.
\n- origin (str):\nSpecify where the data has its origin.
\n- symbol (list):\nA list of symbols that describe the observables to be written. May be empty.
\n- who (str):\nProvide the name of the person that exports the data.
\n- enstags (dict):\nProvide alternative enstag for ensembles in the form enstags = {ename: enstag}\nOtherwise, the ensemble name is used.
\n- gz (bool):\nIf True, the output is a gzipped XML. If False, the output is a XML file.
\n
\n", "signature": "(\tobsl,\tfname,\tname,\tspec='dobs v1.0',\torigin='',\tsymbol=[],\twho=None,\tenstags=None,\tgz=True):", "funcdef": "def"}, "pyerrors.input.hadrons": {"fullname": "pyerrors.input.hadrons", "modulename": "pyerrors.input.hadrons", "kind": "module", "doc": "\n"}, "pyerrors.input.hadrons.read_meson_hd5": {"fullname": "pyerrors.input.hadrons.read_meson_hd5", "modulename": "pyerrors.input.hadrons", "qualname": "read_meson_hd5", "kind": "function", "doc": "Read hadrons meson hdf5 file and extract the meson labeled 'meson'
\n\nParameters
\n\n\n- path (str):\npath to the files to read
\n- filestem (str):\nnamestem of the files to read
\n- ens_id (str):\nname of the ensemble, required for internal bookkeeping
\n- meson (str):\nlabel of the meson to be extracted, standard value meson_0 which\ncorresponds to the pseudoscalar pseudoscalar two-point function.
\n- gammas (tuple of strings):\nInstrad of a meson label one can also provide a tuple of two strings\nindicating the gamma matrices at source and sink.\n(\"Gamma5\", \"Gamma5\") corresponds to the pseudoscalar pseudoscalar\ntwo-point function. The gammas argument dominateds over meson.
\n- idl (range):\nIf specified only configurations in the given range are read in.
\n
\n", "signature": "(path, filestem, ens_id, meson='meson_0', idl=None, gammas=None):", "funcdef": "def"}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"fullname": "pyerrors.input.hadrons.read_DistillationContraction_hd5", "modulename": "pyerrors.input.hadrons", "qualname": "read_DistillationContraction_hd5", "kind": "function", "doc": "Read hadrons DistillationContraction hdf5 files in given directory structure
\n\nParameters
\n\n\n- path (str):\npath to the directories to read
\n- ens_id (str):\nname of the ensemble, required for internal bookkeeping
\n- diagrams (list):\nList of strings of the diagrams to extract, e.g. [\"direct\", \"box\", \"cross\"].
\n- idl (range):\nIf specified only configurations in the given range are read in.
\n
\n", "signature": "(path, ens_id, diagrams=['direct'], idl=None):", "funcdef": "def"}, "pyerrors.input.hadrons.Npr_matrix": {"fullname": "pyerrors.input.hadrons.Npr_matrix", "modulename": "pyerrors.input.hadrons", "qualname": "Npr_matrix", "kind": "class", "doc": "ndarray(shape, dtype=float, buffer=None, offset=0,\n strides=None, order=None)
\n\nAn array object represents a multidimensional, homogeneous array\nof fixed-size items. An associated data-type object describes the\nformat of each element in the array (its byte-order, how many bytes it\noccupies in memory, whether it is an integer, a floating point number,\nor something else, etc.)
\n\nArrays should be constructed using array
, zeros
or empty
(refer\nto the See Also section below). The parameters given here refer to\na low-level method (ndarray(...)
) for instantiating an array.
\n\nFor more information, refer to the numpy
module and examine the\nmethods and attributes of an array.
\n\nParameters
\n\n\n- (for the __new__ method; see Notes below)
\n- shape (tuple of ints):\nShape of created array.
\n- dtype (data-type, optional):\nAny object that can be interpreted as a numpy data type.
\n- buffer (object exposing buffer interface, optional):\nUsed to fill the array with data.
\n- offset (int, optional):\nOffset of array data in buffer.
\n- strides (tuple of ints, optional):\nStrides of data in memory.
\n- order ({'C', 'F'}, optional):\nRow-major (C-style) or column-major (Fortran-style) order.
\n
\n\nAttributes
\n\n\n- T (ndarray):\nTranspose of the array.
\n- data (buffer):\nThe array's elements, in memory.
\n- dtype (dtype object):\nDescribes the format of the elements in the array.
\n- flags (dict):\nDictionary containing information related to memory use, e.g.,\n'C_CONTIGUOUS', 'OWNDATA', 'WRITEABLE', etc.
\n- flat (numpy.flatiter object):\nFlattened version of the array as an iterator. The iterator\nallows assignments, e.g.,
x.flat = 3
(See ndarray.flat
for\nassignment examples; TODO). \n- imag (ndarray):\nImaginary part of the array.
\n- real (ndarray):\nReal part of the array.
\n- size (int):\nNumber of elements in the array.
\n- itemsize (int):\nThe memory use of each array element in bytes.
\n- nbytes (int):\nThe total number of bytes required to store the array data,\ni.e.,
itemsize * size
. \n- ndim (int):\nThe array's number of dimensions.
\n- shape (tuple of ints):\nShape of the array.
\n- strides (tuple of ints):\nThe step-size required to move from one element to the next in\nmemory. For example, a contiguous
(3, 4)
array of type\nint16
in C-order has strides (8, 2)
. This implies that\nto move from element to element in memory requires jumps of 2 bytes.\nTo move from row-to-row, one needs to jump 8 bytes at a time\n(2 * 4
). \n- ctypes (ctypes object):\nClass containing properties of the array needed for interaction\nwith ctypes.
\n- base (ndarray):\nIf the array is a view into another array, that array is its
base
\n(unless that array is also a view). The base
array is where the\narray data is actually stored. \n
\n\nSee Also
\n\narray
: Construct an array.
\nzeros
: Create an array, each element of which is zero.
\nempty
: Create an array, but leave its allocated memory unchanged (i.e.,\nit contains \"garbage\").
\ndtype
: Create a data-type.
\nnumpy.typing.NDArray
: An ndarray alias :term:generic <generic type>
\nw.r.t. its dtype.type <numpy.dtype.type>
.
\n\nNotes
\n\nThere are two modes of creating an array using __new__
:
\n\n\n- If
buffer
is None, then only shape
, dtype
, and order
\nare used. \n- If
buffer
is an object exposing the buffer interface, then\nall keywords are interpreted. \n
\n\nNo __init__
method is needed because the array is fully initialized\nafter the __new__
method.
\n\nExamples
\n\nThese examples illustrate the low-level ndarray
constructor. Refer\nto the See Also
section above for easier ways of constructing an\nndarray.
\n\nFirst mode, buffer
is None:
\n\n\n
>>> np.ndarray(shape=(2,2), dtype=float, order='F')\narray([[0.0e+000, 0.0e+000], # random\n [ nan, 2.5e-323]])\n
\n
\n\nSecond mode:
\n\n\n
>>> np.ndarray((2,), buffer=np.array([1,2,3]),\n... offset=np.int_().itemsize,\n... dtype=int) # offset = 1*itemsize, i.e. skip first element\narray([2, 3])\n
\n
\n", "bases": "numpy.ndarray"}, "pyerrors.input.hadrons.Npr_matrix.__init__": {"fullname": "pyerrors.input.hadrons.Npr_matrix.__init__", "modulename": "pyerrors.input.hadrons", "qualname": "Npr_matrix.__init__", "kind": "function", "doc": "\n", "signature": "()"}, "pyerrors.input.hadrons.Npr_matrix.g5H": {"fullname": "pyerrors.input.hadrons.Npr_matrix.g5H", "modulename": "pyerrors.input.hadrons", "qualname": "Npr_matrix.g5H", "kind": "variable", "doc": "Gamma_5 hermitean conjugate
\n\nUses the fact that the propagator is gamma5 hermitean, so just the\nin and out momenta of the propagator are exchanged.
\n"}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"fullname": "pyerrors.input.hadrons.read_ExternalLeg_hd5", "modulename": "pyerrors.input.hadrons", "qualname": "read_ExternalLeg_hd5", "kind": "function", "doc": "Read hadrons ExternalLeg hdf5 file and output an array of CObs
\n\nParameters
\n\n\n- path (str):\npath to the files to read
\n- filestem (str):\nnamestem of the files to read
\n- ens_id (str):\nname of the ensemble, required for internal bookkeeping
\n- idl (range):\nIf specified only configurations in the given range are read in.
\n
\n", "signature": "(path, filestem, ens_id, idl=None):", "funcdef": "def"}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"fullname": "pyerrors.input.hadrons.read_Bilinear_hd5", "modulename": "pyerrors.input.hadrons", "qualname": "read_Bilinear_hd5", "kind": "function", "doc": "Read hadrons Bilinear hdf5 file and output an array of CObs
\n\nParameters
\n\n\n- path (str):\npath to the files to read
\n- filestem (str):\nnamestem of the files to read
\n- ens_id (str):\nname of the ensemble, required for internal bookkeeping
\n- idl (range):\nIf specified only configurations in the given range are read in.
\n
\n", "signature": "(path, filestem, ens_id, idl=None):", "funcdef": "def"}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"fullname": "pyerrors.input.hadrons.read_Fourquark_hd5", "modulename": "pyerrors.input.hadrons", "qualname": "read_Fourquark_hd5", "kind": "function", "doc": "Read hadrons FourquarkFullyConnected hdf5 file and output an array of CObs
\n\nParameters
\n\n\n- path (str):\npath to the files to read
\n- filestem (str):\nnamestem of the files to read
\n- ens_id (str):\nname of the ensemble, required for internal bookkeeping
\n- idl (range):\nIf specified only configurations in the given range are read in.
\n- vertices (list):\nVertex functions to be extracted.
\n
\n", "signature": "(path, filestem, ens_id, idl=None, vertices=['VA', 'AV']):", "funcdef": "def"}, "pyerrors.input.json": {"fullname": "pyerrors.input.json", "modulename": "pyerrors.input.json", "kind": "module", "doc": "\n"}, "pyerrors.input.json.create_json_string": {"fullname": "pyerrors.input.json.create_json_string", "modulename": "pyerrors.input.json", "qualname": "create_json_string", "kind": "function", "doc": "Generate the string for the export of a list of Obs or structures containing Obs\nto a .json(.gz) file
\n\nParameters
\n\n\n- ol (list):\nList of objects that will be exported. At the moment, these objects can be\neither of: Obs, list, numpy.ndarray, Corr.\nAll Obs inside a structure have to be defined on the same set of configurations.
\n- description (str):\nOptional string that describes the contents of the json file.
\n- indent (int):\nSpecify the indentation level of the json file. None or 0 is permissible and\nsaves disk space.
\n
\n", "signature": "(ol, description='', indent=1):", "funcdef": "def"}, "pyerrors.input.json.dump_to_json": {"fullname": "pyerrors.input.json.dump_to_json", "modulename": "pyerrors.input.json", "qualname": "dump_to_json", "kind": "function", "doc": "Export a list of Obs or structures containing Obs to a .json(.gz) file
\n\nParameters
\n\n\n- ol (list):\nList of objects that will be exported. At the moment, these objects can be\neither of: Obs, list, numpy.ndarray, Corr.\nAll Obs inside a structure have to be defined on the same set of configurations.
\n- fname (str):\nFilename of the output file.
\n- description (str):\nOptional string that describes the contents of the json file.
\n- indent (int):\nSpecify the indentation level of the json file. None or 0 is permissible and\nsaves disk space.
\n- gz (bool):\nIf True, the output is a gzipped json. If False, the output is a json file.
\n
\n", "signature": "(ol, fname, description='', indent=1, gz=True):", "funcdef": "def"}, "pyerrors.input.json.import_json_string": {"fullname": "pyerrors.input.json.import_json_string", "modulename": "pyerrors.input.json", "qualname": "import_json_string", "kind": "function", "doc": "Reconstruct a list of Obs or structures containing Obs from a json string.
\n\nThe following structures are supported: Obs, list, numpy.ndarray, Corr\nIf the list contains only one element, it is unpacked from the list.
\n\nParameters
\n\n\n- json_string (str):\njson string containing the data.
\n- verbose (bool):\nPrint additional information that was written to the file.
\n- full_output (bool):\nIf True, a dict containing auxiliary information and the data is returned.\nIf False, only the data is returned.
\n
\n", "signature": "(json_string, verbose=True, full_output=False):", "funcdef": "def"}, "pyerrors.input.json.load_json": {"fullname": "pyerrors.input.json.load_json", "modulename": "pyerrors.input.json", "qualname": "load_json", "kind": "function", "doc": "Import a list of Obs or structures containing Obs from a .json(.gz) file.
\n\nThe following structures are supported: Obs, list, numpy.ndarray, Corr\nIf the list contains only one element, it is unpacked from the list.
\n\nParameters
\n\n\n- fname (str):\nFilename of the input file.
\n- verbose (bool):\nPrint additional information that was written to the file.
\n- gz (bool):\nIf True, assumes that data is gzipped. If False, assumes JSON file.
\n- full_output (bool):\nIf True, a dict containing auxiliary information and the data is returned.\nIf False, only the data is returned.
\n
\n", "signature": "(fname, verbose=True, gz=True, full_output=False):", "funcdef": "def"}, "pyerrors.input.json.dump_dict_to_json": {"fullname": "pyerrors.input.json.dump_dict_to_json", "modulename": "pyerrors.input.json", "qualname": "dump_dict_to_json", "kind": "function", "doc": "Export a dict of Obs or structures containing Obs to a .json(.gz) file
\n\nParameters
\n\n\n- od (dict):\nDict of JSON valid structures and objects that will be exported.\nAt the moment, these objects can be either of: Obs, list, numpy.ndarray, Corr.\nAll Obs inside a structure have to be defined on the same set of configurations.
\n- fname (str):\nFilename of the output file.
\n- description (str):\nOptional string that describes the contents of the json file.
\n- indent (int):\nSpecify the indentation level of the json file. None or 0 is permissible and\nsaves disk space.
\n- reps (str):\nSpecify the structure of the placeholder in exported dict to be reps[0-9]+.
\n- gz (bool):\nIf True, the output is a gzipped json. If False, the output is a json file.
\n
\n", "signature": "(od, fname, description='', indent=1, reps='DICTOBS', gz=True):", "funcdef": "def"}, "pyerrors.input.json.load_json_dict": {"fullname": "pyerrors.input.json.load_json_dict", "modulename": "pyerrors.input.json", "qualname": "load_json_dict", "kind": "function", "doc": "Import a dict of Obs or structures containing Obs from a .json(.gz) file.
\n\nThe following structures are supported: Obs, list, numpy.ndarray, Corr
\n\nParameters
\n\n\n- fname (str):\nFilename of the input file.
\n- verbose (bool):\nPrint additional information that was written to the file.
\n- gz (bool):\nIf True, assumes that data is gzipped. If False, assumes JSON file.
\n- full_output (bool):\nIf True, a dict containing auxiliary information and the data is returned.\nIf False, only the data is returned.
\n- reps (str):\nSpecify the structure of the placeholder in imported dict to be reps[0-9]+.
\n
\n", "signature": "(fname, verbose=True, gz=True, full_output=False, reps='DICTOBS'):", "funcdef": "def"}, "pyerrors.input.misc": {"fullname": "pyerrors.input.misc", "modulename": "pyerrors.input.misc", "kind": "module", "doc": "\n"}, "pyerrors.input.misc.read_pbp": {"fullname": "pyerrors.input.misc.read_pbp", "modulename": "pyerrors.input.misc", "qualname": "read_pbp", "kind": "function", "doc": "Read pbp format from given folder structure. Returns a list of length nrw
\n\nParameters
\n\n\n- r_start (list):\nlist which contains the first config to be read for each replicum
\n- r_stop (list):\nlist which contains the last config to be read for each replicum
\n
\n", "signature": "(path, prefix, **kwargs):", "funcdef": "def"}, "pyerrors.input.openQCD": {"fullname": "pyerrors.input.openQCD", "modulename": "pyerrors.input.openQCD", "kind": "module", "doc": "\n"}, "pyerrors.input.openQCD.read_rwms": {"fullname": "pyerrors.input.openQCD.read_rwms", "modulename": "pyerrors.input.openQCD", "qualname": "read_rwms", "kind": "function", "doc": "Read rwms format from given folder structure. Returns a list of length nrw
\n\nParameters
\n\n\n- path (str):\npath that contains the data files
\n- prefix (str):\nall files in path that start with prefix are considered as input files.\nMay be used together postfix to consider only special file endings.\nPrefix is ignored, if the keyword 'files' is used.
\n- version (str):\nversion of openQCD, default 2.0
\n- names (list):\nlist of names that is assigned to the data according according\nto the order in the file list. Use careful, if you do not provide file names!
\n- r_start (list):\nlist which contains the first config to be read for each replicum
\n- r_stop (list):\nlist which contains the last config to be read for each replicum
\n- r_step (int):\ninteger that defines a fixed step size between two measurements (in units of configs)\nIf not given, r_step=1 is assumed.
\n- postfix (str):\npostfix of the file to read, e.g. '.ms1' for openQCD-files
\n- files (list):\nlist which contains the filenames to be read. No automatic detection of\nfiles performed if given.
\n- print_err (bool):\nPrint additional information that is useful for debugging.
\n
\n", "signature": "(path, prefix, version='2.0', names=None, **kwargs):", "funcdef": "def"}, "pyerrors.input.openQCD.extract_t0": {"fullname": "pyerrors.input.openQCD.extract_t0", "modulename": "pyerrors.input.openQCD", "qualname": "extract_t0", "kind": "function", "doc": "Extract t0 from given .ms.dat files. Returns t0 as Obs.
\n\nIt is assumed that all boundary effects have\nsufficiently decayed at x0=xmin.\nThe data around the zero crossing of t^2 - 0.3\nis fitted with a linear function\nfrom which the exact root is extracted.
\n\nIt is assumed that one measurement is performed for each config.\nIf this is not the case, the resulting idl, as well as the handling\nof r_start, r_stop and r_step is wrong and the user has to correct\nthis in the resulting observable.
\n\nParameters
\n\n\n- path (str):\nPath to .ms.dat files
\n- prefix (str):\nEnsemble prefix
\n- dtr_read (int):\nDetermines how many trajectories should be skipped\nwhen reading the ms.dat files.\nCorresponds to dtr_cnfg / dtr_ms in the openQCD input file.
\n- xmin (int):\nFirst timeslice where the boundary\neffects have sufficiently decayed.
\n- spatial_extent (int):\nspatial extent of the lattice, required for normalization.
\n- fit_range (int):\nNumber of data points left and right of the zero\ncrossing to be included in the linear fit. (Default: 5)
\n- r_start (list):\nlist which contains the first config to be read for each replicum.
\n- r_stop (list):\nlist which contains the last config to be read for each replicum.
\n- r_step (int):\ninteger that defines a fixed step size between two measurements (in units of configs)\nIf not given, r_step=1 is assumed.
\n- plaquette (bool):\nIf true extract the plaquette estimate of t0 instead.
\n- names (list):\nlist of names that is assigned to the data according according\nto the order in the file list. Use careful, if you do not provide file names!
\n- files (list):\nlist which contains the filenames to be read. No automatic detection of\nfiles performed if given.
\n- plot_fit (bool):\nIf true, the fit for the extraction of t0 is shown together with the data.
\n- assume_thermalization (bool):\nIf True: If the first record divided by the distance between two measurements is larger than\n1, it is assumed that this is due to thermalization and the first measurement belongs\nto the first config (default).\nIf False: The config numbers are assumed to be traj_number // difference
\n
\n", "signature": "(path, prefix, dtr_read, xmin, spatial_extent, fit_range=5, **kwargs):", "funcdef": "def"}, "pyerrors.input.openQCD.read_qtop": {"fullname": "pyerrors.input.openQCD.read_qtop", "modulename": "pyerrors.input.openQCD", "qualname": "read_qtop", "kind": "function", "doc": "Read the topologial charge based on openQCD gradient flow measurements.
\n\nParameters
\n\n\n- path (str):\npath of the measurement files
\n- prefix (str):\nprefix of the measurement files, e.g. _id0_r0.ms.dat.\nIgnored if file names are passed explicitly via keyword files.
\n- c (double):\nSmearing radius in units of the lattice extent, c = sqrt(8 t0) / L.
\n- dtr_cnfg (int):\n(optional) parameter that specifies the number of measurements\nbetween two configs.\nIf it is not set, the distance between two measurements\nin the file is assumed to be the distance between two configurations.
\n- steps (int):\n(optional) Distance between two configurations in units of trajectories /\n cycles. Assumed to be the distance between two measurements * dtr_cnfg if not given
\n- version (str):\nEither openQCD or sfqcd, depending on the data.
\n- L (int):\nspatial length of the lattice in L/a.\nHAS to be set if version != sfqcd, since openQCD does not provide\nthis in the header
\n- r_start (list):\nlist which contains the first config to be read for each replicum.
\n- r_stop (list):\nlist which contains the last config to be read for each replicum.
\n- files (list):\nspecify the exact files that need to be read\nfrom path, practical if e.g. only one replicum is needed
\n- postfix (str):\npostfix of the file to read, e.g. '.gfms.dat' for openQCD-files
\n- names (list):\nAlternative labeling for replicas/ensembles.\nHas to have the appropriate length.
\n- Zeuthen_flow (bool):\n(optional) If True, the Zeuthen flow is used for Qtop. Only possible\nfor version=='sfqcd' If False, the Wilson flow is used.
\n- integer_charge (bool):\nIf True, the charge is rounded towards the nearest integer on each config.
\n
\n", "signature": "(path, prefix, c, dtr_cnfg=1, version='openQCD', **kwargs):", "funcdef": "def"}, "pyerrors.input.openQCD.read_gf_coupling": {"fullname": "pyerrors.input.openQCD.read_gf_coupling", "modulename": "pyerrors.input.openQCD", "qualname": "read_gf_coupling", "kind": "function", "doc": "Read the gradient flow coupling based on sfqcd gradient flow measurements. See 1607.06423 for details.
\n\nNote: The current implementation only works for c=0.3 and T=L. The definition of the coupling in 1607.06423 requires projection to topological charge zero which is not done within this function but has to be performed in a separate step.
\n\nParameters
\n\n\n- path (str):\npath of the measurement files
\n- prefix (str):\nprefix of the measurement files, e.g. _id0_r0.ms.dat.\nIgnored if file names are passed explicitly via keyword files.
\n- c (double):\nSmearing radius in units of the lattice extent, c = sqrt(8 t0) / L.
\n- dtr_cnfg (int):\n(optional) parameter that specifies the number of measurements\nbetween two configs.\nIf it is not set, the distance between two measurements\nin the file is assumed to be the distance between two configurations.
\n- steps (int):\n(optional) Distance between two configurations in units of trajectories /\n cycles. Assumed to be the distance between two measurements * dtr_cnfg if not given
\n- r_start (list):\nlist which contains the first config to be read for each replicum.
\n- r_stop (list):\nlist which contains the last config to be read for each replicum.
\n- files (list):\nspecify the exact files that need to be read\nfrom path, practical if e.g. only one replicum is needed
\n- names (list):\nAlternative labeling for replicas/ensembles.\nHas to have the appropriate length.
\n- postfix (str):\npostfix of the file to read, e.g. '.gfms.dat' for openQCD-files
\n- Zeuthen_flow (bool):\n(optional) If True, the Zeuthen flow is used for the coupling. If False, the Wilson flow is used.
\n
\n", "signature": "(path, prefix, c, dtr_cnfg=1, Zeuthen_flow=True, **kwargs):", "funcdef": "def"}, "pyerrors.input.openQCD.qtop_projection": {"fullname": "pyerrors.input.openQCD.qtop_projection", "modulename": "pyerrors.input.openQCD", "qualname": "qtop_projection", "kind": "function", "doc": "Returns the projection to the topological charge sector defined by target.
\n\nParameters
\n\n\n- path (Obs):\nTopological charge.
\n- target (int):\nSpecifies the topological sector to be reweighted to (default 0)
\n
\n", "signature": "(qtop, target=0):", "funcdef": "def"}, "pyerrors.input.openQCD.read_qtop_sector": {"fullname": "pyerrors.input.openQCD.read_qtop_sector", "modulename": "pyerrors.input.openQCD", "qualname": "read_qtop_sector", "kind": "function", "doc": "Constructs reweighting factors to a specified topological sector.
\n\nParameters
\n\n\n- path (str):\npath of the measurement files
\n- prefix (str):\nprefix of the measurement files, e.g. _id0_r0.ms.dat
\n- c (double):\nSmearing radius in units of the lattice extent, c = sqrt(8 t0) / L
\n- target (int):\nSpecifies the topological sector to be reweighted to (default 0)
\n- dtr_cnfg (int):\n(optional) parameter that specifies the number of trajectories\nbetween two configs.\nif it is not set, the distance between two measurements\nin the file is assumed to be the distance between two configurations.
\n- steps (int):\n(optional) Distance between two configurations in units of trajectories /\n cycles. Assumed to be the distance between two measurements * dtr_cnfg if not given
\n- version (str):\nversion string of the openQCD (sfqcd) version used to create\nthe ensemble. Default is 2.0. May also be set to sfqcd.
\n- L (int):\nspatial length of the lattice in L/a.\nHAS to be set if version != sfqcd, since openQCD does not provide\nthis in the header
\n- r_start (list):\noffset of the first ensemble, making it easier to match\nlater on with other Obs
\n- r_stop (list):\nlast configurations that need to be read (per replicum)
\n- files (list):\nspecify the exact files that need to be read\nfrom path, practical if e.g. only one replicum is needed
\n- names (list):\nAlternative labeling for replicas/ensembles.\nHas to have the appropriate length
\n- Zeuthen_flow (bool):\n(optional) If True, the Zeuthen flow is used for Qtop. Only possible\nfor version=='sfqcd' If False, the Wilson flow is used.
\n
\n", "signature": "(path, prefix, c, target=0, **kwargs):", "funcdef": "def"}, "pyerrors.input.openQCD.read_ms5_xsf": {"fullname": "pyerrors.input.openQCD.read_ms5_xsf", "modulename": "pyerrors.input.openQCD", "qualname": "read_ms5_xsf", "kind": "function", "doc": "Read data from files in the specified directory with the specified prefix and quark combination extension, and return a Corr
object containing the data.
\n\nParameters
\n\n\n\nReturns
\n\n\n- Corr: A complex valued
Corr
object containing the data read from the files. In case of boudary to bulk correlators. \n- or
\n- CObs: A complex valued
CObs
object containing the data read from the files. In case of boudary to boundary correlators. \n
\n\nRaises
\n\n\n- FileNotFoundError: If no files matching the specified prefix and quark combination extension are found in the specified directory.
\n- IOError: If there is an error reading a file.
\n- struct.error: If there is an error unpacking binary data.
\n
\n", "signature": "(path, prefix, qc, corr, sep='r', **kwargs):", "funcdef": "def"}, "pyerrors.input.pandas": {"fullname": "pyerrors.input.pandas", "modulename": "pyerrors.input.pandas", "kind": "module", "doc": "\n"}, "pyerrors.input.pandas.to_sql": {"fullname": "pyerrors.input.pandas.to_sql", "modulename": "pyerrors.input.pandas", "qualname": "to_sql", "kind": "function", "doc": "Write DataFrame including Obs or Corr valued columns to sqlite database.
\n\nParameters
\n\n\n- df (pandas.DataFrame):\nDataframe to be written to the database.
\n- table_name (str):\nName of the table in the database.
\n- db (str):\nPath to the sqlite database.
\n- if exists (str):\nHow to behave if table already exists. Options 'fail', 'replace', 'append'.
\n- gz (bool):\nIf True the json strings are gzipped.
\n
\n", "signature": "(df, table_name, db, if_exists='fail', gz=True, **kwargs):", "funcdef": "def"}, "pyerrors.input.pandas.read_sql": {"fullname": "pyerrors.input.pandas.read_sql", "modulename": "pyerrors.input.pandas", "qualname": "read_sql", "kind": "function", "doc": "Execute SQL query on sqlite database and obtain DataFrame including Obs or Corr valued columns.
\n\nParameters
\n\n\n- sql (str):\nSQL query to be executed.
\n- db (str):\nPath to the sqlite database.
\n- auto_gamma (bool):\nIf True applies the gamma_method to all imported Obs objects with the default parameters for\nthe error analysis. Default False.
\n
\n", "signature": "(sql, db, auto_gamma=False, **kwargs):", "funcdef": "def"}, "pyerrors.input.pandas.dump_df": {"fullname": "pyerrors.input.pandas.dump_df", "modulename": "pyerrors.input.pandas", "qualname": "dump_df", "kind": "function", "doc": "Exports a pandas DataFrame containing Obs valued columns to a (gzipped) csv file.
\n\nBefore making use of pandas to_csv functionality Obs objects are serialized via the standardized\njson format of pyerrors.
\n\nParameters
\n\n\n- df (pandas.DataFrame):\nDataframe to be dumped to a file.
\n- fname (str):\nFilename of the output file.
\n- gz (bool):\nIf True, the output is a gzipped csv file. If False, the output is a csv file.
\n
\n", "signature": "(df, fname, gz=True):", "funcdef": "def"}, "pyerrors.input.pandas.load_df": {"fullname": "pyerrors.input.pandas.load_df", "modulename": "pyerrors.input.pandas", "qualname": "load_df", "kind": "function", "doc": "Imports a pandas DataFrame from a csv.(gz) file in which Obs objects are serialized as json strings.
\n\nParameters
\n\n\n- fname (str):\nFilename of the input file.
\n- auto_gamma (bool):\nIf True applies the gamma_method to all imported Obs objects with the default parameters for\nthe error analysis. Default False.
\n- gz (bool):\nIf True, assumes that data is gzipped. If False, assumes JSON file.
\n
\n", "signature": "(fname, auto_gamma=False, gz=True):", "funcdef": "def"}, "pyerrors.input.sfcf": {"fullname": "pyerrors.input.sfcf", "modulename": "pyerrors.input.sfcf", "kind": "module", "doc": "\n"}, "pyerrors.input.sfcf.read_sfcf": {"fullname": "pyerrors.input.sfcf.read_sfcf", "modulename": "pyerrors.input.sfcf", "qualname": "read_sfcf", "kind": "function", "doc": "Read sfcf c format from given folder structure.
\n\nParameters
\n\n\n- path (str):\nPath to the sfcf files.
\n- prefix (str):\nPrefix of the sfcf files.
\n- name (str):\nName of the correlation function to read.
\n- quarks (str):\nLabel of the quarks used in the sfcf input file. e.g. \"quark quark\"\nfor version 0.0 this does NOT need to be given with the typical \" - \"\nthat is present in the output file,\nthis is done automatically for this version
\n- corr_type (str):\nType of correlation function to read. Can be\n
\n- 'bi' for boundary-inner
\n- 'bb' for boundary-boundary
\n- 'bib' for boundary-inner-boundary
\n
\n- noffset (int):\nOffset of the source (only relevant when wavefunctions are used)
\n- wf (int):\nID of wave function
\n- wf2 (int):\nID of the second wavefunction\n(only relevant for boundary-to-boundary correlation functions)
\n- im (bool):\nif True, read imaginary instead of real part\nof the correlation function.
\n- names (list):\nAlternative labeling for replicas/ensembles.\nHas to have the appropriate length
\n- ens_name (str):\nreplaces the name of the ensemble
\n- version (str):\nversion of SFCF, with which the measurement was done.\nif the compact output option (-c) was specified,\nappend a \"c\" to the version (e.g. \"1.0c\")\nif the append output option (-a) was specified,\nappend an \"a\" to the version
\n- cfg_separator (str):\nString that separates the ensemble identifier from the configuration number (default 'n').
\n- replica (list):\nlist of replica to be read, default is all
\n- files (list):\nlist of files to be read per replica, default is all.\nfor non-compact output format, hand the folders to be read here.
\n- check_configs:: list of list of supposed configs, eg. [range(1,1000)]\nfor one replicum with 1000 configs
\n
\n", "signature": "(\tpath,\tprefix,\tname,\tquarks='.*',\tcorr_type='bi',\tnoffset=0,\twf=0,\twf2=0,\tversion='1.0c',\tcfg_separator='n',\t**kwargs):", "funcdef": "def"}, "pyerrors.input.utils": {"fullname": "pyerrors.input.utils", "modulename": "pyerrors.input.utils", "kind": "module", "doc": "Utilities for the input
\n"}, "pyerrors.input.utils.check_idl": {"fullname": "pyerrors.input.utils.check_idl", "modulename": "pyerrors.input.utils", "qualname": "check_idl", "kind": "function", "doc": "Checks if list of configurations is contained in an idl
\n\nParameters
\n\n\n- idl (range or list):\nidl of the current replicum
\n- che (list):\nlist of configurations to be checked against
\n
\n", "signature": "(idl, che):", "funcdef": "def"}, "pyerrors.linalg": {"fullname": "pyerrors.linalg", "modulename": "pyerrors.linalg", "kind": "module", "doc": "\n"}, "pyerrors.linalg.matmul": {"fullname": "pyerrors.linalg.matmul", "modulename": "pyerrors.linalg", "qualname": "matmul", "kind": "function", "doc": "Matrix multiply all operands.
\n\nParameters
\n\n\n- operands (numpy.ndarray):\nArbitrary number of 2d-numpy arrays which can be real or complex\nObs valued.
\n- This implementation is faster compared to standard multiplication via the @ operator.
\n
\n", "signature": "(*operands):", "funcdef": "def"}, "pyerrors.linalg.jack_matmul": {"fullname": "pyerrors.linalg.jack_matmul", "modulename": "pyerrors.linalg", "qualname": "jack_matmul", "kind": "function", "doc": "Matrix multiply both operands making use of the jackknife approximation.
\n\nParameters
\n\n\n- operands (numpy.ndarray):\nArbitrary number of 2d-numpy arrays which can be real or complex\nObs valued.
\n- For large matrices this is considerably faster compared to matmul.
\n
\n", "signature": "(*operands):", "funcdef": "def"}, "pyerrors.linalg.einsum": {"fullname": "pyerrors.linalg.einsum", "modulename": "pyerrors.linalg", "qualname": "einsum", "kind": "function", "doc": "Wrapper for numpy.einsum
\n\nParameters
\n\n\n- subscripts (str):\nSubscripts for summation (see numpy documentation for details)
\n- operands (numpy.ndarray):\nArbitrary number of 2d-numpy arrays which can be real or complex\nObs valued.
\n
\n", "signature": "(subscripts, *operands):", "funcdef": "def"}, "pyerrors.linalg.inv": {"fullname": "pyerrors.linalg.inv", "modulename": "pyerrors.linalg", "qualname": "inv", "kind": "function", "doc": "Inverse of Obs or CObs valued matrices.
\n", "signature": "(x):", "funcdef": "def"}, "pyerrors.linalg.cholesky": {"fullname": "pyerrors.linalg.cholesky", "modulename": "pyerrors.linalg", "qualname": "cholesky", "kind": "function", "doc": "Cholesky decomposition of Obs valued matrices.
\n", "signature": "(x):", "funcdef": "def"}, "pyerrors.linalg.det": {"fullname": "pyerrors.linalg.det", "modulename": "pyerrors.linalg", "qualname": "det", "kind": "function", "doc": "Determinant of Obs valued matrices.
\n", "signature": "(x):", "funcdef": "def"}, "pyerrors.linalg.eigh": {"fullname": "pyerrors.linalg.eigh", "modulename": "pyerrors.linalg", "qualname": "eigh", "kind": "function", "doc": "Computes the eigenvalues and eigenvectors of a given hermitian matrix of Obs according to np.linalg.eigh.
\n", "signature": "(obs, **kwargs):", "funcdef": "def"}, "pyerrors.linalg.eig": {"fullname": "pyerrors.linalg.eig", "modulename": "pyerrors.linalg", "qualname": "eig", "kind": "function", "doc": "Computes the eigenvalues of a given matrix of Obs according to np.linalg.eig.
\n", "signature": "(obs, **kwargs):", "funcdef": "def"}, "pyerrors.linalg.pinv": {"fullname": "pyerrors.linalg.pinv", "modulename": "pyerrors.linalg", "qualname": "pinv", "kind": "function", "doc": "Computes the Moore-Penrose pseudoinverse of a matrix of Obs.
\n", "signature": "(obs, **kwargs):", "funcdef": "def"}, "pyerrors.linalg.svd": {"fullname": "pyerrors.linalg.svd", "modulename": "pyerrors.linalg", "qualname": "svd", "kind": "function", "doc": "Computes the singular value decomposition of a matrix of Obs.
\n", "signature": "(obs, **kwargs):", "funcdef": "def"}, "pyerrors.misc": {"fullname": "pyerrors.misc", "modulename": "pyerrors.misc", "kind": "module", "doc": "\n"}, "pyerrors.misc.dump_object": {"fullname": "pyerrors.misc.dump_object", "modulename": "pyerrors.misc", "qualname": "dump_object", "kind": "function", "doc": "Dump object into pickle file.
\n\nParameters
\n\n\n- obj (object):\nobject to be saved in the pickle file
\n- name (str):\nname of the file
\n- path (str):\nspecifies a custom path for the file (default '.')
\n
\n", "signature": "(obj, name, **kwargs):", "funcdef": "def"}, "pyerrors.misc.load_object": {"fullname": "pyerrors.misc.load_object", "modulename": "pyerrors.misc", "qualname": "load_object", "kind": "function", "doc": "Load object from pickle file.
\n\nParameters
\n\n\n- path (str):\npath to the file
\n
\n", "signature": "(path):", "funcdef": "def"}, "pyerrors.misc.pseudo_Obs": {"fullname": "pyerrors.misc.pseudo_Obs", "modulename": "pyerrors.misc", "qualname": "pseudo_Obs", "kind": "function", "doc": "Generate an Obs object with given value, dvalue and name for test purposes
\n\nParameters
\n\n\n- value (float):\ncentral value of the Obs to be generated.
\n- dvalue (float):\nerror of the Obs to be generated.
\n- name (str):\nname of the ensemble for which the Obs is to be generated.
\n- samples (int):\nnumber of samples for the Obs (default 1000).
\n
\n", "signature": "(value, dvalue, name, samples=1000):", "funcdef": "def"}, "pyerrors.misc.gen_correlated_data": {"fullname": "pyerrors.misc.gen_correlated_data", "modulename": "pyerrors.misc", "qualname": "gen_correlated_data", "kind": "function", "doc": "Generate observables with given covariance and autocorrelation times.
\n\nParameters
\n\n\n- means (list):\nlist containing the mean value of each observable.
\n- cov (numpy.ndarray):\ncovariance matrix for the data to be generated.
\n- name (str):\nensemble name for the data to be geneated.
\n- tau (float or list):\ncan either be a real number or a list with an entry for\nevery dataset.
\n- samples (int):\nnumber of samples to be generated for each observable.
\n
\n", "signature": "(means, cov, name, tau=0.5, samples=1000):", "funcdef": "def"}, "pyerrors.mpm": {"fullname": "pyerrors.mpm", "modulename": "pyerrors.mpm", "kind": "module", "doc": "\n"}, "pyerrors.mpm.matrix_pencil_method": {"fullname": "pyerrors.mpm.matrix_pencil_method", "modulename": "pyerrors.mpm", "qualname": "matrix_pencil_method", "kind": "function", "doc": "Matrix pencil method to extract k energy levels from data
\n\nImplementation of the matrix pencil method based on\neq. (2.17) of Y. Hua, T. K. Sarkar, IEEE Trans. Acoust. 38, 814-824 (1990)
\n\nParameters
\n\n\n- data (list):\ncan be a list of Obs for the analysis of a single correlator, or a list of lists\nof Obs if several correlators are to analyzed at once.
\n- k (int):\nNumber of states to extract (default 1).
\n- p (int):\nmatrix pencil parameter which filters noise. The optimal value is expected between\nlen(data)/3 and 2*len(data)/3. The computation is more expensive the closer p is\nto len(data)/2 but could possibly suppress more noise (default len(data)//2).
\n
\n", "signature": "(corrs, k=1, p=None, **kwargs):", "funcdef": "def"}, "pyerrors.obs": {"fullname": "pyerrors.obs", "modulename": "pyerrors.obs", "kind": "module", "doc": "\n"}, "pyerrors.obs.Obs": {"fullname": "pyerrors.obs.Obs", "modulename": "pyerrors.obs", "qualname": "Obs", "kind": "class", "doc": "Class for a general observable.
\n\nInstances of Obs are the basic objects of a pyerrors error analysis.\nThey are initialized with a list which contains arrays of samples for\ndifferent ensembles/replica and another list of same length which contains\nthe names of the ensembles/replica. Mathematical operations can be\nperformed on instances. The result is another instance of Obs. The error of\nan instance can be computed with the gamma_method. Also contains additional\nmethods for output and visualization of the error calculation.
\n\nAttributes
\n\n\n- S_global (float):\nStandard value for S (default 2.0)
\n- S_dict (dict):\nDictionary for S values. If an entry for a given ensemble\nexists this overwrites the standard value for that ensemble.
\n- tau_exp_global (float):\nStandard value for tau_exp (default 0.0)
\n- tau_exp_dict (dict):\nDictionary for tau_exp values. If an entry for a given ensemble exists\nthis overwrites the standard value for that ensemble.
\n- N_sigma_global (float):\nStandard value for N_sigma (default 1.0)
\n- N_sigma_dict (dict):\nDictionary for N_sigma values. If an entry for a given ensemble exists\nthis overwrites the standard value for that ensemble.
\n
\n"}, "pyerrors.obs.Obs.__init__": {"fullname": "pyerrors.obs.Obs.__init__", "modulename": "pyerrors.obs", "qualname": "Obs.__init__", "kind": "function", "doc": "Initialize Obs object.
\n\nParameters
\n\n\n- samples (list):\nlist of numpy arrays containing the Monte Carlo samples
\n- names (list):\nlist of strings labeling the individual samples
\n- idl (list, optional):\nlist of ranges or lists on which the samples are defined
\n
\n", "signature": "(samples, names, idl=None, **kwargs)"}, "pyerrors.obs.Obs.gamma_method": {"fullname": "pyerrors.obs.Obs.gamma_method", "modulename": "pyerrors.obs", "qualname": "Obs.gamma_method", "kind": "function", "doc": "Estimate the error and related properties of the Obs.
\n\nParameters
\n\n\n- S (float):\nspecifies a custom value for the parameter S (default 2.0).\nIf set to 0 it is assumed that the data exhibits no\nautocorrelation. In this case the error estimates coincides\nwith the sample standard error.
\n- tau_exp (float):\npositive value triggers the critical slowing down analysis\n(default 0.0).
\n- N_sigma (float):\nnumber of standard deviations from zero until the tail is\nattached to the autocorrelation function (default 1).
\n- fft (bool):\ndetermines whether the fft algorithm is used for the computation\nof the autocorrelation function (default True)
\n
\n", "signature": "(self, **kwargs):", "funcdef": "def"}, "pyerrors.obs.Obs.gm": {"fullname": "pyerrors.obs.Obs.gm", "modulename": "pyerrors.obs", "qualname": "Obs.gm", "kind": "function", "doc": "Estimate the error and related properties of the Obs.
\n\nParameters
\n\n\n- S (float):\nspecifies a custom value for the parameter S (default 2.0).\nIf set to 0 it is assumed that the data exhibits no\nautocorrelation. In this case the error estimates coincides\nwith the sample standard error.
\n- tau_exp (float):\npositive value triggers the critical slowing down analysis\n(default 0.0).
\n- N_sigma (float):\nnumber of standard deviations from zero until the tail is\nattached to the autocorrelation function (default 1).
\n- fft (bool):\ndetermines whether the fft algorithm is used for the computation\nof the autocorrelation function (default True)
\n
\n", "signature": "(self, **kwargs):", "funcdef": "def"}, "pyerrors.obs.Obs.details": {"fullname": "pyerrors.obs.Obs.details", "modulename": "pyerrors.obs", "qualname": "Obs.details", "kind": "function", "doc": "Output detailed properties of the Obs.
\n\nParameters
\n\n\n- ens_content (bool):\nprint details about the ensembles and replica if true.
\n
\n", "signature": "(self, ens_content=True):", "funcdef": "def"}, "pyerrors.obs.Obs.reweight": {"fullname": "pyerrors.obs.Obs.reweight", "modulename": "pyerrors.obs", "qualname": "Obs.reweight", "kind": "function", "doc": "Reweight the obs with given rewighting factors.
\n\nParameters
\n\n\n- weight (Obs):\nReweighting factor. An Observable that has to be defined on a superset of the\nconfigurations in obs[i].idl for all i.
\n- all_configs (bool):\nif True, the reweighted observables are normalized by the average of\nthe reweighting factor on all configurations in weight.idl and not\non the configurations in obs[i].idl. Default False.
\n
\n", "signature": "(self, weight):", "funcdef": "def"}, "pyerrors.obs.Obs.is_zero_within_error": {"fullname": "pyerrors.obs.Obs.is_zero_within_error", "modulename": "pyerrors.obs", "qualname": "Obs.is_zero_within_error", "kind": "function", "doc": "Checks whether the observable is zero within 'sigma' standard errors.
\n\nParameters
\n\n\n- sigma (int):\nNumber of standard errors used for the check.
\n- Works only properly when the gamma method was run.
\n
\n", "signature": "(self, sigma=1):", "funcdef": "def"}, "pyerrors.obs.Obs.is_zero": {"fullname": "pyerrors.obs.Obs.is_zero", "modulename": "pyerrors.obs", "qualname": "Obs.is_zero", "kind": "function", "doc": "Checks whether the observable is zero within a given tolerance.
\n\nParameters
\n\n\n- atol (float):\nAbsolute tolerance (for details see numpy documentation).
\n
\n", "signature": "(self, atol=1e-10):", "funcdef": "def"}, "pyerrors.obs.Obs.plot_tauint": {"fullname": "pyerrors.obs.Obs.plot_tauint", "modulename": "pyerrors.obs", "qualname": "Obs.plot_tauint", "kind": "function", "doc": "Plot integrated autocorrelation time for each ensemble.
\n\nParameters
\n\n\n- save (str):\nsaves the figure to a file named 'save' if.
\n
\n", "signature": "(self, save=None):", "funcdef": "def"}, "pyerrors.obs.Obs.plot_rho": {"fullname": "pyerrors.obs.Obs.plot_rho", "modulename": "pyerrors.obs", "qualname": "Obs.plot_rho", "kind": "function", "doc": "Plot normalized autocorrelation function time for each ensemble.
\n\nParameters
\n\n\n- save (str):\nsaves the figure to a file named 'save' if.
\n
\n", "signature": "(self, save=None):", "funcdef": "def"}, "pyerrors.obs.Obs.plot_rep_dist": {"fullname": "pyerrors.obs.Obs.plot_rep_dist", "modulename": "pyerrors.obs", "qualname": "Obs.plot_rep_dist", "kind": "function", "doc": "Plot replica distribution for each ensemble with more than one replicum.
\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.obs.Obs.plot_history": {"fullname": "pyerrors.obs.Obs.plot_history", "modulename": "pyerrors.obs", "qualname": "Obs.plot_history", "kind": "function", "doc": "Plot derived Monte Carlo history for each ensemble
\n\nParameters
\n\n\n- expand (bool):\nshow expanded history for irregular Monte Carlo chains (default: True).
\n
\n", "signature": "(self, expand=True):", "funcdef": "def"}, "pyerrors.obs.Obs.plot_piechart": {"fullname": "pyerrors.obs.Obs.plot_piechart", "modulename": "pyerrors.obs", "qualname": "Obs.plot_piechart", "kind": "function", "doc": "Plot piechart which shows the fractional contribution of each\nensemble to the error and returns a dictionary containing the fractions.
\n\nParameters
\n\n\n- save (str):\nsaves the figure to a file named 'save' if.
\n
\n", "signature": "(self, save=None):", "funcdef": "def"}, "pyerrors.obs.Obs.dump": {"fullname": "pyerrors.obs.Obs.dump", "modulename": "pyerrors.obs", "qualname": "Obs.dump", "kind": "function", "doc": "Dump the Obs to a file 'name' of chosen format.
\n\nParameters
\n\n\n- filename (str):\nname of the file to be saved.
\n- datatype (str):\nFormat of the exported file. Supported formats include\n\"json.gz\" and \"pickle\"
\n- description (str):\nDescription for output file, only relevant for json.gz format.
\n- path (str):\nspecifies a custom path for the file (default '.')
\n
\n", "signature": "(self, filename, datatype='json.gz', description='', **kwargs):", "funcdef": "def"}, "pyerrors.obs.Obs.export_jackknife": {"fullname": "pyerrors.obs.Obs.export_jackknife", "modulename": "pyerrors.obs", "qualname": "Obs.export_jackknife", "kind": "function", "doc": "Export jackknife samples from the Obs
\n\nReturns
\n\n\n- numpy.ndarray: Returns a numpy array of length N + 1 where N is the number of samples\nfor the given ensemble and replicum. The zeroth entry of the array contains\nthe mean value of the Obs, entries 1 to N contain the N jackknife samples\nderived from the Obs. The current implementation only works for observables\ndefined on exactly one ensemble and replicum. The derived jackknife samples\nshould agree with samples from a full jackknife analysis up to O(1/N).
\n
\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.obs.Obs.sqrt": {"fullname": "pyerrors.obs.Obs.sqrt", "modulename": "pyerrors.obs", "qualname": "Obs.sqrt", "kind": "function", "doc": "\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.obs.Obs.log": {"fullname": "pyerrors.obs.Obs.log", "modulename": "pyerrors.obs", "qualname": "Obs.log", "kind": "function", "doc": "\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.obs.Obs.exp": {"fullname": "pyerrors.obs.Obs.exp", "modulename": "pyerrors.obs", "qualname": "Obs.exp", "kind": "function", "doc": "\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.obs.Obs.sin": {"fullname": "pyerrors.obs.Obs.sin", "modulename": "pyerrors.obs", "qualname": "Obs.sin", "kind": "function", "doc": "\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.obs.Obs.cos": {"fullname": "pyerrors.obs.Obs.cos", "modulename": "pyerrors.obs", "qualname": "Obs.cos", "kind": "function", "doc": "\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.obs.Obs.tan": {"fullname": "pyerrors.obs.Obs.tan", "modulename": "pyerrors.obs", "qualname": "Obs.tan", "kind": "function", "doc": "\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.obs.Obs.arcsin": {"fullname": "pyerrors.obs.Obs.arcsin", "modulename": "pyerrors.obs", "qualname": "Obs.arcsin", "kind": "function", "doc": "\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.obs.Obs.arccos": {"fullname": "pyerrors.obs.Obs.arccos", "modulename": "pyerrors.obs", "qualname": "Obs.arccos", "kind": "function", "doc": "\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.obs.Obs.arctan": {"fullname": "pyerrors.obs.Obs.arctan", "modulename": "pyerrors.obs", "qualname": "Obs.arctan", "kind": "function", "doc": "\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.obs.Obs.sinh": {"fullname": "pyerrors.obs.Obs.sinh", "modulename": "pyerrors.obs", "qualname": "Obs.sinh", "kind": "function", "doc": "\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.obs.Obs.cosh": {"fullname": "pyerrors.obs.Obs.cosh", "modulename": "pyerrors.obs", "qualname": "Obs.cosh", "kind": "function", "doc": "\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.obs.Obs.tanh": {"fullname": "pyerrors.obs.Obs.tanh", "modulename": "pyerrors.obs", "qualname": "Obs.tanh", "kind": "function", "doc": "\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.obs.Obs.arcsinh": {"fullname": "pyerrors.obs.Obs.arcsinh", "modulename": "pyerrors.obs", "qualname": "Obs.arcsinh", "kind": "function", "doc": "\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.obs.Obs.arccosh": {"fullname": "pyerrors.obs.Obs.arccosh", "modulename": "pyerrors.obs", "qualname": "Obs.arccosh", "kind": "function", "doc": "\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.obs.Obs.arctanh": {"fullname": "pyerrors.obs.Obs.arctanh", "modulename": "pyerrors.obs", "qualname": "Obs.arctanh", "kind": "function", "doc": "\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.obs.CObs": {"fullname": "pyerrors.obs.CObs", "modulename": "pyerrors.obs", "qualname": "CObs", "kind": "class", "doc": "Class for a complex valued observable.
\n"}, "pyerrors.obs.CObs.__init__": {"fullname": "pyerrors.obs.CObs.__init__", "modulename": "pyerrors.obs", "qualname": "CObs.__init__", "kind": "function", "doc": "\n", "signature": "(real, imag=0.0)"}, "pyerrors.obs.CObs.gamma_method": {"fullname": "pyerrors.obs.CObs.gamma_method", "modulename": "pyerrors.obs", "qualname": "CObs.gamma_method", "kind": "function", "doc": "Executes the gamma_method for the real and the imaginary part.
\n", "signature": "(self, **kwargs):", "funcdef": "def"}, "pyerrors.obs.CObs.is_zero": {"fullname": "pyerrors.obs.CObs.is_zero", "modulename": "pyerrors.obs", "qualname": "CObs.is_zero", "kind": "function", "doc": "Checks whether both real and imaginary part are zero within machine precision.
\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.obs.CObs.conjugate": {"fullname": "pyerrors.obs.CObs.conjugate", "modulename": "pyerrors.obs", "qualname": "CObs.conjugate", "kind": "function", "doc": "\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.obs.derived_observable": {"fullname": "pyerrors.obs.derived_observable", "modulename": "pyerrors.obs", "qualname": "derived_observable", "kind": "function", "doc": "Construct a derived Obs according to func(data, **kwargs) using automatic differentiation.
\n\nParameters
\n\n\n- func (object):\narbitrary function of the form func(data, **kwargs). For the\nautomatic differentiation to work, all numpy functions have to have\nthe autograd wrapper (use 'import autograd.numpy as anp').
\n- data (list):\nlist of Obs, e.g. [obs1, obs2, obs3].
\n- num_grad (bool):\nif True, numerical derivatives are used instead of autograd\n(default False). To control the numerical differentiation the\nkwargs of numdifftools.step_generators.MaxStepGenerator\ncan be used.
\n- man_grad (list):\nmanually supply a list or an array which contains the jacobian\nof func. Use cautiously, supplying the wrong derivative will\nnot be intercepted.
\n
\n\nNotes
\n\nFor simple mathematical operations it can be practical to use anonymous\nfunctions. For the ratio of two observables one can e.g. use
\n\nnew_obs = derived_observable(lambda x: x[0] / x[1], [obs1, obs2])
\n", "signature": "(func, data, array_mode=False, **kwargs):", "funcdef": "def"}, "pyerrors.obs.reweight": {"fullname": "pyerrors.obs.reweight", "modulename": "pyerrors.obs", "qualname": "reweight", "kind": "function", "doc": "Reweight a list of observables.
\n\nParameters
\n\n\n- weight (Obs):\nReweighting factor. An Observable that has to be defined on a superset of the\nconfigurations in obs[i].idl for all i.
\n- obs (list):\nlist of Obs, e.g. [obs1, obs2, obs3].
\n- all_configs (bool):\nif True, the reweighted observables are normalized by the average of\nthe reweighting factor on all configurations in weight.idl and not\non the configurations in obs[i].idl. Default False.
\n
\n", "signature": "(weight, obs, **kwargs):", "funcdef": "def"}, "pyerrors.obs.correlate": {"fullname": "pyerrors.obs.correlate", "modulename": "pyerrors.obs", "qualname": "correlate", "kind": "function", "doc": "Correlate two observables.
\n\nParameters
\n\n\n- obs_a (Obs):\nFirst observable
\n- obs_b (Obs):\nSecond observable
\n
\n\nNotes
\n\nKeep in mind to only correlate primary observables which have not been reweighted\nyet. The reweighting has to be applied after correlating the observables.\nCurrently only works if ensembles are identical (this is not strictly necessary).
\n", "signature": "(obs_a, obs_b):", "funcdef": "def"}, "pyerrors.obs.covariance": {"fullname": "pyerrors.obs.covariance", "modulename": "pyerrors.obs", "qualname": "covariance", "kind": "function", "doc": "Calculates the error covariance matrix of a set of observables.
\n\nWARNING: This function should be used with care, especially for observables with support on multiple\n ensembles with differing autocorrelations. See the notes below for details.
\n\nThe gamma method has to be applied first to all observables.
\n\nParameters
\n\n\n- obs (list or numpy.ndarray):\nList or one dimensional array of Obs
\n- visualize (bool):\nIf True plots the corresponding normalized correlation matrix (default False).
\n- correlation (bool):\nIf True the correlation matrix instead of the error covariance matrix is returned (default False).
\n- smooth (None or int):\nIf smooth is an integer 'E' between 2 and the dimension of the matrix minus 1 the eigenvalue\nsmoothing procedure of hep-lat/9412087 is applied to the correlation matrix which leaves the\nlargest E eigenvalues essentially unchanged and smoothes the smaller eigenvalues to avoid extremely\nsmall ones.
\n
\n\nNotes
\n\nThe error covariance is defined such that it agrees with the squared standard error for two identical observables\n$$\\operatorname{cov}(a,a)=\\sum_{s=1}^N\\delta_a^s\\delta_a^s/N^2=\\Gamma_{aa}(0)/N=\\operatorname{var}(a)/N=\\sigma_a^2$$\nin the absence of autocorrelation.\nThe error covariance is estimated by calculating the correlation matrix assuming no autocorrelation and then rescaling the correlation matrix by the full errors including the previous gamma method estimate for the autocorrelation of the observables. The covariance at windowsize 0 is guaranteed to be positive semi-definite\n$$\\sum_{i,j}v_i\\Gamma_{ij}(0)v_j=\\frac{1}{N}\\sum_{s=1}^N\\sum_{i,j}v_i\\delta_i^s\\delta_j^s v_j=\\frac{1}{N}\\sum_{s=1}^N\\sum_{i}|v_i\\delta_i^s|^2\\geq 0\\,,$$ for every $v\\in\\mathbb{R}^M$, while such an identity does not hold for larger windows/lags.\nFor observables defined on a single ensemble our approximation is equivalent to assuming that the integrated autocorrelation time of an off-diagonal element is equal to the geometric mean of the integrated autocorrelation times of the corresponding diagonal elements.\n$$\\tau_{\\mathrm{int}, ij}=\\sqrt{\\tau_{\\mathrm{int}, i}\\times \\tau_{\\mathrm{int}, j}}$$\nThis construction ensures that the estimated covariance matrix is positive semi-definite (up to numerical rounding errors).
\n", "signature": "(obs, visualize=False, correlation=False, smooth=None, **kwargs):", "funcdef": "def"}, "pyerrors.obs.import_jackknife": {"fullname": "pyerrors.obs.import_jackknife", "modulename": "pyerrors.obs", "qualname": "import_jackknife", "kind": "function", "doc": "Imports jackknife samples and returns an Obs
\n\nParameters
\n\n\n- jacks (numpy.ndarray):\nnumpy array containing the mean value as zeroth entry and\nthe N jackknife samples as first to Nth entry.
\n- name (str):\nname of the ensemble the samples are defined on.
\n
\n", "signature": "(jacks, name, idl=None):", "funcdef": "def"}, "pyerrors.obs.merge_obs": {"fullname": "pyerrors.obs.merge_obs", "modulename": "pyerrors.obs", "qualname": "merge_obs", "kind": "function", "doc": "Combine all observables in list_of_obs into one new observable
\n\nParameters
\n\n\n- list_of_obs (list):\nlist of the Obs object to be combined
\n
\n\nNotes
\n\nIt is not possible to combine obs which are based on the same replicum
\n", "signature": "(list_of_obs):", "funcdef": "def"}, "pyerrors.obs.cov_Obs": {"fullname": "pyerrors.obs.cov_Obs", "modulename": "pyerrors.obs", "qualname": "cov_Obs", "kind": "function", "doc": "Create an Obs based on mean(s) and a covariance matrix
\n\nParameters
\n\n\n- mean (list of floats or float):\nN mean value(s) of the new Obs
\n- cov (list or array):\n2d (NxN) Covariance matrix, 1d diagonal entries or 0d covariance
\n- name (str):\nidentifier for the covariance matrix
\n- grad (list or array):\nGradient of the Covobs wrt. the means belonging to cov.
\n
\n", "signature": "(means, cov, name, grad=None):", "funcdef": "def"}, "pyerrors.roots": {"fullname": "pyerrors.roots", "modulename": "pyerrors.roots", "kind": "module", "doc": "\n"}, "pyerrors.roots.find_root": {"fullname": "pyerrors.roots.find_root", "modulename": "pyerrors.roots", "qualname": "find_root", "kind": "function", "doc": "Finds the root of the function func(x, d) where d is an Obs
.
\n\nParameters
\n\n\n\nReturns
\n\n\n- Obs:
Obs
valued root of the function. \n
\n", "signature": "(d, func, guess=1.0, **kwargs):", "funcdef": "def"}, "pyerrors.version": {"fullname": "pyerrors.version", "modulename": "pyerrors.version", "kind": "module", "doc": "\n"}}, "docInfo": {"pyerrors": {"qualname": 0, "fullname": 1, "annotation": 0, "default_value": 0, "signature": 0, "bases": 0, "doc": 8007}, "pyerrors.correlators": {"qualname": 0, "fullname": 2, "annotation": 0, "default_value": 0, "signature": 0, "bases": 0, "doc": 3}, "pyerrors.correlators.Corr": {"qualname": 1, "fullname": 3, "annotation": 0, "default_value": 0, "signature": 0, "bases": 0, "doc": 108}, "pyerrors.correlators.Corr.__init__": {"qualname": 3, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 40, "bases": 0, "doc": 94}, "pyerrors.correlators.Corr.gamma_method": {"qualname": 3, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 18, "bases": 0, "doc": 13}, "pyerrors.correlators.Corr.gm": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 18, "bases": 0, "doc": 13}, "pyerrors.correlators.Corr.projected": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 43, "bases": 0, "doc": 64}, "pyerrors.correlators.Corr.item": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 21, "bases": 0, "doc": 53}, "pyerrors.correlators.Corr.plottable": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 31}, "pyerrors.correlators.Corr.symmetric": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 9}, "pyerrors.correlators.Corr.anti_symmetric": {"qualname": 3, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 10}, "pyerrors.correlators.Corr.is_matrix_symmetric": {"qualname": 4, "fullname": 6, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 13}, "pyerrors.correlators.Corr.matrix_symmetric": {"qualname": 3, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 10}, "pyerrors.correlators.Corr.GEVP": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 47, "bases": 0, "doc": 326}, "pyerrors.correlators.Corr.Eigenvalue": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 50, "bases": 0, "doc": 59}, "pyerrors.correlators.Corr.Hankel": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 26, "bases": 0, "doc": 67}, "pyerrors.correlators.Corr.roll": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 16, "bases": 0, "doc": 26}, "pyerrors.correlators.Corr.reverse": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 9}, "pyerrors.correlators.Corr.thin": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 31, "bases": 0, "doc": 43}, "pyerrors.correlators.Corr.correlate": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 16, "bases": 0, "doc": 53}, "pyerrors.correlators.Corr.reweight": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 23, "bases": 0, "doc": 79}, "pyerrors.correlators.Corr.T_symmetry": {"qualname": 3, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 26, "bases": 0, "doc": 51}, "pyerrors.correlators.Corr.deriv": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 25, "bases": 0, "doc": 47}, "pyerrors.correlators.Corr.second_deriv": {"qualname": 3, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 25, "bases": 0, "doc": 45}, "pyerrors.correlators.Corr.m_eff": {"qualname": 3, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 36, "bases": 0, "doc": 148}, "pyerrors.correlators.Corr.fit": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 43, "bases": 0, "doc": 110}, "pyerrors.correlators.Corr.plateau": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 47, "bases": 0, "doc": 92}, "pyerrors.correlators.Corr.set_prange": {"qualname": 3, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 16, "bases": 0, "doc": 11}, "pyerrors.correlators.Corr.show": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 149, "bases": 0, "doc": 241}, "pyerrors.correlators.Corr.spaghetti_plot": {"qualname": 3, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 21, "bases": 0, "doc": 42}, "pyerrors.correlators.Corr.dump": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 38, "bases": 0, "doc": 69}, "pyerrors.correlators.Corr.print": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 22, "bases": 0, "doc": 3}, "pyerrors.correlators.Corr.sqrt": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 3}, "pyerrors.correlators.Corr.log": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 3}, "pyerrors.correlators.Corr.exp": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 3}, "pyerrors.correlators.Corr.sin": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 3}, "pyerrors.correlators.Corr.cos": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 3}, "pyerrors.correlators.Corr.tan": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 3}, "pyerrors.correlators.Corr.sinh": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 3}, "pyerrors.correlators.Corr.cosh": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 3}, "pyerrors.correlators.Corr.tanh": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 3}, "pyerrors.correlators.Corr.arcsin": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 3}, "pyerrors.correlators.Corr.arccos": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 3}, "pyerrors.correlators.Corr.arctan": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 3}, "pyerrors.correlators.Corr.arcsinh": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 3}, "pyerrors.correlators.Corr.arccosh": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 3}, "pyerrors.correlators.Corr.arctanh": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 3}, "pyerrors.correlators.Corr.prune": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 46, "bases": 0, "doc": 325}, "pyerrors.covobs": {"qualname": 0, "fullname": 2, "annotation": 0, "default_value": 0, "signature": 0, "bases": 0, "doc": 3}, "pyerrors.covobs.Covobs": {"qualname": 1, "fullname": 3, "annotation": 0, "default_value": 0, "signature": 0, "bases": 0, "doc": 3}, "pyerrors.covobs.Covobs.__init__": {"qualname": 3, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 39, "bases": 0, "doc": 100}, "pyerrors.covobs.Covobs.errsq": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 12}, "pyerrors.dirac": {"qualname": 0, "fullname": 2, "annotation": 0, "default_value": 0, "signature": 0, "bases": 0, "doc": 3}, "pyerrors.dirac.epsilon_tensor": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 21, "bases": 0, "doc": 15}, "pyerrors.dirac.epsilon_tensor_rank4": {"qualname": 3, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 26, "bases": 0, "doc": 15}, "pyerrors.dirac.Grid_gamma": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 12, "bases": 0, "doc": 9}, "pyerrors.fits": {"qualname": 0, "fullname": 2, "annotation": 0, "default_value": 0, "signature": 0, "bases": 0, "doc": 3}, "pyerrors.fits.Fit_result": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 0, "bases": 3, "doc": 75}, "pyerrors.fits.Fit_result.__init__": {"qualname": 4, "fullname": 6, "annotation": 0, "default_value": 0, "signature": 4, "bases": 0, "doc": 3}, "pyerrors.fits.Fit_result.gamma_method": {"qualname": 4, "fullname": 6, "annotation": 0, "default_value": 0, "signature": 18, "bases": 0, "doc": 10}, "pyerrors.fits.Fit_result.gm": {"qualname": 3, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 18, "bases": 0, "doc": 10}, "pyerrors.fits.least_squares": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 48, "bases": 0, "doc": 659}, "pyerrors.fits.total_least_squares": {"qualname": 3, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 38, "bases": 0, "doc": 444}, "pyerrors.fits.fit_lin": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 23, "bases": 0, "doc": 90}, "pyerrors.fits.qqplot": {"qualname": 1, "fullname": 3, "annotation": 0, "default_value": 0, "signature": 27, "bases": 0, "doc": 27}, "pyerrors.fits.residual_plot": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 27, "bases": 0, "doc": 17}, "pyerrors.fits.error_band": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 21, "bases": 0, "doc": 23}, "pyerrors.fits.ks_test": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 16, "bases": 0, "doc": 40}, "pyerrors.input": {"qualname": 0, "fullname": 2, "annotation": 0, "default_value": 0, "signature": 0, "bases": 0, "doc": 81}, "pyerrors.input.bdio": {"qualname": 0, "fullname": 3, "annotation": 0, "default_value": 0, "signature": 0, "bases": 0, "doc": 3}, "pyerrors.input.bdio.read_ADerrors": {"qualname": 2, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 35, "bases": 0, "doc": 106}, "pyerrors.input.bdio.write_ADerrors": {"qualname": 2, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 41, "bases": 0, "doc": 108}, "pyerrors.input.bdio.read_mesons": {"qualname": 2, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 35, "bases": 0, "doc": 194}, "pyerrors.input.bdio.read_dSdm": {"qualname": 2, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 35, "bases": 0, "doc": 191}, "pyerrors.input.dobs": {"qualname": 0, "fullname": 3, "annotation": 0, "default_value": 0, "signature": 0, "bases": 0, "doc": 3}, "pyerrors.input.dobs.create_pobs_string": {"qualname": 3, "fullname": 6, "annotation": 0, "default_value": 0, "signature": 62, "bases": 0, "doc": 164}, "pyerrors.input.dobs.write_pobs": {"qualname": 2, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 85, "bases": 0, "doc": 202}, "pyerrors.input.dobs.read_pobs": {"qualname": 2, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 43, "bases": 0, "doc": 131}, "pyerrors.input.dobs.import_dobs_string": {"qualname": 3, "fullname": 6, "annotation": 0, "default_value": 0, "signature": 43, "bases": 0, "doc": 185}, "pyerrors.input.dobs.read_dobs": {"qualname": 2, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 58, "bases": 0, "doc": 204}, "pyerrors.input.dobs.create_dobs_string": {"qualname": 3, "fullname": 6, "annotation": 0, "default_value": 0, "signature": 82, "bases": 0, "doc": 208}, "pyerrors.input.dobs.write_dobs": {"qualname": 2, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 99, "bases": 0, "doc": 240}, "pyerrors.input.hadrons": {"qualname": 0, "fullname": 3, "annotation": 0, "default_value": 0, "signature": 0, "bases": 0, "doc": 3}, "pyerrors.input.hadrons.read_meson_hd5": {"qualname": 3, "fullname": 6, "annotation": 0, "default_value": 0, "signature": 57, "bases": 0, "doc": 158}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"qualname": 3, "fullname": 6, "annotation": 0, "default_value": 0, "signature": 45, "bases": 0, "doc": 89}, "pyerrors.input.hadrons.Npr_matrix": {"qualname": 2, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 0, "bases": 2, "doc": 1069}, "pyerrors.input.hadrons.Npr_matrix.__init__": {"qualname": 4, "fullname": 7, "annotation": 0, "default_value": 0, "signature": 4, "bases": 0, "doc": 3}, "pyerrors.input.hadrons.Npr_matrix.g5H": {"qualname": 3, "fullname": 6, "annotation": 0, "default_value": 0, "signature": 0, "bases": 0, "doc": 30}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"qualname": 3, "fullname": 6, "annotation": 0, "default_value": 0, "signature": 32, "bases": 0, "doc": 81}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"qualname": 3, "fullname": 6, "annotation": 0, "default_value": 0, "signature": 32, "bases": 0, "doc": 81}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"qualname": 3, "fullname": 6, "annotation": 0, "default_value": 0, "signature": 60, "bases": 0, "doc": 94}, "pyerrors.input.json": {"qualname": 0, "fullname": 3, "annotation": 0, "default_value": 0, "signature": 0, "bases": 0, "doc": 3}, "pyerrors.input.json.create_json_string": {"qualname": 3, "fullname": 6, "annotation": 0, "default_value": 0, "signature": 34, "bases": 0, "doc": 116}, "pyerrors.input.json.dump_to_json": {"qualname": 3, "fullname": 6, "annotation": 0, "default_value": 0, "signature": 49, "bases": 0, "doc": 147}, "pyerrors.input.json.import_json_string": {"qualname": 3, "fullname": 6, "annotation": 0, "default_value": 0, "signature": 33, "bases": 0, "doc": 108}, "pyerrors.input.json.load_json": {"qualname": 2, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 42, "bases": 0, "doc": 128}, "pyerrors.input.json.dump_dict_to_json": {"qualname": 4, "fullname": 7, "annotation": 0, "default_value": 0, "signature": 63, "bases": 0, "doc": 172}, "pyerrors.input.json.load_json_dict": {"qualname": 3, "fullname": 6, "annotation": 0, "default_value": 0, "signature": 56, "bases": 0, "doc": 135}, "pyerrors.input.misc": {"qualname": 0, "fullname": 3, "annotation": 0, "default_value": 0, "signature": 0, "bases": 0, "doc": 3}, "pyerrors.input.misc.read_pbp": {"qualname": 2, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 23, "bases": 0, "doc": 62}, "pyerrors.input.openQCD": {"qualname": 0, "fullname": 3, "annotation": 0, "default_value": 0, "signature": 0, "bases": 0, "doc": 3}, "pyerrors.input.openQCD.read_rwms": {"qualname": 2, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 48, "bases": 0, "doc": 254}, "pyerrors.input.openQCD.extract_t0": {"qualname": 2, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 51, "bases": 0, "doc": 457}, "pyerrors.input.openQCD.read_qtop": {"qualname": 2, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 53, "bases": 0, "doc": 366}, "pyerrors.input.openQCD.read_gf_coupling": {"qualname": 3, "fullname": 6, "annotation": 0, "default_value": 0, "signature": 50, "bases": 0, "doc": 345}, "pyerrors.input.openQCD.qtop_projection": {"qualname": 2, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 21, "bases": 0, "doc": 49}, "pyerrors.input.openQCD.read_qtop_sector": {"qualname": 3, "fullname": 6, "annotation": 0, "default_value": 0, "signature": 38, "bases": 0, "doc": 340}, "pyerrors.input.openQCD.read_ms5_xsf": {"qualname": 3, "fullname": 6, "annotation": 0, "default_value": 0, "signature": 47, "bases": 0, "doc": 276}, "pyerrors.input.pandas": {"qualname": 0, "fullname": 3, "annotation": 0, "default_value": 0, "signature": 0, "bases": 0, "doc": 3}, "pyerrors.input.pandas.to_sql": {"qualname": 2, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 54, "bases": 0, "doc": 101}, "pyerrors.input.pandas.read_sql": {"qualname": 2, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 34, "bases": 0, "doc": 81}, "pyerrors.input.pandas.dump_df": {"qualname": 2, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 26, "bases": 0, "doc": 99}, "pyerrors.input.pandas.load_df": {"qualname": 2, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 32, "bases": 0, "doc": 91}, "pyerrors.input.sfcf": {"qualname": 0, "fullname": 3, "annotation": 0, "default_value": 0, "signature": 0, "bases": 0, "doc": 3}, "pyerrors.input.sfcf.read_sfcf": {"qualname": 2, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 128, "bases": 0, "doc": 390}, "pyerrors.input.utils": {"qualname": 0, "fullname": 3, "annotation": 0, "default_value": 0, "signature": 0, "bases": 0, "doc": 6}, "pyerrors.input.utils.check_idl": {"qualname": 2, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 16, "bases": 0, "doc": 47}, "pyerrors.linalg": {"qualname": 0, "fullname": 2, "annotation": 0, "default_value": 0, "signature": 0, "bases": 0, "doc": 3}, "pyerrors.linalg.matmul": {"qualname": 1, "fullname": 3, "annotation": 0, "default_value": 0, "signature": 13, "bases": 0, "doc": 54}, "pyerrors.linalg.jack_matmul": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 13, "bases": 0, "doc": 58}, "pyerrors.linalg.einsum": {"qualname": 1, "fullname": 3, "annotation": 0, "default_value": 0, "signature": 18, "bases": 0, "doc": 52}, "pyerrors.linalg.inv": {"qualname": 1, "fullname": 3, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 10}, "pyerrors.linalg.cholesky": {"qualname": 1, "fullname": 3, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 9}, "pyerrors.linalg.det": {"qualname": 1, "fullname": 3, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 8}, "pyerrors.linalg.eigh": {"qualname": 1, "fullname": 3, "annotation": 0, "default_value": 0, "signature": 18, "bases": 0, "doc": 20}, "pyerrors.linalg.eig": {"qualname": 1, "fullname": 3, "annotation": 0, "default_value": 0, "signature": 18, "bases": 0, "doc": 17}, "pyerrors.linalg.pinv": {"qualname": 1, "fullname": 3, "annotation": 0, "default_value": 0, "signature": 18, "bases": 0, "doc": 13}, "pyerrors.linalg.svd": {"qualname": 1, "fullname": 3, "annotation": 0, "default_value": 0, "signature": 18, "bases": 0, "doc": 13}, "pyerrors.misc": {"qualname": 0, "fullname": 2, "annotation": 0, "default_value": 0, "signature": 0, "bases": 0, "doc": 3}, "pyerrors.misc.dump_object": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 23, "bases": 0, "doc": 57}, "pyerrors.misc.load_object": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 26}, "pyerrors.misc.pseudo_Obs": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 31, "bases": 0, "doc": 89}, "pyerrors.misc.gen_correlated_data": {"qualname": 3, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 42, "bases": 0, "doc": 109}, "pyerrors.mpm": {"qualname": 0, "fullname": 2, "annotation": 0, "default_value": 0, "signature": 0, "bases": 0, "doc": 3}, "pyerrors.mpm.matrix_pencil_method": {"qualname": 3, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 38, "bases": 0, "doc": 147}, "pyerrors.obs": {"qualname": 0, "fullname": 2, "annotation": 0, "default_value": 0, "signature": 0, "bases": 0, "doc": 3}, "pyerrors.obs.Obs": {"qualname": 1, "fullname": 3, "annotation": 0, "default_value": 0, "signature": 0, "bases": 0, "doc": 238}, "pyerrors.obs.Obs.__init__": {"qualname": 3, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 31, "bases": 0, "doc": 62}, "pyerrors.obs.Obs.gamma_method": {"qualname": 3, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 18, "bases": 0, "doc": 133}, "pyerrors.obs.Obs.gm": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 18, "bases": 0, "doc": 133}, "pyerrors.obs.Obs.details": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 22, "bases": 0, "doc": 34}, "pyerrors.obs.Obs.reweight": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 16, "bases": 0, "doc": 85}, "pyerrors.obs.Obs.is_zero_within_error": {"qualname": 5, "fullname": 7, "annotation": 0, "default_value": 0, "signature": 21, "bases": 0, "doc": 50}, "pyerrors.obs.Obs.is_zero": {"qualname": 3, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 22, "bases": 0, "doc": 35}, "pyerrors.obs.Obs.plot_tauint": {"qualname": 3, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 21, "bases": 0, "doc": 34}, "pyerrors.obs.Obs.plot_rho": {"qualname": 3, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 21, "bases": 0, "doc": 35}, "pyerrors.obs.Obs.plot_rep_dist": {"qualname": 4, "fullname": 6, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 14}, "pyerrors.obs.Obs.plot_history": {"qualname": 3, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 21, "bases": 0, "doc": 35}, "pyerrors.obs.Obs.plot_piechart": {"qualname": 3, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 21, "bases": 0, "doc": 47}, "pyerrors.obs.Obs.dump": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 51, "bases": 0, "doc": 89}, "pyerrors.obs.Obs.export_jackknife": {"qualname": 3, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 101}, "pyerrors.obs.Obs.sqrt": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 3}, "pyerrors.obs.Obs.log": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 3}, "pyerrors.obs.Obs.exp": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 3}, "pyerrors.obs.Obs.sin": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 3}, "pyerrors.obs.Obs.cos": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 3}, "pyerrors.obs.Obs.tan": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 3}, "pyerrors.obs.Obs.arcsin": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 3}, "pyerrors.obs.Obs.arccos": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 3}, "pyerrors.obs.Obs.arctan": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 3}, "pyerrors.obs.Obs.sinh": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 3}, "pyerrors.obs.Obs.cosh": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 3}, "pyerrors.obs.Obs.tanh": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 3}, "pyerrors.obs.Obs.arcsinh": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 3}, "pyerrors.obs.Obs.arccosh": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 3}, "pyerrors.obs.Obs.arctanh": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 3}, "pyerrors.obs.CObs": {"qualname": 1, "fullname": 3, "annotation": 0, "default_value": 0, "signature": 0, "bases": 0, "doc": 9}, "pyerrors.obs.CObs.__init__": {"qualname": 3, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 20, "bases": 0, "doc": 3}, "pyerrors.obs.CObs.gamma_method": {"qualname": 3, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 18, "bases": 0, "doc": 14}, "pyerrors.obs.CObs.is_zero": {"qualname": 3, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 15}, "pyerrors.obs.CObs.conjugate": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 3}, "pyerrors.obs.derived_observable": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 34, "bases": 0, "doc": 184}, "pyerrors.obs.reweight": {"qualname": 1, "fullname": 3, "annotation": 0, "default_value": 0, "signature": 23, "bases": 0, "doc": 99}, "pyerrors.obs.correlate": {"qualname": 1, "fullname": 3, "annotation": 0, "default_value": 0, "signature": 18, "bases": 0, "doc": 75}, "pyerrors.obs.covariance": {"qualname": 1, "fullname": 3, "annotation": 0, "default_value": 0, "signature": 48, "bases": 0, "doc": 374}, "pyerrors.obs.import_jackknife": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 26, "bases": 0, "doc": 61}, "pyerrors.obs.merge_obs": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 13, "bases": 0, "doc": 56}, "pyerrors.obs.cov_Obs": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 31, "bases": 0, "doc": 90}, "pyerrors.roots": {"qualname": 0, "fullname": 2, "annotation": 0, "default_value": 0, "signature": 0, "bases": 0, "doc": 3}, "pyerrors.roots.find_root": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 34, "bases": 0, "doc": 179}, "pyerrors.version": {"qualname": 0, "fullname": 2, "annotation": 0, "default_value": 0, "signature": 0, "bases": 0, "doc": 3}}, "length": 181, "save": true}, "index": {"qualname": {"root": {"docs": {"pyerrors.correlators.Corr.__init__": {"tf": 1}, "pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.fits.Fit_result.__init__": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix.__init__": {"tf": 1}, "pyerrors.obs.Obs.__init__": {"tf": 1}, "pyerrors.obs.CObs.__init__": {"tf": 1}}, "df": 6, "c": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.correlators.Corr.__init__": {"tf": 1}, "pyerrors.correlators.Corr.gamma_method": {"tf": 1}, "pyerrors.correlators.Corr.gm": {"tf": 1}, "pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.correlators.Corr.item": {"tf": 1}, "pyerrors.correlators.Corr.plottable": {"tf": 1}, "pyerrors.correlators.Corr.symmetric": {"tf": 1}, "pyerrors.correlators.Corr.anti_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.is_matrix_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.matrix_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 1}, "pyerrors.correlators.Corr.Hankel": {"tf": 1}, "pyerrors.correlators.Corr.roll": {"tf": 1}, "pyerrors.correlators.Corr.reverse": {"tf": 1}, "pyerrors.correlators.Corr.thin": {"tf": 1}, "pyerrors.correlators.Corr.correlate": {"tf": 1}, "pyerrors.correlators.Corr.reweight": {"tf": 1}, "pyerrors.correlators.Corr.T_symmetry": {"tf": 1}, "pyerrors.correlators.Corr.deriv": {"tf": 1}, "pyerrors.correlators.Corr.second_deriv": {"tf": 1}, "pyerrors.correlators.Corr.m_eff": {"tf": 1}, "pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.correlators.Corr.set_prange": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.correlators.Corr.spaghetti_plot": {"tf": 1}, "pyerrors.correlators.Corr.dump": {"tf": 1}, "pyerrors.correlators.Corr.print": {"tf": 1}, "pyerrors.correlators.Corr.sqrt": {"tf": 1}, "pyerrors.correlators.Corr.log": {"tf": 1}, "pyerrors.correlators.Corr.exp": {"tf": 1}, "pyerrors.correlators.Corr.sin": {"tf": 1}, "pyerrors.correlators.Corr.cos": {"tf": 1}, "pyerrors.correlators.Corr.tan": {"tf": 1}, "pyerrors.correlators.Corr.sinh": {"tf": 1}, "pyerrors.correlators.Corr.cosh": {"tf": 1}, "pyerrors.correlators.Corr.tanh": {"tf": 1}, "pyerrors.correlators.Corr.arcsin": {"tf": 1}, "pyerrors.correlators.Corr.arccos": {"tf": 1}, "pyerrors.correlators.Corr.arctan": {"tf": 1}, "pyerrors.correlators.Corr.arcsinh": {"tf": 1}, "pyerrors.correlators.Corr.arccosh": {"tf": 1}, "pyerrors.correlators.Corr.arctanh": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 46, "e": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.correlate": {"tf": 1}, "pyerrors.obs.correlate": {"tf": 1}}, "df": 2, "d": {"docs": {"pyerrors.misc.gen_correlated_data": {"tf": 1}}, "df": 1}}}}}}}}, "s": {"docs": {"pyerrors.correlators.Corr.cos": {"tf": 1}, "pyerrors.obs.Obs.cos": {"tf": 1}}, "df": 2, "h": {"docs": {"pyerrors.correlators.Corr.cosh": {"tf": 1}, "pyerrors.obs.Obs.cosh": {"tf": 1}}, "df": 2}}, "v": {"docs": {"pyerrors.obs.cov_Obs": {"tf": 1}}, "df": 1, "o": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.covobs.Covobs": {"tf": 1}, "pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.covobs.Covobs.errsq": {"tf": 1}}, "df": 3}}}, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}}}}}}, "u": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}}, "df": 1}}}}}}, "b": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.obs.CObs": {"tf": 1}, "pyerrors.obs.CObs.__init__": {"tf": 1}, "pyerrors.obs.CObs.gamma_method": {"tf": 1}, "pyerrors.obs.CObs.is_zero": {"tf": 1}, "pyerrors.obs.CObs.conjugate": {"tf": 1}}, "df": 5}}, "n": {"docs": {}, "df": 0, "j": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.obs.CObs.conjugate": {"tf": 1}}, "df": 1}}}}}}}}, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1}}, "df": 3}}}}}, "h": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "k": {"docs": {"pyerrors.input.utils.check_idl": {"tf": 1}}, "df": 1}}}, "o": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "k": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.linalg.cholesky": {"tf": 1}}, "df": 1}}}}}}}}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.__init__": {"tf": 1}, "pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.fits.Fit_result.__init__": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix.__init__": {"tf": 1}, "pyerrors.obs.Obs.__init__": {"tf": 1}, "pyerrors.obs.CObs.__init__": {"tf": 1}}, "df": 6}}, "v": {"docs": {"pyerrors.linalg.inv": {"tf": 1}}, "df": 1}}, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "m": {"docs": {"pyerrors.correlators.Corr.item": {"tf": 1}}, "df": 1}}}, "s": {"docs": {"pyerrors.correlators.Corr.is_matrix_symmetric": {"tf": 1}, "pyerrors.obs.Obs.is_zero_within_error": {"tf": 1}, "pyerrors.obs.Obs.is_zero": {"tf": 1}, "pyerrors.obs.CObs.is_zero": {"tf": 1}}, "df": 4}, "m": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.obs.import_jackknife": {"tf": 1}}, "df": 3}}}}}, "d": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.input.utils.check_idl": {"tf": 1}}, "df": 1}}}, "g": {"5": {"docs": {}, "df": 0, "h": {"docs": {"pyerrors.input.hadrons.Npr_matrix.g5H": {"tf": 1}}, "df": 1}}, "docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "a": {"docs": {"pyerrors.correlators.Corr.gamma_method": {"tf": 1}, "pyerrors.dirac.Grid_gamma": {"tf": 1}, "pyerrors.fits.Fit_result.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.CObs.gamma_method": {"tf": 1}}, "df": 5}}}}, "m": {"docs": {"pyerrors.correlators.Corr.gm": {"tf": 1}, "pyerrors.fits.Fit_result.gm": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}}, "df": 3}, "e": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, "p": {"docs": {"pyerrors.correlators.Corr.GEVP": {"tf": 1}}, "df": 1}}, "n": {"docs": {"pyerrors.misc.gen_correlated_data": {"tf": 1}}, "df": 1}}, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.dirac.Grid_gamma": {"tf": 1}}, "df": 1}}}, "f": {"docs": {"pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}}, "df": 1}}, "m": {"docs": {"pyerrors.correlators.Corr.m_eff": {"tf": 1}}, "df": 1, "e": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.gamma_method": {"tf": 1}, "pyerrors.fits.Fit_result.gamma_method": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.CObs.gamma_method": {"tf": 1}}, "df": 5}}}}, "s": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}}, "df": 1, "s": {"docs": {"pyerrors.input.bdio.read_mesons": {"tf": 1}}, "df": 1}}}}, "r": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.obs.merge_obs": {"tf": 1}}, "df": 1}}}}, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "x": {"docs": {"pyerrors.correlators.Corr.is_matrix_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.matrix_symmetric": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix.__init__": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix.g5H": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 6}}}, "m": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.linalg.matmul": {"tf": 1}, "pyerrors.linalg.jack_matmul": {"tf": 1}}, "df": 2}}}}}, "s": {"5": {"docs": {"pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}}, "p": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "j": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.projected": {"tf": 1}}, "df": 1}}, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.openQCD.qtop_projection": {"tf": 1}}, "df": 1}}}}}}}}, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.set_prange": {"tf": 1}}, "df": 1}}}}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.print": {"tf": 1}}, "df": 1}}}, "u": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 1}}}}, "l": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.spaghetti_plot": {"tf": 1}, "pyerrors.fits.residual_plot": {"tf": 1}, "pyerrors.obs.Obs.plot_tauint": {"tf": 1}, "pyerrors.obs.Obs.plot_rho": {"tf": 1}, "pyerrors.obs.Obs.plot_rep_dist": {"tf": 1}, "pyerrors.obs.Obs.plot_history": {"tf": 1}, "pyerrors.obs.Obs.plot_piechart": {"tf": 1}}, "df": 7, "t": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.plottable": {"tf": 1}}, "df": 1}}}}}}}, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "u": {"docs": {"pyerrors.correlators.Corr.plateau": {"tf": 1}}, "df": 1}}}}}}, "o": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}}, "df": 3}}}, "b": {"docs": {}, "df": 0, "p": {"docs": {"pyerrors.input.misc.read_pbp": {"tf": 1}}, "df": 1}}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "v": {"docs": {"pyerrors.linalg.pinv": {"tf": 1}}, "df": 1}}, "e": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.obs.Obs.plot_piechart": {"tf": 1}}, "df": 1}}}}}}}, "s": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "o": {"docs": {"pyerrors.misc.pseudo_Obs": {"tf": 1}}, "df": 1}}}}}, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 1}}}}}}, "s": {"docs": {}, "df": 0, "y": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "c": {"docs": {"pyerrors.correlators.Corr.symmetric": {"tf": 1}, "pyerrors.correlators.Corr.anti_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.is_matrix_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.matrix_symmetric": {"tf": 1}}, "df": 4}}, "y": {"docs": {"pyerrors.correlators.Corr.T_symmetry": {"tf": 1}}, "df": 1}}}}}}}, "e": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.second_deriv": {"tf": 1}}, "df": 1}}}, "t": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 1}}}}, "t": {"docs": {"pyerrors.correlators.Corr.set_prange": {"tf": 1}}, "df": 1}}, "h": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "w": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1}}, "df": 1}}}, "p": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {"pyerrors.correlators.Corr.spaghetti_plot": {"tf": 1}}, "df": 1}}}}}}}}, "q": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.sqrt": {"tf": 1}, "pyerrors.obs.Obs.sqrt": {"tf": 1}}, "df": 2}}, "u": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}}, "df": 2}}}}}, "l": {"docs": {"pyerrors.input.pandas.to_sql": {"tf": 1}, "pyerrors.input.pandas.read_sql": {"tf": 1}}, "df": 2}}, "i": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.correlators.Corr.sin": {"tf": 1}, "pyerrors.obs.Obs.sin": {"tf": 1}}, "df": 2, "h": {"docs": {"pyerrors.correlators.Corr.sinh": {"tf": 1}, "pyerrors.obs.Obs.sinh": {"tf": 1}}, "df": 2}}}, "t": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.import_json_string": {"tf": 1}}, "df": 5}}}}}, "f": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "f": {"docs": {"pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 1}}}, "v": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.linalg.svd": {"tf": 1}}, "df": 1}}}, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {"pyerrors.correlators.Corr.anti_symmetric": {"tf": 1}}, "df": 1}}}, "r": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.correlators.Corr.arcsin": {"tf": 1}, "pyerrors.obs.Obs.arcsin": {"tf": 1}}, "df": 2, "h": {"docs": {"pyerrors.correlators.Corr.arcsinh": {"tf": 1}, "pyerrors.obs.Obs.arcsinh": {"tf": 1}}, "df": 2}}}}, "c": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.correlators.Corr.arccos": {"tf": 1}, "pyerrors.obs.Obs.arccos": {"tf": 1}}, "df": 2, "h": {"docs": {"pyerrors.correlators.Corr.arccosh": {"tf": 1}, "pyerrors.obs.Obs.arccosh": {"tf": 1}}, "df": 2}}}}, "t": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.correlators.Corr.arctan": {"tf": 1}, "pyerrors.obs.Obs.arctan": {"tf": 1}}, "df": 2, "h": {"docs": {"pyerrors.correlators.Corr.arctanh": {"tf": 1}, "pyerrors.obs.Obs.arctanh": {"tf": 1}}, "df": 2}}}}}}, "d": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}}, "df": 2}}}}}}}}, "e": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.linalg.eig": {"tf": 1}}, "df": 1, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.Eigenvalue": {"tf": 1}}, "df": 1}}}}}}}, "h": {"docs": {"pyerrors.linalg.eigh": {"tf": 1}}, "df": 1}}, "n": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "m": {"docs": {"pyerrors.linalg.einsum": {"tf": 1}}, "df": 1}}}}}, "f": {"docs": {}, "df": 0, "f": {"docs": {"pyerrors.correlators.Corr.m_eff": {"tf": 1}}, "df": 1}}, "x": {"docs": {}, "df": 0, "p": {"docs": {"pyerrors.correlators.Corr.exp": {"tf": 1}, "pyerrors.obs.Obs.exp": {"tf": 1}}, "df": 2, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.obs.Obs.export_jackknife": {"tf": 1}}, "df": 1}}}}, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}}, "df": 1}}}}}}}}, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 1}}}}}}, "r": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "q": {"docs": {"pyerrors.covobs.Covobs.errsq": {"tf": 1}}, "df": 1}}, "o": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.fits.error_band": {"tf": 1}, "pyerrors.obs.Obs.is_zero_within_error": {"tf": 1}}, "df": 2}}}}, "p": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.dirac.epsilon_tensor": {"tf": 1}, "pyerrors.dirac.epsilon_tensor_rank4": {"tf": 1}}, "df": 2}}}}}}}, "h": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "k": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.correlators.Corr.Hankel": {"tf": 1}}, "df": 1}}}}}, "d": {"5": {"docs": {"pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}}, "df": 5}, "docs": {}, "df": 0}, "i": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.obs.Obs.plot_history": {"tf": 1}}, "df": 1}}}}}}}, "r": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.correlators.Corr.roll": {"tf": 1}}, "df": 1}}, "o": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.roots.find_root": {"tf": 1}}, "df": 1}}}, "e": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.reverse": {"tf": 1}}, "df": 1}}}}}, "w": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.reweight": {"tf": 1}, "pyerrors.obs.Obs.reweight": {"tf": 1}, "pyerrors.obs.reweight": {"tf": 1}}, "df": 3}}}}}}, "s": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.fits.Fit_result": {"tf": 1}, "pyerrors.fits.Fit_result.__init__": {"tf": 1}, "pyerrors.fits.Fit_result.gamma_method": {"tf": 1}, "pyerrors.fits.Fit_result.gm": {"tf": 1}}, "df": 4}}}, "i": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.fits.residual_plot": {"tf": 1}}, "df": 1}}}}}}, "a": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}, "pyerrors.input.misc.read_pbp": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}, "pyerrors.input.pandas.read_sql": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 18}}, "p": {"docs": {"pyerrors.obs.Obs.plot_rep_dist": {"tf": 1}}, "df": 1}}, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "k": {"4": {"docs": {"pyerrors.dirac.epsilon_tensor_rank4": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}}}, "w": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.openQCD.read_rwms": {"tf": 1}}, "df": 1}}}, "h": {"docs": {}, "df": 0, "o": {"docs": {"pyerrors.obs.Obs.plot_rho": {"tf": 1}}, "df": 1}}}, "t": {"0": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 1}, "docs": {"pyerrors.correlators.Corr.T_symmetry": {"tf": 1}}, "df": 1, "h": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.correlators.Corr.thin": {"tf": 1}}, "df": 1}}}, "a": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.correlators.Corr.tan": {"tf": 1}, "pyerrors.obs.Obs.tan": {"tf": 1}}, "df": 2, "h": {"docs": {"pyerrors.correlators.Corr.tanh": {"tf": 1}, "pyerrors.obs.Obs.tanh": {"tf": 1}}, "df": 2}}, "u": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.obs.Obs.plot_tauint": {"tf": 1}}, "df": 1}}}}}, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.dirac.epsilon_tensor": {"tf": 1}, "pyerrors.dirac.epsilon_tensor_rank4": {"tf": 1}}, "df": 2}}}}, "s": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.fits.ks_test": {"tf": 1}}, "df": 1}}}, "o": {"docs": {"pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.pandas.to_sql": {"tf": 1}}, "df": 3, "t": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.fits.total_least_squares": {"tf": 1}}, "df": 1}}}}}, "d": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "v": {"docs": {"pyerrors.correlators.Corr.deriv": {"tf": 1}, "pyerrors.correlators.Corr.second_deriv": {"tf": 1}}, "df": 2, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.obs.derived_observable": {"tf": 1}}, "df": 1}}}}}, "t": {"docs": {"pyerrors.linalg.det": {"tf": 1}}, "df": 1, "a": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.obs.Obs.details": {"tf": 1}}, "df": 1}}}}}}, "u": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "p": {"docs": {"pyerrors.correlators.Corr.dump": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.pandas.dump_df": {"tf": 1}, "pyerrors.misc.dump_object": {"tf": 1}, "pyerrors.obs.Obs.dump": {"tf": 1}}, "df": 6}}}, "s": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "m": {"docs": {"pyerrors.input.bdio.read_dSdm": {"tf": 1}}, "df": 1}}}, "o": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}}, "df": 4}}}, "i": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.obs.Obs.plot_rep_dist": {"tf": 1}}, "df": 1, "i": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}}, "df": 1}}}}}}}}}}}}}}}}}}}}}, "c": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}}, "df": 2}}}, "f": {"docs": {"pyerrors.input.pandas.dump_df": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1}}, "df": 2}, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "a": {"docs": {"pyerrors.misc.gen_correlated_data": {"tf": 1}}, "df": 1}}}}, "f": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.fits.Fit_result": {"tf": 1}, "pyerrors.fits.Fit_result.__init__": {"tf": 1}, "pyerrors.fits.Fit_result.gamma_method": {"tf": 1}, "pyerrors.fits.Fit_result.gm": {"tf": 1}, "pyerrors.fits.fit_lin": {"tf": 1}}, "df": 6}, "n": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.roots.find_root": {"tf": 1}}, "df": 1}}}, "o": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "q": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "k": {"docs": {"pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}}, "df": 1}}}}}}}}}, "l": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.correlators.Corr.log": {"tf": 1}, "pyerrors.obs.Obs.log": {"tf": 1}}, "df": 2}, "a": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1}, "pyerrors.misc.load_object": {"tf": 1}}, "df": 4}}}, "e": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}}, "df": 2}}}}, "i": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.fits.fit_lin": {"tf": 1}}, "df": 1}}}, "q": {"docs": {}, "df": 0, "q": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.fits.qqplot": {"tf": 1}}, "df": 1}}}}}, "t": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "p": {"docs": {"pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.qtop_projection": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 3}}}}, "b": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.fits.error_band": {"tf": 1}}, "df": 1}}}, "i": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}}, "df": 1}}}}}}}}, "k": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.fits.ks_test": {"tf": 1}}, "df": 1}}, "w": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}}, "df": 3}}}}, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.obs.Obs.is_zero_within_error": {"tf": 1}}, "df": 1}}}}}}, "n": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix.__init__": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix.g5H": {"tf": 1}}, "df": 3}}}, "j": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}}, "df": 6}}}, "a": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "k": {"docs": {"pyerrors.linalg.jack_matmul": {"tf": 1}}, "df": 1, "k": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "f": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.obs.Obs.export_jackknife": {"tf": 1}, "pyerrors.obs.import_jackknife": {"tf": 1}}, "df": 2}}}}}}}}}, "x": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "f": {"docs": {"pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}}, "df": 1}}}, "o": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "j": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.misc.dump_object": {"tf": 1}, "pyerrors.misc.load_object": {"tf": 1}}, "df": 2}}}}, "s": {"docs": {"pyerrors.misc.pseudo_Obs": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1}, "pyerrors.obs.Obs.__init__": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}, "pyerrors.obs.Obs.details": {"tf": 1}, "pyerrors.obs.Obs.reweight": {"tf": 1}, "pyerrors.obs.Obs.is_zero_within_error": {"tf": 1}, "pyerrors.obs.Obs.is_zero": {"tf": 1}, "pyerrors.obs.Obs.plot_tauint": {"tf": 1}, "pyerrors.obs.Obs.plot_rho": {"tf": 1}, "pyerrors.obs.Obs.plot_rep_dist": {"tf": 1}, "pyerrors.obs.Obs.plot_history": {"tf": 1}, "pyerrors.obs.Obs.plot_piechart": {"tf": 1}, "pyerrors.obs.Obs.dump": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1}, "pyerrors.obs.Obs.sqrt": {"tf": 1}, "pyerrors.obs.Obs.log": {"tf": 1}, "pyerrors.obs.Obs.exp": {"tf": 1}, "pyerrors.obs.Obs.sin": {"tf": 1}, "pyerrors.obs.Obs.cos": {"tf": 1}, "pyerrors.obs.Obs.tan": {"tf": 1}, "pyerrors.obs.Obs.arcsin": {"tf": 1}, "pyerrors.obs.Obs.arccos": {"tf": 1}, "pyerrors.obs.Obs.arctan": {"tf": 1}, "pyerrors.obs.Obs.sinh": {"tf": 1}, "pyerrors.obs.Obs.cosh": {"tf": 1}, "pyerrors.obs.Obs.tanh": {"tf": 1}, "pyerrors.obs.Obs.arcsinh": {"tf": 1}, "pyerrors.obs.Obs.arccosh": {"tf": 1}, "pyerrors.obs.Obs.arctanh": {"tf": 1}, "pyerrors.obs.merge_obs": {"tf": 1}, "pyerrors.obs.cov_Obs": {"tf": 1}}, "df": 33, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.obs.derived_observable": {"tf": 1}}, "df": 1}}}}}}}}}}, "z": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "o": {"docs": {"pyerrors.obs.Obs.is_zero_within_error": {"tf": 1}, "pyerrors.obs.Obs.is_zero": {"tf": 1}, "pyerrors.obs.CObs.is_zero": {"tf": 1}}, "df": 3}}}}}}, "fullname": {"root": {"docs": {"pyerrors.correlators.Corr.__init__": {"tf": 1}, "pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.fits.Fit_result.__init__": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix.__init__": {"tf": 1}, "pyerrors.obs.Obs.__init__": {"tf": 1}, "pyerrors.obs.CObs.__init__": {"tf": 1}}, "df": 6, "p": {"docs": {}, "df": 0, "y": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators": {"tf": 1}, "pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.correlators.Corr.__init__": {"tf": 1}, "pyerrors.correlators.Corr.gamma_method": {"tf": 1}, "pyerrors.correlators.Corr.gm": {"tf": 1}, "pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.correlators.Corr.item": {"tf": 1}, "pyerrors.correlators.Corr.plottable": {"tf": 1}, "pyerrors.correlators.Corr.symmetric": {"tf": 1}, "pyerrors.correlators.Corr.anti_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.is_matrix_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.matrix_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 1}, "pyerrors.correlators.Corr.Hankel": {"tf": 1}, "pyerrors.correlators.Corr.roll": {"tf": 1}, "pyerrors.correlators.Corr.reverse": {"tf": 1}, "pyerrors.correlators.Corr.thin": {"tf": 1}, "pyerrors.correlators.Corr.correlate": {"tf": 1}, "pyerrors.correlators.Corr.reweight": {"tf": 1}, "pyerrors.correlators.Corr.T_symmetry": {"tf": 1}, "pyerrors.correlators.Corr.deriv": {"tf": 1}, "pyerrors.correlators.Corr.second_deriv": {"tf": 1}, "pyerrors.correlators.Corr.m_eff": {"tf": 1}, "pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.correlators.Corr.set_prange": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.correlators.Corr.spaghetti_plot": {"tf": 1}, "pyerrors.correlators.Corr.dump": {"tf": 1}, "pyerrors.correlators.Corr.print": {"tf": 1}, "pyerrors.correlators.Corr.sqrt": {"tf": 1}, "pyerrors.correlators.Corr.log": {"tf": 1}, "pyerrors.correlators.Corr.exp": {"tf": 1}, "pyerrors.correlators.Corr.sin": {"tf": 1}, "pyerrors.correlators.Corr.cos": {"tf": 1}, "pyerrors.correlators.Corr.tan": {"tf": 1}, "pyerrors.correlators.Corr.sinh": {"tf": 1}, "pyerrors.correlators.Corr.cosh": {"tf": 1}, "pyerrors.correlators.Corr.tanh": {"tf": 1}, "pyerrors.correlators.Corr.arcsin": {"tf": 1}, "pyerrors.correlators.Corr.arccos": {"tf": 1}, "pyerrors.correlators.Corr.arctan": {"tf": 1}, "pyerrors.correlators.Corr.arcsinh": {"tf": 1}, "pyerrors.correlators.Corr.arccosh": {"tf": 1}, "pyerrors.correlators.Corr.arctanh": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1}, "pyerrors.covobs": {"tf": 1}, "pyerrors.covobs.Covobs": {"tf": 1}, "pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.covobs.Covobs.errsq": {"tf": 1}, "pyerrors.dirac": {"tf": 1}, "pyerrors.dirac.epsilon_tensor": {"tf": 1}, "pyerrors.dirac.epsilon_tensor_rank4": {"tf": 1}, "pyerrors.dirac.Grid_gamma": {"tf": 1}, "pyerrors.fits": {"tf": 1}, "pyerrors.fits.Fit_result": {"tf": 1}, "pyerrors.fits.Fit_result.__init__": {"tf": 1}, "pyerrors.fits.Fit_result.gamma_method": {"tf": 1}, "pyerrors.fits.Fit_result.gm": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.fits.fit_lin": {"tf": 1}, "pyerrors.fits.qqplot": {"tf": 1}, "pyerrors.fits.residual_plot": {"tf": 1}, "pyerrors.fits.error_band": {"tf": 1}, "pyerrors.fits.ks_test": {"tf": 1}, "pyerrors.input": {"tf": 1}, "pyerrors.input.bdio": {"tf": 1}, "pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}, "pyerrors.input.dobs": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.hadrons": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix.__init__": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix.g5H": {"tf": 1}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}, "pyerrors.input.json": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}, "pyerrors.input.misc": {"tf": 1}, "pyerrors.input.misc.read_pbp": {"tf": 1}, "pyerrors.input.openQCD": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.qtop_projection": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}, "pyerrors.input.pandas": {"tf": 1}, "pyerrors.input.pandas.to_sql": {"tf": 1}, "pyerrors.input.pandas.read_sql": {"tf": 1}, "pyerrors.input.pandas.dump_df": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1}, "pyerrors.input.sfcf": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.input.utils": {"tf": 1}, "pyerrors.input.utils.check_idl": {"tf": 1}, "pyerrors.linalg": {"tf": 1}, "pyerrors.linalg.matmul": {"tf": 1}, "pyerrors.linalg.jack_matmul": {"tf": 1}, "pyerrors.linalg.einsum": {"tf": 1}, "pyerrors.linalg.inv": {"tf": 1}, "pyerrors.linalg.cholesky": {"tf": 1}, "pyerrors.linalg.det": {"tf": 1}, "pyerrors.linalg.eigh": {"tf": 1}, "pyerrors.linalg.eig": {"tf": 1}, "pyerrors.linalg.pinv": {"tf": 1}, "pyerrors.linalg.svd": {"tf": 1}, "pyerrors.misc": {"tf": 1}, "pyerrors.misc.dump_object": {"tf": 1}, "pyerrors.misc.load_object": {"tf": 1}, "pyerrors.misc.pseudo_Obs": {"tf": 1}, "pyerrors.misc.gen_correlated_data": {"tf": 1}, "pyerrors.mpm": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}, "pyerrors.obs": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1}, "pyerrors.obs.Obs.__init__": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}, "pyerrors.obs.Obs.details": {"tf": 1}, "pyerrors.obs.Obs.reweight": {"tf": 1}, "pyerrors.obs.Obs.is_zero_within_error": {"tf": 1}, "pyerrors.obs.Obs.is_zero": {"tf": 1}, "pyerrors.obs.Obs.plot_tauint": {"tf": 1}, "pyerrors.obs.Obs.plot_rho": {"tf": 1}, "pyerrors.obs.Obs.plot_rep_dist": {"tf": 1}, "pyerrors.obs.Obs.plot_history": {"tf": 1}, "pyerrors.obs.Obs.plot_piechart": {"tf": 1}, "pyerrors.obs.Obs.dump": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1}, "pyerrors.obs.Obs.sqrt": {"tf": 1}, "pyerrors.obs.Obs.log": {"tf": 1}, "pyerrors.obs.Obs.exp": {"tf": 1}, "pyerrors.obs.Obs.sin": {"tf": 1}, "pyerrors.obs.Obs.cos": {"tf": 1}, "pyerrors.obs.Obs.tan": {"tf": 1}, "pyerrors.obs.Obs.arcsin": {"tf": 1}, "pyerrors.obs.Obs.arccos": {"tf": 1}, "pyerrors.obs.Obs.arctan": {"tf": 1}, "pyerrors.obs.Obs.sinh": {"tf": 1}, "pyerrors.obs.Obs.cosh": {"tf": 1}, "pyerrors.obs.Obs.tanh": {"tf": 1}, "pyerrors.obs.Obs.arcsinh": {"tf": 1}, "pyerrors.obs.Obs.arccosh": {"tf": 1}, "pyerrors.obs.Obs.arctanh": {"tf": 1}, "pyerrors.obs.CObs": {"tf": 1}, "pyerrors.obs.CObs.__init__": {"tf": 1}, "pyerrors.obs.CObs.gamma_method": {"tf": 1}, "pyerrors.obs.CObs.is_zero": {"tf": 1}, "pyerrors.obs.CObs.conjugate": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}, "pyerrors.obs.reweight": {"tf": 1}, "pyerrors.obs.correlate": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}, "pyerrors.obs.import_jackknife": {"tf": 1}, "pyerrors.obs.merge_obs": {"tf": 1}, "pyerrors.obs.cov_Obs": {"tf": 1}, "pyerrors.roots": {"tf": 1}, "pyerrors.roots.find_root": {"tf": 1}, "pyerrors.version": {"tf": 1}}, "df": 181}}}}}}}, "r": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "j": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.projected": {"tf": 1}}, "df": 1}}, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.openQCD.qtop_projection": {"tf": 1}}, "df": 1}}}}}}}}, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.set_prange": {"tf": 1}}, "df": 1}}}}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.print": {"tf": 1}}, "df": 1}}}, "u": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 1}}}}, "l": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.spaghetti_plot": {"tf": 1}, "pyerrors.fits.residual_plot": {"tf": 1}, "pyerrors.obs.Obs.plot_tauint": {"tf": 1}, "pyerrors.obs.Obs.plot_rho": {"tf": 1}, "pyerrors.obs.Obs.plot_rep_dist": {"tf": 1}, "pyerrors.obs.Obs.plot_history": {"tf": 1}, "pyerrors.obs.Obs.plot_piechart": {"tf": 1}}, "df": 7, "t": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.plottable": {"tf": 1}}, "df": 1}}}}}}}, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "u": {"docs": {"pyerrors.correlators.Corr.plateau": {"tf": 1}}, "df": 1}}}}}}, "o": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}}, "df": 3}}}, "b": {"docs": {}, "df": 0, "p": {"docs": {"pyerrors.input.misc.read_pbp": {"tf": 1}}, "df": 1}}, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.pandas": {"tf": 1}, "pyerrors.input.pandas.to_sql": {"tf": 1}, "pyerrors.input.pandas.read_sql": {"tf": 1}, "pyerrors.input.pandas.dump_df": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1}}, "df": 5}}}}}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "v": {"docs": {"pyerrors.linalg.pinv": {"tf": 1}}, "df": 1}}, "e": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.obs.Obs.plot_piechart": {"tf": 1}}, "df": 1}}}}}}}, "s": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "o": {"docs": {"pyerrors.misc.pseudo_Obs": {"tf": 1}}, "df": 1}}}}}, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 1}}}}}}, "c": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.correlators.Corr.__init__": {"tf": 1}, "pyerrors.correlators.Corr.gamma_method": {"tf": 1}, "pyerrors.correlators.Corr.gm": {"tf": 1}, "pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.correlators.Corr.item": {"tf": 1}, "pyerrors.correlators.Corr.plottable": {"tf": 1}, "pyerrors.correlators.Corr.symmetric": {"tf": 1}, "pyerrors.correlators.Corr.anti_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.is_matrix_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.matrix_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 1}, "pyerrors.correlators.Corr.Hankel": {"tf": 1}, "pyerrors.correlators.Corr.roll": {"tf": 1}, "pyerrors.correlators.Corr.reverse": {"tf": 1}, "pyerrors.correlators.Corr.thin": {"tf": 1}, "pyerrors.correlators.Corr.correlate": {"tf": 1}, "pyerrors.correlators.Corr.reweight": {"tf": 1}, "pyerrors.correlators.Corr.T_symmetry": {"tf": 1}, "pyerrors.correlators.Corr.deriv": {"tf": 1}, "pyerrors.correlators.Corr.second_deriv": {"tf": 1}, "pyerrors.correlators.Corr.m_eff": {"tf": 1}, "pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.correlators.Corr.set_prange": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.correlators.Corr.spaghetti_plot": {"tf": 1}, "pyerrors.correlators.Corr.dump": {"tf": 1}, "pyerrors.correlators.Corr.print": {"tf": 1}, "pyerrors.correlators.Corr.sqrt": {"tf": 1}, "pyerrors.correlators.Corr.log": {"tf": 1}, "pyerrors.correlators.Corr.exp": {"tf": 1}, "pyerrors.correlators.Corr.sin": {"tf": 1}, "pyerrors.correlators.Corr.cos": {"tf": 1}, "pyerrors.correlators.Corr.tan": {"tf": 1}, "pyerrors.correlators.Corr.sinh": {"tf": 1}, "pyerrors.correlators.Corr.cosh": {"tf": 1}, "pyerrors.correlators.Corr.tanh": {"tf": 1}, "pyerrors.correlators.Corr.arcsin": {"tf": 1}, "pyerrors.correlators.Corr.arccos": {"tf": 1}, "pyerrors.correlators.Corr.arctan": {"tf": 1}, "pyerrors.correlators.Corr.arcsinh": {"tf": 1}, "pyerrors.correlators.Corr.arccosh": {"tf": 1}, "pyerrors.correlators.Corr.arctanh": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 46, "e": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.correlators": {"tf": 1}, "pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.correlators.Corr.__init__": {"tf": 1}, "pyerrors.correlators.Corr.gamma_method": {"tf": 1}, "pyerrors.correlators.Corr.gm": {"tf": 1}, "pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.correlators.Corr.item": {"tf": 1}, "pyerrors.correlators.Corr.plottable": {"tf": 1}, "pyerrors.correlators.Corr.symmetric": {"tf": 1}, "pyerrors.correlators.Corr.anti_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.is_matrix_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.matrix_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 1}, "pyerrors.correlators.Corr.Hankel": {"tf": 1}, "pyerrors.correlators.Corr.roll": {"tf": 1}, "pyerrors.correlators.Corr.reverse": {"tf": 1}, "pyerrors.correlators.Corr.thin": {"tf": 1}, "pyerrors.correlators.Corr.correlate": {"tf": 1}, "pyerrors.correlators.Corr.reweight": {"tf": 1}, "pyerrors.correlators.Corr.T_symmetry": {"tf": 1}, "pyerrors.correlators.Corr.deriv": {"tf": 1}, "pyerrors.correlators.Corr.second_deriv": {"tf": 1}, "pyerrors.correlators.Corr.m_eff": {"tf": 1}, "pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.correlators.Corr.set_prange": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.correlators.Corr.spaghetti_plot": {"tf": 1}, "pyerrors.correlators.Corr.dump": {"tf": 1}, "pyerrors.correlators.Corr.print": {"tf": 1}, "pyerrors.correlators.Corr.sqrt": {"tf": 1}, "pyerrors.correlators.Corr.log": {"tf": 1}, "pyerrors.correlators.Corr.exp": {"tf": 1}, "pyerrors.correlators.Corr.sin": {"tf": 1}, "pyerrors.correlators.Corr.cos": {"tf": 1}, "pyerrors.correlators.Corr.tan": {"tf": 1}, "pyerrors.correlators.Corr.sinh": {"tf": 1}, "pyerrors.correlators.Corr.cosh": {"tf": 1}, "pyerrors.correlators.Corr.tanh": {"tf": 1}, "pyerrors.correlators.Corr.arcsin": {"tf": 1}, "pyerrors.correlators.Corr.arccos": {"tf": 1}, "pyerrors.correlators.Corr.arctan": {"tf": 1}, "pyerrors.correlators.Corr.arcsinh": {"tf": 1}, "pyerrors.correlators.Corr.arccosh": {"tf": 1}, "pyerrors.correlators.Corr.arctanh": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 47}}}, "e": {"docs": {"pyerrors.correlators.Corr.correlate": {"tf": 1}, "pyerrors.obs.correlate": {"tf": 1}}, "df": 2, "d": {"docs": {"pyerrors.misc.gen_correlated_data": {"tf": 1}}, "df": 1}}}}}}}}, "s": {"docs": {"pyerrors.correlators.Corr.cos": {"tf": 1}, "pyerrors.obs.Obs.cos": {"tf": 1}}, "df": 2, "h": {"docs": {"pyerrors.correlators.Corr.cosh": {"tf": 1}, "pyerrors.obs.Obs.cosh": {"tf": 1}}, "df": 2}}, "v": {"docs": {"pyerrors.obs.cov_Obs": {"tf": 1}}, "df": 1, "o": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.covobs": {"tf": 1}, "pyerrors.covobs.Covobs": {"tf": 1.4142135623730951}, "pyerrors.covobs.Covobs.__init__": {"tf": 1.4142135623730951}, "pyerrors.covobs.Covobs.errsq": {"tf": 1.4142135623730951}}, "df": 4}}}, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}}}}}}, "u": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}}, "df": 1}}}}}}, "b": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.obs.CObs": {"tf": 1}, "pyerrors.obs.CObs.__init__": {"tf": 1}, "pyerrors.obs.CObs.gamma_method": {"tf": 1}, "pyerrors.obs.CObs.is_zero": {"tf": 1}, "pyerrors.obs.CObs.conjugate": {"tf": 1}}, "df": 5}}, "n": {"docs": {}, "df": 0, "j": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.obs.CObs.conjugate": {"tf": 1}}, "df": 1}}}}}}}}, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1}}, "df": 3}}}}}, "h": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "k": {"docs": {"pyerrors.input.utils.check_idl": {"tf": 1}}, "df": 1}}}, "o": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "k": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.linalg.cholesky": {"tf": 1}}, "df": 1}}}}}}}}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.__init__": {"tf": 1}, "pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.fits.Fit_result.__init__": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix.__init__": {"tf": 1}, "pyerrors.obs.Obs.__init__": {"tf": 1}, "pyerrors.obs.CObs.__init__": {"tf": 1}}, "df": 6}}, "p": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.input": {"tf": 1}, "pyerrors.input.bdio": {"tf": 1}, "pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}, "pyerrors.input.dobs": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.hadrons": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix.__init__": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix.g5H": {"tf": 1}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}, "pyerrors.input.json": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}, "pyerrors.input.misc": {"tf": 1}, "pyerrors.input.misc.read_pbp": {"tf": 1}, "pyerrors.input.openQCD": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.qtop_projection": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}, "pyerrors.input.pandas": {"tf": 1}, "pyerrors.input.pandas.to_sql": {"tf": 1}, "pyerrors.input.pandas.read_sql": {"tf": 1}, "pyerrors.input.pandas.dump_df": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1}, "pyerrors.input.sfcf": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.input.utils": {"tf": 1}, "pyerrors.input.utils.check_idl": {"tf": 1}}, "df": 49}}}, "v": {"docs": {"pyerrors.linalg.inv": {"tf": 1}}, "df": 1}}, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "m": {"docs": {"pyerrors.correlators.Corr.item": {"tf": 1}}, "df": 1}}}, "s": {"docs": {"pyerrors.correlators.Corr.is_matrix_symmetric": {"tf": 1}, "pyerrors.obs.Obs.is_zero_within_error": {"tf": 1}, "pyerrors.obs.Obs.is_zero": {"tf": 1}, "pyerrors.obs.CObs.is_zero": {"tf": 1}}, "df": 4}, "m": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.obs.import_jackknife": {"tf": 1}}, "df": 3}}}}}, "d": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.input.utils.check_idl": {"tf": 1}}, "df": 1}}}, "g": {"5": {"docs": {}, "df": 0, "h": {"docs": {"pyerrors.input.hadrons.Npr_matrix.g5H": {"tf": 1}}, "df": 1}}, "docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "a": {"docs": {"pyerrors.correlators.Corr.gamma_method": {"tf": 1}, "pyerrors.dirac.Grid_gamma": {"tf": 1}, "pyerrors.fits.Fit_result.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.CObs.gamma_method": {"tf": 1}}, "df": 5}}}}, "m": {"docs": {"pyerrors.correlators.Corr.gm": {"tf": 1}, "pyerrors.fits.Fit_result.gm": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}}, "df": 3}, "e": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, "p": {"docs": {"pyerrors.correlators.Corr.GEVP": {"tf": 1}}, "df": 1}}, "n": {"docs": {"pyerrors.misc.gen_correlated_data": {"tf": 1}}, "df": 1}}, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.dirac.Grid_gamma": {"tf": 1}}, "df": 1}}}, "f": {"docs": {"pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}}, "df": 1}}, "m": {"docs": {"pyerrors.correlators.Corr.m_eff": {"tf": 1}}, "df": 1, "e": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.gamma_method": {"tf": 1}, "pyerrors.fits.Fit_result.gamma_method": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.CObs.gamma_method": {"tf": 1}}, "df": 5}}}}, "s": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}}, "df": 1, "s": {"docs": {"pyerrors.input.bdio.read_mesons": {"tf": 1}}, "df": 1}}}}, "r": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.obs.merge_obs": {"tf": 1}}, "df": 1}}}}, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "x": {"docs": {"pyerrors.correlators.Corr.is_matrix_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.matrix_symmetric": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix.__init__": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix.g5H": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 6}}}, "m": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.linalg.matmul": {"tf": 1}, "pyerrors.linalg.jack_matmul": {"tf": 1}}, "df": 2}}}}}, "i": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "c": {"docs": {"pyerrors.input.misc": {"tf": 1}, "pyerrors.input.misc.read_pbp": {"tf": 1}, "pyerrors.misc": {"tf": 1}, "pyerrors.misc.dump_object": {"tf": 1}, "pyerrors.misc.load_object": {"tf": 1}, "pyerrors.misc.pseudo_Obs": {"tf": 1}, "pyerrors.misc.gen_correlated_data": {"tf": 1}}, "df": 7}}}, "s": {"5": {"docs": {"pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "p": {"docs": {}, "df": 0, "m": {"docs": {"pyerrors.mpm": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 2}}}, "s": {"docs": {}, "df": 0, "y": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "c": {"docs": {"pyerrors.correlators.Corr.symmetric": {"tf": 1}, "pyerrors.correlators.Corr.anti_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.is_matrix_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.matrix_symmetric": {"tf": 1}}, "df": 4}}, "y": {"docs": {"pyerrors.correlators.Corr.T_symmetry": {"tf": 1}}, "df": 1}}}}}}}, "e": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.second_deriv": {"tf": 1}}, "df": 1}}}, "t": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 1}}}}, "t": {"docs": {"pyerrors.correlators.Corr.set_prange": {"tf": 1}}, "df": 1}}, "h": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "w": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1}}, "df": 1}}}, "p": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {"pyerrors.correlators.Corr.spaghetti_plot": {"tf": 1}}, "df": 1}}}}}}}}, "q": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.sqrt": {"tf": 1}, "pyerrors.obs.Obs.sqrt": {"tf": 1}}, "df": 2}}, "u": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}}, "df": 2}}}}}, "l": {"docs": {"pyerrors.input.pandas.to_sql": {"tf": 1}, "pyerrors.input.pandas.read_sql": {"tf": 1}}, "df": 2}}, "i": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.correlators.Corr.sin": {"tf": 1}, "pyerrors.obs.Obs.sin": {"tf": 1}}, "df": 2, "h": {"docs": {"pyerrors.correlators.Corr.sinh": {"tf": 1}, "pyerrors.obs.Obs.sinh": {"tf": 1}}, "df": 2}}}, "t": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.import_json_string": {"tf": 1}}, "df": 5}}}}}, "f": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "f": {"docs": {"pyerrors.input.sfcf": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.4142135623730951}}, "df": 2}}}, "v": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.linalg.svd": {"tf": 1}}, "df": 1}}}, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {"pyerrors.correlators.Corr.anti_symmetric": {"tf": 1}}, "df": 1}}}, "r": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.correlators.Corr.arcsin": {"tf": 1}, "pyerrors.obs.Obs.arcsin": {"tf": 1}}, "df": 2, "h": {"docs": {"pyerrors.correlators.Corr.arcsinh": {"tf": 1}, "pyerrors.obs.Obs.arcsinh": {"tf": 1}}, "df": 2}}}}, "c": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.correlators.Corr.arccos": {"tf": 1}, "pyerrors.obs.Obs.arccos": {"tf": 1}}, "df": 2, "h": {"docs": {"pyerrors.correlators.Corr.arccosh": {"tf": 1}, "pyerrors.obs.Obs.arccosh": {"tf": 1}}, "df": 2}}}}, "t": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.correlators.Corr.arctan": {"tf": 1}, "pyerrors.obs.Obs.arctan": {"tf": 1}}, "df": 2, "h": {"docs": {"pyerrors.correlators.Corr.arctanh": {"tf": 1}, "pyerrors.obs.Obs.arctanh": {"tf": 1}}, "df": 2}}}}}}, "d": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}}, "df": 2}}}}}}}}, "e": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.linalg.eig": {"tf": 1}}, "df": 1, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.Eigenvalue": {"tf": 1}}, "df": 1}}}}}}}, "h": {"docs": {"pyerrors.linalg.eigh": {"tf": 1}}, "df": 1}}, "n": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "m": {"docs": {"pyerrors.linalg.einsum": {"tf": 1}}, "df": 1}}}}}, "f": {"docs": {}, "df": 0, "f": {"docs": {"pyerrors.correlators.Corr.m_eff": {"tf": 1}}, "df": 1}}, "x": {"docs": {}, "df": 0, "p": {"docs": {"pyerrors.correlators.Corr.exp": {"tf": 1}, "pyerrors.obs.Obs.exp": {"tf": 1}}, "df": 2, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.obs.Obs.export_jackknife": {"tf": 1}}, "df": 1}}}}, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}}, "df": 1}}}}}}}}, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 1}}}}}}, "r": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "q": {"docs": {"pyerrors.covobs.Covobs.errsq": {"tf": 1}}, "df": 1}}, "o": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.fits.error_band": {"tf": 1}, "pyerrors.obs.Obs.is_zero_within_error": {"tf": 1}}, "df": 2}}}}, "p": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.dirac.epsilon_tensor": {"tf": 1}, "pyerrors.dirac.epsilon_tensor_rank4": {"tf": 1}}, "df": 2}}}}}}}, "h": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "k": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.correlators.Corr.Hankel": {"tf": 1}}, "df": 1}}}}, "d": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.hadrons": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix.__init__": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix.g5H": {"tf": 1}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}}, "df": 9}}}}}}, "d": {"5": {"docs": {"pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}}, "df": 5}, "docs": {}, "df": 0}, "i": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.obs.Obs.plot_history": {"tf": 1}}, "df": 1}}}}}}}, "r": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.correlators.Corr.roll": {"tf": 1}}, "df": 1}}, "o": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.roots.find_root": {"tf": 1}}, "df": 1, "s": {"docs": {"pyerrors.roots": {"tf": 1}, "pyerrors.roots.find_root": {"tf": 1}}, "df": 2}}}}, "e": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.reverse": {"tf": 1}}, "df": 1}}}}}, "w": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.reweight": {"tf": 1}, "pyerrors.obs.Obs.reweight": {"tf": 1}, "pyerrors.obs.reweight": {"tf": 1}}, "df": 3}}}}}}, "s": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.fits.Fit_result": {"tf": 1}, "pyerrors.fits.Fit_result.__init__": {"tf": 1}, "pyerrors.fits.Fit_result.gamma_method": {"tf": 1}, "pyerrors.fits.Fit_result.gm": {"tf": 1}}, "df": 4}}}, "i": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.fits.residual_plot": {"tf": 1}}, "df": 1}}}}}}, "a": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}, "pyerrors.input.misc.read_pbp": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}, "pyerrors.input.pandas.read_sql": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 18}}, "p": {"docs": {"pyerrors.obs.Obs.plot_rep_dist": {"tf": 1}}, "df": 1}}, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "k": {"4": {"docs": {"pyerrors.dirac.epsilon_tensor_rank4": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}}}, "w": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.openQCD.read_rwms": {"tf": 1}}, "df": 1}}}, "h": {"docs": {}, "df": 0, "o": {"docs": {"pyerrors.obs.Obs.plot_rho": {"tf": 1}}, "df": 1}}}, "t": {"0": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 1}, "docs": {"pyerrors.correlators.Corr.T_symmetry": {"tf": 1}}, "df": 1, "h": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.correlators.Corr.thin": {"tf": 1}}, "df": 1}}}, "a": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.correlators.Corr.tan": {"tf": 1}, "pyerrors.obs.Obs.tan": {"tf": 1}}, "df": 2, "h": {"docs": {"pyerrors.correlators.Corr.tanh": {"tf": 1}, "pyerrors.obs.Obs.tanh": {"tf": 1}}, "df": 2}}, "u": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.obs.Obs.plot_tauint": {"tf": 1}}, "df": 1}}}}}, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.dirac.epsilon_tensor": {"tf": 1}, "pyerrors.dirac.epsilon_tensor_rank4": {"tf": 1}}, "df": 2}}}}, "s": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.fits.ks_test": {"tf": 1}}, "df": 1}}}, "o": {"docs": {"pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.pandas.to_sql": {"tf": 1}}, "df": 3, "t": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.fits.total_least_squares": {"tf": 1}}, "df": 1}}}}}, "d": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "v": {"docs": {"pyerrors.correlators.Corr.deriv": {"tf": 1}, "pyerrors.correlators.Corr.second_deriv": {"tf": 1}}, "df": 2, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.obs.derived_observable": {"tf": 1}}, "df": 1}}}}}, "t": {"docs": {"pyerrors.linalg.det": {"tf": 1}}, "df": 1, "a": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.obs.Obs.details": {"tf": 1}}, "df": 1}}}}}}, "u": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "p": {"docs": {"pyerrors.correlators.Corr.dump": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.pandas.dump_df": {"tf": 1}, "pyerrors.misc.dump_object": {"tf": 1}, "pyerrors.obs.Obs.dump": {"tf": 1}}, "df": 6}}}, "i": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "c": {"docs": {"pyerrors.dirac": {"tf": 1}, "pyerrors.dirac.epsilon_tensor": {"tf": 1}, "pyerrors.dirac.epsilon_tensor_rank4": {"tf": 1}, "pyerrors.dirac.Grid_gamma": {"tf": 1}}, "df": 4}}}, "s": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.obs.Obs.plot_rep_dist": {"tf": 1}}, "df": 1, "i": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}}, "df": 1}}}}}}}}}}}}}}}}}}}}}, "c": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}}, "df": 2}}}, "s": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "m": {"docs": {"pyerrors.input.bdio.read_dSdm": {"tf": 1}}, "df": 1}}}, "o": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.dobs": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.read_dobs": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.write_dobs": {"tf": 1.4142135623730951}}, "df": 8}}}, "f": {"docs": {"pyerrors.input.pandas.dump_df": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1}}, "df": 2}, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "a": {"docs": {"pyerrors.misc.gen_correlated_data": {"tf": 1}}, "df": 1}}}}, "f": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.fits.Fit_result": {"tf": 1}, "pyerrors.fits.Fit_result.__init__": {"tf": 1}, "pyerrors.fits.Fit_result.gamma_method": {"tf": 1}, "pyerrors.fits.Fit_result.gm": {"tf": 1}, "pyerrors.fits.fit_lin": {"tf": 1}}, "df": 6, "s": {"docs": {"pyerrors.fits": {"tf": 1}, "pyerrors.fits.Fit_result": {"tf": 1}, "pyerrors.fits.Fit_result.__init__": {"tf": 1}, "pyerrors.fits.Fit_result.gamma_method": {"tf": 1}, "pyerrors.fits.Fit_result.gm": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.fits.fit_lin": {"tf": 1}, "pyerrors.fits.qqplot": {"tf": 1}, "pyerrors.fits.residual_plot": {"tf": 1}, "pyerrors.fits.error_band": {"tf": 1}, "pyerrors.fits.ks_test": {"tf": 1}}, "df": 12}}, "n": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.roots.find_root": {"tf": 1}}, "df": 1}}}, "o": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "q": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "k": {"docs": {"pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}}, "df": 1}}}}}}}}}, "l": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.correlators.Corr.log": {"tf": 1}, "pyerrors.obs.Obs.log": {"tf": 1}}, "df": 2}, "a": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1}, "pyerrors.misc.load_object": {"tf": 1}}, "df": 4}}}, "e": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}}, "df": 2}}}}, "i": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.fits.fit_lin": {"tf": 1}}, "df": 1, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.linalg": {"tf": 1}, "pyerrors.linalg.matmul": {"tf": 1}, "pyerrors.linalg.jack_matmul": {"tf": 1}, "pyerrors.linalg.einsum": {"tf": 1}, "pyerrors.linalg.inv": {"tf": 1}, "pyerrors.linalg.cholesky": {"tf": 1}, "pyerrors.linalg.det": {"tf": 1}, "pyerrors.linalg.eigh": {"tf": 1}, "pyerrors.linalg.eig": {"tf": 1}, "pyerrors.linalg.pinv": {"tf": 1}, "pyerrors.linalg.svd": {"tf": 1}}, "df": 11}}}}}}, "q": {"docs": {}, "df": 0, "q": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.fits.qqplot": {"tf": 1}}, "df": 1}}}}}, "t": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "p": {"docs": {"pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.qtop_projection": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 3}}}}, "b": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.fits.error_band": {"tf": 1}}, "df": 1}}}, "d": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {"pyerrors.input.bdio": {"tf": 1}, "pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}}, "df": 5}}}, "i": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}}, "df": 1}}}}}}}}, "k": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.fits.ks_test": {"tf": 1}}, "df": 1}}, "w": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}}, "df": 3}}}}, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.obs.Obs.is_zero_within_error": {"tf": 1}}, "df": 1}}}}}}, "n": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix.__init__": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix.g5H": {"tf": 1}}, "df": 3}}}, "j": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.json": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1.4142135623730951}, "pyerrors.input.json.dump_to_json": {"tf": 1.4142135623730951}, "pyerrors.input.json.import_json_string": {"tf": 1.4142135623730951}, "pyerrors.input.json.load_json": {"tf": 1.4142135623730951}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1.4142135623730951}, "pyerrors.input.json.load_json_dict": {"tf": 1.4142135623730951}}, "df": 7}}}, "a": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "k": {"docs": {"pyerrors.linalg.jack_matmul": {"tf": 1}}, "df": 1, "k": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "f": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.obs.Obs.export_jackknife": {"tf": 1}, "pyerrors.obs.import_jackknife": {"tf": 1}}, "df": 2}}}}}}}}}, "o": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "q": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.openQCD": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.qtop_projection": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}}, "df": 8}}}}}}, "b": {"docs": {}, "df": 0, "j": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.misc.dump_object": {"tf": 1}, "pyerrors.misc.load_object": {"tf": 1}}, "df": 2}}}}, "s": {"docs": {"pyerrors.misc.pseudo_Obs": {"tf": 1}, "pyerrors.obs": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.__init__": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.gamma_method": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.gm": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.details": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.reweight": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.is_zero_within_error": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.is_zero": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.plot_tauint": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.plot_rho": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.plot_rep_dist": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.plot_history": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.plot_piechart": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.dump": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.sqrt": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.log": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.exp": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.sin": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.cos": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.tan": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.arcsin": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.arccos": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.arctan": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.sinh": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.cosh": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.tanh": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.arcsinh": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.arccosh": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.arctanh": {"tf": 1.4142135623730951}, "pyerrors.obs.CObs": {"tf": 1}, "pyerrors.obs.CObs.__init__": {"tf": 1}, "pyerrors.obs.CObs.gamma_method": {"tf": 1}, "pyerrors.obs.CObs.is_zero": {"tf": 1}, "pyerrors.obs.CObs.conjugate": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}, "pyerrors.obs.reweight": {"tf": 1}, "pyerrors.obs.correlate": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}, "pyerrors.obs.import_jackknife": {"tf": 1}, "pyerrors.obs.merge_obs": {"tf": 1.4142135623730951}, "pyerrors.obs.cov_Obs": {"tf": 1.4142135623730951}}, "df": 44, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.obs.derived_observable": {"tf": 1}}, "df": 1}}}}}}}}}}, "x": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "f": {"docs": {"pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}}, "df": 1}}}, "u": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.utils": {"tf": 1}, "pyerrors.input.utils.check_idl": {"tf": 1}}, "df": 2}}}}}, "z": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "o": {"docs": {"pyerrors.obs.Obs.is_zero_within_error": {"tf": 1}, "pyerrors.obs.Obs.is_zero": {"tf": 1}, "pyerrors.obs.CObs.is_zero": {"tf": 1}}, "df": 3}}}}, "v": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.version": {"tf": 1}}, "df": 1}}}}}}}}}, "annotation": {"root": {"docs": {}, "df": 0}}, "default_value": {"root": {"docs": {}, "df": 0}}, "signature": {"root": {"0": {"docs": {"pyerrors.correlators.Corr.__init__": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 1}, "pyerrors.correlators.Corr.thin": {"tf": 1}, "pyerrors.correlators.Corr.m_eff": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.qtop_projection": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.7320508075688772}, "pyerrors.misc.gen_correlated_data": {"tf": 1}, "pyerrors.obs.CObs.__init__": {"tf": 1.4142135623730951}, "pyerrors.roots.find_root": {"tf": 1}}, "df": 14, "c": {"docs": {"pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 1}}, "1": {"0": {"0": {"0": {"docs": {"pyerrors.misc.pseudo_Obs": {"tf": 1}, "pyerrors.misc.gen_correlated_data": {"tf": 1}}, "df": 2}, "docs": {}, "df": 0}, "docs": {"pyerrors.obs.Obs.is_zero": {"tf": 1}}, "df": 1}, "docs": {"pyerrors.correlators.Corr.T_symmetry": {"tf": 1}, "pyerrors.correlators.Corr.m_eff": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}, "pyerrors.obs.Obs.is_zero_within_error": {"tf": 1}, "pyerrors.roots.find_root": {"tf": 1}}, "df": 11, "e": {"docs": {"pyerrors.obs.Obs.is_zero": {"tf": 1}}, "df": 1}}, "2": {"docs": {"pyerrors.correlators.Corr.thin": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}}, "df": 3}, "3": {"9": {"docs": {"pyerrors.correlators.Corr.GEVP": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.deriv": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.second_deriv": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.m_eff": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.plateau": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.dump": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.read_ADerrors": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.read_mesons": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.read_dSdm": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.create_pobs_string": {"tf": 2}, "pyerrors.input.dobs.write_pobs": {"tf": 2}, "pyerrors.input.dobs.create_dobs_string": {"tf": 2}, "pyerrors.input.dobs.write_dobs": {"tf": 2}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 2}, "pyerrors.input.json.create_json_string": {"tf": 1.4142135623730951}, "pyerrors.input.json.dump_to_json": {"tf": 1.4142135623730951}, "pyerrors.input.json.dump_dict_to_json": {"tf": 2}, "pyerrors.input.json.load_json_dict": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_rwms": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1.4142135623730951}, "pyerrors.input.pandas.to_sql": {"tf": 1.4142135623730951}, "pyerrors.input.sfcf.read_sfcf": {"tf": 2.8284271247461903}, "pyerrors.obs.Obs.dump": {"tf": 2}}, "df": 28}, "docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 1}, "5": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.misc.gen_correlated_data": {"tf": 1}}, "df": 2}, "docs": {"pyerrors.correlators.Corr.__init__": {"tf": 5.744562646538029}, "pyerrors.correlators.Corr.gamma_method": {"tf": 4}, "pyerrors.correlators.Corr.gm": {"tf": 4}, "pyerrors.correlators.Corr.projected": {"tf": 5.830951894845301}, "pyerrors.correlators.Corr.item": {"tf": 4.242640687119285}, "pyerrors.correlators.Corr.plottable": {"tf": 3.1622776601683795}, "pyerrors.correlators.Corr.symmetric": {"tf": 3.1622776601683795}, "pyerrors.correlators.Corr.anti_symmetric": {"tf": 3.1622776601683795}, "pyerrors.correlators.Corr.is_matrix_symmetric": {"tf": 3.1622776601683795}, "pyerrors.correlators.Corr.matrix_symmetric": {"tf": 3.1622776601683795}, "pyerrors.correlators.Corr.GEVP": {"tf": 6.164414002968976}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 6.324555320336759}, "pyerrors.correlators.Corr.Hankel": {"tf": 4.69041575982343}, "pyerrors.correlators.Corr.roll": {"tf": 3.7416573867739413}, "pyerrors.correlators.Corr.reverse": {"tf": 3.1622776601683795}, "pyerrors.correlators.Corr.thin": {"tf": 5.0990195135927845}, "pyerrors.correlators.Corr.correlate": {"tf": 3.7416573867739413}, "pyerrors.correlators.Corr.reweight": {"tf": 4.47213595499958}, "pyerrors.correlators.Corr.T_symmetry": {"tf": 4.69041575982343}, "pyerrors.correlators.Corr.deriv": {"tf": 4.47213595499958}, "pyerrors.correlators.Corr.second_deriv": {"tf": 4.47213595499958}, "pyerrors.correlators.Corr.m_eff": {"tf": 5.291502622129181}, "pyerrors.correlators.Corr.fit": {"tf": 6}, "pyerrors.correlators.Corr.plateau": {"tf": 6}, "pyerrors.correlators.Corr.set_prange": {"tf": 3.7416573867739413}, "pyerrors.correlators.Corr.show": {"tf": 10.908712114635714}, "pyerrors.correlators.Corr.spaghetti_plot": {"tf": 4.242640687119285}, "pyerrors.correlators.Corr.dump": {"tf": 5.477225575051661}, "pyerrors.correlators.Corr.print": {"tf": 4.242640687119285}, "pyerrors.correlators.Corr.sqrt": {"tf": 3.1622776601683795}, "pyerrors.correlators.Corr.log": {"tf": 3.1622776601683795}, "pyerrors.correlators.Corr.exp": {"tf": 3.1622776601683795}, "pyerrors.correlators.Corr.sin": {"tf": 3.1622776601683795}, "pyerrors.correlators.Corr.cos": {"tf": 3.1622776601683795}, "pyerrors.correlators.Corr.tan": {"tf": 3.1622776601683795}, "pyerrors.correlators.Corr.sinh": {"tf": 3.1622776601683795}, "pyerrors.correlators.Corr.cosh": {"tf": 3.1622776601683795}, "pyerrors.correlators.Corr.tanh": {"tf": 3.1622776601683795}, "pyerrors.correlators.Corr.arcsin": {"tf": 3.1622776601683795}, "pyerrors.correlators.Corr.arccos": {"tf": 3.1622776601683795}, "pyerrors.correlators.Corr.arctan": {"tf": 3.1622776601683795}, "pyerrors.correlators.Corr.arcsinh": {"tf": 3.1622776601683795}, "pyerrors.correlators.Corr.arccosh": {"tf": 3.1622776601683795}, "pyerrors.correlators.Corr.arctanh": {"tf": 3.1622776601683795}, "pyerrors.correlators.Corr.prune": {"tf": 6.164414002968976}, "pyerrors.covobs.Covobs.__init__": {"tf": 5.656854249492381}, "pyerrors.covobs.Covobs.errsq": {"tf": 3.1622776601683795}, "pyerrors.dirac.epsilon_tensor": {"tf": 4.242640687119285}, "pyerrors.dirac.epsilon_tensor_rank4": {"tf": 4.69041575982343}, "pyerrors.dirac.Grid_gamma": {"tf": 3.1622776601683795}, "pyerrors.fits.Fit_result.__init__": {"tf": 2}, "pyerrors.fits.Fit_result.gamma_method": {"tf": 4}, "pyerrors.fits.Fit_result.gm": {"tf": 4}, "pyerrors.fits.least_squares": {"tf": 6.324555320336759}, "pyerrors.fits.total_least_squares": {"tf": 5.656854249492381}, "pyerrors.fits.fit_lin": {"tf": 4.47213595499958}, "pyerrors.fits.qqplot": {"tf": 4.69041575982343}, "pyerrors.fits.residual_plot": {"tf": 4.69041575982343}, "pyerrors.fits.error_band": {"tf": 4.242640687119285}, "pyerrors.fits.ks_test": {"tf": 3.7416573867739413}, "pyerrors.input.bdio.read_ADerrors": {"tf": 5.0990195135927845}, "pyerrors.input.bdio.write_ADerrors": {"tf": 5.477225575051661}, "pyerrors.input.bdio.read_mesons": {"tf": 5.0990195135927845}, "pyerrors.input.bdio.read_dSdm": {"tf": 5.0990195135927845}, "pyerrors.input.dobs.create_pobs_string": {"tf": 7.14142842854285}, "pyerrors.input.dobs.write_pobs": {"tf": 8.426149773176359}, "pyerrors.input.dobs.read_pobs": {"tf": 5.830951894845301}, "pyerrors.input.dobs.import_dobs_string": {"tf": 5.830951894845301}, "pyerrors.input.dobs.read_dobs": {"tf": 6.855654600401044}, "pyerrors.input.dobs.create_dobs_string": {"tf": 8.12403840463596}, "pyerrors.input.dobs.write_dobs": {"tf": 8.94427190999916}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 6.6332495807108}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 6}, "pyerrors.input.hadrons.Npr_matrix.__init__": {"tf": 2}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 5.0990195135927845}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 5.0990195135927845}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 6.855654600401044}, "pyerrors.input.json.create_json_string": {"tf": 5.291502622129181}, "pyerrors.input.json.dump_to_json": {"tf": 6.324555320336759}, "pyerrors.input.json.import_json_string": {"tf": 5.0990195135927845}, "pyerrors.input.json.load_json": {"tf": 5.830951894845301}, "pyerrors.input.json.dump_dict_to_json": {"tf": 7.0710678118654755}, "pyerrors.input.json.load_json_dict": {"tf": 6.6332495807108}, "pyerrors.input.misc.read_pbp": {"tf": 4.47213595499958}, "pyerrors.input.openQCD.read_rwms": {"tf": 6.164414002968976}, "pyerrors.input.openQCD.extract_t0": {"tf": 6.324555320336759}, "pyerrors.input.openQCD.read_qtop": {"tf": 6.48074069840786}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 6.324555320336759}, "pyerrors.input.openQCD.qtop_projection": {"tf": 4.242640687119285}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 5.656854249492381}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 6.164414002968976}, "pyerrors.input.pandas.to_sql": {"tf": 6.48074069840786}, "pyerrors.input.pandas.read_sql": {"tf": 5.291502622129181}, "pyerrors.input.pandas.dump_df": {"tf": 4.69041575982343}, "pyerrors.input.pandas.load_df": {"tf": 5.0990195135927845}, "pyerrors.input.sfcf.read_sfcf": {"tf": 10}, "pyerrors.input.utils.check_idl": {"tf": 3.7416573867739413}, "pyerrors.linalg.matmul": {"tf": 3.4641016151377544}, "pyerrors.linalg.jack_matmul": {"tf": 3.4641016151377544}, "pyerrors.linalg.einsum": {"tf": 4}, "pyerrors.linalg.inv": {"tf": 3.1622776601683795}, "pyerrors.linalg.cholesky": {"tf": 3.1622776601683795}, "pyerrors.linalg.det": {"tf": 3.1622776601683795}, "pyerrors.linalg.eigh": {"tf": 4}, "pyerrors.linalg.eig": {"tf": 4}, "pyerrors.linalg.pinv": {"tf": 4}, "pyerrors.linalg.svd": {"tf": 4}, "pyerrors.misc.dump_object": {"tf": 4.47213595499958}, "pyerrors.misc.load_object": {"tf": 3.1622776601683795}, "pyerrors.misc.pseudo_Obs": {"tf": 5.0990195135927845}, "pyerrors.misc.gen_correlated_data": {"tf": 5.830951894845301}, "pyerrors.mpm.matrix_pencil_method": {"tf": 5.656854249492381}, "pyerrors.obs.Obs.__init__": {"tf": 5.0990195135927845}, "pyerrors.obs.Obs.gamma_method": {"tf": 4}, "pyerrors.obs.Obs.gm": {"tf": 4}, "pyerrors.obs.Obs.details": {"tf": 4.242640687119285}, "pyerrors.obs.Obs.reweight": {"tf": 3.7416573867739413}, "pyerrors.obs.Obs.is_zero_within_error": {"tf": 4.242640687119285}, "pyerrors.obs.Obs.is_zero": {"tf": 4.242640687119285}, "pyerrors.obs.Obs.plot_tauint": {"tf": 4.242640687119285}, "pyerrors.obs.Obs.plot_rho": {"tf": 4.242640687119285}, "pyerrors.obs.Obs.plot_rep_dist": {"tf": 3.1622776601683795}, "pyerrors.obs.Obs.plot_history": {"tf": 4.242640687119285}, "pyerrors.obs.Obs.plot_piechart": {"tf": 4.242640687119285}, "pyerrors.obs.Obs.dump": {"tf": 6.324555320336759}, "pyerrors.obs.Obs.export_jackknife": {"tf": 3.1622776601683795}, "pyerrors.obs.Obs.sqrt": {"tf": 3.1622776601683795}, "pyerrors.obs.Obs.log": {"tf": 3.1622776601683795}, "pyerrors.obs.Obs.exp": {"tf": 3.1622776601683795}, "pyerrors.obs.Obs.sin": {"tf": 3.1622776601683795}, "pyerrors.obs.Obs.cos": {"tf": 3.1622776601683795}, "pyerrors.obs.Obs.tan": {"tf": 3.1622776601683795}, "pyerrors.obs.Obs.arcsin": {"tf": 3.1622776601683795}, "pyerrors.obs.Obs.arccos": {"tf": 3.1622776601683795}, "pyerrors.obs.Obs.arctan": {"tf": 3.1622776601683795}, "pyerrors.obs.Obs.sinh": {"tf": 3.1622776601683795}, "pyerrors.obs.Obs.cosh": {"tf": 3.1622776601683795}, "pyerrors.obs.Obs.tanh": {"tf": 3.1622776601683795}, "pyerrors.obs.Obs.arcsinh": {"tf": 3.1622776601683795}, "pyerrors.obs.Obs.arccosh": {"tf": 3.1622776601683795}, "pyerrors.obs.Obs.arctanh": {"tf": 3.1622776601683795}, "pyerrors.obs.CObs.__init__": {"tf": 4}, "pyerrors.obs.CObs.gamma_method": {"tf": 4}, "pyerrors.obs.CObs.is_zero": {"tf": 3.1622776601683795}, "pyerrors.obs.CObs.conjugate": {"tf": 3.1622776601683795}, "pyerrors.obs.derived_observable": {"tf": 5.291502622129181}, "pyerrors.obs.reweight": {"tf": 4.47213595499958}, "pyerrors.obs.correlate": {"tf": 3.7416573867739413}, "pyerrors.obs.covariance": {"tf": 6.324555320336759}, "pyerrors.obs.import_jackknife": {"tf": 4.69041575982343}, "pyerrors.obs.merge_obs": {"tf": 3.1622776601683795}, "pyerrors.obs.cov_Obs": {"tf": 5.0990195135927845}, "pyerrors.roots.find_root": {"tf": 5.291502622129181}}, "df": 153, "d": {"docs": {"pyerrors.roots.find_root": {"tf": 1}}, "df": 1, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "a": {"docs": {"pyerrors.correlators.Corr.__init__": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}}, "df": 2, "t": {"docs": {}, "df": 0, "y": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.dump": {"tf": 1}, "pyerrors.obs.Obs.dump": {"tf": 1}}, "df": 2}}}}}}}, "t": {"docs": {"pyerrors.correlators.Corr.roll": {"tf": 1}}, "df": 1, "r": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}}, "df": 3}}, "o": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}}, "df": 2}}}, "i": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}}, "df": 1}}}}}}, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}}, "df": 1}}}}, "c": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}}, "df": 2}}}}}}, "e": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.obs.Obs.dump": {"tf": 1}}, "df": 4}}}}}}}}}}, "f": {"docs": {"pyerrors.input.pandas.to_sql": {"tf": 1}, "pyerrors.input.pandas.dump_df": {"tf": 1}}, "df": 2}, "b": {"docs": {"pyerrors.input.pandas.to_sql": {"tf": 1}, "pyerrors.input.pandas.read_sql": {"tf": 1}}, "df": 2}, "v": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.misc.pseudo_Obs": {"tf": 1}}, "df": 1}}}}}}, "i": {"docs": {"pyerrors.correlators.Corr.item": {"tf": 1}, "pyerrors.dirac.epsilon_tensor": {"tf": 1}, "pyerrors.dirac.epsilon_tensor_rank4": {"tf": 1}}, "df": 3, "n": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.__init__": {"tf": 1}}, "df": 1}}}, "s": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}}, "df": 3}}}}}}}, "d": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}}, "df": 3}}}}}, "d": {"docs": {"pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}}, "df": 5, "l": {"docs": {"pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}, "pyerrors.input.utils.check_idl": {"tf": 1}, "pyerrors.obs.Obs.__init__": {"tf": 1}, "pyerrors.obs.import_jackknife": {"tf": 1}}, "df": 8}}, "f": {"docs": {"pyerrors.input.pandas.to_sql": {"tf": 1}}, "df": 1}, "m": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.obs.CObs.__init__": {"tf": 1}}, "df": 1}}}}, "p": {"docs": {"pyerrors.fits.qqplot": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 2, "a": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.correlators.Corr.__init__": {"tf": 1}}, "df": 1}}}}}, "r": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.correlators.Corr.correlate": {"tf": 1}, "pyerrors.correlators.Corr.T_symmetry": {"tf": 1}}, "df": 2}}}}, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.correlators.Corr.T_symmetry": {"tf": 1}}, "df": 1}}}}, "t": {"docs": {}, "df": 0, "h": {"docs": {"pyerrors.input.bdio.read_ADerrors": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.read_mesons": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.read_dSdm": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}, "pyerrors.input.misc.read_pbp": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.misc.load_object": {"tf": 1}}, "df": 18}}}, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.__init__": {"tf": 1}, "pyerrors.correlators.Corr.set_prange": {"tf": 1}}, "df": 2}}}}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.print": {"tf": 1}}, "df": 1}}, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}}, "df": 1}}}}, "e": {"docs": {}, "df": 0, "f": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "x": {"docs": {"pyerrors.input.misc.read_pbp": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 8}}}}}, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "c": {"docs": {"pyerrors.correlators.Corr.Hankel": {"tf": 1}}, "df": 1}}}}}}}, "l": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "u": {"docs": {"pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1}}, "df": 2}}}}}}, "o": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.covobs.Covobs.__init__": {"tf": 1}}, "df": 1}}}, "n": {"docs": {"pyerrors.correlators.Corr.Hankel": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 2, "o": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.__init__": {"tf": 1}, "pyerrors.correlators.Corr.projected": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 1}, "pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 3.1622776601683795}, "pyerrors.correlators.Corr.print": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1}, "pyerrors.covobs.Covobs.__init__": {"tf": 1.4142135623730951}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.ks_test": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.write_dobs": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}, "pyerrors.obs.Obs.__init__": {"tf": 1}, "pyerrors.obs.Obs.plot_tauint": {"tf": 1}, "pyerrors.obs.Obs.plot_rho": {"tf": 1}, "pyerrors.obs.Obs.plot_piechart": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}, "pyerrors.obs.import_jackknife": {"tf": 1}, "pyerrors.obs.cov_Obs": {"tf": 1}}, "df": 31}}, "r": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "z": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.projected": {"tf": 1}}, "df": 1}}}}}}}, "e": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}}, "df": 2}}}}}, "f": {"docs": {}, "df": 0, "f": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 1}}}}}}, "t": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "c": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 1}}}}}, "a": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.pandas.to_sql": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.misc.dump_object": {"tf": 1}, "pyerrors.misc.pseudo_Obs": {"tf": 1}, "pyerrors.misc.gen_correlated_data": {"tf": 1}, "pyerrors.obs.import_jackknife": {"tf": 1}, "pyerrors.obs.cov_Obs": {"tf": 1}}, "df": 12, "s": {"docs": {"pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.obs.Obs.__init__": {"tf": 1}}, "df": 2}}}}}, "s": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "f": {"docs": {"pyerrors.correlators.Corr.gamma_method": {"tf": 1}, "pyerrors.correlators.Corr.gm": {"tf": 1}, "pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.correlators.Corr.item": {"tf": 1}, "pyerrors.correlators.Corr.plottable": {"tf": 1}, "pyerrors.correlators.Corr.symmetric": {"tf": 1}, "pyerrors.correlators.Corr.anti_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.is_matrix_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.matrix_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 1}, "pyerrors.correlators.Corr.Hankel": {"tf": 1}, "pyerrors.correlators.Corr.roll": {"tf": 1}, "pyerrors.correlators.Corr.reverse": {"tf": 1}, "pyerrors.correlators.Corr.thin": {"tf": 1}, "pyerrors.correlators.Corr.correlate": {"tf": 1}, "pyerrors.correlators.Corr.reweight": {"tf": 1}, "pyerrors.correlators.Corr.T_symmetry": {"tf": 1}, "pyerrors.correlators.Corr.deriv": {"tf": 1}, "pyerrors.correlators.Corr.second_deriv": {"tf": 1}, "pyerrors.correlators.Corr.m_eff": {"tf": 1}, "pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.correlators.Corr.set_prange": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.correlators.Corr.spaghetti_plot": {"tf": 1}, "pyerrors.correlators.Corr.dump": {"tf": 1}, "pyerrors.correlators.Corr.print": {"tf": 1}, "pyerrors.correlators.Corr.sqrt": {"tf": 1}, "pyerrors.correlators.Corr.log": {"tf": 1}, "pyerrors.correlators.Corr.exp": {"tf": 1}, "pyerrors.correlators.Corr.sin": {"tf": 1}, "pyerrors.correlators.Corr.cos": {"tf": 1}, "pyerrors.correlators.Corr.tan": {"tf": 1}, "pyerrors.correlators.Corr.sinh": {"tf": 1}, "pyerrors.correlators.Corr.cosh": {"tf": 1}, "pyerrors.correlators.Corr.tanh": {"tf": 1}, "pyerrors.correlators.Corr.arcsin": {"tf": 1}, "pyerrors.correlators.Corr.arccos": {"tf": 1}, "pyerrors.correlators.Corr.arctan": {"tf": 1}, "pyerrors.correlators.Corr.arcsinh": {"tf": 1}, "pyerrors.correlators.Corr.arccosh": {"tf": 1}, "pyerrors.correlators.Corr.arctanh": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1}, "pyerrors.covobs.Covobs.errsq": {"tf": 1}, "pyerrors.fits.Fit_result.gamma_method": {"tf": 1}, "pyerrors.fits.Fit_result.gm": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}, "pyerrors.obs.Obs.details": {"tf": 1}, "pyerrors.obs.Obs.reweight": {"tf": 1}, "pyerrors.obs.Obs.is_zero_within_error": {"tf": 1}, "pyerrors.obs.Obs.is_zero": {"tf": 1}, "pyerrors.obs.Obs.plot_tauint": {"tf": 1}, "pyerrors.obs.Obs.plot_rho": {"tf": 1}, "pyerrors.obs.Obs.plot_rep_dist": {"tf": 1}, "pyerrors.obs.Obs.plot_history": {"tf": 1}, "pyerrors.obs.Obs.plot_piechart": {"tf": 1}, "pyerrors.obs.Obs.dump": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1}, "pyerrors.obs.Obs.sqrt": {"tf": 1}, "pyerrors.obs.Obs.log": {"tf": 1}, "pyerrors.obs.Obs.exp": {"tf": 1}, "pyerrors.obs.Obs.sin": {"tf": 1}, "pyerrors.obs.Obs.cos": {"tf": 1}, "pyerrors.obs.Obs.tan": {"tf": 1}, "pyerrors.obs.Obs.arcsin": {"tf": 1}, "pyerrors.obs.Obs.arccos": {"tf": 1}, "pyerrors.obs.Obs.arctan": {"tf": 1}, "pyerrors.obs.Obs.sinh": {"tf": 1}, "pyerrors.obs.Obs.cosh": {"tf": 1}, "pyerrors.obs.Obs.tanh": {"tf": 1}, "pyerrors.obs.Obs.arcsinh": {"tf": 1}, "pyerrors.obs.Obs.arccosh": {"tf": 1}, "pyerrors.obs.Obs.arctanh": {"tf": 1}, "pyerrors.obs.CObs.gamma_method": {"tf": 1}, "pyerrors.obs.CObs.is_zero": {"tf": 1}, "pyerrors.obs.CObs.conjugate": {"tf": 1}}, "df": 78}}, "p": {"docs": {"pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}}, "df": 1, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 4}}}}}}}}, "o": {"docs": {"pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}}, "df": 4, "r": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 1}}, "df": 2}}}, "t": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.Eigenvalue": {"tf": 1}}, "df": 1}}}, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.json.import_json_string": {"tf": 1}}, "df": 1}}}}}, "p": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.correlators.Corr.thin": {"tf": 1}}, "df": 1}}}}, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 1}}}}}, "e": {"docs": {}, "df": 0, "c": {"docs": {"pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}}, "df": 4}}}, "y": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "c": {"docs": {"pyerrors.correlators.Corr.deriv": {"tf": 1}, "pyerrors.correlators.Corr.second_deriv": {"tf": 1}}, "df": 2}}}}}}, "b": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}}, "df": 4}}}}}, "i": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}}, "df": 3}}}}, "g": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "a": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.obs.Obs.is_zero_within_error": {"tf": 1}}, "df": 2}}}}, "a": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.obs.Obs.plot_tauint": {"tf": 1}, "pyerrors.obs.Obs.plot_rho": {"tf": 1}, "pyerrors.obs.Obs.plot_piechart": {"tf": 1}}, "df": 4}}, "m": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.misc.pseudo_Obs": {"tf": 1}, "pyerrors.misc.gen_correlated_data": {"tf": 1}, "pyerrors.obs.Obs.__init__": {"tf": 1}}, "df": 3}}}}}}, "q": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.input.pandas.read_sql": {"tf": 1}}, "df": 1}}, "u": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.linalg.einsum": {"tf": 1}}, "df": 1}}}}}}}}}, "m": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "h": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}}}}, "k": {"docs": {"pyerrors.dirac.epsilon_tensor": {"tf": 1}, "pyerrors.dirac.epsilon_tensor_rank4": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 3, "w": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.correlators.Corr.gamma_method": {"tf": 1}, "pyerrors.correlators.Corr.gm": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.reweight": {"tf": 1}, "pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.correlators.Corr.dump": {"tf": 1}, "pyerrors.fits.Fit_result.gamma_method": {"tf": 1}, "pyerrors.fits.Fit_result.gm": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.fits.fit_lin": {"tf": 1}, "pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}, "pyerrors.input.misc.read_pbp": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}, "pyerrors.input.pandas.to_sql": {"tf": 1}, "pyerrors.input.pandas.read_sql": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.linalg.eigh": {"tf": 1}, "pyerrors.linalg.eig": {"tf": 1}, "pyerrors.linalg.pinv": {"tf": 1}, "pyerrors.linalg.svd": {"tf": 1}, "pyerrors.misc.dump_object": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}, "pyerrors.obs.Obs.__init__": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}, "pyerrors.obs.Obs.dump": {"tf": 1}, "pyerrors.obs.CObs.gamma_method": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}, "pyerrors.obs.reweight": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}, "pyerrors.roots.find_root": {"tf": 1}}, "df": 40}}}}}}, "v": {"1": {"docs": {"pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}}, "df": 2}, "docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.correlators.Corr.projected": {"tf": 1.4142135623730951}}, "df": 1}}}}, "r": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}}, "df": 1}}}}}, "b": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}}, "df": 3}}}}, "s": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 3}}}}}}, "a": {"docs": {"pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}}, "df": 1, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.deriv": {"tf": 1}, "pyerrors.correlators.Corr.second_deriv": {"tf": 1}, "pyerrors.correlators.Corr.m_eff": {"tf": 1}}, "df": 3}}}}}, "l": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.misc.pseudo_Obs": {"tf": 1}}, "df": 1}}}}, "i": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "z": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}}}}}}}, "l": {"docs": {"pyerrors.correlators.Corr.projected": {"tf": 1}}, "df": 1, "o": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.correlators.Corr.m_eff": {"tf": 1}}, "df": 1, "s": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.correlators.Corr.spaghetti_plot": {"tf": 1}}, "df": 2}}}}}}}, "i": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {"pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}}, "df": 4}}}}}, "s": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.obs.merge_obs": {"tf": 1}}, "df": 2}}}}, "r": {"docs": {"pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}}, "df": 2, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.print": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 4}}}}, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.fits.residual_plot": {"tf": 1}}, "df": 2}, "f": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1}}, "df": 1}}}}}}}}, "p": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}}, "df": 2}}, "a": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 1}, "l": {"docs": {"pyerrors.obs.CObs.__init__": {"tf": 1}}, "df": 1}}}}, "f": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.correlators.Corr.Hankel": {"tf": 1}, "pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1.4142135623730951}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.read_dobs": {"tf": 1.4142135623730951}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}, "pyerrors.input.pandas.read_sql": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1.4142135623730951}}, "df": 17}}}, "i": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.input.pandas.to_sql": {"tf": 1}}, "df": 1}}}, "u": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "c": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.fits.qqplot": {"tf": 1}, "pyerrors.fits.residual_plot": {"tf": 1}, "pyerrors.fits.error_band": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}, "pyerrors.roots.find_root": {"tf": 1}}, "df": 7, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.correlators.Corr.fit": {"tf": 1}}, "df": 1}}}}}}, "l": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}}, "df": 6}}}, "i": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.fits.residual_plot": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 4, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.fit": {"tf": 1}}, "df": 1}}}}}}, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}}, "df": 4, "n": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.dump": {"tf": 1}, "pyerrors.obs.Obs.dump": {"tf": 1}}, "df": 2}}}}, "s": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "m": {"docs": {"pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}}, "df": 4}}}}}}}, "n": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}, "pyerrors.input.pandas.dump_df": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1}}, "df": 10}}}}, "l": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "w": {"docs": {"pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}}, "df": 1}}}}, "j": {"docs": {"pyerrors.correlators.Corr.item": {"tf": 1}, "pyerrors.dirac.epsilon_tensor": {"tf": 1}, "pyerrors.dirac.epsilon_tensor_rank4": {"tf": 1}}, "df": 3, "s": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.correlators.Corr.dump": {"tf": 1}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.obs.Obs.dump": {"tf": 1}}, "df": 3}}}, "a": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "k": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.obs.import_jackknife": {"tf": 1}}, "df": 1}}}}}, "t": {"0": {"docs": {"pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 1}}, "df": 2, "p": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "j": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 1}}}}}, "docs": {}, "df": 0, "s": {"docs": {"pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 1}}, "df": 2}, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1}}, "df": 1}}}}, "r": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.spaghetti_plot": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1.4142135623730951}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.pandas.to_sql": {"tf": 1}, "pyerrors.input.pandas.dump_df": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1}, "pyerrors.obs.Obs.details": {"tf": 1}, "pyerrors.obs.Obs.plot_history": {"tf": 1}}, "df": 17}}}, "p": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "j": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 1}}}}, "a": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.dirac.Grid_gamma": {"tf": 1}}, "df": 1}, "r": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.input.openQCD.qtop_projection": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 2}}}}, "b": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.pandas.to_sql": {"tf": 1}}, "df": 1}}}, "u": {"docs": {"pyerrors.misc.gen_correlated_data": {"tf": 1}}, "df": 1}}, "y": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 1}}}}, "e": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 1}}, "df": 2}}}}}}}}}, "n": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}, "pyerrors.obs.Obs.details": {"tf": 1}}, "df": 6, "t": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}}, "df": 2, "s": {"docs": {"pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}}, "df": 2}}}}}}, "x": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 1}}}}, "i": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.pandas.to_sql": {"tf": 1}}, "df": 1}}}}, "p": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.obs.Obs.plot_history": {"tf": 1}}, "df": 1}}}}}}, "o": {"docs": {"pyerrors.dirac.epsilon_tensor_rank4": {"tf": 1}, "pyerrors.fits.qqplot": {"tf": 1}}, "df": 2, "f": {"docs": {"pyerrors.obs.merge_obs": {"tf": 1}}, "df": 1, "f": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.thin": {"tf": 1}}, "df": 1}}}}}, "b": {"docs": {}, "df": 0, "j": {"docs": {"pyerrors.misc.dump_object": {"tf": 1}}, "df": 1, "e": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.fits.ks_test": {"tf": 1}}, "df": 1}}}}}, "s": {"docs": {"pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.linalg.eigh": {"tf": 1}, "pyerrors.linalg.eig": {"tf": 1}, "pyerrors.linalg.pinv": {"tf": 1}, "pyerrors.linalg.svd": {"tf": 1}, "pyerrors.obs.reweight": {"tf": 1}, "pyerrors.obs.correlate": {"tf": 1.4142135623730951}, "pyerrors.obs.covariance": {"tf": 1}, "pyerrors.obs.merge_obs": {"tf": 1}}, "df": 9, "l": {"docs": {"pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}}, "df": 4}}}, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}}, "df": 4}}}}}, "u": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}}, "df": 6}}}}}, "l": {"docs": {"pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}}, "df": 2}, "d": {"docs": {"pyerrors.input.json.dump_dict_to_json": {"tf": 1}}, "df": 1}, "p": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "q": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.openQCD.read_qtop": {"tf": 1}}, "df": 1}}}}, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.linalg.matmul": {"tf": 1}, "pyerrors.linalg.jack_matmul": {"tf": 1}, "pyerrors.linalg.einsum": {"tf": 1}}, "df": 3}}}}}}}}, "w": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.reweight": {"tf": 1}, "pyerrors.obs.Obs.reweight": {"tf": 1}, "pyerrors.obs.reweight": {"tf": 1}}, "df": 3}}}}}, "h": {"docs": {}, "df": 0, "o": {"docs": {"pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}}, "df": 2}}, "f": {"2": {"docs": {"pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 1}, "docs": {"pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 1}}, "g": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.correlators.Corr.m_eff": {"tf": 1}, "pyerrors.roots.find_root": {"tf": 1}}, "df": 2}}}}, "a": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "a": {"docs": {"pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.dirac.Grid_gamma": {"tf": 1}, "pyerrors.input.pandas.read_sql": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1}}, "df": 5, "s": {"docs": {"pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}}, "df": 1}}}}}, "z": {"docs": {"pyerrors.correlators.Corr.dump": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}, "pyerrors.input.pandas.to_sql": {"tf": 1}, "pyerrors.input.pandas.dump_df": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1}, "pyerrors.obs.Obs.dump": {"tf": 1}}, "df": 13}, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.obs.cov_Obs": {"tf": 1}}, "df": 2}}}}, "m": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.plateau": {"tf": 1}}, "df": 1}}}}, "a": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.covobs.Covobs.__init__": {"tf": 1}}, "df": 1, "s": {"docs": {"pyerrors.misc.gen_correlated_data": {"tf": 1}, "pyerrors.obs.cov_Obs": {"tf": 1}}, "df": 2}}}, "s": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.hadrons.read_meson_hd5": {"tf": 1.4142135623730951}}, "df": 1}}}}, "o": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.obs.derived_observable": {"tf": 1}}, "df": 1}}}}, "a": {"docs": {"pyerrors.obs.correlate": {"tf": 1}}, "df": 1, "u": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "o": {"docs": {"pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.input.pandas.read_sql": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1}}, "df": 4}}}, "v": {"docs": {"pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}}, "df": 1}, "t": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.obs.Obs.is_zero": {"tf": 1}}, "df": 1}}}, "r": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.obs.derived_observable": {"tf": 1}}, "df": 1}}}}}, "x": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.fits.fit_lin": {"tf": 1}, "pyerrors.fits.qqplot": {"tf": 1}, "pyerrors.fits.residual_plot": {"tf": 1}, "pyerrors.fits.error_band": {"tf": 1}, "pyerrors.linalg.inv": {"tf": 1}, "pyerrors.linalg.cholesky": {"tf": 1}, "pyerrors.linalg.det": {"tf": 1}}, "df": 10, "m": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 1}}}}, "c": {"docs": {"pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 3, "o": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "p": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1}}, "df": 1}}, "v": {"docs": {"pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.misc.gen_correlated_data": {"tf": 1}, "pyerrors.obs.cov_Obs": {"tf": 1}}, "df": 3}, "n": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.obs.Obs.details": {"tf": 1}}, "df": 2}}}}}, "r": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 2, "s": {"docs": {"pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 1}, "e": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}}}}}}}}, "n": {"docs": {}, "df": 0, "f": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}}, "df": 2}}}, "f": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 1}}, "h": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.utils.check_idl": {"tf": 1}}, "df": 1}}}, "y": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.fits.fit_lin": {"tf": 1}, "pyerrors.fits.qqplot": {"tf": 1}, "pyerrors.fits.residual_plot": {"tf": 1}}, "df": 6, "l": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1}}, "df": 1}}}}}}, "h": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1}}, "df": 1}}}}, "b": {"docs": {"pyerrors.obs.correlate": {"tf": 1}}, "df": 1, "a": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "x": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 1}}}}}}}}}, "e": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "a": {"docs": {"pyerrors.fits.error_band": {"tf": 1}}, "df": 1}}}, "d": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {"pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}}, "df": 4}}}, "i": {"docs": {"pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 1}}, "z": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}}, "df": 1}}}}}}}, "q": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "p": {"docs": {"pyerrors.input.openQCD.qtop_projection": {"tf": 1}}, "df": 1}}}, "c": {"docs": {"pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}}, "df": 1}, "u": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "k": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 1}}}}}}}}, "bases": {"root": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.fits.Fit_result": {"tf": 1}}, "df": 1}}}}}}}}}}}, "a": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "c": {"docs": {"pyerrors.fits.Fit_result": {"tf": 1}}, "df": 1}}}, "s": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "q": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.fits.Fit_result": {"tf": 1}}, "df": 1}}}}}}}}, "n": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}}}, "d": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}}}}}}}}, "doc": {"root": {"0": {"0": {"0": {"0": {"0": {"0": {"0": {"0": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1, "+": {"0": {"0": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "docs": {}, "df": 0}}}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "5": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {"pyerrors": {"tf": 1}}, "df": 1}, "6": {"9": {"7": {"9": {"5": {"8": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "+": {"0": {"0": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}, "docs": {}, "df": 0}, "docs": {}, "df": 0}}}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}, "1": {"2": {"8": {"9": {"docs": {"pyerrors": {"tf": 2}}, "df": 1}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "7": {"1": {"8": {"0": {"6": {"4": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "+": {"0": {"0": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "docs": {}, "df": 0}}}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {"pyerrors": {"tf": 3}}, "df": 1}, "2": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}, "3": {"4": {"4": {"5": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "4": {"5": {"8": {"5": {"6": {"5": {"0": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "7": {"5": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}, "5": {"4": {"8": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "6": {"4": {"2": {"3": {"docs": {"pyerrors.input.openQCD.read_gf_coupling": {"tf": 1.4142135623730951}}, "df": 1}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "7": {"5": {"6": {"0": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "9": {"9": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "docs": {"pyerrors": {"tf": 6.164414002968976}, "pyerrors.correlators.Corr.projected": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.symmetric": {"tf": 1}, "pyerrors.correlators.Corr.anti_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.prune": {"tf": 2.6457513110645907}, "pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 2.23606797749979}, "pyerrors.fits.total_least_squares": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.7320508075688772}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.qtop_projection": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.4142135623730951}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs": {"tf": 2}, "pyerrors.obs.Obs.gamma_method": {"tf": 2}, "pyerrors.obs.Obs.gm": {"tf": 2}, "pyerrors.obs.covariance": {"tf": 2}}, "df": 24, "e": {"docs": {}, "df": 0, "+": {"0": {"0": {"0": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}}, "df": 1}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}}, "c": {"docs": {"pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 1}, "d": {"docs": {"pyerrors.obs.cov_Obs": {"tf": 1}}, "df": 1}}, "1": {"0": {"0": {"0": {"docs": {"pyerrors": {"tf": 3.605551275463989}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.4142135623730951}, "pyerrors.misc.pseudo_Obs": {"tf": 1}}, "df": 3}, "3": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}, "5": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "4": {"7": {"2": {"docs": {"pyerrors.correlators.Corr.GEVP": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {"pyerrors": {"tf": 1}}, "df": 1}, "1": {"9": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "2": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}, "3": {"4": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}, "docs": {}, "df": 0}, "4": {"3": {"7": {"1": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "5": {"0": {"0": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}, "docs": {}, "df": 0}, "3": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "6": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "6": {"0": {"7": {"docs": {"pyerrors.input.openQCD.read_gf_coupling": {"tf": 1.4142135623730951}}, "df": 1}, "docs": {}, "df": 0}, "docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}, "7": {"6": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {"pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 1}, "9": {"0": {"6": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "8": {"8": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "9": {"0": {"docs": {"pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "docs": {"pyerrors": {"tf": 1.7320508075688772}}, "df": 1}, "docs": {"pyerrors": {"tf": 6.164414002968976}, "pyerrors.correlators.Corr.Hankel": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.T_symmetry": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.m_eff": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.prune": {"tf": 1.4142135623730951}, "pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.total_least_squares": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.read_mesons": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.read_dSdm": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.4142135623730951}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.4142135623730951}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1.4142135623730951}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 20, "}": {"docs": {}, "df": 0, "^": {"docs": {}, "df": 0, "{": {"docs": {}, "df": 0, "w": {"docs": {}, "df": 0, "}": {"docs": {}, "df": 0, "\\": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "o": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}}, "n": {"docs": {}, "df": 0, "\\": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "a": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}}}, "s": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "m": {"docs": {"pyerrors.obs.covariance": {"tf": 1.4142135623730951}}, "df": 1}}}}}}}, "\\": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}, "+": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "d": {"docs": {"pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.obs.cov_Obs": {"tf": 1}}, "df": 2}, "*": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "z": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}}}}}}}}, "/": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.obs.Obs.export_jackknife": {"tf": 1}}, "df": 1}}}, "2": {"0": {"0": {"4": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "7": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "1": {"1": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "9": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "4": {"1": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}, "docs": {}, "df": 0}, "docs": {"pyerrors": {"tf": 2}}, "df": 1}, "1": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}, "3": {"8": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {"pyerrors": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}}, "df": 2}, "5": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "7": {"2": {"1": {"8": {"6": {"6": {"7": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "+": {"0": {"0": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "docs": {}, "df": 0}}}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "8": {"0": {"9": {"7": {"7": {"6": {"2": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {"pyerrors": {"tf": 2}}, "df": 1}, "9": {"9": {"0": {"9": {"7": {"0": {"3": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "+": {"0": {"0": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "docs": {}, "df": 0}}}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}, "docs": {"pyerrors": {"tf": 5}, "pyerrors.correlators.Corr.GEVP": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.m_eff": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.total_least_squares": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 3}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}, "pyerrors.roots.find_root": {"tf": 1}}, "df": 15, "x": {"2": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "d": {"docs": {"pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.linalg.matmul": {"tf": 1}, "pyerrors.linalg.jack_matmul": {"tf": 1}, "pyerrors.linalg.einsum": {"tf": 1}, "pyerrors.obs.cov_Obs": {"tf": 1}}, "df": 5}, "*": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 1}}}}}, "3": {"0": {"6": {"7": {"5": {"2": {"0": {"1": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "1": {"4": {"9": {"8": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "2": {"2": {"7": {"docs": {"pyerrors": {"tf": 1.7320508075688772}}, "df": 1}, "docs": {}, "df": 0}, "3": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "3": {"docs": {"pyerrors": {"tf": 2}}, "df": 1}, "4": {"9": {"7": {"6": {"8": {"0": {"0": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "+": {"0": {"2": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "docs": {}, "df": 0}}}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "5": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "6": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "7": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "8": {"3": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {"pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 1}, "9": {"docs": {"pyerrors": {"tf": 7.745966692414834}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}}, "df": 2}, "docs": {"pyerrors": {"tf": 3.605551275463989}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1}, "pyerrors.dirac.epsilon_tensor": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 2}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1.4142135623730951}}, "df": 8, "a": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "4": {"0": {"3": {"2": {"0": {"9": {"8": {"3": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {"pyerrors.fits.least_squares": {"tf": 1}}, "df": 1}, "9": {"5": {"9": {"1": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.fit": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.dirac.epsilon_tensor_rank4": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}}, "df": 6, "x": {"4": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}}, "5": {"0": {"0": {"docs": {"pyerrors": {"tf": 2.8284271247461903}, "pyerrors.fits.least_squares": {"tf": 1}}, "df": 2}, "1": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}, "docs": {}, "df": 0}, "1": {"5": {"6": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}, "docs": {}, "df": 0}, "9": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}, "docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}, "2": {"0": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "2": {"8": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "3": {"8": {"0": {"docs": {"pyerrors.correlators.Corr.m_eff": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "4": {"8": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "6": {"7": {"3": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "4": {"6": {"5": {"9": {"8": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "7": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "8": {"3": {"4": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {"pyerrors": {"tf": 3.605551275463989}, "pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix.g5H": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 5, "e": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}, "6": {"4": {"docs": {"pyerrors": {"tf": 1.7320508075688772}}, "df": 1}, "5": {"0": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "9": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "6": {"8": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.correlators.Corr.fit": {"tf": 1.4142135623730951}}, "df": 2}, "7": {"0": {"0": {"0": {"0": {"0": {"0": {"0": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "+": {"0": {"0": {"docs": {"pyerrors": {"tf": 1.7320508075688772}}, "df": 1}, "docs": {}, "df": 0}, "docs": {}, "df": 0}}}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {"pyerrors": {"tf": 1}}, "df": 1}, "1": {"4": {"2": {"2": {"9": {"0": {"0": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "+": {"0": {"0": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "docs": {}, "df": 0}}}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "2": {"0": {"4": {"6": {"6": {"5": {"8": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {"pyerrors": {"tf": 1}}, "df": 1}, "4": {"5": {"7": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "7": {"3": {"1": {"0": {"1": {"0": {"2": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "9": {"0": {"7": {"7": {"5": {"2": {"4": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "5": {"7": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {"pyerrors": {"tf": 2.23606797749979}}, "df": 1}, "8": {"0": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}, "1": {"4": {"docs": {"pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "2": {"4": {"docs": {"pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "4": {"5": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 6}, "9": {"1": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "3": {"3": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {"pyerrors": {"tf": 1}}, "df": 1}, "4": {"7": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "5": {"9": {"3": {"0": {"3": {"5": {"7": {"8": {"5": {"1": {"6": {"0": {"9": {"3": {"6": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "4": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "7": {"6": {"8": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "8": {"3": {"1": {"9": {"8": {"8": {"1": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "9": {"1": {"0": {"0": {"7": {"1": {"2": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "5": {"8": {"3": {"6": {"5": {"4": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "9": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {"pyerrors": {"tf": 2.449489742783178}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}}, "df": 3}, "docs": {"pyerrors": {"tf": 62.928530890209096}, "pyerrors.correlators": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr": {"tf": 3}, "pyerrors.correlators.Corr.__init__": {"tf": 5.0990195135927845}, "pyerrors.correlators.Corr.gamma_method": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.gm": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.projected": {"tf": 2.23606797749979}, "pyerrors.correlators.Corr.item": {"tf": 4.58257569495584}, "pyerrors.correlators.Corr.plottable": {"tf": 2.449489742783178}, "pyerrors.correlators.Corr.symmetric": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.anti_symmetric": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.is_matrix_symmetric": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.matrix_symmetric": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.GEVP": {"tf": 10.535653752852738}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 4.358898943540674}, "pyerrors.correlators.Corr.Hankel": {"tf": 4.58257569495584}, "pyerrors.correlators.Corr.roll": {"tf": 3.605551275463989}, "pyerrors.correlators.Corr.reverse": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.thin": {"tf": 4.242640687119285}, "pyerrors.correlators.Corr.correlate": {"tf": 3.7416573867739413}, "pyerrors.correlators.Corr.reweight": {"tf": 4.58257569495584}, "pyerrors.correlators.Corr.T_symmetry": {"tf": 4.242640687119285}, "pyerrors.correlators.Corr.deriv": {"tf": 3.7416573867739413}, "pyerrors.correlators.Corr.second_deriv": {"tf": 3.7416573867739413}, "pyerrors.correlators.Corr.m_eff": {"tf": 5.830951894845301}, "pyerrors.correlators.Corr.fit": {"tf": 5.291502622129181}, "pyerrors.correlators.Corr.plateau": {"tf": 5}, "pyerrors.correlators.Corr.set_prange": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.show": {"tf": 8.660254037844387}, "pyerrors.correlators.Corr.spaghetti_plot": {"tf": 3.872983346207417}, "pyerrors.correlators.Corr.dump": {"tf": 5.0990195135927845}, "pyerrors.correlators.Corr.print": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.sqrt": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.log": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.exp": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.sin": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.cos": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.tan": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.sinh": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.cosh": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.tanh": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.arcsin": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.arccos": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.arctan": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.arcsinh": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.arccosh": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.arctanh": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.prune": {"tf": 6.855654600401044}, "pyerrors.covobs": {"tf": 1.7320508075688772}, "pyerrors.covobs.Covobs": {"tf": 1.7320508075688772}, "pyerrors.covobs.Covobs.__init__": {"tf": 5.916079783099616}, "pyerrors.covobs.Covobs.errsq": {"tf": 1.4142135623730951}, "pyerrors.dirac": {"tf": 1.7320508075688772}, "pyerrors.dirac.epsilon_tensor": {"tf": 2.449489742783178}, "pyerrors.dirac.epsilon_tensor_rank4": {"tf": 2.449489742783178}, "pyerrors.dirac.Grid_gamma": {"tf": 1.7320508075688772}, "pyerrors.fits": {"tf": 1.7320508075688772}, "pyerrors.fits.Fit_result": {"tf": 5.656854249492381}, "pyerrors.fits.Fit_result.__init__": {"tf": 1.7320508075688772}, "pyerrors.fits.Fit_result.gamma_method": {"tf": 1.4142135623730951}, "pyerrors.fits.Fit_result.gm": {"tf": 1.4142135623730951}, "pyerrors.fits.least_squares": {"tf": 16.06237840420901}, "pyerrors.fits.total_least_squares": {"tf": 15}, "pyerrors.fits.fit_lin": {"tf": 4.795831523312719}, "pyerrors.fits.qqplot": {"tf": 1.7320508075688772}, "pyerrors.fits.residual_plot": {"tf": 1.4142135623730951}, "pyerrors.fits.error_band": {"tf": 1.7320508075688772}, "pyerrors.fits.ks_test": {"tf": 3.872983346207417}, "pyerrors.input": {"tf": 4.69041575982343}, "pyerrors.input.bdio": {"tf": 1.7320508075688772}, "pyerrors.input.bdio.read_ADerrors": {"tf": 5.196152422706632}, "pyerrors.input.bdio.write_ADerrors": {"tf": 5.196152422706632}, "pyerrors.input.bdio.read_mesons": {"tf": 7.416198487095663}, "pyerrors.input.bdio.read_dSdm": {"tf": 7.416198487095663}, "pyerrors.input.dobs": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.create_pobs_string": {"tf": 7}, "pyerrors.input.dobs.write_pobs": {"tf": 7.810249675906654}, "pyerrors.input.dobs.read_pobs": {"tf": 5.744562646538029}, "pyerrors.input.dobs.import_dobs_string": {"tf": 6.244997998398398}, "pyerrors.input.dobs.read_dobs": {"tf": 6.782329983125268}, "pyerrors.input.dobs.create_dobs_string": {"tf": 7.3484692283495345}, "pyerrors.input.dobs.write_dobs": {"tf": 8.18535277187245}, "pyerrors.input.hadrons": {"tf": 1.7320508075688772}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 6.48074069840786}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 5.656854249492381}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 20.904544960366874}, "pyerrors.input.hadrons.Npr_matrix.__init__": {"tf": 1.7320508075688772}, "pyerrors.input.hadrons.Npr_matrix.g5H": {"tf": 2.23606797749979}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 5.385164807134504}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 5.385164807134504}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 5.916079783099616}, "pyerrors.input.json": {"tf": 1.7320508075688772}, "pyerrors.input.json.create_json_string": {"tf": 5.0990195135927845}, "pyerrors.input.json.dump_to_json": {"tf": 6.164414002968976}, "pyerrors.input.json.import_json_string": {"tf": 5.477225575051661}, "pyerrors.input.json.load_json": {"tf": 6}, "pyerrors.input.json.dump_dict_to_json": {"tf": 6.6332495807108}, "pyerrors.input.json.load_json_dict": {"tf": 6.4031242374328485}, "pyerrors.input.misc": {"tf": 1.7320508075688772}, "pyerrors.input.misc.read_pbp": {"tf": 4.242640687119285}, "pyerrors.input.openQCD": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_rwms": {"tf": 7.874007874011811}, "pyerrors.input.openQCD.extract_t0": {"tf": 9.899494936611665}, "pyerrors.input.openQCD.read_qtop": {"tf": 9.695359714832659}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 8.888194417315589}, "pyerrors.input.openQCD.qtop_projection": {"tf": 4.58257569495584}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 9.219544457292887}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 10}, "pyerrors.input.pandas": {"tf": 1.7320508075688772}, "pyerrors.input.pandas.to_sql": {"tf": 6.244997998398398}, "pyerrors.input.pandas.read_sql": {"tf": 5.196152422706632}, "pyerrors.input.pandas.dump_df": {"tf": 5.477225575051661}, "pyerrors.input.pandas.load_df": {"tf": 5.196152422706632}, "pyerrors.input.sfcf": {"tf": 1.7320508075688772}, "pyerrors.input.sfcf.read_sfcf": {"tf": 10.488088481701515}, "pyerrors.input.utils": {"tf": 1.4142135623730951}, "pyerrors.input.utils.check_idl": {"tf": 4.242640687119285}, "pyerrors.linalg": {"tf": 1.7320508075688772}, "pyerrors.linalg.matmul": {"tf": 4.58257569495584}, "pyerrors.linalg.jack_matmul": {"tf": 4.47213595499958}, "pyerrors.linalg.einsum": {"tf": 4.47213595499958}, "pyerrors.linalg.inv": {"tf": 1.7320508075688772}, "pyerrors.linalg.cholesky": {"tf": 1.7320508075688772}, "pyerrors.linalg.det": {"tf": 1.7320508075688772}, "pyerrors.linalg.eigh": {"tf": 1.7320508075688772}, "pyerrors.linalg.eig": {"tf": 1.7320508075688772}, "pyerrors.linalg.pinv": {"tf": 1.7320508075688772}, "pyerrors.linalg.svd": {"tf": 1.7320508075688772}, "pyerrors.misc": {"tf": 1.7320508075688772}, "pyerrors.misc.dump_object": {"tf": 5}, "pyerrors.misc.load_object": {"tf": 3.7416573867739413}, "pyerrors.misc.pseudo_Obs": {"tf": 5.656854249492381}, "pyerrors.misc.gen_correlated_data": {"tf": 6.244997998398398}, "pyerrors.mpm": {"tf": 1.7320508075688772}, "pyerrors.mpm.matrix_pencil_method": {"tf": 5.385164807134504}, "pyerrors.obs": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs": {"tf": 6.928203230275509}, "pyerrors.obs.Obs.__init__": {"tf": 4.898979485566356}, "pyerrors.obs.Obs.gamma_method": {"tf": 5.744562646538029}, "pyerrors.obs.Obs.gm": {"tf": 5.744562646538029}, "pyerrors.obs.Obs.details": {"tf": 3.872983346207417}, "pyerrors.obs.Obs.reweight": {"tf": 4.58257569495584}, "pyerrors.obs.Obs.is_zero_within_error": {"tf": 4.47213595499958}, "pyerrors.obs.Obs.is_zero": {"tf": 3.872983346207417}, "pyerrors.obs.Obs.plot_tauint": {"tf": 3.872983346207417}, "pyerrors.obs.Obs.plot_rho": {"tf": 3.872983346207417}, "pyerrors.obs.Obs.plot_rep_dist": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.plot_history": {"tf": 3.7416573867739413}, "pyerrors.obs.Obs.plot_piechart": {"tf": 3.872983346207417}, "pyerrors.obs.Obs.dump": {"tf": 5.744562646538029}, "pyerrors.obs.Obs.export_jackknife": {"tf": 3.7416573867739413}, "pyerrors.obs.Obs.sqrt": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.log": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.exp": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.sin": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.cos": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.tan": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.arcsin": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.arccos": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.arctan": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.sinh": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.cosh": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.tanh": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.arcsinh": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.arccosh": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.arctanh": {"tf": 1.7320508075688772}, "pyerrors.obs.CObs": {"tf": 1.7320508075688772}, "pyerrors.obs.CObs.__init__": {"tf": 1.7320508075688772}, "pyerrors.obs.CObs.gamma_method": {"tf": 1.7320508075688772}, "pyerrors.obs.CObs.is_zero": {"tf": 1.7320508075688772}, "pyerrors.obs.CObs.conjugate": {"tf": 1.7320508075688772}, "pyerrors.obs.derived_observable": {"tf": 6.4031242374328485}, "pyerrors.obs.reweight": {"tf": 5.196152422706632}, "pyerrors.obs.correlate": {"tf": 4.898979485566356}, "pyerrors.obs.covariance": {"tf": 6.6332495807108}, "pyerrors.obs.import_jackknife": {"tf": 4.47213595499958}, "pyerrors.obs.merge_obs": {"tf": 4.123105625617661}, "pyerrors.obs.cov_Obs": {"tf": 5.385164807134504}, "pyerrors.roots": {"tf": 1.7320508075688772}, "pyerrors.roots.find_root": {"tf": 10.44030650891055}, "pyerrors.version": {"tf": 1.7320508075688772}}, "df": 181, "w": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 2, "h": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}}, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 2.449489742783178}, "pyerrors.correlators.Corr.__init__": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 2}, "pyerrors.fits.fit_lin": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1}, "pyerrors.roots.find_root": {"tf": 1}}, "df": 12}}, "n": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.obs.Obs.is_zero_within_error": {"tf": 1}}, "df": 10}, "t": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.is_matrix_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.Hankel": {"tf": 1}, "pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.correlators.Corr.spaghetti_plot": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}, "pyerrors.obs.Obs.is_zero_within_error": {"tf": 1}, "pyerrors.obs.Obs.is_zero": {"tf": 1}, "pyerrors.obs.CObs.is_zero": {"tf": 1}}, "df": 11}}}}}, "i": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "h": {"docs": {"pyerrors": {"tf": 3.1622776601683795}, "pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.correlate": {"tf": 1}, "pyerrors.correlators.Corr.deriv": {"tf": 1}, "pyerrors.correlators.Corr.second_deriv": {"tf": 1}, "pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1.7320508075688772}, "pyerrors.fits.least_squares": {"tf": 2}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.fits.fit_lin": {"tf": 1}, "pyerrors.fits.qqplot": {"tf": 1}, "pyerrors.fits.residual_plot": {"tf": 1}, "pyerrors.input": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.misc.read_pbp": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_rwms": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.extract_t0": {"tf": 2}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1.7320508075688772}, "pyerrors.input.pandas.load_df": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.linalg.matmul": {"tf": 1}, "pyerrors.linalg.jack_matmul": {"tf": 1}, "pyerrors.linalg.einsum": {"tf": 1}, "pyerrors.misc.pseudo_Obs": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.__init__": {"tf": 1}, "pyerrors.obs.Obs.plot_piechart": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}, "pyerrors.obs.correlate": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}, "pyerrors.obs.merge_obs": {"tf": 1}}, "df": 35}}, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 2}}}, "o": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}}, "df": 3}}, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "h": {"docs": {"pyerrors": {"tf": 5.830951894845301}, "pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.correlators.Corr.__init__": {"tf": 1}, "pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.correlators.Corr.correlate": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.deriv": {"tf": 1}, "pyerrors.correlators.Corr.second_deriv": {"tf": 1}, "pyerrors.correlators.Corr.plateau": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.show": {"tf": 1.4142135623730951}, "pyerrors.fits.least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.fits.error_band": {"tf": 1}, "pyerrors.input": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}, "pyerrors.input.pandas.read_sql": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.7320508075688772}, "pyerrors.misc.pseudo_Obs": {"tf": 1}, "pyerrors.misc.gen_correlated_data": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}, "pyerrors.obs.Obs.reweight": {"tf": 1}, "pyerrors.obs.Obs.plot_rep_dist": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 2}}, "df": 34, "i": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.obs.Obs.is_zero_within_error": {"tf": 1}, "pyerrors.obs.Obs.is_zero": {"tf": 1}, "pyerrors.obs.CObs.is_zero": {"tf": 1}}, "df": 9}}, "o": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}, "n": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "w": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors": {"tf": 1.7320508075688772}}, "df": 1}}}, "s": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "z": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}, "/": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}}}}}}}}, "l": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}}, "df": 12}, "s": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 3}}}}}, "o": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "f": {"docs": {}, "df": 0, "f": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}, "r": {"docs": {}, "df": 0, "k": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}}, "df": 4, "s": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.obs.Obs.is_zero_within_error": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1}, "pyerrors.obs.correlate": {"tf": 1}}, "df": 6}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}, "f": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "w": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input": {"tf": 1}}, "df": 1}}}}}}}}, "e": {"docs": {"pyerrors": {"tf": 2}, "pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1.4142135623730951}}, "df": 3, "l": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 2}}, "i": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.reweight": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.reweight": {"tf": 1.4142135623730951}, "pyerrors.obs.reweight": {"tf": 1.4142135623730951}}, "df": 3}}}}}, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 2}}}}}, "y": {"docs": {"pyerrors": {"tf": 1}}, "df": 1, "s": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}, "s": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.is_zero_within_error": {"tf": 1}}, "df": 6}, "n": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.correlators.Corr": {"tf": 1}}, "df": 1}}}, "v": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 1, "f": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 1, "s": {"docs": {"pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 1}}}}}}}}}}}}, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.linalg.einsum": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}, "pyerrors.roots.find_root": {"tf": 1}}, "df": 3, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.write_pobs": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.read_dobs": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.write_dobs": {"tf": 1.4142135623730951}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}, "pyerrors.input.pandas.to_sql": {"tf": 1}}, "df": 12}}}, "e": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.pandas.to_sql": {"tf": 1}}, "df": 3, "a": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}}}}}}, "t": {"docs": {"pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.obs.cov_Obs": {"tf": 1}}, "df": 2}, "o": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}}, "df": 2}}}}, "l": {"docs": {"pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}}, "df": 4}, "f": {"2": {"docs": {"pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 1}, "docs": {"pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 1}}, "i": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.item": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.GEVP": {"tf": 2.449489742783178}, "pyerrors.correlators.Corr.reweight": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.reweight": {"tf": 1}, "pyerrors.obs.reweight": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1.4142135623730951}}, "df": 9, "s": {"docs": {"pyerrors": {"tf": 8}, "pyerrors.correlators.Corr": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.is_matrix_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 2.6457513110645907}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.Hankel": {"tf": 1}, "pyerrors.correlators.Corr.correlate": {"tf": 1}, "pyerrors.correlators.Corr.deriv": {"tf": 1}, "pyerrors.correlators.Corr.second_deriv": {"tf": 1}, "pyerrors.correlators.Corr.fit": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.correlators.Corr.spaghetti_plot": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 3.4641016151377544}, "pyerrors.covobs.Covobs.__init__": {"tf": 1.4142135623730951}, "pyerrors.fits.least_squares": {"tf": 3.4641016151377544}, "pyerrors.fits.total_least_squares": {"tf": 1.7320508075688772}, "pyerrors.fits.fit_lin": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.write_pobs": {"tf": 2.23606797749979}, "pyerrors.input.dobs.read_pobs": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.import_dobs_string": {"tf": 2.23606797749979}, "pyerrors.input.dobs.read_dobs": {"tf": 2.449489742783178}, "pyerrors.input.dobs.create_dobs_string": {"tf": 2}, "pyerrors.input.dobs.write_dobs": {"tf": 2.449489742783178}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 3.4641016151377544}, "pyerrors.input.hadrons.Npr_matrix.g5H": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1.7320508075688772}, "pyerrors.input.json.import_json_string": {"tf": 1.7320508075688772}, "pyerrors.input.json.load_json": {"tf": 2}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1.7320508075688772}, "pyerrors.input.json.load_json_dict": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_rwms": {"tf": 2.23606797749979}, "pyerrors.input.openQCD.extract_t0": {"tf": 3.605551275463989}, "pyerrors.input.openQCD.read_qtop": {"tf": 2.449489742783178}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 2.449489742783178}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 2.449489742783178}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1.4142135623730951}, "pyerrors.input.pandas.dump_df": {"tf": 1.4142135623730951}, "pyerrors.input.pandas.load_df": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 2}, "pyerrors.input.utils.check_idl": {"tf": 1}, "pyerrors.linalg.matmul": {"tf": 1}, "pyerrors.linalg.jack_matmul": {"tf": 1}, "pyerrors.misc.pseudo_Obs": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.gm": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.is_zero_within_error": {"tf": 1}, "pyerrors.obs.Obs.is_zero": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1}, "pyerrors.obs.correlate": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 3}, "pyerrors.obs.merge_obs": {"tf": 1}, "pyerrors.roots.find_root": {"tf": 1}}, "df": 56}, "t": {"docs": {"pyerrors": {"tf": 3.1622776601683795}, "pyerrors.correlators.Corr": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.projected": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.GEVP": {"tf": 1.7320508075688772}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.7320508075688772}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}, "pyerrors.obs.merge_obs": {"tf": 1}}, "df": 20, "s": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.T_symmetry": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 2}}, "df": 8, "e": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "f": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 2}}}}, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1.7320508075688772}}, "df": 1}}}, "e": {"docs": {"pyerrors.correlators.Corr": {"tf": 1}}, "df": 1}, "o": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}}, "df": 1}}}}}, "m": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1, "i": {"docs": {}, "df": 0, "z": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1.7320508075688772}}, "df": 1}}}}}}}, "n": {"docs": {"pyerrors": {"tf": 8.18535277187245}, "pyerrors.correlators.Corr.plottable": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 1}, "pyerrors.correlators.Corr.reweight": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.show": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.prune": {"tf": 1.4142135623730951}, "pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.dirac.Grid_gamma": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 2}, "pyerrors.fits.fit_lin": {"tf": 1}, "pyerrors.fits.ks_test": {"tf": 1}, "pyerrors.input": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.read_dobs": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1.7320508075688772}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 3.3166247903554}, "pyerrors.input.hadrons.Npr_matrix.g5H": {"tf": 1}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1.4142135623730951}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.extract_t0": {"tf": 2.23606797749979}, "pyerrors.input.openQCD.read_qtop": {"tf": 2.23606797749979}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 2.23606797749979}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 2.23606797749979}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 2.23606797749979}, "pyerrors.input.pandas.to_sql": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.4142135623730951}, "pyerrors.input.utils.check_idl": {"tf": 1}, "pyerrors.misc.dump_object": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}, "pyerrors.obs.Obs.reweight": {"tf": 1.7320508075688772}, "pyerrors.obs.reweight": {"tf": 1.7320508075688772}, "pyerrors.obs.correlate": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}, "pyerrors.obs.merge_obs": {"tf": 1}}, "df": 45, "t": {"1": {"6": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.item": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.GEVP": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 1}, "pyerrors.correlators.Corr.Hankel": {"tf": 1}, "pyerrors.correlators.Corr.roll": {"tf": 1}, "pyerrors.correlators.Corr.thin": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.T_symmetry": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1.7320508075688772}, "pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1.7320508075688772}, "pyerrors.input.bdio.read_dSdm": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.read_pobs": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.read_dobs": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 2.6457513110645907}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 2.23606797749979}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.qtop_projection": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 2}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.7320508075688772}, "pyerrors.misc.pseudo_Obs": {"tf": 1}, "pyerrors.misc.gen_correlated_data": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.is_zero_within_error": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 31, "r": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 2}}, "df": 1}, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}}}}}, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1}}, "df": 1, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 1}}, "df": 2}}}}}, "p": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}}}}}}}, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1}}, "df": 1, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}}, "df": 1}}}}}}, "n": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}}, "df": 5}}}, "f": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}}, "df": 1}}}}, "a": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}}}}}, "c": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.obs.derived_observable": {"tf": 1}}, "df": 1}}}}}}}, "g": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.plot_tauint": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1.4142135623730951}}, "df": 3}}}}}, "e": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.4142135623730951}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 6, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}, "o": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.dump": {"tf": 1}, "pyerrors.input": {"tf": 1}, "pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.misc.dump_object": {"tf": 1}, "pyerrors.obs.merge_obs": {"tf": 1}}, "df": 10}, "s": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 2}}, "df": 1}}, "v": {"docs": {"pyerrors": {"tf": 1}}, "df": 1, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.linalg.inv": {"tf": 1}}, "df": 3}}, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}}, "df": 2}}}}}}, "i": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1, "i": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.fits.least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.total_least_squares": {"tf": 1.4142135623730951}, "pyerrors.roots.find_root": {"tf": 1}}, "df": 3, "i": {"docs": {}, "df": 0, "z": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.__init__": {"tf": 1}, "pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.obs.Obs.__init__": {"tf": 1}}, "df": 4, "d": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1}}, "df": 3}}}}}}}}}, "d": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 2}}, "df": 1, "l": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}}}}, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}}, "df": 3, "i": {"docs": {}, "df": 0, "f": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.__init__": {"tf": 1}}, "df": 1}}}}}, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}}, "df": 3}}}}}}}, "x": {"docs": {"pyerrors.correlators.Corr.item": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.plottable": {"tf": 1}}, "df": 2, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.correlators.Corr.fit": {"tf": 1}}, "df": 1}}}}}, "i": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.fits.Fit_result": {"tf": 1}, "pyerrors.obs.Obs.__init__": {"tf": 1}}, "df": 4, "l": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.correlators.Corr.GEVP": {"tf": 1}}, "df": 1}}}}}}}}, "c": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}}, "df": 2}}}}}, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.fits.Fit_result": {"tf": 1}}, "df": 1}}}}}, "p": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 2.449489742783178}, "pyerrors.correlators.Corr.__init__": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.total_least_squares": {"tf": 1.4142135623730951}, "pyerrors.input": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.input.utils": {"tf": 1}}, "df": 14}}}, "f": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 3.1622776601683795}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}, "pyerrors.input.json.import_json_string": {"tf": 1.4142135623730951}, "pyerrors.input.json.load_json": {"tf": 1.4142135623730951}, "pyerrors.input.json.load_json_dict": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}}, "df": 9}}}}}}}}}, "s": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 7}}}, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}}, "df": 1}}}, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}}}}}}, "c": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.obs.Obs": {"tf": 1.4142135623730951}}, "df": 1, "s": {"docs": {"pyerrors.obs.Obs": {"tf": 1.4142135623730951}}, "df": 1}}}}}}, "p": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}, "i": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 2}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}}, "df": 8}}}, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}}, "df": 3, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.dobs.read_pobs": {"tf": 2}, "pyerrors.input.dobs.import_dobs_string": {"tf": 2}, "pyerrors.input.dobs.read_dobs": {"tf": 2}}, "df": 3}}}, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.dobs.import_dobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.read_dobs": {"tf": 1.4142135623730951}}, "df": 2}}}}}}, "c": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.dump": {"tf": 1}, "pyerrors.fits.ks_test": {"tf": 1}, "pyerrors.obs.Obs.dump": {"tf": 1}}, "df": 4, "s": {"docs": {"pyerrors.input": {"tf": 1}}, "df": 1}, "d": {"docs": {"pyerrors.input.dobs.import_dobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.read_dobs": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 3}}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.pandas.to_sql": {"tf": 1}, "pyerrors.input.pandas.read_sql": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 3}}}}, "s": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.fit": {"tf": 1}}, "df": 1}}}}}}, "o": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr": {"tf": 1}}, "df": 1}}}}}}}}}, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}}, "df": 2}}}}}}}, "n": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.input.sfcf.read_sfcf": {"tf": 1.4142135623730951}}, "df": 1}}}}, "f": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 2}, "pyerrors.correlators.Corr.reweight": {"tf": 1}, "pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.prune": {"tf": 1}, "pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 2.449489742783178}, "pyerrors.fits.total_least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.qqplot": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.read_pobs": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.import_dobs_string": {"tf": 2}, "pyerrors.input.dobs.read_dobs": {"tf": 2.449489742783178}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1.7320508075688772}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.7320508075688772}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1.4142135623730951}, "pyerrors.input.json.import_json_string": {"tf": 1.7320508075688772}, "pyerrors.input.json.load_json": {"tf": 2.23606797749979}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1.4142135623730951}, "pyerrors.input.json.load_json_dict": {"tf": 2}, "pyerrors.input.openQCD.read_rwms": {"tf": 2}, "pyerrors.input.openQCD.extract_t0": {"tf": 3}, "pyerrors.input.openQCD.read_qtop": {"tf": 2.8284271247461903}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 2.449489742783178}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 2.449489742783178}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1.7320508075688772}, "pyerrors.input.pandas.to_sql": {"tf": 1.7320508075688772}, "pyerrors.input.pandas.read_sql": {"tf": 1}, "pyerrors.input.pandas.dump_df": {"tf": 1.4142135623730951}, "pyerrors.input.pandas.load_df": {"tf": 1.7320508075688772}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.7320508075688772}, "pyerrors.input.utils.check_idl": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}, "pyerrors.obs.Obs.details": {"tf": 1}, "pyerrors.obs.Obs.reweight": {"tf": 1}, "pyerrors.obs.Obs.plot_tauint": {"tf": 1}, "pyerrors.obs.Obs.plot_rho": {"tf": 1}, "pyerrors.obs.Obs.plot_piechart": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}, "pyerrors.obs.reweight": {"tf": 1}, "pyerrors.obs.correlate": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1.7320508075688772}}, "df": 54}, "m": {"docs": {"pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 1, "p": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.linalg.matmul": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1}}, "df": 4, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}}}}}}}, "i": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}}}, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 2.6457513110645907}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.input": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}, "pyerrors.roots.find_root": {"tf": 1}}, "df": 11, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}}, "df": 4}}}, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.json.load_json_dict": {"tf": 1}, "pyerrors.input.pandas.read_sql": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1}}, "df": 3}}, "s": {"docs": {"pyerrors.input.pandas.load_df": {"tf": 1}, "pyerrors.obs.import_jackknife": {"tf": 1}}, "df": 2}}}}, "r": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.deriv": {"tf": 1}, "pyerrors.correlators.Corr.second_deriv": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 3}}}}}}, "a": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 2, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.obs.CObs.gamma_method": {"tf": 1}, "pyerrors.obs.CObs.is_zero": {"tf": 1}}, "df": 5}}}}}}}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}}, "df": 1}}}}}}, "d": {"0": {"docs": {"pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 3}, "docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.4142135623730951}}, "df": 7, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "f": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}, "i": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}}, "df": 2}, "s": {"docs": {"pyerrors": {"tf": 1.7320508075688772}}, "df": 1}, "r": {"docs": {"pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.obs.cov_Obs": {"tf": 1}}, "df": 3}}}}, "c": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 1}, "pyerrors.obs.correlate": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 4}}}, "t": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}}}}, "l": {"docs": {"pyerrors": {"tf": 2}, "pyerrors.correlators.Corr.reweight": {"tf": 1.7320508075688772}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.utils.check_idl": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.__init__": {"tf": 1}, "pyerrors.obs.Obs.reweight": {"tf": 1.7320508075688772}, "pyerrors.obs.reweight": {"tf": 1.7320508075688772}}, "df": 12}}, "a": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "z": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "o": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}}}}}}, "r": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.obs.Obs.plot_history": {"tf": 1}}, "df": 2}}}}}, "l": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 1}}}}}}}}}, "/": {"docs": {}, "df": 0, "o": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "l": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}}}}}}}}, "g": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}}, "df": 3}}}}}}, "o": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}}, "df": 1}}}}}}, "e": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 1}}}, "\\": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "a": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}}}, "d": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "a": {"docs": {"pyerrors.obs.covariance": {"tf": 1.4142135623730951}}, "df": 1}}}}}}, "j": {"docs": {"pyerrors.obs.covariance": {"tf": 1.4142135623730951}}, "df": 1}, "^": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "\\": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "a": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}}}}, "|": {"docs": {}, "df": 0, "^": {"2": {"docs": {}, "df": 0, "\\": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "q": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}}}, "docs": {}, "df": 0}}}}, "}": {"docs": {}, "df": 0, "|": {"docs": {}, "df": 0, "v": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}, "\\": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}}}}}}, "p": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.fits.Fit_result": {"tf": 2}, "pyerrors.fits.ks_test": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1.4142135623730951}}, "df": 4, "y": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 6.6332495807108}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.input": {"tf": 2}, "pyerrors.input.pandas.dump_df": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1}}, "df": 5}}}}}}, "t": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.correlators.Corr.fit": {"tf": 1}}, "df": 2}}}}}, "a": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "k": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}}}}}, "r": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.obs.CObs.gamma_method": {"tf": 1}, "pyerrors.obs.CObs.is_zero": {"tf": 1}}, "df": 5, "l": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "n": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.correlate": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.T_symmetry": {"tf": 1.7320508075688772}}, "df": 3}}}, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.correlators.Corr.T_symmetry": {"tf": 1}}, "df": 1}}}}, "a": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors": {"tf": 2.6457513110645907}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}}, "df": 8, "s": {"docs": {"pyerrors": {"tf": 3}, "pyerrors.correlators.Corr.__init__": {"tf": 1}, "pyerrors.correlators.Corr.item": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.Hankel": {"tf": 1}, "pyerrors.correlators.Corr.roll": {"tf": 1}, "pyerrors.correlators.Corr.thin": {"tf": 1}, "pyerrors.correlators.Corr.correlate": {"tf": 1}, "pyerrors.correlators.Corr.reweight": {"tf": 1}, "pyerrors.correlators.Corr.T_symmetry": {"tf": 1}, "pyerrors.correlators.Corr.deriv": {"tf": 1}, "pyerrors.correlators.Corr.second_deriv": {"tf": 1}, "pyerrors.correlators.Corr.m_eff": {"tf": 1}, "pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.correlators.Corr.plateau": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.show": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.spaghetti_plot": {"tf": 1}, "pyerrors.correlators.Corr.dump": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1}, "pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.fits.Fit_result": {"tf": 1.4142135623730951}, "pyerrors.fits.Fit_result.gamma_method": {"tf": 1}, "pyerrors.fits.Fit_result.gm": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1.7320508075688772}, "pyerrors.fits.total_least_squares": {"tf": 2}, "pyerrors.fits.fit_lin": {"tf": 1}, "pyerrors.fits.error_band": {"tf": 1}, "pyerrors.fits.ks_test": {"tf": 1}, "pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}, "pyerrors.input.misc.read_pbp": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.qtop_projection": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}, "pyerrors.input.pandas.to_sql": {"tf": 1}, "pyerrors.input.pandas.read_sql": {"tf": 1.4142135623730951}, "pyerrors.input.pandas.dump_df": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1.4142135623730951}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.input.utils.check_idl": {"tf": 1}, "pyerrors.linalg.matmul": {"tf": 1}, "pyerrors.linalg.jack_matmul": {"tf": 1}, "pyerrors.linalg.einsum": {"tf": 1}, "pyerrors.misc.dump_object": {"tf": 1}, "pyerrors.misc.load_object": {"tf": 1}, "pyerrors.misc.pseudo_Obs": {"tf": 1}, "pyerrors.misc.gen_correlated_data": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}, "pyerrors.obs.Obs.__init__": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}, "pyerrors.obs.Obs.details": {"tf": 1}, "pyerrors.obs.Obs.reweight": {"tf": 1}, "pyerrors.obs.Obs.is_zero_within_error": {"tf": 1}, "pyerrors.obs.Obs.is_zero": {"tf": 1}, "pyerrors.obs.Obs.plot_tauint": {"tf": 1}, "pyerrors.obs.Obs.plot_rho": {"tf": 1}, "pyerrors.obs.Obs.plot_history": {"tf": 1}, "pyerrors.obs.Obs.plot_piechart": {"tf": 1}, "pyerrors.obs.Obs.dump": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}, "pyerrors.obs.reweight": {"tf": 1}, "pyerrors.obs.correlate": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}, "pyerrors.obs.import_jackknife": {"tf": 1}, "pyerrors.obs.merge_obs": {"tf": 1}, "pyerrors.obs.cov_Obs": {"tf": 1}, "pyerrors.roots.find_root": {"tf": 1}}, "df": 94}}}}}}}, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.T_symmetry": {"tf": 1}}, "df": 2}}}, "s": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}}, "df": 1, "r": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input": {"tf": 1}}, "df": 1}}}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}}, "df": 1}}}}}, "s": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.roots.find_root": {"tf": 1}}, "df": 4}}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}}}}}, "t": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}, "h": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.correlators.Corr.dump": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.read_ADerrors": {"tf": 2}, "pyerrors.input.bdio.write_ADerrors": {"tf": 2}, "pyerrors.input.bdio.read_mesons": {"tf": 2}, "pyerrors.input.bdio.read_dSdm": {"tf": 2}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_rwms": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.qtop_projection": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}, "pyerrors.input.pandas.to_sql": {"tf": 1}, "pyerrors.input.pandas.read_sql": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.4142135623730951}, "pyerrors.misc.dump_object": {"tf": 1.4142135623730951}, "pyerrors.misc.load_object": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.dump": {"tf": 1.4142135623730951}}, "df": 24}}, "d": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.__init__": {"tf": 1.7320508075688772}}, "df": 2}}}}}, "n": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.pandas.to_sql": {"tf": 1}, "pyerrors.input.pandas.dump_df": {"tf": 1.7320508075688772}, "pyerrors.input.pandas.load_df": {"tf": 1}}, "df": 3}}}}}, "r": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 3.872983346207417}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}}, "df": 3}}}, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1.7320508075688772}}, "df": 1}}, "o": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.input.hadrons.Npr_matrix.g5H": {"tf": 1.4142135623730951}}, "df": 1}}}}}}, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.is_zero_within_error": {"tf": 1}}, "df": 2}}, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}, "pyerrors.obs.Obs.details": {"tf": 1}}, "df": 4}}}}}}}, "v": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.write_dobs": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 10, "d": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "s": {"docs": {"pyerrors": {"tf": 1.7320508075688772}}, "df": 1}}}}}, "c": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.prune": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 4}}}}}}, "b": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "m": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}}, "df": 2}}}}, "j": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 2, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.correlators.Corr.Eigenvalue": {"tf": 1}}, "df": 1}}, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.qtop_projection": {"tf": 1}}, "df": 2}}}, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 1}}}}}}, "g": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "m": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1, "s": {"docs": {"pyerrors.input": {"tf": 1}}, "df": 1}}}}}, "d": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.correlators.Corr.spaghetti_plot": {"tf": 1}}, "df": 1}}}}}}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 3.4641016151377544}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.details": {"tf": 1}}, "df": 6, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.fit": {"tf": 1}}, "df": 1}}, "s": {"docs": {"pyerrors.fits.total_least_squares": {"tf": 1}}, "df": 1}}}, "m": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.obs.correlate": {"tf": 1}}, "df": 2}}}}, "o": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}}, "df": 1, "s": {"docs": {"pyerrors.fits.least_squares": {"tf": 1.4142135623730951}}, "df": 1}}}}, "e": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.obs.CObs.is_zero": {"tf": 1}}, "df": 2}}}}}, "o": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}}, "df": 1}}}}}}}}}}}}, "f": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}, "i": {"docs": {}, "df": 0, "x": {"docs": {"pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 2}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.4142135623730951}}, "df": 9}}}, "v": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 2}}}}}, "s": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 1}}}}}, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.__init__": {"tf": 1}, "pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.correlators.Corr.set_prange": {"tf": 1}}, "df": 3}}}, "c": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}}, "df": 1}, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}}, "df": 4}}}}}}}}, "u": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1}}, "df": 1, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}, "l": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}, "s": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}}}}}}}, "r": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.misc.pseudo_Obs": {"tf": 1}}, "df": 1}}}}}}}, "l": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}, "o": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 2}, "pyerrors.correlators.Corr.spaghetti_plot": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.qqplot": {"tf": 1}, "pyerrors.fits.residual_plot": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.obs.Obs.plot_tauint": {"tf": 1}, "pyerrors.obs.Obs.plot_rho": {"tf": 1}, "pyerrors.obs.Obs.plot_rep_dist": {"tf": 1}, "pyerrors.obs.Obs.plot_history": {"tf": 1}, "pyerrors.obs.Obs.plot_piechart": {"tf": 1}}, "df": 11, "a": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.plottable": {"tf": 1}}, "df": 1}}}}, "s": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 2}, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1}}, "df": 1}}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1}}, "df": 1}}}}}}, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "u": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.__init__": {"tf": 1}, "pyerrors.correlators.Corr.plateau": {"tf": 2.23606797749979}, "pyerrors.correlators.Corr.show": {"tf": 1.7320508075688772}}, "df": 4}}}}, "c": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}}, "df": 2}}}}}}}}, "q": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1.4142135623730951}}, "df": 1}}}}}}}}, "h": {"docs": {}, "df": 0, "y": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 2}}, "df": 1, "i": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}}, "e": {"docs": {"pyerrors": {"tf": 5.477225575051661}, "pyerrors.correlators.Corr": {"tf": 1}}, "df": 2, "r": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.item": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 5, "f": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "m": {"docs": {"pyerrors.fits.least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.total_least_squares": {"tf": 1}}, "df": 2, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 2}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1}}, "df": 5}}, "s": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.fits.fit_lin": {"tf": 1}, "pyerrors.fits.ks_test": {"tf": 1}}, "df": 4}}}}}, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "c": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.Hankel": {"tf": 1}, "pyerrors.correlators.Corr.m_eff": {"tf": 1}}, "df": 3, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.Hankel": {"tf": 1}, "pyerrors.correlators.Corr.roll": {"tf": 1}}, "df": 3}}}}, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.correlators.Corr.m_eff": {"tf": 1.4142135623730951}}, "df": 1}}}}}}}}}, "m": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}}, "df": 4}}}}}}}}, "s": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}}, "df": 2}}}}, "n": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.linalg.pinv": {"tf": 1}}, "df": 2}}}}, "c": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.mpm.matrix_pencil_method": {"tf": 1.7320508075688772}}, "df": 1}}}}}, "i": {"docs": {"pyerrors": {"tf": 1}}, "df": 1, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "^": {"0": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}, "docs": {}, "df": 0}, "n": {"docs": {}, "df": 0, "v": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "c": {"docs": {}, "df": 0, "k": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.correlators.Corr.item": {"tf": 1}}, "df": 1}, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.item": {"tf": 1.4142135623730951}}, "df": 1}}, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.dump": {"tf": 1}, "pyerrors.misc.dump_object": {"tf": 1.4142135623730951}, "pyerrors.misc.load_object": {"tf": 1}, "pyerrors.obs.Obs.dump": {"tf": 1}}, "df": 4}}}}, "e": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.obs.Obs.plot_piechart": {"tf": 1}}, "df": 1}}}}}}}, "o": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 5, "s": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 2}}}}, "s": {"docs": {"pyerrors.covobs.Covobs.__init__": {"tf": 1}}, "df": 1, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}, "f": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "x": {"docs": {"pyerrors.input.openQCD.read_rwms": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1.4142135623730951}}, "df": 3}}}}, "s": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 2}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.obs.merge_obs": {"tf": 1}}, "df": 5}, "y": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 3}}}}}, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}}, "df": 5}}, "v": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1.4142135623730951}}, "df": 3}}}}}}, "w": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}}, "df": 1}}}}, "b": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.dobs.create_pobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.write_pobs": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.read_pobs": {"tf": 1}}, "df": 3}}}, "s": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.linalg.pinv": {"tf": 1}}, "df": 2}}}}}}}, "s": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.input.hadrons.read_meson_hd5": {"tf": 2}}, "df": 1}}}}}}}}}}}, "b": {"docs": {}, "df": 0, "p": {"docs": {"pyerrors.input.misc.read_pbp": {"tf": 1}}, "df": 1}}}, "a": {"docs": {"pyerrors": {"tf": 8.366600265340756}, "pyerrors.correlators.Corr": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.__init__": {"tf": 1}, "pyerrors.correlators.Corr.projected": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.item": {"tf": 1}, "pyerrors.correlators.Corr.plottable": {"tf": 1}, "pyerrors.correlators.Corr.is_matrix_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.thin": {"tf": 1}, "pyerrors.correlators.Corr.correlate": {"tf": 1}, "pyerrors.correlators.Corr.reweight": {"tf": 1}, "pyerrors.correlators.Corr.plateau": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.spaghetti_plot": {"tf": 1}, "pyerrors.correlators.Corr.dump": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.prune": {"tf": 2}, "pyerrors.fits.least_squares": {"tf": 3.605551275463989}, "pyerrors.fits.total_least_squares": {"tf": 3.3166247903554}, "pyerrors.fits.fit_lin": {"tf": 1.7320508075688772}, "pyerrors.fits.qqplot": {"tf": 1}, "pyerrors.fits.residual_plot": {"tf": 1}, "pyerrors.fits.ks_test": {"tf": 1}, "pyerrors.input.bdio.read_ADerrors": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.read_mesons": {"tf": 2}, "pyerrors.input.bdio.read_dSdm": {"tf": 2}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.write_pobs": {"tf": 2.23606797749979}, "pyerrors.input.dobs.read_pobs": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.import_dobs_string": {"tf": 2}, "pyerrors.input.dobs.read_dobs": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.create_dobs_string": {"tf": 2}, "pyerrors.input.dobs.write_dobs": {"tf": 2.449489742783178}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 3}, "pyerrors.input.json.create_json_string": {"tf": 1.7320508075688772}, "pyerrors.input.json.dump_to_json": {"tf": 2.23606797749979}, "pyerrors.input.json.import_json_string": {"tf": 1.7320508075688772}, "pyerrors.input.json.load_json": {"tf": 1.7320508075688772}, "pyerrors.input.json.dump_dict_to_json": {"tf": 2.23606797749979}, "pyerrors.input.json.load_json_dict": {"tf": 1.7320508075688772}, "pyerrors.input.misc.read_pbp": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 2.23606797749979}, "pyerrors.input.pandas.dump_df": {"tf": 2.23606797749979}, "pyerrors.input.pandas.load_df": {"tf": 1.4142135623730951}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.7320508075688772}, "pyerrors.linalg.eigh": {"tf": 1}, "pyerrors.linalg.eig": {"tf": 1}, "pyerrors.linalg.pinv": {"tf": 1}, "pyerrors.linalg.svd": {"tf": 1}, "pyerrors.misc.dump_object": {"tf": 1}, "pyerrors.misc.gen_correlated_data": {"tf": 1.4142135623730951}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs": {"tf": 2.449489742783178}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}, "pyerrors.obs.Obs.reweight": {"tf": 1}, "pyerrors.obs.Obs.is_zero": {"tf": 1}, "pyerrors.obs.Obs.plot_tauint": {"tf": 1}, "pyerrors.obs.Obs.plot_rho": {"tf": 1}, "pyerrors.obs.Obs.plot_piechart": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.dump": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1.4142135623730951}, "pyerrors.obs.CObs": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1.4142135623730951}, "pyerrors.obs.reweight": {"tf": 1.4142135623730951}, "pyerrors.obs.correlate": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 2.23606797749979}, "pyerrors.obs.cov_Obs": {"tf": 1}}, "df": 72, "n": {"docs": {"pyerrors": {"tf": 5.0990195135927845}, "pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.correlators.Corr.Hankel": {"tf": 1}, "pyerrors.correlators.Corr.correlate": {"tf": 1}, "pyerrors.correlators.Corr.reweight": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 2.23606797749979}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.fits.error_band": {"tf": 1}, "pyerrors.input": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 3.605551275463989}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1.4142135623730951}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.input.utils.check_idl": {"tf": 1}, "pyerrors.misc.pseudo_Obs": {"tf": 1}, "pyerrors.misc.gen_correlated_data": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 2}, "pyerrors.obs.Obs.reweight": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}, "pyerrors.obs.reweight": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1.7320508075688772}, "pyerrors.obs.import_jackknife": {"tf": 1}, "pyerrors.obs.cov_Obs": {"tf": 1}, "pyerrors.roots.find_root": {"tf": 1}}, "df": 31, "d": {"docs": {"pyerrors": {"tf": 7}, "pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.correlators.Corr.__init__": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.item": {"tf": 1}, "pyerrors.correlators.Corr.plottable": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 1}, "pyerrors.correlators.Corr.reweight": {"tf": 1}, "pyerrors.correlators.Corr.T_symmetry": {"tf": 1}, "pyerrors.correlators.Corr.fit": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.correlators.Corr.dump": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1.7320508075688772}, "pyerrors.fits.least_squares": {"tf": 2.449489742783178}, "pyerrors.fits.total_least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.fit_lin": {"tf": 1}, "pyerrors.fits.residual_plot": {"tf": 1}, "pyerrors.input": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.read_dSdm": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.7320508075688772}, "pyerrors.input.hadrons.Npr_matrix.g5H": {"tf": 1}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1.4142135623730951}, "pyerrors.input.json.load_json_dict": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 2}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1.7320508075688772}, "pyerrors.input.pandas.read_sql": {"tf": 1}, "pyerrors.linalg.eigh": {"tf": 1}, "pyerrors.misc.pseudo_Obs": {"tf": 1}, "pyerrors.misc.gen_correlated_data": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}, "pyerrors.obs.Obs.details": {"tf": 1}, "pyerrors.obs.Obs.reweight": {"tf": 1}, "pyerrors.obs.Obs.plot_piechart": {"tf": 1}, "pyerrors.obs.Obs.dump": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1.4142135623730951}, "pyerrors.obs.CObs.gamma_method": {"tf": 1}, "pyerrors.obs.CObs.is_zero": {"tf": 1}, "pyerrors.obs.reweight": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1.7320508075688772}, "pyerrors.obs.import_jackknife": {"tf": 1.4142135623730951}, "pyerrors.obs.cov_Obs": {"tf": 1}}, "df": 60}, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "y": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 2.8284271247461903}, "pyerrors.fits.ks_test": {"tf": 1}, "pyerrors.input": {"tf": 1}, "pyerrors.input.pandas.read_sql": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1}}, "df": 10}}}, "z": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 1}}}}}}, "y": {"docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.roots.find_root": {"tf": 1}}, "df": 3}, "t": {"docs": {}, "df": 0, "i": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.anti_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.m_eff": {"tf": 1}}, "df": 3}}, "o": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.correlate": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1.4142135623730951}}, "df": 4}}}}, "n": {"docs": {}, "df": 0, "y": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.obs.derived_observable": {"tf": 1}}, "df": 1}}}}}}}, "p": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.fits.least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.total_least_squares": {"tf": 1.4142135623730951}, "pyerrors.obs.derived_observable": {"tf": 1}, "pyerrors.roots.find_root": {"tf": 1.4142135623730951}}, "df": 5}}, "r": {"docs": {}, "df": 0, "x": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, ":": {"1": {"0": {"0": {"9": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "2": {"0": {"5": {"docs": {"pyerrors.correlators.Corr.m_eff": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "8": {"0": {"9": {"docs": {"pyerrors": {"tf": 2}}, "df": 1}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "2": {"0": {"0": {"4": {"docs": {"pyerrors.correlators.Corr.GEVP": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "2": {"0": {"9": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "p": {"docs": {"pyerrors": {"tf": 2}}, "df": 1}}}}}}}, "e": {"docs": {"pyerrors": {"tf": 5.477225575051661}, "pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 1}, "pyerrors.correlators.Corr.reweight": {"tf": 1}, "pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.prune": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.fits.fit_lin": {"tf": 1.4142135623730951}, "pyerrors.fits.qqplot": {"tf": 1}, "pyerrors.input": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.read_dobs": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.7320508075688772}, "pyerrors.input.hadrons.Npr_matrix.g5H": {"tf": 1}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1.4142135623730951}, "pyerrors.input.pandas.to_sql": {"tf": 1}, "pyerrors.input.pandas.dump_df": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.__init__": {"tf": 1}, "pyerrors.obs.Obs.reweight": {"tf": 1}, "pyerrors.obs.CObs.is_zero": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}, "pyerrors.obs.reweight": {"tf": 1}, "pyerrors.obs.correlate": {"tf": 1}, "pyerrors.obs.import_jackknife": {"tf": 1}, "pyerrors.obs.merge_obs": {"tf": 1}}, "df": 50}, "g": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}}, "df": 3, "s": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1.4142135623730951}}, "df": 2}}}}}}}, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 4.47213595499958}, "pyerrors.correlators.Corr.__init__": {"tf": 1.4142135623730951}, "pyerrors.covobs.Covobs.__init__": {"tf": 1.4142135623730951}, "pyerrors.fits.error_band": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 6.082762530298219}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1.4142135623730951}, "pyerrors.obs.derived_observable": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}, "pyerrors.obs.import_jackknife": {"tf": 1}, "pyerrors.obs.cov_Obs": {"tf": 1.4142135623730951}}, "df": 13, "s": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.correlators.Corr.__init__": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.linalg.matmul": {"tf": 1}, "pyerrors.linalg.jack_matmul": {"tf": 1}, "pyerrors.linalg.einsum": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1}, "pyerrors.obs.Obs.__init__": {"tf": 1}}, "df": 9}}, "n": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}, "i": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}, "o": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.symmetric": {"tf": 1}, "pyerrors.correlators.Corr.anti_symmetric": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 3}}}}, "c": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "h": {"docs": {"pyerrors.correlators.Corr.m_eff": {"tf": 1}}, "df": 1}}}}}, "b": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.linalg.matmul": {"tf": 1}, "pyerrors.linalg.jack_matmul": {"tf": 1}, "pyerrors.linalg.einsum": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}}, "df": 4}}}}}}}}, "u": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "o": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.input.pandas.read_sql": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1}}, "df": 5, "m": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "c": {"docs": {"pyerrors": {"tf": 3.1622776601683795}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1.4142135623730951}}, "df": 6, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 9}}}}}}}}}, "g": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.fits.least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.total_least_squares": {"tf": 1.4142135623730951}, "pyerrors.obs.derived_observable": {"tf": 1.7320508075688772}, "pyerrors.roots.find_root": {"tf": 1.4142135623730951}}, "df": 5}}}}, "c": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 2.8284271247461903}, "pyerrors.misc.gen_correlated_data": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.gm": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.plot_tauint": {"tf": 1}, "pyerrors.obs.Obs.plot_rho": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 2.23606797749979}}, "df": 7, "s": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}}}}}}}}}}}}, "x": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}}, "df": 7}}}}}}}}, "s": {"docs": {"pyerrors": {"tf": 6}, "pyerrors.correlators.Corr.m_eff": {"tf": 1}, "pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1.7320508075688772}, "pyerrors.fits.least_squares": {"tf": 1.7320508075688772}, "pyerrors.fits.total_least_squares": {"tf": 1.7320508075688772}, "pyerrors.fits.fit_lin": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.read_dobs": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.7320508075688772}, "pyerrors.input.pandas.load_df": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}, "pyerrors.obs.import_jackknife": {"tf": 1.4142135623730951}, "pyerrors.roots.find_root": {"tf": 1}}, "df": 19, "s": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 1, "d": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.fits.fit_lin": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 2.23606797749979}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}}, "df": 9}, "s": {"docs": {"pyerrors.input.dobs.read_dobs": {"tf": 1.4142135623730951}, "pyerrors.input.json.load_json": {"tf": 1.4142135623730951}, "pyerrors.input.json.load_json_dict": {"tf": 1.4142135623730951}, "pyerrors.input.pandas.load_df": {"tf": 1.4142135623730951}}, "df": 4}}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1.4142135623730951}}, "df": 4}}}}}, "o": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}}}}}}, "i": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1, "s": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}}}}, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 2}}}}}}}, "l": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "o": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}, "p": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "a": {"docs": {"pyerrors": {"tf": 2.23606797749979}}, "df": 1, "^": {"docs": {}, "df": 0, "i": {"docs": {"pyerrors": {"tf": 1.7320508075688772}}, "df": 1}}}}}, "l": {"docs": {"pyerrors": {"tf": 2.449489742783178}, "pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 1}, "pyerrors.correlators.Corr.correlate": {"tf": 1}, "pyerrors.correlators.Corr.reweight": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.fits.Fit_result.gamma_method": {"tf": 1}, "pyerrors.fits.Fit_result.gm": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.total_least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.ks_test": {"tf": 1}, "pyerrors.input.bdio.read_ADerrors": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.read_mesons": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.read_dSdm": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.pandas.read_sql": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.4142135623730951}, "pyerrors.linalg.matmul": {"tf": 1}, "pyerrors.obs.Obs.reweight": {"tf": 1.7320508075688772}, "pyerrors.obs.derived_observable": {"tf": 1}, "pyerrors.obs.reweight": {"tf": 1.7320508075688772}, "pyerrors.obs.covariance": {"tf": 1}, "pyerrors.obs.merge_obs": {"tf": 1}}, "df": 34, "o": {"docs": {}, "df": 0, "w": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 1, "s": {"docs": {"pyerrors": {"tf": 2}, "pyerrors.correlators.Corr.prune": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 3}}, "c": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}}}}}}, "s": {"docs": {}, "df": 0, "o": {"docs": {"pyerrors": {"tf": 2.449489742783178}, "pyerrors.fits.Fit_result": {"tf": 1}, "pyerrors.input": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 2}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1}}, "df": 7}}, "g": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "m": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}}, "df": 3, "s": {"docs": {"pyerrors": {"tf": 1.7320508075688772}}, "df": 1}}}}}}}}, "w": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "y": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 9, "s": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}}, "df": 1}}}}}}}}}}, "i": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}}, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.input.pandas.to_sql": {"tf": 1}}, "df": 1}}}}}}, "p": {"docs": {"pyerrors": {"tf": 1}}, "df": 1, "p": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.input.pandas.read_sql": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1}}, "df": 3}, "d": {"docs": {"pyerrors.obs.correlate": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1.4142135623730951}}, "df": 2}}}, "y": {"docs": {"pyerrors.correlators.Corr.gamma_method": {"tf": 1}, "pyerrors.correlators.Corr.gm": {"tf": 1}, "pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.fits.Fit_result.gamma_method": {"tf": 1}, "pyerrors.fits.Fit_result.gm": {"tf": 1}}, "df": 6}}, "r": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 4}}}}}}, "x": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.linalg.jack_matmul": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 2}}}}}}}}}}, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.pandas.to_sql": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.7320508075688772}}, "df": 2}}}}, "i": {"docs": {"pyerrors": {"tf": 2.23606797749979}}, "df": 1}}, "v": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.correlators.Corr.deriv": {"tf": 1}, "pyerrors.correlators.Corr.second_deriv": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1.4142135623730951}}, "df": 4}}}}}}}, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.reweight": {"tf": 1}, "pyerrors.correlators.Corr.T_symmetry": {"tf": 1}, "pyerrors.correlators.Corr.plateau": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.reweight": {"tf": 1}, "pyerrors.obs.reweight": {"tf": 1}}, "df": 5, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}, "g": {"docs": {"pyerrors.correlators.Corr.plateau": {"tf": 1}}, "df": 1}, "o": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}}, "f": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.obs.correlate": {"tf": 1}}, "df": 5}}}, "f": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}, "t": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.__init__": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1.4142135623730951}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 18, "t": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}}, "df": 2}}}}}, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.set_prange": {"tf": 1}}, "df": 1, "s": {"docs": {"pyerrors.fits.Fit_result": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs": {"tf": 1}}, "df": 3}}}}}}}}, "o": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.obs.Obs.is_zero": {"tf": 1}}, "df": 1}}}, "m": {"docs": {}, "df": 0, "p": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "c": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}, "r": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.correlators.Corr.GEVP": {"tf": 1.7320508075688772}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.4142135623730951}, "pyerrors.linalg.eigh": {"tf": 1}, "pyerrors.linalg.eig": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}}, "df": 11}}}}}}, "e": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}}, "df": 4}}, "i": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.fits.Fit_result": {"tf": 1}}, "df": 1}}}}}}}}, "h": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}}, "df": 6}}}}}}, "t": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1, "l": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}}}}}, "o": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 1}}}}}, "b": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1, "o": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.obs.Obs.is_zero": {"tf": 1}}, "df": 2}}}}}, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}}}, "o": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 2}}, "u": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.obs.Obs.details": {"tf": 1}}, "df": 3}}}}, "i": {"docs": {}, "df": 0, "m": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "d": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}, "d": {"docs": {"pyerrors.correlators.Corr": {"tf": 1}}, "df": 1, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1}}, "df": 7}}}}}}, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}}, "df": 4}}}, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 1}}}, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}}, "df": 3}}}}}}}, "x": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.spaghetti_plot": {"tf": 1}}, "df": 2}}}, "g": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1.4142135623730951}, "pyerrors.input.utils.check_idl": {"tf": 1}}, "df": 2}}}}}, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.obs.Obs.export_jackknife": {"tf": 1}}, "df": 1, "s": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}}}, "^": {"2": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "\\": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "a": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}}}}, "/": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "^": {"2": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}}}}}, "a": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}, "f": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}}, "df": 2, "o": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors": {"tf": 6.855654600401044}, "pyerrors.correlators.Corr": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.__init__": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 2.23606797749979}, "pyerrors.correlators.Corr.reweight": {"tf": 1}, "pyerrors.correlators.Corr.m_eff": {"tf": 2}, "pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.dump": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1}, "pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.fits.Fit_result": {"tf": 1.4142135623730951}, "pyerrors.fits.least_squares": {"tf": 3}, "pyerrors.fits.total_least_squares": {"tf": 2}, "pyerrors.fits.fit_lin": {"tf": 1.4142135623730951}, "pyerrors.fits.error_band": {"tf": 1.4142135623730951}, "pyerrors.fits.ks_test": {"tf": 1}, "pyerrors.input": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 2.6457513110645907}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.misc.read_pbp": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_rwms": {"tf": 2}, "pyerrors.input.openQCD.extract_t0": {"tf": 2.23606797749979}, "pyerrors.input.openQCD.read_qtop": {"tf": 2.449489742783178}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 2.6457513110645907}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1.7320508075688772}, "pyerrors.input.pandas.read_sql": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 3}, "pyerrors.input.utils": {"tf": 1}, "pyerrors.linalg.jack_matmul": {"tf": 1}, "pyerrors.linalg.einsum": {"tf": 1.7320508075688772}, "pyerrors.misc.dump_object": {"tf": 1}, "pyerrors.misc.pseudo_Obs": {"tf": 1.7320508075688772}, "pyerrors.misc.gen_correlated_data": {"tf": 2}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 3.872983346207417}, "pyerrors.obs.Obs.gamma_method": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.gm": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.reweight": {"tf": 1}, "pyerrors.obs.Obs.is_zero_within_error": {"tf": 1}, "pyerrors.obs.Obs.is_zero": {"tf": 1}, "pyerrors.obs.Obs.plot_tauint": {"tf": 1}, "pyerrors.obs.Obs.plot_rho": {"tf": 1}, "pyerrors.obs.Obs.plot_rep_dist": {"tf": 1}, "pyerrors.obs.Obs.plot_history": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.dump": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1.4142135623730951}, "pyerrors.obs.CObs": {"tf": 1}, "pyerrors.obs.CObs.gamma_method": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1.7320508075688772}, "pyerrors.obs.reweight": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 2.6457513110645907}, "pyerrors.obs.cov_Obs": {"tf": 1}, "pyerrors.roots.find_root": {"tf": 1}}, "df": 63, "t": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 2}}}}, "m": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.m_eff": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.total_least_squares": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}}, "df": 7, "u": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "a": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1, "e": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}, "a": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 3.605551275463989}, "pyerrors.correlators.Corr.plottable": {"tf": 1}, "pyerrors.correlators.Corr.dump": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}, "pyerrors.input.misc.read_pbp": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.pandas.dump_df": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.dump": {"tf": 1.7320508075688772}}, "df": 16, "s": {"docs": {"pyerrors.correlators.Corr.dump": {"tf": 1}, "pyerrors.obs.Obs.dump": {"tf": 1}}, "df": 2}, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}}, "df": 1}}}}}}, "w": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.deriv": {"tf": 1}}, "df": 1}}}}}, "u": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}}, "df": 2}}, "r": {"docs": {}, "df": 0, "q": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "k": {"docs": {}, "df": 0, "f": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "y": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}}, "df": 1}}}}}}}}}}}}}}}}}}}}}, "l": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "w": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors": {"tf": 3}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}}, "df": 5}}}}}}, "d": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.input.misc.read_pbp": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 3, "s": {"docs": {"pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 1}}}}}}, "e": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}}, "r": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "m": {"docs": {"pyerrors": {"tf": 4.358898943540674}, "pyerrors.correlators.Corr.item": {"tf": 1}, "pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.input": {"tf": 1}, "pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.7320508075688772}, "pyerrors.input.json.import_json_string": {"tf": 1.4142135623730951}, "pyerrors.input.json.load_json": {"tf": 1.4142135623730951}, "pyerrors.input.json.load_json_dict": {"tf": 1}, "pyerrors.input.misc.read_pbp": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1.7320508075688772}, "pyerrors.input.pandas.load_df": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.4142135623730951}, "pyerrors.misc.load_object": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1.7320508075688772}}, "df": 34}, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.__init__": {"tf": 1}}, "df": 1}}}, "a": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "w": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "k": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}, "n": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "o": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}, "c": {"docs": {}, "df": 0, "{": {"1": {"docs": {}, "df": 0, "}": {"docs": {}, "df": 0, "{": {"2": {"docs": {"pyerrors": {"tf": 1}}, "df": 1, "}": {"docs": {}, "df": 0, "+": {"docs": {}, "df": 0, "\\": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "m": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}}, "docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "}": {"docs": {}, "df": 0, "\\": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "m": {"docs": {"pyerrors.obs.covariance": {"tf": 1.4142135623730951}}, "df": 1}}}}}}}}}, "docs": {}, "df": 0}, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.obs.Obs.plot_piechart": {"tf": 1}}, "df": 1}}, "s": {"docs": {"pyerrors.obs.Obs.plot_piechart": {"tf": 1}}, "df": 1}}}}}}}}, "i": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 3.7416573867739413}, "pyerrors.correlators.Corr.fit": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1.7320508075688772}, "pyerrors.fits.Fit_result": {"tf": 2}, "pyerrors.fits.Fit_result.gamma_method": {"tf": 1}, "pyerrors.fits.Fit_result.gm": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 3}, "pyerrors.fits.total_least_squares": {"tf": 1.7320508075688772}, "pyerrors.fits.fit_lin": {"tf": 1.7320508075688772}, "pyerrors.fits.qqplot": {"tf": 1.4142135623730951}, "pyerrors.fits.residual_plot": {"tf": 1}, "pyerrors.fits.error_band": {"tf": 1}, "pyerrors.fits.ks_test": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.extract_t0": {"tf": 2}}, "df": 15, "s": {"docs": {"pyerrors": {"tf": 3.7416573867739413}, "pyerrors.correlators.Corr.fit": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.fits.Fit_result": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.total_least_squares": {"tf": 1}}, "df": 6}, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 2}}}, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.fit": {"tf": 1.4142135623730951}}, "df": 2}}}}}}, "r": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.correlators.Corr.__init__": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.item": {"tf": 1}, "pyerrors.correlators.Corr.deriv": {"tf": 1}, "pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}, "pyerrors.input.misc.read_pbp": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 2.23606797749979}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.obs.correlate": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}, "pyerrors.obs.import_jackknife": {"tf": 1}}, "df": 18}}}, "x": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 7}}}, "n": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.GEVP": {"tf": 1}}, "df": 1, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}}}, "e": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.correlators.Corr.m_eff": {"tf": 1}}, "df": 1}}, "s": {"docs": {"pyerrors.roots.find_root": {"tf": 1}}, "df": 1}}, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.deriv": {"tf": 1}, "pyerrors.correlators.Corr.second_deriv": {"tf": 1}}, "df": 2}}}}, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 4.358898943540674}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.correlators.Corr.dump": {"tf": 2}, "pyerrors.input.bdio.read_ADerrors": {"tf": 1.7320508075688772}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1.7320508075688772}, "pyerrors.input.bdio.read_mesons": {"tf": 1.7320508075688772}, "pyerrors.input.bdio.read_dSdm": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 2}, "pyerrors.input.dobs.read_pobs": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 2}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.write_dobs": {"tf": 2}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1.7320508075688772}, "pyerrors.input.json.dump_to_json": {"tf": 2.23606797749979}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 2}, "pyerrors.input.json.dump_dict_to_json": {"tf": 2.23606797749979}, "pyerrors.input.json.load_json_dict": {"tf": 2}, "pyerrors.input.openQCD.read_rwms": {"tf": 2}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}, "pyerrors.input.pandas.dump_df": {"tf": 2.23606797749979}, "pyerrors.input.pandas.load_df": {"tf": 1.7320508075688772}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.4142135623730951}, "pyerrors.misc.dump_object": {"tf": 2}, "pyerrors.misc.load_object": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.plot_tauint": {"tf": 1}, "pyerrors.obs.Obs.plot_rho": {"tf": 1}, "pyerrors.obs.Obs.plot_piechart": {"tf": 1}, "pyerrors.obs.Obs.dump": {"tf": 2.23606797749979}}, "df": 39, "s": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_rwms": {"tf": 2.6457513110645907}, "pyerrors.input.openQCD.extract_t0": {"tf": 2.23606797749979}, "pyerrors.input.openQCD.read_qtop": {"tf": 2.449489742783178}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 2.449489742783178}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 2}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 2.6457513110645907}, "pyerrors.input.sfcf.read_sfcf": {"tf": 2}}, "df": 13, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "m": {"docs": {"pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}}, "df": 4}}}}, "n": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.dump": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}, "pyerrors.input.pandas.dump_df": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1}, "pyerrors.obs.Obs.dump": {"tf": 1}}, "df": 12, "s": {"docs": {"pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 2}}}}, "o": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "f": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}}, "df": 1}}}}}}}}}}}}}}, "l": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 1}}}}}, "e": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 2}}, "df": 1}}}, "g": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.plot_tauint": {"tf": 1}, "pyerrors.obs.Obs.plot_rho": {"tf": 1}, "pyerrors.obs.Obs.plot_piechart": {"tf": 1}}, "df": 4}}}}}, "a": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}, "c": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}}, "t": {"docs": {"pyerrors.input.hadrons.Npr_matrix.g5H": {"tf": 1}}, "df": 1, "o": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.correlators.Corr.reweight": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.reweight": {"tf": 1.4142135623730951}, "pyerrors.obs.reweight": {"tf": 1.4142135623730951}}, "df": 3, "s": {"docs": {"pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.obs.Obs.reweight": {"tf": 1}}, "df": 2}}}}}, "l": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.fits.least_squares": {"tf": 2.23606797749979}, "pyerrors.fits.total_least_squares": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.read_dobs": {"tf": 2}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1.4142135623730951}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.pandas.read_sql": {"tf": 1}, "pyerrors.input.pandas.dump_df": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.reweight": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}, "pyerrors.obs.reweight": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1.4142135623730951}}, "df": 23}}}, "i": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.input.pandas.to_sql": {"tf": 1}}, "df": 1}}, "s": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.linalg.matmul": {"tf": 1}, "pyerrors.linalg.jack_matmul": {"tf": 1}}, "df": 2}}}}}, "^": {"docs": {}, "df": 0, "i": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "u": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "c": {"docs": {"pyerrors": {"tf": 2}, "pyerrors.fits.least_squares": {"tf": 2.23606797749979}, "pyerrors.fits.total_least_squares": {"tf": 2.449489742783178}, "pyerrors.fits.error_band": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 2}, "pyerrors.roots.find_root": {"tf": 1.7320508075688772}}, "df": 6, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.correlators.Corr.fit": {"tf": 1.7320508075688772}, "pyerrors.fits.least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.error_band": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 2}, "pyerrors.obs.Obs.gamma_method": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.gm": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.plot_rho": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}, "pyerrors.roots.find_root": {"tf": 2}}, "df": 14, "s": {"docs": {"pyerrors": {"tf": 3.605551275463989}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1.4142135623730951}, "pyerrors.roots.find_root": {"tf": 1}}, "df": 7}, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.input.pandas.dump_df": {"tf": 1}}, "df": 1}}}}}}}}}}}, "l": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 11, "y": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}}, "r": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}}}}}, "l": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.m_eff": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.fits.Fit_result": {"tf": 1.7320508075688772}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}, "pyerrors.misc.pseudo_Obs": {"tf": 1.4142135623730951}, "pyerrors.misc.gen_correlated_data": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.gamma_method": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.gm": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.is_zero": {"tf": 1}, "pyerrors.obs.cov_Obs": {"tf": 1}, "pyerrors.roots.find_root": {"tf": 1}}, "df": 13, "s": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.fit_lin": {"tf": 1}, "pyerrors.obs.cov_Obs": {"tf": 1}}, "df": 4}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 2}}}}}, "w": {"docs": {"pyerrors.input.openQCD.read_qtop": {"tf": 2}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 2.23606797749979}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.7320508075688772}}, "df": 3}}, "u": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}}}}}, "a": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}}, "df": 4, "s": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}, "t": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1.7320508075688772}}, "df": 1, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}}}, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}}}}}}}, "t": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}, "p": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "c": {"docs": {"pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}}, "df": 4}}}, "n": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}, "pyerrors.input.pandas.dump_df": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1}}, "df": 10}}}}, "f": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.obs.Obs.gamma_method": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.gm": {"tf": 1.4142135623730951}}, "df": 2}}}, "e": {"docs": {"pyerrors": {"tf": 2}, "pyerrors.correlators.Corr.m_eff": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 2.23606797749979}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.4142135623730951}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.4142135623730951}, "pyerrors.obs.derived_observable": {"tf": 1.4142135623730951}, "pyerrors.obs.reweight": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1.4142135623730951}}, "df": 14, "r": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.input.openQCD.read_rwms": {"tf": 1}}, "df": 1, "o": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors": {"tf": 5.916079783099616}, "pyerrors.correlators.Corr.plottable": {"tf": 1}, "pyerrors.covobs.Covobs.errsq": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.fits.error_band": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1.7320508075688772}, "pyerrors.input.pandas.read_sql": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1}, "pyerrors.misc.pseudo_Obs": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.gamma_method": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.gm": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.plot_piechart": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 2.23606797749979}}, "df": 15, "s": {"docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.obs.Obs.is_zero_within_error": {"tf": 1.4142135623730951}, "pyerrors.obs.covariance": {"tf": 1.4142135623730951}}, "df": 4}}}, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "m": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}, "x": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 5, "l": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1}}, "df": 2}}}}, "m": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 2.449489742783178}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.roots.find_root": {"tf": 1}}, "df": 3, "s": {"docs": {"pyerrors": {"tf": 2}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.7320508075688772}}, "df": 2, "/": {"docs": {}, "df": 0, "j": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}}}}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}}}}, "i": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1}}, "df": 1, "s": {"docs": {"pyerrors.input.pandas.to_sql": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs": {"tf": 1.7320508075688772}}, "df": 2}}}}, "p": {"docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.obs.Obs": {"tf": 2}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}, "pyerrors.roots.find_root": {"tf": 1}}, "df": 5, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}, "d": {"docs": {"pyerrors.obs.Obs.plot_history": {"tf": 1}}, "df": 1, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.obs.Obs.plot_history": {"tf": 1}}, "df": 1}}}}}, "o": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}}}}}}}, "r": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.input": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1}}, "df": 10, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.dump": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.dump": {"tf": 1}}, "df": 10}}, "s": {"docs": {"pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.pandas.dump_df": {"tf": 1}}, "df": 3}}}, "s": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}}, "df": 1}}}}}, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}}}}}, "c": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.fits.least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.total_least_squares": {"tf": 1.4142135623730951}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 3}}}}, "n": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 1}}}}}}, "l": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.m_eff": {"tf": 1}}, "df": 1, "l": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}}, "df": 2}}}}}}}}, "h": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1}}, "df": 1, "s": {"docs": {"pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}}, "df": 2}}}}}}, "t": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.plateau": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1.4142135623730951}}, "df": 10, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 3}}, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 1}}}}}}, "e": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}}}}, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors": {"tf": 2.8284271247461903}}, "df": 1, "l": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}}, "df": 1}}}}}}}, "n": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.Hankel": {"tf": 1}}, "df": 1}}}, "s": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.dirac.epsilon_tensor_rank4": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1.7320508075688772}}, "df": 2}}}}, "t": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 4}}}}, "e": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}, "c": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.pandas.read_sql": {"tf": 1}}, "df": 1, "d": {"docs": {"pyerrors.input.pandas.read_sql": {"tf": 1}}, "df": 1}, "s": {"docs": {"pyerrors.obs.CObs.gamma_method": {"tf": 1}}, "df": 1}}}}}}, "c": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.correlators.Corr.spaghetti_plot": {"tf": 1}}, "df": 1}}}}}}}}, "h": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.hadrons.Npr_matrix.g5H": {"tf": 1}}, "df": 1}}}}}}}}, "i": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.linalg.eig": {"tf": 1}}, "df": 2, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 2.23606797749979}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 4, "s": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.linalg.eigh": {"tf": 1}, "pyerrors.linalg.eig": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1.4142135623730951}}, "df": 5}}}}}, "e": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.correlators.Corr.GEVP": {"tf": 2.23606797749979}}, "df": 1, "s": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.GEVP": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.prune": {"tf": 1.7320508075688772}, "pyerrors.linalg.eigh": {"tf": 1}}, "df": 4}}}}}}}}}, "h": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.linalg.eigh": {"tf": 1}}, "df": 2}}, "t": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.correlate": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.fit_lin": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.misc.gen_correlated_data": {"tf": 1}}, "df": 9}}}}, "n": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "m": {"docs": {"pyerrors.linalg.einsum": {"tf": 1}}, "df": 1}}}}}, "n": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.obs.Obs.details": {"tf": 1}}, "df": 7, "e": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"1": {"docs": {"pyerrors": {"tf": 3.4641016151377544}}, "df": 1, "|": {"docs": {}, "df": 0, "r": {"0": {"1": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "2": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "docs": {}, "df": 0}}}, "2": {"docs": {"pyerrors": {"tf": 1.7320508075688772}}, "df": 1}, "3": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {"pyerrors": {"tf": 5.5677643628300215}, "pyerrors.input.bdio.read_mesons": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.read_dSdm": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.write_pobs": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.4142135623730951}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.4142135623730951}, "pyerrors.misc.pseudo_Obs": {"tf": 1}, "pyerrors.misc.gen_correlated_data": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 2.449489742783178}, "pyerrors.obs.Obs.plot_tauint": {"tf": 1}, "pyerrors.obs.Obs.plot_rho": {"tf": 1}, "pyerrors.obs.Obs.plot_rep_dist": {"tf": 1}, "pyerrors.obs.Obs.plot_history": {"tf": 1}, "pyerrors.obs.Obs.plot_piechart": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1.4142135623730951}, "pyerrors.obs.covariance": {"tf": 1}, "pyerrors.obs.import_jackknife": {"tf": 1}}, "df": 28, "s": {"docs": {"pyerrors": {"tf": 2.449489742783178}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.read_dobs": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.obs.Obs.details": {"tf": 1}, "pyerrors.obs.correlate": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 8, "/": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "a": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1.4142135623730951}}, "df": 2}}}}}}}}}}}}}}, "t": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.dobs.create_pobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.write_pobs": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.write_dobs": {"tf": 1.4142135623730951}}, "df": 4, "s": {"docs": {"pyerrors.input.dobs.create_dobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.write_dobs": {"tf": 1.4142135623730951}}, "df": 2}}}}, "n": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}}, "df": 2}}}}, "u": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}}}, "t": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 3.1622776601683795}, "pyerrors.correlators.Corr.__init__": {"tf": 1}, "pyerrors.correlators.Corr.correlate": {"tf": 1}, "pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1}, "pyerrors.obs.cov_Obs": {"tf": 1}}, "df": 10}}}, "y": {"docs": {"pyerrors": {"tf": 3.3166247903554}, "pyerrors.correlators.Corr.thin": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.misc.gen_correlated_data": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1}, "pyerrors.obs.import_jackknife": {"tf": 1.4142135623730951}}, "df": 7}}, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}}}, "i": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 3}}}}, "a": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}}, "df": 2}}}, "d": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.openQCD.read_rwms": {"tf": 1}}, "df": 1}}}}}}, "s": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 5, "d": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1.4142135623730951}}, "df": 3}, "s": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}}, "df": 3}}, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 2.23606797749979}}, "df": 1}}}}}}}, "a": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}}}}, "p": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 2}}}}}}}}, "s": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}}}}}}}}, "v": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 2}}}}}}}, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr": {"tf": 2}, "pyerrors.correlators.Corr.item": {"tf": 1}, "pyerrors.correlators.Corr.is_matrix_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.matrix_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.thin": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.misc.gen_correlated_data": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 10, "t": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.correlators.Corr": {"tf": 1}}, "df": 1}}}}}}}, "n": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}}}, "f": {"docs": {}, "df": 0, "f": {"docs": {"pyerrors": {"tf": 2}}, "df": 1, "e": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.m_eff": {"tf": 1.7320508075688772}}, "df": 2}}}, "s": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.4142135623730951}}, "df": 3}}}}}}, "a": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "h": {"docs": {"pyerrors": {"tf": 3.605551275463989}, "pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.correlators.Corr.plottable": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.7320508075688772}, "pyerrors.input.misc.read_pbp": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_rwms": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1.4142135623730951}, "pyerrors.misc.gen_correlated_data": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.plot_tauint": {"tf": 1}, "pyerrors.obs.Obs.plot_rho": {"tf": 1}, "pyerrors.obs.Obs.plot_rep_dist": {"tf": 1}, "pyerrors.obs.Obs.plot_history": {"tf": 1}, "pyerrors.obs.Obs.plot_piechart": {"tf": 1}}, "df": 15}}, "r": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 1}}}, "s": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 2}}}}}, "l": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.item": {"tf": 1}, "pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 2.6457513110645907}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 6, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}, "s": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1.7320508075688772}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 2}}}}}}, "s": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}}, "q": {"docs": {"pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 1, "u": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "v": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1.4142135623730951}}, "df": 1, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 2, "l": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}}}}, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.correlators.Corr.thin": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 2}}}}, "g": {"docs": {"pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 2}, "p": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.dirac.epsilon_tensor": {"tf": 1}, "pyerrors.dirac.epsilon_tensor_rank4": {"tf": 1}}, "df": 2}}}}}}, "m": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}}, "df": 5}}}}, "t": {"docs": {}, "df": 0, "c": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}}, "df": 1}}}, "c": {"docs": {"pyerrors.correlators.Corr.GEVP": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.Hankel": {"tf": 3}, "pyerrors.correlators.Corr.m_eff": {"tf": 2.8284271247461903}, "pyerrors.correlators.Corr.prune": {"tf": 2}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 2}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.4142135623730951}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.7320508075688772}}, "df": 9, "o": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "p": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1}}, "df": 1, "u": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1.7320508075688772}}, "df": 1, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}}, "df": 4}}}}}, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs": {"tf": 1}}, "df": 2}, "s": {"docs": {"pyerrors.linalg.eigh": {"tf": 1}, "pyerrors.linalg.eig": {"tf": 1}, "pyerrors.linalg.pinv": {"tf": 1}, "pyerrors.linalg.svd": {"tf": 1}}, "df": 4}}}}, "l": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "x": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1.4142135623730951}, "pyerrors.linalg.matmul": {"tf": 1}, "pyerrors.linalg.jack_matmul": {"tf": 1}, "pyerrors.linalg.einsum": {"tf": 1}, "pyerrors.obs.CObs": {"tf": 1}}, "df": 6}, "t": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}}, "df": 1, "d": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}, "pyerrors.linalg.matmul": {"tf": 1}, "pyerrors.linalg.jack_matmul": {"tf": 1}}, "df": 3}, "s": {"docs": {"pyerrors.fits.residual_plot": {"tf": 1}}, "df": 1}}, "i": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.input": {"tf": 1}}, "df": 2}}}}}, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}, "c": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.input.sfcf.read_sfcf": {"tf": 1.4142135623730951}}, "df": 1}}}, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}, "i": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}}, "df": 4}}}}}, "m": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1.7320508075688772}}, "df": 1}}}, "b": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.obs.merge_obs": {"tf": 1.4142135623730951}}, "df": 1, "d": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.obs.merge_obs": {"tf": 1}}, "df": 2}}, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1.7320508075688772}}, "df": 1}}}}}}}}, "e": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "/": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "/": {"1": {"6": {"0": {"3": {"7": {"5": {"docs": {"pyerrors.dirac.epsilon_tensor": {"tf": 1}, "pyerrors.dirac.epsilon_tensor_rank4": {"tf": 1}}, "df": 2}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}}}}, "h": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}, "n": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}}, "df": 2, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.openQCD.read_rwms": {"tf": 1}}, "df": 1}}, "a": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.linalg.jack_matmul": {"tf": 1}}, "df": 1}}}}}}}, "s": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1}}, "df": 1, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1}}, "df": 1}}}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}}, "df": 2}}}}}}, "t": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}}, "df": 3, "s": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.Hankel": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 3}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 4}}, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}, "o": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": null}}, "df": 1}}}}}}, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.plateau": {"tf": 1}}, "df": 1, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}, "o": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}}, "df": 2}}}}, "t": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.obs.Obs.export_jackknife": {"tf": 1}}, "df": 1, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.__init__": {"tf": 1}, "pyerrors.correlators.Corr.item": {"tf": 1}, "pyerrors.correlators.Corr.plottable": {"tf": 1}, "pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.import_json_string": {"tf": 1.7320508075688772}, "pyerrors.input.json.load_json": {"tf": 1.4142135623730951}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1.7320508075688772}, "pyerrors.input.pandas.dump_df": {"tf": 1}, "pyerrors.misc.gen_correlated_data": {"tf": 1}, "pyerrors.obs.Obs.__init__": {"tf": 1}, "pyerrors.obs.Obs.plot_piechart": {"tf": 1}, "pyerrors.obs.import_jackknife": {"tf": 1}}, "df": 26}}}, "s": {"docs": {"pyerrors": {"tf": 5}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.misc.read_pbp": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_rwms": {"tf": 2}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}}, "df": 12}, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.input": {"tf": 1}, "pyerrors.input.utils.check_idl": {"tf": 1}}, "df": 3}}}}}, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.obs.Obs.plot_piechart": {"tf": 1}}, "df": 4}}}, "e": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}, "o": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.obs.derived_observable": {"tf": 1}}, "df": 1}}}, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.correlators.Corr.gamma_method": {"tf": 1}, "pyerrors.correlators.Corr.gm": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.obs.Obs.details": {"tf": 1}}, "df": 6, "s": {"docs": {"pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}}, "df": 7}}}}, "i": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}}, "df": 1}}}}}}}, "f": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.misc.read_pbp": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_rwms": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.extract_t0": {"tf": 2.23606797749979}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1.4142135623730951}}, "df": 5, "u": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 2.449489742783178}, "pyerrors.input.bdio.read_mesons": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.read_dSdm": {"tf": 1.4142135623730951}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 4, "s": {"docs": {"pyerrors": {"tf": 3.605551275463989}, "pyerrors.correlators.Corr.reweight": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.spaghetti_plot": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.7320508075688772}, "pyerrors.input.utils.check_idl": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.reweight": {"tf": 1.7320508075688772}, "pyerrors.obs.reweight": {"tf": 1.7320508075688772}}, "df": 19}}}}}}}}, "s": {"docs": {"pyerrors.correlators.Corr.reweight": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.reweight": {"tf": 1}, "pyerrors.obs.reweight": {"tf": 1}}, "df": 9}}}}, "d": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 2, "s": {"docs": {"pyerrors": {"tf": 2.23606797749979}}, "df": 1}, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}}, "df": 1}}}}}}}}, "j": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix.g5H": {"tf": 1}}, "df": 2}}}}}}, "c": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "r": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}}, "v": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr": {"tf": 1}}, "df": 1}}}}, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.bdio.write_ADerrors": {"tf": 1}}, "df": 1}}}}}}}}}, "r": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors": {"tf": 6.6332495807108}, "pyerrors.correlators.Corr.__init__": {"tf": 1}, "pyerrors.correlators.Corr.gamma_method": {"tf": 1}, "pyerrors.correlators.Corr.gm": {"tf": 1}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 1}, "pyerrors.correlators.Corr.reverse": {"tf": 1}, "pyerrors.correlators.Corr.correlate": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.T_symmetry": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.plateau": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.set_prange": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.dump": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1.4142135623730951}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 2}, "pyerrors.input.pandas.to_sql": {"tf": 1}, "pyerrors.input.pandas.read_sql": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 23, "e": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.correlate": {"tf": 1.4142135623730951}, "pyerrors.obs.correlate": {"tf": 1.4142135623730951}}, "df": 3, "d": {"docs": {"pyerrors": {"tf": 2.449489742783178}, "pyerrors.correlators.Corr.correlate": {"tf": 1}, "pyerrors.fits.Fit_result": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 2}, "pyerrors.fits.total_least_squares": {"tf": 1}}, "df": 5}}, "o": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors": {"tf": 3.3166247903554}, "pyerrors.correlators.Corr": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.__init__": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.correlators.Corr.item": {"tf": 1}, "pyerrors.correlators.Corr.plottable": {"tf": 1}, "pyerrors.correlators.Corr.symmetric": {"tf": 1}, "pyerrors.correlators.Corr.anti_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.is_matrix_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.matrix_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 1}, "pyerrors.correlators.Corr.roll": {"tf": 1}, "pyerrors.correlators.Corr.thin": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.correlate": {"tf": 2}, "pyerrors.correlators.Corr.reweight": {"tf": 1}, "pyerrors.correlators.Corr.T_symmetry": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.deriv": {"tf": 1}, "pyerrors.correlators.Corr.second_deriv": {"tf": 1}, "pyerrors.correlators.Corr.m_eff": {"tf": 2.23606797749979}, "pyerrors.correlators.Corr.show": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.spaghetti_plot": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 24, "s": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.show": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1.4142135623730951}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 5}}}, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.correlators.Corr.prune": {"tf": 2.23606797749979}, "pyerrors.fits.least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 2}, "pyerrors.obs.covariance": {"tf": 2.449489742783178}}, "df": 6, "s": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.thin": {"tf": 1}}, "df": 2}}}, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.obs.correlate": {"tf": 1}}, "df": 1}}}}}}, "c": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 2, "n": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 2}}}}, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}}, "df": 2}}}}, "s": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.GEVP": {"tf": 1}}, "df": 1, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.fits.residual_plot": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1.4142135623730951}}, "df": 5}}}, "s": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 5}}}}}}}, "s": {"docs": {"pyerrors.correlators.Corr.__init__": {"tf": 1}}, "df": 1}}}, "b": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 3.605551275463989}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1.4142135623730951}, "pyerrors.linalg.inv": {"tf": 1}}, "df": 6}}, "v": {"docs": {"pyerrors": {"tf": 2}, "pyerrors.covobs.Covobs.__init__": {"tf": 2}, "pyerrors.misc.gen_correlated_data": {"tf": 1}, "pyerrors.obs.cov_Obs": {"tf": 1.4142135623730951}}, "df": 4, "o": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 3.1622776601683795}, "pyerrors.covobs.Covobs.__init__": {"tf": 1.4142135623730951}, "pyerrors.covobs.Covobs.errsq": {"tf": 1}, "pyerrors.obs.cov_Obs": {"tf": 1}}, "df": 4}}}, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 4}, "pyerrors.covobs.Covobs.__init__": {"tf": 1.4142135623730951}, "pyerrors.fits.least_squares": {"tf": 2}, "pyerrors.misc.gen_correlated_data": {"tf": 1.4142135623730951}, "pyerrors.obs.covariance": {"tf": 2.449489742783178}, "pyerrors.obs.cov_Obs": {"tf": 2}}, "df": 6}}}}}}}}, "l": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}}}}}, "u": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1, "s": {"docs": {"pyerrors.input.pandas.to_sql": {"tf": 1}, "pyerrors.input.pandas.read_sql": {"tf": 1}, "pyerrors.input.pandas.dump_df": {"tf": 1}}, "df": 3}}}}}, "s": {"docs": {}, "df": 0, "h": {"docs": {"pyerrors.correlators.Corr.m_eff": {"tf": 1.7320508075688772}}, "df": 1}, "t": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}}, "df": 1}}, "u": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.openQCD.read_gf_coupling": {"tf": 1.7320508075688772}}, "df": 1}}}}}, "l": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 1}}}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}}, "df": 2}}}}}}}}, "h": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1, "s": {"docs": {"pyerrors": {"tf": 3.1622776601683795}, "pyerrors.obs.Obs.plot_history": {"tf": 1}}, "df": 2}, "/": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}}}}}}, "n": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}}, "df": 4}}}}}, "r": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.openQCD.read_qtop": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.qtop_projection": {"tf": 1.4142135623730951}}, "df": 3}}}}, "o": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "k": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.linalg.cholesky": {"tf": 1}}, "df": 3}}}}, "s": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "k": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}, "s": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.dump": {"tf": 1}, "pyerrors.obs.Obs.dump": {"tf": 1}}, "df": 4}}}, "i": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.deriv": {"tf": 1}, "pyerrors.correlators.Corr.second_deriv": {"tf": 1}}, "df": 2}}}, "o": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 1}}}, "e": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}}, "df": 1}}}}, "e": {"docs": {"pyerrors.input.utils.check_idl": {"tf": 1}}, "df": 1, "c": {"docs": {}, "df": 0, "k": {"docs": {"pyerrors": {"tf": 2}, "pyerrors.fits.qqplot": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.obs.Obs.is_zero_within_error": {"tf": 1}}, "df": 4, "s": {"docs": {"pyerrors.correlators.Corr.is_matrix_symmetric": {"tf": 1}, "pyerrors.input.utils.check_idl": {"tf": 1}, "pyerrors.obs.Obs.is_zero_within_error": {"tf": 1}, "pyerrors.obs.Obs.is_zero": {"tf": 1}, "pyerrors.obs.CObs.is_zero": {"tf": 1}}, "df": 5}, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.utils.check_idl": {"tf": 1}}, "df": 1}}}}}, "i": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "q": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.fits.Fit_result": {"tf": 1.4142135623730951}, "pyerrors.fits.least_squares": {"tf": 2}, "pyerrors.fits.total_least_squares": {"tf": 1.4142135623730951}}, "df": 4}}}}}}}}, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "o": {"docs": {"pyerrors": {"tf": 3.4641016151377544}, "pyerrors.obs.Obs.__init__": {"tf": 1}, "pyerrors.obs.Obs.plot_history": {"tf": 1.4142135623730951}}, "df": 3}}, "e": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 2, "f": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 2}}}}}, "l": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1}}, "df": 2}}, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}}, "df": 2}, "s": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}}}}}, "l": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}}}}, "n": {"docs": {"pyerrors": {"tf": 5.744562646538029}, "pyerrors.correlators.Corr": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.correlate": {"tf": 1}, "pyerrors.correlators.Corr.T_symmetry": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 2.23606797749979}, "pyerrors.fits.total_least_squares": {"tf": 1.7320508075688772}, "pyerrors.fits.fit_lin": {"tf": 1}, "pyerrors.fits.qqplot": {"tf": 1}, "pyerrors.input": {"tf": 1}, "pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.read_dSdm": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.linalg.matmul": {"tf": 1}, "pyerrors.linalg.jack_matmul": {"tf": 1}, "pyerrors.linalg.einsum": {"tf": 1}, "pyerrors.misc.gen_correlated_data": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1.4142135623730951}, "pyerrors.obs.derived_observable": {"tf": 1.7320508075688772}}, "df": 29, "n": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 1}}}}, "s": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.correlators.Corr.prune": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.fit_lin": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}}, "df": 10}}, "u": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.correlators.Corr.fit": {"tf": 1}}, "df": 1}, "u": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.obs.derived_observable": {"tf": 1}}, "df": 1}}}}}}}, "s": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}}, "df": 2}}}}}, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}}, "df": 3}}}}}}, "o": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}}, "df": 2, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1.4142135623730951}}, "df": 1}}}}}}, "e": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}}, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}}, "o": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "e": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.obs.cov_Obs": {"tf": 1}}, "df": 3, "d": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}}}}}, "l": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 3.1622776601683795}, "pyerrors.correlators.Corr": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1}, "pyerrors.obs.CObs": {"tf": 1}}, "df": 5}}}, "o": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 1}}}}}, "u": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.utils.check_idl": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1}}, "df": 3, "l": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.obs.correlate": {"tf": 1}}, "df": 2}}}}}}}, "s": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "m": {"docs": {"pyerrors.correlators.Corr.dump": {"tf": 1}, "pyerrors.misc.dump_object": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}, "pyerrors.obs.Obs.dump": {"tf": 1}}, "df": 5}}}}}, "d": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "a": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}}}}, "c": {"docs": {"pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}}, "df": 4}, "p": {"docs": {"pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}}, "df": 4}, "t": {"docs": {}, "df": 0, "y": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1.7320508075688772}}, "df": 1}}}}}, "n": {"docs": {}, "df": 0, "f": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.4142135623730951}}, "df": 4}}}, "y": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 3}}}}}, "s": {"docs": {}, "df": 0, "v": {"docs": {"pyerrors.input.pandas.dump_df": {"tf": 2}, "pyerrors.input.pandas.load_df": {"tf": 1}}, "df": 2}}, "f": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 1}}, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.misc.pseudo_Obs": {"tf": 1}}, "df": 1}}}}}}}, "o": {"docs": {"pyerrors": {"tf": 2}, "pyerrors.correlators.Corr.prune": {"tf": 1}, "pyerrors.input.bdio.read_ADerrors": {"tf": 1.7320508075688772}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1.7320508075688772}, "pyerrors.input.bdio.read_mesons": {"tf": 1.7320508075688772}, "pyerrors.input.bdio.read_dSdm": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1}}, "df": 7, "f": {"docs": {"pyerrors": {"tf": 10.344080432788601}, "pyerrors.correlators.Corr": {"tf": 2}, "pyerrors.correlators.Corr.__init__": {"tf": 2.449489742783178}, "pyerrors.correlators.Corr.gamma_method": {"tf": 1}, "pyerrors.correlators.Corr.gm": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.Hankel": {"tf": 1}, "pyerrors.correlators.Corr.roll": {"tf": 1}, "pyerrors.correlators.Corr.reverse": {"tf": 1}, "pyerrors.correlators.Corr.thin": {"tf": 1}, "pyerrors.correlators.Corr.correlate": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.reweight": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.T_symmetry": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.deriv": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.second_deriv": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.m_eff": {"tf": 2}, "pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.correlators.Corr.set_prange": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 2.8284271247461903}, "pyerrors.correlators.Corr.spaghetti_plot": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.dump": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.prune": {"tf": 2.6457513110645907}, "pyerrors.covobs.Covobs.__init__": {"tf": 1.7320508075688772}, "pyerrors.covobs.Covobs.errsq": {"tf": 1.4142135623730951}, "pyerrors.dirac.epsilon_tensor_rank4": {"tf": 1}, "pyerrors.fits.Fit_result": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 3.3166247903554}, "pyerrors.fits.total_least_squares": {"tf": 3.1622776601683795}, "pyerrors.fits.fit_lin": {"tf": 2.23606797749979}, "pyerrors.fits.qqplot": {"tf": 1.4142135623730951}, "pyerrors.fits.error_band": {"tf": 1}, "pyerrors.fits.ks_test": {"tf": 1.4142135623730951}, "pyerrors.input": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 2.23606797749979}, "pyerrors.input.dobs.write_pobs": {"tf": 2.449489742783178}, "pyerrors.input.dobs.read_pobs": {"tf": 2}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.read_dobs": {"tf": 2}, "pyerrors.input.dobs.create_dobs_string": {"tf": 2.8284271247461903}, "pyerrors.input.dobs.write_dobs": {"tf": 2.8284271247461903}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 2.449489742783178}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1.7320508075688772}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 5.0990195135927845}, "pyerrors.input.hadrons.Npr_matrix.g5H": {"tf": 1}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1.7320508075688772}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1.7320508075688772}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1.7320508075688772}, "pyerrors.input.json.create_json_string": {"tf": 2.6457513110645907}, "pyerrors.input.json.dump_to_json": {"tf": 2.6457513110645907}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1.4142135623730951}, "pyerrors.input.json.dump_dict_to_json": {"tf": 2.8284271247461903}, "pyerrors.input.json.load_json_dict": {"tf": 1.7320508075688772}, "pyerrors.input.misc.read_pbp": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 2.449489742783178}, "pyerrors.input.openQCD.extract_t0": {"tf": 3.1622776601683795}, "pyerrors.input.openQCD.read_qtop": {"tf": 2.6457513110645907}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 2.6457513110645907}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 2.8284271247461903}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1.7320508075688772}, "pyerrors.input.pandas.to_sql": {"tf": 1}, "pyerrors.input.pandas.dump_df": {"tf": 1.7320508075688772}, "pyerrors.input.pandas.load_df": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 3.872983346207417}, "pyerrors.input.utils.check_idl": {"tf": 1.7320508075688772}, "pyerrors.linalg.matmul": {"tf": 1}, "pyerrors.linalg.jack_matmul": {"tf": 1.4142135623730951}, "pyerrors.linalg.einsum": {"tf": 1}, "pyerrors.linalg.inv": {"tf": 1}, "pyerrors.linalg.cholesky": {"tf": 1}, "pyerrors.linalg.det": {"tf": 1}, "pyerrors.linalg.eigh": {"tf": 1.4142135623730951}, "pyerrors.linalg.eig": {"tf": 1.4142135623730951}, "pyerrors.linalg.pinv": {"tf": 1.4142135623730951}, "pyerrors.linalg.svd": {"tf": 1.4142135623730951}, "pyerrors.misc.dump_object": {"tf": 1}, "pyerrors.misc.pseudo_Obs": {"tf": 2}, "pyerrors.misc.gen_correlated_data": {"tf": 1.4142135623730951}, "pyerrors.mpm.matrix_pencil_method": {"tf": 2.6457513110645907}, "pyerrors.obs.Obs": {"tf": 2.8284271247461903}, "pyerrors.obs.Obs.__init__": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.gamma_method": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.gm": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.details": {"tf": 1}, "pyerrors.obs.Obs.reweight": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.is_zero_within_error": {"tf": 1}, "pyerrors.obs.Obs.plot_piechart": {"tf": 1}, "pyerrors.obs.Obs.dump": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.export_jackknife": {"tf": 2}, "pyerrors.obs.derived_observable": {"tf": 2.449489742783178}, "pyerrors.obs.reweight": {"tf": 2}, "pyerrors.obs.covariance": {"tf": 3.3166247903554}, "pyerrors.obs.import_jackknife": {"tf": 1}, "pyerrors.obs.merge_obs": {"tf": 1.7320508075688772}, "pyerrors.obs.cov_Obs": {"tf": 1.7320508075688772}, "pyerrors.roots.find_root": {"tf": 1.4142135623730951}}, "df": 99, "f": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}, "s": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.thin": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 2.23606797749979}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 4}}}}, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.correlators.Corr": {"tf": 1}}, "df": 1}}}}, "n": {"docs": {"pyerrors": {"tf": 5.291502622129181}, "pyerrors.correlators.Corr.plottable": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.is_matrix_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.matrix_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.reweight": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1.4142135623730951}, "pyerrors.dirac.epsilon_tensor": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.pandas.read_sql": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1}, "pyerrors.obs.Obs.__init__": {"tf": 1}, "pyerrors.obs.Obs.reweight": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1}, "pyerrors.obs.reweight": {"tf": 1.7320508075688772}, "pyerrors.obs.covariance": {"tf": 1.4142135623730951}, "pyerrors.obs.import_jackknife": {"tf": 1}, "pyerrors.obs.merge_obs": {"tf": 1}, "pyerrors.obs.cov_Obs": {"tf": 1}}, "df": 32, "e": {"docs": {"pyerrors": {"tf": 3}, "pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.correlators.Corr.item": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.obs.Obs.plot_rep_dist": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}, "pyerrors.obs.merge_obs": {"tf": 1}}, "df": 20, "s": {"docs": {"pyerrors.correlators.Corr.Eigenvalue": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 3}}, "l": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.GEVP": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.thin": {"tf": 1}, "pyerrors.correlators.Corr.m_eff": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}, "pyerrors.input.json.import_json_string": {"tf": 1.4142135623730951}, "pyerrors.input.json.load_json": {"tf": 1.4142135623730951}, "pyerrors.input.json.load_json_dict": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.4142135623730951}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.is_zero_within_error": {"tf": 1}, "pyerrors.obs.Obs.dump": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1}, "pyerrors.obs.correlate": {"tf": 1.4142135623730951}}, "df": 26}}, "t": {"docs": {}, "df": 0, "o": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 1}}, "c": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 1}}}, "p": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.correlators.Corr": {"tf": 1}}, "df": 1, "s": {"docs": {"pyerrors": {"tf": 2.449489742783178}, "pyerrors.obs.Obs": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}}, "df": 3}}}}, "o": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.linalg.matmul": {"tf": 1}}, "df": 1, "s": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 1}, "n": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "{": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "v": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}, "v": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}}}}}}}}}, "n": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.linalg.matmul": {"tf": 1.4142135623730951}, "pyerrors.linalg.jack_matmul": {"tf": 1.4142135623730951}, "pyerrors.linalg.einsum": {"tf": 1}}, "df": 3}}}}}, "n": {"docs": {}, "df": 0, "q": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.openQCD.read_rwms": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 2}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.4142135623730951}}, "df": 5}}}}}, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.4142135623730951}}, "df": 2, "s": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.input.pandas.to_sql": {"tf": 1}}, "df": 2}, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.__init__": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.Hankel": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1.7320508075688772}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.fits.ks_test": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 2.23606797749979}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}, "pyerrors.obs.Obs.__init__": {"tf": 1}}, "df": 20}}}}, "m": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "z": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}}, "df": 1, "d": {"docs": {"pyerrors.fits.error_band": {"tf": 1}}, "df": 1}}}}, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 1}}}}}, "p": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.fit": {"tf": 1}}, "df": 1}}}}}}, "b": {"docs": {}, "df": 0, "s": {"1": {"docs": {"pyerrors": {"tf": 3}, "pyerrors.obs.derived_observable": {"tf": 1.4142135623730951}, "pyerrors.obs.reweight": {"tf": 1}}, "df": 3}, "2": {"docs": {"pyerrors": {"tf": 3}, "pyerrors.obs.derived_observable": {"tf": 1.4142135623730951}, "pyerrors.obs.reweight": {"tf": 1}}, "df": 3}, "3": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.obs.derived_observable": {"tf": 1}, "pyerrors.obs.reweight": {"tf": 1}}, "df": 3}, "docs": {"pyerrors": {"tf": 9.591663046625438}, "pyerrors.correlators.Corr": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.__init__": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.item": {"tf": 1}, "pyerrors.correlators.Corr.correlate": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.reweight": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1.7320508075688772}, "pyerrors.fits.total_least_squares": {"tf": 2.23606797749979}, "pyerrors.fits.fit_lin": {"tf": 2.23606797749979}, "pyerrors.input": {"tf": 2.23606797749979}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 2}, "pyerrors.input.dobs.write_pobs": {"tf": 2}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.read_dobs": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.create_dobs_string": {"tf": 2}, "pyerrors.input.dobs.write_dobs": {"tf": 2}, "pyerrors.input.json.create_json_string": {"tf": 2}, "pyerrors.input.json.dump_to_json": {"tf": 2}, "pyerrors.input.json.import_json_string": {"tf": 1.7320508075688772}, "pyerrors.input.json.load_json": {"tf": 1.7320508075688772}, "pyerrors.input.json.dump_dict_to_json": {"tf": 2}, "pyerrors.input.json.load_json_dict": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.openQCD.qtop_projection": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.pandas.to_sql": {"tf": 1}, "pyerrors.input.pandas.read_sql": {"tf": 1.4142135623730951}, "pyerrors.input.pandas.dump_df": {"tf": 1.4142135623730951}, "pyerrors.input.pandas.load_df": {"tf": 1.4142135623730951}, "pyerrors.linalg.matmul": {"tf": 1}, "pyerrors.linalg.jack_matmul": {"tf": 1}, "pyerrors.linalg.einsum": {"tf": 1}, "pyerrors.linalg.inv": {"tf": 1}, "pyerrors.linalg.cholesky": {"tf": 1}, "pyerrors.linalg.det": {"tf": 1}, "pyerrors.linalg.eigh": {"tf": 1}, "pyerrors.linalg.eig": {"tf": 1}, "pyerrors.linalg.pinv": {"tf": 1}, "pyerrors.linalg.svd": {"tf": 1}, "pyerrors.misc.pseudo_Obs": {"tf": 2.23606797749979}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.__init__": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}, "pyerrors.obs.Obs.details": {"tf": 1}, "pyerrors.obs.Obs.reweight": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.dump": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1.7320508075688772}, "pyerrors.obs.derived_observable": {"tf": 1.7320508075688772}, "pyerrors.obs.reweight": {"tf": 1.7320508075688772}, "pyerrors.obs.correlate": {"tf": 2}, "pyerrors.obs.covariance": {"tf": 1.4142135623730951}, "pyerrors.obs.import_jackknife": {"tf": 1}, "pyerrors.obs.merge_obs": {"tf": 2}, "pyerrors.obs.cov_Obs": {"tf": 1.4142135623730951}, "pyerrors.roots.find_root": {"tf": 2.23606797749979}}, "df": 61, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 3.4641016151377544}, "pyerrors.correlators.Corr.reweight": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.misc.gen_correlated_data": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs": {"tf": 1}, "pyerrors.obs.Obs.reweight": {"tf": 1}, "pyerrors.obs.Obs.is_zero_within_error": {"tf": 1}, "pyerrors.obs.Obs.is_zero": {"tf": 1}, "pyerrors.obs.CObs": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}, "pyerrors.obs.reweight": {"tf": 1}, "pyerrors.obs.correlate": {"tf": 1.4142135623730951}, "pyerrors.obs.merge_obs": {"tf": 1}}, "df": 17, "s": {"docs": {"pyerrors": {"tf": 3}, "pyerrors.correlators.Corr.reweight": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.misc.gen_correlated_data": {"tf": 1}, "pyerrors.obs.Obs.reweight": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}, "pyerrors.obs.reweight": {"tf": 1.4142135623730951}, "pyerrors.obs.correlate": {"tf": 1.7320508075688772}, "pyerrors.obs.covariance": {"tf": 2.449489742783178}, "pyerrors.obs.merge_obs": {"tf": 1}}, "df": 14}}}}}}}}, "d": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "a": {"docs": {"pyerrors": {"tf": 1.7320508075688772}}, "df": 1}}}}, "[": {"docs": {}, "df": 0, "i": {"docs": {"pyerrors.correlators.Corr.reweight": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.reweight": {"tf": 1.4142135623730951}, "pyerrors.obs.reweight": {"tf": 1.4142135623730951}}, "df": 3}}, "l": {"docs": {"pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}}, "df": 4}}, "j": {"docs": {"pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.misc.dump_object": {"tf": 1}}, "df": 2, "e": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.correlators.Corr.__init__": {"tf": 1}, "pyerrors.correlators.Corr.m_eff": {"tf": 1}, "pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.correlators.Corr.set_prange": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.fits.ks_test": {"tf": 1}, "pyerrors.input": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 2.8284271247461903}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1.7320508075688772}, "pyerrors.misc.dump_object": {"tf": 1.7320508075688772}, "pyerrors.misc.load_object": {"tf": 1}, "pyerrors.misc.pseudo_Obs": {"tf": 1}, "pyerrors.obs.Obs.__init__": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}, "pyerrors.obs.merge_obs": {"tf": 1}, "pyerrors.roots.find_root": {"tf": 1}}, "df": 21, "s": {"docs": {"pyerrors": {"tf": 3.605551275463989}, "pyerrors.fits.ks_test": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1.4142135623730951}, "pyerrors.input.json.dump_to_json": {"tf": 1.4142135623730951}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1.4142135623730951}, "pyerrors.input.pandas.read_sql": {"tf": 1}, "pyerrors.input.pandas.dump_df": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs": {"tf": 1}}, "df": 9}}}}}, "t": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.pandas.read_sql": {"tf": 1}}, "df": 1, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}}}}}}}, "r": {"docs": {"pyerrors": {"tf": 4.123105625617661}, "pyerrors.correlators.Corr": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.__init__": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.correlators.Corr.correlate": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.T_symmetry": {"tf": 1}, "pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.spaghetti_plot": {"tf": 1}, "pyerrors.covobs.Covobs.__init__": {"tf": 1.7320508075688772}, "pyerrors.fits.least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.fits.fit_lin": {"tf": 1}, "pyerrors.input": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.write_pobs": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.read_pobs": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.read_dobs": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.write_dobs": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.7320508075688772}, "pyerrors.input.json.create_json_string": {"tf": 1.4142135623730951}, "pyerrors.input.json.dump_to_json": {"tf": 1.4142135623730951}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1.4142135623730951}, "pyerrors.input.json.load_json_dict": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}, "pyerrors.input.pandas.to_sql": {"tf": 1}, "pyerrors.input.pandas.read_sql": {"tf": 1}, "pyerrors.input.utils.check_idl": {"tf": 1}, "pyerrors.linalg.matmul": {"tf": 1}, "pyerrors.linalg.jack_matmul": {"tf": 1}, "pyerrors.linalg.einsum": {"tf": 1}, "pyerrors.linalg.inv": {"tf": 1}, "pyerrors.misc.gen_correlated_data": {"tf": 1.4142135623730951}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}, "pyerrors.obs.Obs.__init__": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1.7320508075688772}, "pyerrors.obs.cov_Obs": {"tf": 2}}, "df": 44, "d": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 2.6457513110645907}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 5, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.reverse": {"tf": 1}}, "df": 2}}}, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.Eigenvalue": {"tf": 1}}, "df": 1}}}}}, "t": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}}, "df": 2}}}}}}}}, "i": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.dobs.create_pobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.write_pobs": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.write_dobs": {"tf": 1.4142135623730951}}, "df": 4}}}}}, "t": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors": {"tf": 2}, "pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 1}, "pyerrors.input": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 6, "w": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}}, "df": 4}}}}}}}}, "v": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}}, "df": 4, "l": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}}}}}}, "w": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}}, "df": 2, "s": {"docs": {"pyerrors.obs.Obs": {"tf": 1.7320508075688772}}, "df": 1}}}}}}}}}, "d": {"docs": {"pyerrors.input.json.dump_dict_to_json": {"tf": 1}}, "df": 1, "d": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}}, "u": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.thin": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix.g5H": {"tf": 1}}, "df": 3, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}, "p": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.fit": {"tf": 1.4142135623730951}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.input": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1.7320508075688772}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1.7320508075688772}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1.7320508075688772}, "pyerrors.input.json.load_json_dict": {"tf": 1}, "pyerrors.input.pandas.dump_df": {"tf": 1.7320508075688772}, "pyerrors.input.sfcf.read_sfcf": {"tf": 2}, "pyerrors.obs.Obs": {"tf": 1}, "pyerrors.obs.Obs.details": {"tf": 1}, "pyerrors.obs.Obs.dump": {"tf": 1}}, "df": 23, "s": {"docs": {"pyerrors.correlators.Corr.plottable": {"tf": 1.4142135623730951}}, "df": 1}}}}}, "r": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}, "m": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}}, "df": 2}}}}}}, "c": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}}, "df": 3}}}}}}, "p": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}}}}}}, "w": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "a": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}}}}}, "l": {"docs": {"pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}}, "df": 2}}, "m": {"docs": {"pyerrors": {"tf": 3.872983346207417}, "pyerrors.correlators.Corr.m_eff": {"tf": 2.449489742783178}, "pyerrors.fits.fit_lin": {"tf": 1.4142135623730951}}, "df": 3, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "k": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "v": {"docs": {"pyerrors": {"tf": 2}}, "df": 1}}}, "q": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}}, "df": 1, "t": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}}}}}}}, "t": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "x": {"docs": {"pyerrors": {"tf": 4.795831523312719}, "pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.correlators.Corr.item": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.Hankel": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.prune": {"tf": 2.8284271247461903}, "pyerrors.covobs.Covobs.__init__": {"tf": 1.4142135623730951}, "pyerrors.dirac.Grid_gamma": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 2.23606797749979}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.linalg.matmul": {"tf": 1}, "pyerrors.linalg.jack_matmul": {"tf": 1}, "pyerrors.linalg.eigh": {"tf": 1}, "pyerrors.linalg.eig": {"tf": 1}, "pyerrors.linalg.pinv": {"tf": 1}, "pyerrors.linalg.svd": {"tf": 1}, "pyerrors.misc.gen_correlated_data": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1.7320508075688772}, "pyerrors.obs.covariance": {"tf": 3}, "pyerrors.obs.cov_Obs": {"tf": 1.7320508075688772}}, "df": 20}, "c": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.correlators.Corr.is_matrix_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.matrix_symmetric": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.linalg.jack_matmul": {"tf": 1}, "pyerrors.linalg.inv": {"tf": 1}, "pyerrors.linalg.cholesky": {"tf": 1}, "pyerrors.linalg.det": {"tf": 1}}, "df": 9}}}}}, "h": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors": {"tf": 2}, "pyerrors.obs.Obs": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}}, "df": 3}}}}}}}}, "r": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "{": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.obs.covariance": {"tf": 1.7320508075688772}}, "df": 2}}}, "e": {"docs": {}, "df": 0, "x": {"docs": {}, "df": 0, "p": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}, "p": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "j": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1.4142135623730951}}, "df": 1}}}}, "t": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "c": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 1}}}}}}}}}, "c": {"docs": {}, "df": 0, "h": {"docs": {"pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1.4142135623730951}}, "df": 2, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}}, "df": 1}}}}}, "m": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.linalg.jack_matmul": {"tf": 1}}, "df": 1}}}}, "c": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 2}, "pyerrors.obs.CObs.is_zero": {"tf": 1}}, "df": 2}}}}}, "d": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}, "k": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.pandas.dump_df": {"tf": 1}, "pyerrors.linalg.jack_matmul": {"tf": 1}}, "df": 3}}}}, "n": {"docs": {"pyerrors.obs.derived_observable": {"tf": 1}}, "df": 1, "i": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}}}}, "y": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 5}, "u": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}}, "df": 3}}}}}}, "s": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 2}, "pyerrors.correlators.Corr.m_eff": {"tf": 1.7320508075688772}}, "df": 2, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}}}}, "y": {"docs": {"pyerrors": {"tf": 2.6457513110645907}, "pyerrors.correlators.Corr.prune": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 8}, "j": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}}, "df": 2}}}, "x": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.obs.derived_observable": {"tf": 1}}, "df": 1}}}}}}}}}}}}}}}, "o": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 3.4641016151377544}, "pyerrors.obs.Obs.__init__": {"tf": 1}, "pyerrors.obs.Obs.plot_history": {"tf": 1.4142135623730951}}, "df": 3}}, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.correlators.Corr.spaghetti_plot": {"tf": 1}}, "df": 1, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}}, "d": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}}, "df": 1, "s": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 2}}, "u": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 2}}}}, "r": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.plot_rep_dist": {"tf": 1}}, "df": 6}}, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.linalg.pinv": {"tf": 1}}, "df": 2}}}, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 1}}}}}}}, "m": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}}, "df": 4, "a": {"docs": {"pyerrors.input.hadrons.Npr_matrix.g5H": {"tf": 1}}, "df": 1}}}}}, "v": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1.7320508075688772}}, "df": 1}}}, "e": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 4.69041575982343}, "pyerrors.correlators.Corr.gamma_method": {"tf": 1}, "pyerrors.correlators.Corr.gm": {"tf": 1}, "pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.plateau": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1.4142135623730951}, "pyerrors.fits.Fit_result.gamma_method": {"tf": 1}, "pyerrors.fits.Fit_result.gm": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 2}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 2}, "pyerrors.input.pandas.read_sql": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs": {"tf": 1}, "pyerrors.obs.Obs.is_zero_within_error": {"tf": 1}, "pyerrors.obs.CObs.gamma_method": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1.4142135623730951}}, "df": 19, "s": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1}}, "df": 4}}}}}, "s": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.hadrons.read_meson_hd5": {"tf": 2.8284271247461903}}, "df": 1, "s": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1.4142135623730951}}, "df": 3}}}}, "a": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.covobs.Covobs.__init__": {"tf": 1.7320508075688772}, "pyerrors.misc.gen_correlated_data": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}, "pyerrors.obs.import_jackknife": {"tf": 1}, "pyerrors.obs.cov_Obs": {"tf": 1.7320508075688772}}, "df": 8, "s": {"docs": {"pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.misc.gen_correlated_data": {"tf": 1}, "pyerrors.obs.cov_Obs": {"tf": 1}}, "df": 4}}, "s": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.4142135623730951}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 5, "s": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop": {"tf": 2}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 2}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.4142135623730951}}, "df": 8}}}}}}}}}, "d": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}}, "df": 1}}, "m": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 2.8284271247461903}}, "df": 1}}}}}, "y": {"docs": {"pyerrors": {"tf": 7.681145747868608}}, "df": 1}, "u": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {"pyerrors": {"tf": 1.7320508075688772}}, "df": 1, "p": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 4}, "y": {"docs": {"pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.linalg.matmul": {"tf": 1}, "pyerrors.linalg.jack_matmul": {"tf": 1}}, "df": 3}, "i": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.linalg.matmul": {"tf": 1}}, "df": 1}}}}}}}}}, "d": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}}}}}}}}}}}}}, "c": {"docs": {}, "df": 0, "h": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}, "i": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.projected": {"tf": 1}}, "df": 2}}, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.fits.least_squares": {"tf": 1.4142135623730951}}, "df": 1}}}}, "n": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "z": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.roots.find_root": {"tf": 1}}, "df": 2}}}}}, "e": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}}, "df": 1, "d": {"docs": {"pyerrors.roots.find_root": {"tf": 1}}, "df": 1}}}}}}, "d": {"docs": {"pyerrors.obs.correlate": {"tf": 1}}, "df": 1}, "u": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}}, "\\": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1.7320508075688772}}, "df": 1}}}}}}, "p": {"docs": {}, "df": 0, "i": {"docs": {"pyerrors": {"tf": 2}}, "df": 1}}, "c": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1, "m": {"docs": {}, "df": 0, "c": {"docs": {"pyerrors.input.bdio.read_ADerrors": {"tf": 1}}, "df": 1}}}, "t": {"docs": {}, "df": 0, "h": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "s": {"1": {"docs": {"pyerrors.input.openQCD.read_rwms": {"tf": 1}}, "df": 1}, "docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 2}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 4}}, "d": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.roots.find_root": {"tf": 2.23606797749979}}, "df": 2, "a": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 4, "a": {"docs": {"pyerrors": {"tf": 4.898979485566356}, "pyerrors.correlators.Corr.__init__": {"tf": 1}, "pyerrors.correlators.Corr.fit": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.fits.residual_plot": {"tf": 1}, "pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.read_dobs": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.write_dobs": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 3.1622776601683795}, "pyerrors.input.json.import_json_string": {"tf": 1.7320508075688772}, "pyerrors.input.json.load_json": {"tf": 1.7320508075688772}, "pyerrors.input.json.load_json_dict": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_rwms": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.extract_t0": {"tf": 2}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 2.449489742783178}, "pyerrors.input.pandas.load_df": {"tf": 1}, "pyerrors.misc.gen_correlated_data": {"tf": 1.4142135623730951}, "pyerrors.mpm.matrix_pencil_method": {"tf": 2.449489742783178}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1.7320508075688772}}, "df": 31, "t": {"docs": {}, "df": 0, "y": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.dump": {"tf": 1}, "pyerrors.obs.Obs.dump": {"tf": 1}}, "df": 3}}}}, "s": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.misc.gen_correlated_data": {"tf": 1}}, "df": 2}}}, "f": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.pandas.to_sql": {"tf": 1.7320508075688772}, "pyerrors.input.pandas.read_sql": {"tf": 1}, "pyerrors.input.pandas.dump_df": {"tf": 1.7320508075688772}, "pyerrors.input.pandas.load_df": {"tf": 1}}, "df": 4}}}}}, "b": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.pandas.to_sql": {"tf": 2}, "pyerrors.input.pandas.read_sql": {"tf": 1.4142135623730951}}, "df": 2}}}}}, "e": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}}}, "i": {"docs": {}, "df": 0, "f": {"docs": {}, "df": 0, "f": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 2.8284271247461903}, "pyerrors.obs.Obs": {"tf": 1}}, "df": 2, "i": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 2.6457513110645907}, "pyerrors.fits.least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.total_least_squares": {"tf": 1.4142135623730951}, "pyerrors.obs.derived_observable": {"tf": 1.7320508075688772}}, "df": 4}}}}}}, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}}, "df": 2}}}}}}, "c": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 2, "s": {"docs": {"pyerrors.correlators.Corr.deriv": {"tf": 1}, "pyerrors.correlators.Corr.second_deriv": {"tf": 1}}, "df": 2}}}}}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}}}}}, "c": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 2}, "pyerrors.input.json.load_json_dict": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs": {"tf": 2.449489742783178}}, "df": 12, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1.7320508075688772}}, "df": 1}}}, "y": {"docs": {"pyerrors.input.bdio.read_mesons": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.read_dSdm": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.plot_piechart": {"tf": 1}}, "df": 5}}}}}}, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}, "s": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "z": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}}}}}}}}}}}, "o": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}}, "u": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 1}}}}}}}}, "t": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 2}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 2}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 2}}, "df": 6}}}}, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.fits.qqplot": {"tf": 1}}, "df": 1}}, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.obs.Obs.plot_rep_dist": {"tf": 1}}, "df": 1}}}}}}}}, "i": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}}, "df": 1}}}}}}}}}}}}}}}}}}}}, "p": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "y": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1}}, "df": 1}}, "s": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.residual_plot": {"tf": 1}}, "df": 2}}}}}, "k": {"docs": {"pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}}, "df": 3}}, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}}, "df": 2, "l": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}}, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1.7320508075688772}}, "df": 2}, "i": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}}, "df": 1}}}}}}}}}, "d": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "m": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.Hankel": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 3, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors": {"tf": 2}, "pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 3}}, "s": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 2}}}}}}}}, "a": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1.4142135623730951}, "pyerrors.obs.cov_Obs": {"tf": 1}}, "df": 4}}}}, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1.4142135623730951}}, "df": 1}}}}}}, "v": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 1}}}}}}, "e": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 2}, "pyerrors.linalg.cholesky": {"tf": 1}, "pyerrors.linalg.svd": {"tf": 1}}, "df": 3}}}}}, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}}, "df": 1}}}}}}}, "i": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.correlators.Corr.deriv": {"tf": 1}, "pyerrors.correlators.Corr.second_deriv": {"tf": 1}, "pyerrors.correlators.Corr.fit": {"tf": 1}}, "df": 3}}}}, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 1}}}}}}, "a": {"docs": {}, "df": 0, "y": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1.4142135623730951}}, "df": 1}}}}}, "t": {"docs": {"pyerrors": {"tf": 1}}, "df": 1, "a": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 2}, "pyerrors.obs.Obs.details": {"tf": 1}}, "df": 2}}, "s": {"docs": {"pyerrors": {"tf": 3.3166247903554}, "pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.input": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.linalg.einsum": {"tf": 1}, "pyerrors.obs.Obs.details": {"tf": 1}, "pyerrors.obs.Obs.is_zero": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 9}}}}, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 1, "d": {"docs": {"pyerrors": {"tf": 1.7320508075688772}}, "df": 1}, "s": {"docs": {"pyerrors.correlators.Corr.Eigenvalue": {"tf": 1}, "pyerrors.correlators.Corr.Hankel": {"tf": 1}, "pyerrors.correlators.Corr.spaghetti_plot": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}}, "df": 6}}, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.linalg.det": {"tf": 1}}, "df": 2}}}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1}}, "df": 1}}}}}}}, "c": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 2}}}}}}}, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "v": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 2.8284271247461903}, "pyerrors.obs.Obs.plot_history": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1.4142135623730951}, "pyerrors.obs.derived_observable": {"tf": 1.4142135623730951}}, "df": 4}}, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.correlators.Corr.deriv": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.second_deriv": {"tf": 1.4142135623730951}, "pyerrors.obs.derived_observable": {"tf": 1}}, "df": 4, "s": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.obs.derived_observable": {"tf": 1}}, "df": 2}}}}}}}}}, "l": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "a": {"docs": {"pyerrors": {"tf": 2}}, "df": 1, "s": {"docs": {"pyerrors": {"tf": 1.7320508075688772}}, "df": 1}}}}, "s": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}, "c": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}}, "df": 4, "d": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}}, "df": 2}, "s": {"docs": {"pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}}, "df": 8}}}, "p": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.obs.Obs.dump": {"tf": 1.4142135623730951}}, "df": 5}}, "v": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}}, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.correlators.Corr.GEVP": {"tf": 1}}, "df": 1}}}}}}}}, "f": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.fits.least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.total_least_squares": {"tf": 1.4142135623730951}, "pyerrors.roots.find_root": {"tf": 1}}, "df": 4, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 2.449489742783178}, "pyerrors.correlators.Corr.deriv": {"tf": 1}, "pyerrors.correlators.Corr.second_deriv": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}}, "df": 5}}}, "e": {"docs": {"pyerrors.obs.covariance": {"tf": 1.4142135623730951}}, "df": 1}}}, "e": {"docs": {"pyerrors": {"tf": 2}}, "df": 1, "d": {"docs": {"pyerrors": {"tf": 3.3166247903554}, "pyerrors.correlators.Corr.reweight": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.openQCD.qtop_projection": {"tf": 1}, "pyerrors.obs.Obs.__init__": {"tf": 1}, "pyerrors.obs.Obs.reweight": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1}, "pyerrors.obs.reweight": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1.4142135623730951}, "pyerrors.obs.import_jackknife": {"tf": 1}}, "df": 16}, "s": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 3}}}}, "a": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.correlators.Corr.deriv": {"tf": 1}, "pyerrors.correlators.Corr.second_deriv": {"tf": 1}, "pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.correlators.Corr.dump": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1.4142135623730951}, "pyerrors.fits.least_squares": {"tf": 2.23606797749979}, "pyerrors.fits.total_least_squares": {"tf": 1.7320508075688772}, "pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 2}, "pyerrors.input.bdio.read_dSdm": {"tf": 2}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.qtop_projection": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.4142135623730951}, "pyerrors.input.pandas.read_sql": {"tf": 1.4142135623730951}, "pyerrors.input.pandas.load_df": {"tf": 1.4142135623730951}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.7320508075688772}, "pyerrors.misc.dump_object": {"tf": 1}, "pyerrors.misc.pseudo_Obs": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.gamma_method": {"tf": 2}, "pyerrors.obs.Obs.gm": {"tf": 2}, "pyerrors.obs.Obs.reweight": {"tf": 1}, "pyerrors.obs.Obs.plot_history": {"tf": 1}, "pyerrors.obs.Obs.dump": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}, "pyerrors.obs.reweight": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1.4142135623730951}}, "df": 35, "s": {"docs": {"pyerrors.correlators.Corr.plateau": {"tf": 1}}, "df": 1}}}}}}, "p": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1}}, "df": 1, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr": {"tf": 1}}, "df": 2}, "c": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.correlators.Corr": {"tf": 1}}, "df": 1}}}}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.openQCD.read_qtop": {"tf": 1}}, "df": 1}}}}}}}, "n": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}, "v": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1}}, "df": 1, "s": {"docs": {"pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}}, "df": 2}}}}}}}}, "d": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr": {"tf": 1}}, "df": 1}}}}}}}, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.correlators.Corr": {"tf": 1}}, "df": 1}}}}}, "b": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.openQCD.read_rwms": {"tf": 1}}, "df": 1}}}}}}}}, "o": {"docs": {"pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 4, "w": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}}, "df": 3}}, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 7}}, "m": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "d": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}}, "df": 1}}}}}}}}, "b": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}}, "df": 5}}, "t": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 1}}, "f": {"docs": {"pyerrors.fits.Fit_result": {"tf": 1}}, "df": 1}, "u": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 3}}}}, "n": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.4142135623730951}}, "df": 2}}, "c": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.linalg.einsum": {"tf": 1}, "pyerrors.obs.Obs.is_zero": {"tf": 1}}, "df": 2}}}}}}}}}}}}, "u": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "p": {"docs": {"pyerrors": {"tf": 2}, "pyerrors.misc.dump_object": {"tf": 1}, "pyerrors.obs.Obs.dump": {"tf": 1}}, "df": 3, "s": {"docs": {"pyerrors.correlators.Corr.dump": {"tf": 1}}, "df": 1}, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.pandas.dump_df": {"tf": 1}}, "df": 1}}}}, "e": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 1}}, "t": {"docs": {"pyerrors.correlators.Corr.roll": {"tf": 1.4142135623730951}}, "df": 1, "y": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 3.1622776601683795}}, "df": 1}}}, "r": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.4142135623730951}}, "df": 4}}, "v": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.misc.pseudo_Obs": {"tf": 1.4142135623730951}}, "df": 1, "s": {"docs": {"pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.fits.fit_lin": {"tf": 1.4142135623730951}}, "df": 2}}}}}}, "s": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "m": {"docs": {"pyerrors.input.bdio.read_dSdm": {"tf": 1.4142135623730951}}, "df": 1}}}, "f": {"docs": {"pyerrors.input.pandas.to_sql": {"tf": 1}, "pyerrors.input.pandas.dump_df": {"tf": 1}}, "df": 2}, "b": {"docs": {"pyerrors.input.pandas.to_sql": {"tf": 1}, "pyerrors.input.pandas.read_sql": {"tf": 1}}, "df": 2}}, "b": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.obs.correlate": {"tf": 1}}, "df": 2, "a": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1.7320508075688772}}, "df": 1, "d": {"docs": {"pyerrors": {"tf": 3}, "pyerrors.correlators.Corr.prune": {"tf": 1}, "pyerrors.dirac.epsilon_tensor": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}, "pyerrors.obs.merge_obs": {"tf": 1}, "pyerrors.obs.cov_Obs": {"tf": 1}}, "df": 9}, "m": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "x": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1.7320508075688772}}, "df": 1}}}}}}}, "i": {"docs": {}, "df": 0, "c": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1}}, "df": 2}}}, "r": {"docs": {"pyerrors": {"tf": 1}}, "df": 1, "{": {"docs": {}, "df": 0, "f": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}, "a": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}, "c": {"docs": {}, "df": 0, "k": {"docs": {"pyerrors.correlators.Corr.__init__": {"tf": 1}}, "df": 1, "w": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.deriv": {"tf": 1}}, "df": 1}}}}}}, "n": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.fits.error_band": {"tf": 1}}, "df": 1}}}, "e": {"docs": {"pyerrors": {"tf": 6.082762530298219}, "pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.correlators.Corr.item": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.correlate": {"tf": 1}, "pyerrors.correlators.Corr.reweight": {"tf": 1}, "pyerrors.correlators.Corr.T_symmetry": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.dump": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1.7320508075688772}, "pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 2.6457513110645907}, "pyerrors.fits.total_least_squares": {"tf": 1.7320508075688772}, "pyerrors.fits.fit_lin": {"tf": 1}, "pyerrors.fits.qqplot": {"tf": 1}, "pyerrors.input.bdio.read_ADerrors": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.read_mesons": {"tf": 2.23606797749979}, "pyerrors.input.bdio.read_dSdm": {"tf": 2.23606797749979}, "pyerrors.input.dobs.create_pobs_string": {"tf": 2}, "pyerrors.input.dobs.write_pobs": {"tf": 2}, "pyerrors.input.dobs.create_dobs_string": {"tf": 2}, "pyerrors.input.dobs.write_dobs": {"tf": 2}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1.7320508075688772}, "pyerrors.input.json.dump_to_json": {"tf": 1.7320508075688772}, "pyerrors.input.json.dump_dict_to_json": {"tf": 2}, "pyerrors.input.json.load_json_dict": {"tf": 1}, "pyerrors.input.misc.read_pbp": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_rwms": {"tf": 2}, "pyerrors.input.openQCD.extract_t0": {"tf": 2.449489742783178}, "pyerrors.input.openQCD.read_qtop": {"tf": 2.449489742783178}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 2.449489742783178}, "pyerrors.input.openQCD.qtop_projection": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 2.6457513110645907}, "pyerrors.input.pandas.to_sql": {"tf": 1}, "pyerrors.input.pandas.read_sql": {"tf": 1}, "pyerrors.input.pandas.dump_df": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 2.23606797749979}, "pyerrors.input.utils.check_idl": {"tf": 1}, "pyerrors.linalg.matmul": {"tf": 1}, "pyerrors.linalg.jack_matmul": {"tf": 1}, "pyerrors.linalg.einsum": {"tf": 1}, "pyerrors.misc.dump_object": {"tf": 1}, "pyerrors.misc.pseudo_Obs": {"tf": 1.7320508075688772}, "pyerrors.misc.gen_correlated_data": {"tf": 2}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.reweight": {"tf": 1}, "pyerrors.obs.Obs.dump": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1.7320508075688772}, "pyerrors.obs.reweight": {"tf": 1}, "pyerrors.obs.correlate": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1.7320508075688772}, "pyerrors.obs.merge_obs": {"tf": 1}, "pyerrors.roots.find_root": {"tf": 1}}, "df": 59, "t": {"docs": {}, "df": 0, "w": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop": {"tf": 2.23606797749979}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 2.23606797749979}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 2.23606797749979}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 10}}}}, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}}, "df": 1}}}, "a": {"docs": {"pyerrors.fits.error_band": {"tf": 1}}, "df": 1}}, "c": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 2}}}}}, "l": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "w": {"docs": {"pyerrors": {"tf": 2}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 3}, "n": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 3}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.covobs.Covobs.__init__": {"tf": 1.4142135623730951}, "pyerrors.obs.cov_Obs": {"tf": 1}}, "df": 3}}}}}}}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}, "e": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.obs.correlate": {"tf": 1}}, "df": 2}}, "f": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.input.pandas.dump_df": {"tf": 1}}, "df": 2}}}}, "h": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.pandas.to_sql": {"tf": 1}}, "df": 1}}}}}, "y": {"docs": {"pyerrors": {"tf": 3}, "pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.roll": {"tf": 1}, "pyerrors.correlators.Corr.reweight": {"tf": 1}, "pyerrors.correlators.Corr.m_eff": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.prune": {"tf": 1.7320508075688772}, "pyerrors.fits.Fit_result": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.total_least_squares": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.read_dobs": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.openQCD.qtop_projection": {"tf": 1}, "pyerrors.obs.Obs.reweight": {"tf": 1}, "pyerrors.obs.reweight": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1.4142135623730951}}, "df": 23, "t": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1, "s": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 2.23606797749979}}, "df": 1}}}}, "u": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 8}, "i": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "d": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.input.bdio.read_ADerrors": {"tf": 1.7320508075688772}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1.7320508075688772}, "pyerrors.input.bdio.read_mesons": {"tf": 1.7320508075688772}, "pyerrors.input.bdio.read_dSdm": {"tf": 1.7320508075688772}}, "df": 4}}}}}}, "f": {"docs": {}, "df": 0, "f": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 3.1622776601683795}}, "df": 1}}}}, "l": {"docs": {}, "df": 0, "k": {"docs": {"pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}}, "df": 1}}}, "o": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 1, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 2.6457513110645907}}, "df": 4}}}}}, "d": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1.4142135623730951}}, "df": 1}}}}}, "t": {"docs": {}, "df": 0, "h": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.linalg.jack_matmul": {"tf": 1}, "pyerrors.obs.CObs.is_zero": {"tf": 1}}, "df": 3}}, "o": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.Hankel": {"tf": 1}, "pyerrors.correlators.Corr.reweight": {"tf": 1}, "pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.spaghetti_plot": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 2.449489742783178}, "pyerrors.fits.total_least_squares": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.read_dobs": {"tf": 2}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.import_json_string": {"tf": 1.4142135623730951}, "pyerrors.input.json.load_json": {"tf": 1.7320508075688772}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.pandas.to_sql": {"tf": 1}, "pyerrors.input.pandas.read_sql": {"tf": 1}, "pyerrors.input.pandas.dump_df": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1.4142135623730951}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}, "pyerrors.obs.Obs.details": {"tf": 1}, "pyerrors.obs.Obs.reweight": {"tf": 1}, "pyerrors.obs.Obs.plot_history": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}, "pyerrors.obs.reweight": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1.4142135623730951}}, "df": 37}, "k": {"docs": {}, "df": 0, "k": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}}, "df": 5}}}}}}}}}, "x": {"docs": {"pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}}, "df": 1}}, "d": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {"pyerrors.input.bdio.read_ADerrors": {"tf": 2.6457513110645907}, "pyerrors.input.bdio.write_ADerrors": {"tf": 2.6457513110645907}, "pyerrors.input.bdio.read_mesons": {"tf": 2.6457513110645907}, "pyerrors.input.bdio.read_dSdm": {"tf": 2.6457513110645907}}, "df": 4}}}, "i": {"docs": {"pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 1, "l": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}}, "df": 1}}}}}}, "n": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}}, "df": 1}}}}, "b": {"docs": {"pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 1}}, "b": {"docs": {"pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 1}}, "t": {"0": {"docs": {"pyerrors.correlators.Corr.GEVP": {"tf": 2}, "pyerrors.input.openQCD.extract_t0": {"tf": 2}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 5, "p": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "j": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1.4142135623730951}}, "df": 1}}}}}, "2": {"docs": {"pyerrors.fits.Fit_result": {"tf": 1}}, "df": 1}, "docs": {"pyerrors": {"tf": 2.8284271247461903}, "pyerrors.correlators.Corr.GEVP": {"tf": 2.8284271247461903}, "pyerrors.correlators.Corr.Hankel": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.m_eff": {"tf": 2.8284271247461903}, "pyerrors.correlators.Corr.prune": {"tf": 4.47213595499958}, "pyerrors.fits.Fit_result": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 9, "h": {"docs": {"pyerrors.correlators.Corr.thin": {"tf": 1}}, "df": 1, "e": {"docs": {"pyerrors": {"tf": 16.15549442140351}, "pyerrors.correlators.Corr": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.__init__": {"tf": 3}, "pyerrors.correlators.Corr.gamma_method": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.gm": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.projected": {"tf": 2}, "pyerrors.correlators.Corr.item": {"tf": 1}, "pyerrors.correlators.Corr.plottable": {"tf": 2}, "pyerrors.correlators.Corr.symmetric": {"tf": 1}, "pyerrors.correlators.Corr.anti_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.matrix_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 4.58257569495584}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 2.449489742783178}, "pyerrors.correlators.Corr.Hankel": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.roll": {"tf": 1}, "pyerrors.correlators.Corr.reverse": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.thin": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.correlate": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.reweight": {"tf": 2.449489742783178}, "pyerrors.correlators.Corr.T_symmetry": {"tf": 2}, "pyerrors.correlators.Corr.deriv": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.second_deriv": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.m_eff": {"tf": 3.1622776601683795}, "pyerrors.correlators.Corr.fit": {"tf": 2.449489742783178}, "pyerrors.correlators.Corr.plateau": {"tf": 2.6457513110645907}, "pyerrors.correlators.Corr.set_prange": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.show": {"tf": 3.4641016151377544}, "pyerrors.correlators.Corr.spaghetti_plot": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.dump": {"tf": 2}, "pyerrors.correlators.Corr.prune": {"tf": 4.795831523312719}, "pyerrors.covobs.Covobs.__init__": {"tf": 2.23606797749979}, "pyerrors.covobs.Covobs.errsq": {"tf": 1.7320508075688772}, "pyerrors.fits.Fit_result": {"tf": 1.4142135623730951}, "pyerrors.fits.Fit_result.gamma_method": {"tf": 1}, "pyerrors.fits.Fit_result.gm": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 4.795831523312719}, "pyerrors.fits.total_least_squares": {"tf": 3.605551275463989}, "pyerrors.fits.fit_lin": {"tf": 2.449489742783178}, "pyerrors.fits.qqplot": {"tf": 1.7320508075688772}, "pyerrors.fits.residual_plot": {"tf": 1.7320508075688772}, "pyerrors.fits.error_band": {"tf": 1}, "pyerrors.fits.ks_test": {"tf": 1.4142135623730951}, "pyerrors.input": {"tf": 1}, "pyerrors.input.bdio.read_ADerrors": {"tf": 2}, "pyerrors.input.bdio.write_ADerrors": {"tf": 2}, "pyerrors.input.bdio.read_mesons": {"tf": 2.6457513110645907}, "pyerrors.input.bdio.read_dSdm": {"tf": 2.6457513110645907}, "pyerrors.input.dobs.create_pobs_string": {"tf": 3.4641016151377544}, "pyerrors.input.dobs.write_pobs": {"tf": 3.872983346207417}, "pyerrors.input.dobs.read_pobs": {"tf": 3}, "pyerrors.input.dobs.import_dobs_string": {"tf": 3.605551275463989}, "pyerrors.input.dobs.read_dobs": {"tf": 3.605551275463989}, "pyerrors.input.dobs.create_dobs_string": {"tf": 4.47213595499958}, "pyerrors.input.dobs.write_dobs": {"tf": 4.58257569495584}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 3.1622776601683795}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 2}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 5.830951894845301}, "pyerrors.input.hadrons.Npr_matrix.g5H": {"tf": 2}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 2}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 2}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 2}, "pyerrors.input.json.create_json_string": {"tf": 2.8284271247461903}, "pyerrors.input.json.dump_to_json": {"tf": 3}, "pyerrors.input.json.import_json_string": {"tf": 2.6457513110645907}, "pyerrors.input.json.load_json": {"tf": 2.6457513110645907}, "pyerrors.input.json.dump_dict_to_json": {"tf": 3.3166247903554}, "pyerrors.input.json.load_json_dict": {"tf": 2.6457513110645907}, "pyerrors.input.misc.read_pbp": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_rwms": {"tf": 3}, "pyerrors.input.openQCD.extract_t0": {"tf": 5.385164807134504}, "pyerrors.input.openQCD.read_qtop": {"tf": 4.58257569495584}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 4.47213595499958}, "pyerrors.input.openQCD.qtop_projection": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 4.242640687119285}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 4.47213595499958}, "pyerrors.input.pandas.to_sql": {"tf": 2.23606797749979}, "pyerrors.input.pandas.read_sql": {"tf": 2}, "pyerrors.input.pandas.dump_df": {"tf": 2}, "pyerrors.input.pandas.load_df": {"tf": 2}, "pyerrors.input.sfcf.read_sfcf": {"tf": 4.58257569495584}, "pyerrors.input.utils": {"tf": 1}, "pyerrors.input.utils.check_idl": {"tf": 1}, "pyerrors.linalg.matmul": {"tf": 1}, "pyerrors.linalg.jack_matmul": {"tf": 1}, "pyerrors.linalg.eigh": {"tf": 1}, "pyerrors.linalg.eig": {"tf": 1}, "pyerrors.linalg.pinv": {"tf": 1}, "pyerrors.linalg.svd": {"tf": 1}, "pyerrors.misc.dump_object": {"tf": 1.7320508075688772}, "pyerrors.misc.load_object": {"tf": 1}, "pyerrors.misc.pseudo_Obs": {"tf": 2.23606797749979}, "pyerrors.misc.gen_correlated_data": {"tf": 1.7320508075688772}, "pyerrors.mpm.matrix_pencil_method": {"tf": 2.23606797749979}, "pyerrors.obs.Obs": {"tf": 3.1622776601683795}, "pyerrors.obs.Obs.__init__": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.gamma_method": {"tf": 3.4641016151377544}, "pyerrors.obs.Obs.gm": {"tf": 3.4641016151377544}, "pyerrors.obs.Obs.details": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.reweight": {"tf": 2.449489742783178}, "pyerrors.obs.Obs.is_zero_within_error": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.is_zero": {"tf": 1}, "pyerrors.obs.Obs.plot_tauint": {"tf": 1}, "pyerrors.obs.Obs.plot_rho": {"tf": 1}, "pyerrors.obs.Obs.plot_piechart": {"tf": 2}, "pyerrors.obs.Obs.dump": {"tf": 2}, "pyerrors.obs.Obs.export_jackknife": {"tf": 3.3166247903554}, "pyerrors.obs.CObs.gamma_method": {"tf": 1.7320508075688772}, "pyerrors.obs.derived_observable": {"tf": 2.8284271247461903}, "pyerrors.obs.reweight": {"tf": 2.23606797749979}, "pyerrors.obs.correlate": {"tf": 1.4142135623730951}, "pyerrors.obs.covariance": {"tf": 5.291502622129181}, "pyerrors.obs.import_jackknife": {"tf": 2}, "pyerrors.obs.merge_obs": {"tf": 1.4142135623730951}, "pyerrors.obs.cov_Obs": {"tf": 2}, "pyerrors.roots.find_root": {"tf": 2.449489742783178}}, "df": 115, "i": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}}, "r": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1.4142135623730951}}, "df": 3, "f": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}, "m": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "z": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1.4142135623730951}}, "df": 1}}}}}}}}}}}, "s": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}}, "df": 6}}, "n": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 4}, "y": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1}}, "df": 2}, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 1}}}, "e": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 1}}}}}}}}}}}, "a": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 6.244997998398398}, "pyerrors.correlators.Corr.reweight": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1.7320508075688772}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 2}, "pyerrors.input.dobs.write_pobs": {"tf": 2}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.create_dobs_string": {"tf": 2}, "pyerrors.input.dobs.write_dobs": {"tf": 2}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 2}, "pyerrors.input.hadrons.Npr_matrix.g5H": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1.4142135623730951}, "pyerrors.input.json.dump_to_json": {"tf": 1.4142135623730951}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1.4142135623730951}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1.4142135623730951}, "pyerrors.input.json.load_json_dict": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_rwms": {"tf": 2.23606797749979}, "pyerrors.input.openQCD.extract_t0": {"tf": 2.23606797749979}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.7320508075688772}, "pyerrors.input.pandas.load_df": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}, "pyerrors.obs.Obs.reweight": {"tf": 1}, "pyerrors.obs.reweight": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1.7320508075688772}}, "df": 33}, "n": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.obs.Obs.plot_rep_dist": {"tf": 1}}, "df": 2}}, "i": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 4}, "pyerrors.correlators.Corr": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.__init__": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 2}, "pyerrors.fits.least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.7320508075688772}, "pyerrors.linalg.matmul": {"tf": 1}, "pyerrors.linalg.jack_matmul": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}, "pyerrors.obs.correlate": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1.4142135623730951}}, "df": 27}, "n": {"docs": {"pyerrors.correlators.Corr.thin": {"tf": 1}}, "df": 1}}, "r": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "h": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}}}}, "e": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.plottable": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.fit": {"tf": 1}}, "df": 3}}}}, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}, "u": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.correlators.Corr.reweight": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 2.23606797749979}, "pyerrors.fits.total_least_squares": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.read_dobs": {"tf": 2}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1.4142135623730951}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.pandas.to_sql": {"tf": 1}, "pyerrors.input.pandas.read_sql": {"tf": 1}, "pyerrors.input.pandas.dump_df": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1.4142135623730951}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}, "pyerrors.obs.Obs.details": {"tf": 1}, "pyerrors.obs.Obs.reweight": {"tf": 1}, "pyerrors.obs.Obs.plot_history": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}, "pyerrors.obs.reweight": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1.4142135623730951}}, "df": 32}}, "a": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "k": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "n": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 1, "p": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}}}}}, "j": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 1, "e": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.4142135623730951}}, "df": 4}}}}}}}}}}, "i": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}}, "s": {"docs": {"pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}}, "df": 2}}}}}}}, "o": {"docs": {"pyerrors": {"tf": 8.660254037844387}, "pyerrors.correlators.Corr": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.gamma_method": {"tf": 1}, "pyerrors.correlators.Corr.gm": {"tf": 1}, "pyerrors.correlators.Corr.projected": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.item": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.GEVP": {"tf": 2.6457513110645907}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 1}, "pyerrors.correlators.Corr.thin": {"tf": 1}, "pyerrors.correlators.Corr.correlate": {"tf": 1}, "pyerrors.correlators.Corr.reweight": {"tf": 1}, "pyerrors.correlators.Corr.deriv": {"tf": 1}, "pyerrors.correlators.Corr.second_deriv": {"tf": 1}, "pyerrors.correlators.Corr.fit": {"tf": 2.6457513110645907}, "pyerrors.correlators.Corr.plateau": {"tf": 2}, "pyerrors.correlators.Corr.show": {"tf": 2.23606797749979}, "pyerrors.correlators.Corr.spaghetti_plot": {"tf": 1}, "pyerrors.correlators.Corr.dump": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 3.1622776601683795}, "pyerrors.covobs.Covobs.__init__": {"tf": 1.7320508075688772}, "pyerrors.fits.Fit_result.gamma_method": {"tf": 1}, "pyerrors.fits.Fit_result.gm": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 2.8284271247461903}, "pyerrors.fits.total_least_squares": {"tf": 2.6457513110645907}, "pyerrors.fits.fit_lin": {"tf": 1}, "pyerrors.fits.qqplot": {"tf": 1}, "pyerrors.fits.residual_plot": {"tf": 1}, "pyerrors.fits.ks_test": {"tf": 1}, "pyerrors.input.bdio.read_ADerrors": {"tf": 2.23606797749979}, "pyerrors.input.bdio.write_ADerrors": {"tf": 2.6457513110645907}, "pyerrors.input.bdio.read_mesons": {"tf": 2.6457513110645907}, "pyerrors.input.bdio.read_dSdm": {"tf": 2.6457513110645907}, "pyerrors.input.dobs.create_pobs_string": {"tf": 2.23606797749979}, "pyerrors.input.dobs.write_pobs": {"tf": 2.23606797749979}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.read_dobs": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.create_dobs_string": {"tf": 2}, "pyerrors.input.dobs.write_dobs": {"tf": 2}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 2.449489742783178}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1.7320508075688772}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 3.7416573867739413}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1.7320508075688772}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1.7320508075688772}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 2}, "pyerrors.input.json.create_json_string": {"tf": 1.4142135623730951}, "pyerrors.input.json.dump_to_json": {"tf": 1.4142135623730951}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1.7320508075688772}, "pyerrors.input.json.load_json_dict": {"tf": 1.4142135623730951}, "pyerrors.input.misc.read_pbp": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_rwms": {"tf": 2.6457513110645907}, "pyerrors.input.openQCD.extract_t0": {"tf": 3.4641016151377544}, "pyerrors.input.openQCD.read_qtop": {"tf": 2.8284271247461903}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 3}, "pyerrors.input.openQCD.qtop_projection": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 3.4641016151377544}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 2.8284271247461903}, "pyerrors.input.pandas.to_sql": {"tf": 2.23606797749979}, "pyerrors.input.pandas.read_sql": {"tf": 1.7320508075688772}, "pyerrors.input.pandas.dump_df": {"tf": 2}, "pyerrors.input.pandas.load_df": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 3.3166247903554}, "pyerrors.input.utils.check_idl": {"tf": 1}, "pyerrors.linalg.matmul": {"tf": 1}, "pyerrors.linalg.jack_matmul": {"tf": 1}, "pyerrors.linalg.eigh": {"tf": 1}, "pyerrors.linalg.eig": {"tf": 1}, "pyerrors.misc.dump_object": {"tf": 1}, "pyerrors.misc.load_object": {"tf": 1}, "pyerrors.misc.pseudo_Obs": {"tf": 1.7320508075688772}, "pyerrors.misc.gen_correlated_data": {"tf": 1.7320508075688772}, "pyerrors.mpm.matrix_pencil_method": {"tf": 2}, "pyerrors.obs.Obs.gamma_method": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.gm": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.reweight": {"tf": 1}, "pyerrors.obs.Obs.plot_tauint": {"tf": 1}, "pyerrors.obs.Obs.plot_rho": {"tf": 1}, "pyerrors.obs.Obs.plot_piechart": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.dump": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1.4142135623730951}, "pyerrors.obs.derived_observable": {"tf": 2.23606797749979}, "pyerrors.obs.reweight": {"tf": 1}, "pyerrors.obs.correlate": {"tf": 1.4142135623730951}, "pyerrors.obs.covariance": {"tf": 2.8284271247461903}, "pyerrors.obs.import_jackknife": {"tf": 1}, "pyerrors.obs.merge_obs": {"tf": 1.4142135623730951}, "pyerrors.obs.cov_Obs": {"tf": 1}, "pyerrors.roots.find_root": {"tf": 1.7320508075688772}}, "df": 89, "g": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 3}}}}}}, "t": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 2}}}, "d": {"docs": {}, "df": 0, "o": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}, "p": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.input.openQCD.read_qtop": {"tf": 1}}, "df": 1}}, "c": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.qtop_projection": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.4142135623730951}}, "df": 3}}}}}}}}}, "w": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.openQCD.read_qtop": {"tf": 1}}, "df": 1}}}}}, "l": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.obs.Obs.is_zero": {"tf": 1.4142135623730951}}, "df": 1}}}}}}}}, "w": {"docs": {}, "df": 0, "o": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.correlators.Corr.__init__": {"tf": 1}, "pyerrors.correlators.Corr.projected": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.fits.fit_lin": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1.7320508075688772}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop": {"tf": 2.23606797749979}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 2.23606797749979}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 2.23606797749979}, "pyerrors.obs.derived_observable": {"tf": 1}, "pyerrors.obs.correlate": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 20}}, "a": {"docs": {}, "df": 0, "k": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.fits.total_least_squares": {"tf": 1}}, "df": 1, "n": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.prune": {"tf": 1}, "pyerrors.covobs.Covobs.__init__": {"tf": 1}}, "df": 3}, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}, "y": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}, "u": {"docs": {"pyerrors": {"tf": 2.6457513110645907}, "pyerrors.misc.gen_correlated_data": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 2}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1.4142135623730951}}, "df": 6, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}}}}, "i": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}}, "df": 3, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}, "g": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.show": {"tf": 1}}, "df": 2, "s": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}}, "df": 8}}, "r": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1.7320508075688772}, "pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}, "pyerrors.input.openQCD.qtop_projection": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 7}}}}, "b": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.pandas.to_sql": {"tf": 1.7320508075688772}}, "df": 1}}}}, "i": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 2.6457513110645907}, "pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.reverse": {"tf": 1}, "pyerrors.correlators.Corr.T_symmetry": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.prune": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.obs.Obs.plot_tauint": {"tf": 1}, "pyerrors.obs.Obs.plot_rho": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 10, "s": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1.4142135623730951}, "pyerrors.misc.gen_correlated_data": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 4, "l": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.__init__": {"tf": 1}, "pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.correlators.Corr.item": {"tf": 1}, "pyerrors.correlators.Corr.plottable": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.is_matrix_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.matrix_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 10, "s": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.correlators.Corr.roll": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.fit": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.plateau": {"tf": 1}}, "df": 5}}}}}}}}, "t": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1.4142135623730951}}, "df": 1}}}}, "e": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}, "r": {"docs": {}, "df": 0, "m": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}}}}, "s": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 2.449489742783178}, "pyerrors.fits.ks_test": {"tf": 1}, "pyerrors.misc.pseudo_Obs": {"tf": 1}}, "df": 3}}, "n": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.dirac.epsilon_tensor": {"tf": 1}, "pyerrors.dirac.epsilon_tensor_rank4": {"tf": 1}}, "df": 2}}}}}, "y": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 2}, "pyerrors.correlators.Corr.dump": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 2.8284271247461903}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.4142135623730951}}, "df": 6, "s": {"docs": {"pyerrors.correlators.Corr": {"tf": 1}}, "df": 1}}, "i": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 1, "l": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}}, "df": 1}}}}}, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}}}}, "s": {"docs": {"pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 2}, "+": {"1": {"docs": {"pyerrors.correlators.Corr.Hankel": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.m_eff": {"tf": 2}}, "df": 2}, "2": {"docs": {"pyerrors.correlators.Corr.Hankel": {"tf": 1.4142135623730951}}, "df": 1}, "docs": {}, "df": 0, "n": {"docs": {"pyerrors.correlators.Corr.Hankel": {"tf": 1.7320508075688772}}, "df": 1}}, "/": {"2": {"docs": {"pyerrors.correlators.Corr.m_eff": {"tf": 2}, "pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 2}, "docs": {}, "df": 0}, "p": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "j": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 1}}}}, "u": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 2}}, "df": 5}}}}, "^": {"2": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}}, "g": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.GEVP": {"tf": 2}, "pyerrors.correlators.Corr.m_eff": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 2}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.4142135623730951}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.4142135623730951}, "pyerrors.obs.derived_observable": {"tf": 1.4142135623730951}, "pyerrors.obs.reweight": {"tf": 1}}, "df": 15, "a": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "a": {"5": {"docs": {"pyerrors.input.hadrons.read_meson_hd5": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.Npr_matrix.g5H": {"tf": 1}}, "df": 2}, "docs": {"pyerrors": {"tf": 4.242640687119285}, "pyerrors.correlators.Corr.gamma_method": {"tf": 1}, "pyerrors.correlators.Corr.gm": {"tf": 1}, "pyerrors.correlators.Corr.plateau": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.show": {"tf": 1.4142135623730951}, "pyerrors.dirac.Grid_gamma": {"tf": 1}, "pyerrors.fits.Fit_result.gamma_method": {"tf": 1}, "pyerrors.fits.Fit_result.gm": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix.g5H": {"tf": 1}, "pyerrors.input.pandas.read_sql": {"tf": 1.4142135623730951}, "pyerrors.input.pandas.load_df": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs": {"tf": 1}, "pyerrors.obs.Obs.is_zero_within_error": {"tf": 1}, "pyerrors.obs.CObs.gamma_method": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1.7320508075688772}}, "df": 16, "s": {"docs": {"pyerrors.input.hadrons.read_meson_hd5": {"tf": 1.4142135623730951}}, "df": 1}}}}, "p": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "u": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.fits.qqplot": {"tf": 1}}, "df": 1}}}}}}, "r": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}}}}}, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "b": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}, "v": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.error_band": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}, "pyerrors.input.misc.read_pbp": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.4142135623730951}, "pyerrors.linalg.eigh": {"tf": 1}, "pyerrors.linalg.eig": {"tf": 1}, "pyerrors.misc.pseudo_Obs": {"tf": 1}, "pyerrors.misc.gen_correlated_data": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.reweight": {"tf": 1}, "pyerrors.obs.Obs.is_zero": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1}}, "df": 28}, "s": {"docs": {"pyerrors.correlators.Corr.GEVP": {"tf": 1}}, "df": 1}}}}, "t": {"docs": {"pyerrors": {"tf": 8.306623862918075}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 2.8284271247461903}}, "df": 2}, "e": {"docs": {}, "df": 0, "q": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "n": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1}}, "df": 2, "i": {"docs": {}, "df": 0, "z": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}}, "df": 2}}}}}, "t": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.misc.pseudo_Obs": {"tf": 1}, "pyerrors.misc.gen_correlated_data": {"tf": 1}}, "df": 5, "d": {"docs": {"pyerrors.fits.least_squares": {"tf": 1.4142135623730951}, "pyerrors.misc.pseudo_Obs": {"tf": 1.7320508075688772}, "pyerrors.misc.gen_correlated_data": {"tf": 1.4142135623730951}}, "df": 3}, "s": {"docs": {"pyerrors.fits.qqplot": {"tf": 1}, "pyerrors.fits.residual_plot": {"tf": 1}}, "df": 2}}, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.obs.derived_observable": {"tf": 1}}, "df": 1}}}}}, "i": {"docs": {}, "df": 0, "c": {"docs": {"pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}}, "df": 2}}}, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.misc.gen_correlated_data": {"tf": 1}}, "df": 1}}}}}}, "v": {"docs": {}, "df": 0, "p": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.prune": {"tf": 2.23606797749979}}, "df": 5}}, "t": {"docs": {"pyerrors.correlators.Corr.projected": {"tf": 1}}, "df": 1}, "o": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "c": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}}}}}}, "l": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1.7320508075688772}}, "df": 2}}}}}, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1.4142135623730951}, "pyerrors.obs.cov_Obs": {"tf": 1}}, "df": 6, "i": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1.4142135623730951}, "pyerrors.obs.cov_Obs": {"tf": 1}}, "df": 5, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}, "n": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 1}}}}}, "o": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "n": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.GEVP": {"tf": 1}}, "df": 1}}}}, "i": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.dirac.Grid_gamma": {"tf": 1}}, "df": 1}}}, "o": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}}, "z": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.dump": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1.4142135623730951}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1.4142135623730951}, "pyerrors.input.json.load_json": {"tf": 1.4142135623730951}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1.4142135623730951}, "pyerrors.input.json.load_json_dict": {"tf": 1.4142135623730951}, "pyerrors.input.pandas.to_sql": {"tf": 1}, "pyerrors.input.pandas.dump_df": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.dump": {"tf": 1.4142135623730951}}, "df": 16, "i": {"docs": {}, "df": 0, "p": {"docs": {"pyerrors": {"tf": 1}}, "df": 1, "p": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}, "pyerrors.input.pandas.to_sql": {"tf": 1}, "pyerrors.input.pandas.dump_df": {"tf": 1.4142135623730951}, "pyerrors.input.pandas.load_df": {"tf": 1}}, "df": 10}}}}}}, "u": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.correlators.Corr.m_eff": {"tf": 1.4142135623730951}, "pyerrors.fits.least_squares": {"tf": 2}, "pyerrors.fits.total_least_squares": {"tf": 1.4142135623730951}, "pyerrors.roots.find_root": {"tf": 1.4142135623730951}}, "df": 4}}}, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}}}}}}}, "^": {"docs": {}, "df": 0, "\\": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 1}}}}}}}, "c": {"docs": {}, "df": 0, "c": {"docs": {"pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}}, "df": 4}}, "f": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}}, "df": 2}}}}, "l": {"docs": {"pyerrors.input.openQCD.read_qtop": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.4142135623730951}}, "df": 3, "a": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1}}, "df": 1, "/": {"0": {"3": {"0": {"6": {"0": {"1": {"7": {"docs": {"pyerrors": {"tf": 2}}, "df": 1}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "9": {"4": {"1": {"2": {"0": {"8": {"7": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.4142135623730951}}, "df": 5}}}}, "e": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 1, "x": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "r": {"docs": {"pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 1}}}, "y": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1.7320508075688772}}, "df": 1}}}}, "b": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1.4142135623730951}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 3, "s": {"docs": {"pyerrors.correlators.Corr.__init__": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1}}, "df": 2}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.dirac.Grid_gamma": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.obs.Obs.__init__": {"tf": 1}}, "df": 6}}}, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}}, "df": 1}}}}}, "s": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.__init__": {"tf": 1}, "pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}, "pyerrors.input.misc.read_pbp": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 10}}, "r": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1.4142135623730951}, "pyerrors.linalg.jack_matmul": {"tf": 1}}, "df": 2, "s": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 2}}, "r": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 2}}}}, "m": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "a": {"docs": {"pyerrors.correlators.Corr.GEVP": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.prune": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}}, "df": 3}}}}}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors": {"tf": 2.449489742783178}, "pyerrors.fits.least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.total_least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.fit_lin": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.4142135623730951}}, "df": 5}}, "s": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1}}, "df": 1}}, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.linalg.eigh": {"tf": 1}, "pyerrors.linalg.eig": {"tf": 1}}, "df": 3}}}}, "s": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 3.1622776601683795}, "pyerrors.correlators.Corr.__init__": {"tf": 2.6457513110645907}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.fit": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.plateau": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.show": {"tf": 2.449489742783178}, "pyerrors.covobs.Covobs.__init__": {"tf": 1.4142135623730951}, "pyerrors.fits.Fit_result": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 2.6457513110645907}, "pyerrors.fits.total_least_squares": {"tf": 2.449489742783178}, "pyerrors.fits.fit_lin": {"tf": 2.23606797749979}, "pyerrors.fits.ks_test": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.create_pobs_string": {"tf": 2.23606797749979}, "pyerrors.input.dobs.write_pobs": {"tf": 2.23606797749979}, "pyerrors.input.dobs.read_pobs": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.read_dobs": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.create_dobs_string": {"tf": 2.23606797749979}, "pyerrors.input.dobs.write_dobs": {"tf": 2.23606797749979}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 2}, "pyerrors.input.json.dump_to_json": {"tf": 2}, "pyerrors.input.json.import_json_string": {"tf": 2}, "pyerrors.input.json.load_json": {"tf": 2}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}, "pyerrors.input.misc.read_pbp": {"tf": 2.23606797749979}, "pyerrors.input.openQCD.read_rwms": {"tf": 3.1622776601683795}, "pyerrors.input.openQCD.extract_t0": {"tf": 3}, "pyerrors.input.openQCD.read_qtop": {"tf": 2.449489742783178}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 2.449489742783178}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 2}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 2.6457513110645907}, "pyerrors.input.utils.check_idl": {"tf": 2}, "pyerrors.misc.gen_correlated_data": {"tf": 2}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.__init__": {"tf": 2.449489742783178}, "pyerrors.obs.derived_observable": {"tf": 2}, "pyerrors.obs.reweight": {"tf": 1.7320508075688772}, "pyerrors.obs.covariance": {"tf": 1.4142135623730951}, "pyerrors.obs.merge_obs": {"tf": 2}, "pyerrors.obs.cov_Obs": {"tf": 1.7320508075688772}}, "df": 45, "s": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.correlators.Corr.plottable": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}, "pyerrors.obs.Obs.__init__": {"tf": 1}}, "df": 6}, "[": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}}, "df": 1}}}}}}, "k": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}}, "df": 2}}, "b": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.input.bdio.read_ADerrors": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.read_mesons": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.read_dSdm": {"tf": 1.4142135623730951}}, "df": 4}}}}, "d": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.input.bdio.read_ADerrors": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.read_mesons": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.read_dSdm": {"tf": 1.4142135623730951}}, "df": 4}}}, "b": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {"pyerrors.input.bdio.read_ADerrors": {"tf": 2.23606797749979}, "pyerrors.input.bdio.write_ADerrors": {"tf": 2.23606797749979}, "pyerrors.input.bdio.read_mesons": {"tf": 2.23606797749979}, "pyerrors.input.bdio.read_dSdm": {"tf": 2.23606797749979}}, "df": 4}}}}}}, "e": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "s": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 3.1622776601683795}, "pyerrors.correlators.Corr.fit": {"tf": 1}}, "df": 2}}, "v": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1, "s": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}}, "s": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "v": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.fits.least_squares": {"tf": 1}}, "df": 2}}}}}, "l": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}}, "df": 4, "s": {"docs": {"pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 1}}}}, "n": {"docs": {"pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1.7320508075688772}}, "df": 3, "g": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "h": {"docs": {"pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.correlators.Corr.correlate": {"tf": 1}, "pyerrors.input.misc.read_pbp": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.4142135623730951}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1}}, "df": 10}}}}, "q": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 1}, "f": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 1}}}, "o": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.deriv": {"tf": 1}, "pyerrors.correlators.Corr.second_deriv": {"tf": 1}, "pyerrors.correlators.Corr.m_eff": {"tf": 1.7320508075688772}}, "df": 4, "s": {"docs": {}, "df": 0, "y": {"docs": {}, "df": 0, "m": {"docs": {"pyerrors.correlators.Corr.m_eff": {"tf": 1}}, "df": 1}}, "c": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.spaghetti_plot": {"tf": 1}}, "df": 2}}}}}, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "c": {"docs": {"pyerrors.correlators.Corr.spaghetti_plot": {"tf": 1}}, "df": 1}}}}}}}}}, "n": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}, "a": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.misc.load_object": {"tf": 1}}, "df": 2}}, "w": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}}, "df": 1, "e": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1.4142135623730951}}, "df": 4}}}}}, "t": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}}, "df": 1}, "/": {"docs": {}, "df": 0, "a": {"docs": {"pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 2}}}, "s": {"docs": {"pyerrors": {"tf": 3.1622776601683795}, "pyerrors.correlators.Corr.GEVP": {"tf": 2.6457513110645907}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs": {"tf": 2}, "pyerrors.obs.Obs.gamma_method": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.gm": {"tf": 1.4142135623730951}, "pyerrors.obs.covariance": {"tf": 1.7320508075688772}, "pyerrors.obs.cov_Obs": {"tf": 1.4142135623730951}}, "df": 11, "o": {"docs": {"pyerrors.input.bdio.read_ADerrors": {"tf": 2.23606797749979}, "pyerrors.input.bdio.write_ADerrors": {"tf": 2.23606797749979}, "pyerrors.input.bdio.read_mesons": {"tf": 2.23606797749979}, "pyerrors.input.bdio.read_dSdm": {"tf": 2.23606797749979}, "pyerrors.input.hadrons.Npr_matrix.g5H": {"tf": 1}}, "df": 5, "m": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 2, "t": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}}}}}, "m": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}, "k": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}, "l": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.GEVP": {"tf": 1}}, "df": 1, "r": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.correlators.Corr.Eigenvalue": {"tf": 1}, "pyerrors.correlators.Corr.m_eff": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.prune": {"tf": 1.4142135623730951}}, "df": 3}}}}}, "u": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 4}}}}, "r": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.GEVP": {"tf": 1.7320508075688772}}, "df": 1, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.GEVP": {"tf": 1}}, "df": 1}}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.correlators.Corr.GEVP": {"tf": 1}}, "df": 1}}}}}, "n": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}}, "df": 4}}}}}, "u": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 2}}, "df": 1}}}}}}}, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "d": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.spaghetti_plot": {"tf": 1}}, "df": 2}}}}, "m": {"docs": {"pyerrors": {"tf": 3.605551275463989}, "pyerrors.obs.covariance": {"tf": 1.4142135623730951}}, "df": 2, "m": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.linalg.einsum": {"tf": 1}}, "df": 1}}}}}}}, "c": {"docs": {}, "df": 0, "h": {"docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.correlators.Corr.prune": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1.4142135623730951}}, "df": 3}}, "r": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "p": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1, "s": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.correlators.Corr.dump": {"tf": 1}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}, "pyerrors.obs.Obs.dump": {"tf": 1}}, "df": 7}}}}, "s": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 2}}}}, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.correlators.Corr.thin": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 2}}}}, "l": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.obs.derived_observable": {"tf": 1}}, "df": 1, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.obs.derived_observable": {"tf": 1}}, "df": 1}}}}}}, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.reweight": {"tf": 1}, "pyerrors.obs.Obs.reweight": {"tf": 1}, "pyerrors.obs.reweight": {"tf": 1}}, "df": 3}}}}}}, "b": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 1}}}}, "c": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.linalg.einsum": {"tf": 1.4142135623730951}}, "df": 1}}}}}}}, "m": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input": {"tf": 1}}, "df": 1}}}}}}}, "f": {"docs": {}, "df": 0, "f": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1.4142135623730951}}, "df": 1}}}}}}}}}}}, "l": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "w": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}}, "df": 3}}}}}, "i": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}}, "df": 2}}}, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}, "i": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1.7320508075688772}}, "df": 1, "s": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}}}}}}}}, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "i": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1, "l": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}, "p": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "f": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}}}, "y": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}, "y": {"docs": {"pyerrors.correlators.Corr": {"tf": 1}}, "df": 1}, "e": {"docs": {"pyerrors.obs.derived_observable": {"tf": 1}}, "df": 1}}}}, "n": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.linalg.svd": {"tf": 1}}, "df": 2}}}}, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 2.449489742783178}, "pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 4}}}, "c": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.prune": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 4}}, "h": {"docs": {"pyerrors.correlators.Corr.m_eff": {"tf": 1.7320508075688772}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}}, "df": 3}, "k": {"docs": {"pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}}, "df": 1}}, "z": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 2}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 7}}, "g": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "a": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.show": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs": {"tf": 2}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}, "pyerrors.obs.Obs.is_zero_within_error": {"tf": 1.4142135623730951}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 7}}}, "d": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.GEVP": {"tf": 1}}, "df": 1}}, "l": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}}, "df": 3}}}}}, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "f": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}, "p": {"docs": {"pyerrors.input.bdio.read_mesons": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.read_dSdm": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.extract_t0": {"tf": 2}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}}, "df": 7, "s": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 4}}}, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}}, "e": {"docs": {"pyerrors.correlators.Corr.GEVP": {"tf": 2.6457513110645907}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 1.7320508075688772}}, "df": 2, "s": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.prune": {"tf": 1.4142135623730951}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 4}}}, "n": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.correlators.Corr.m_eff": {"tf": 1}, "pyerrors.correlators.Corr.fit": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.show": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.spaghetti_plot": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.linalg.matmul": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 2.449489742783178}, "pyerrors.obs.Obs.gamma_method": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.gm": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.is_zero_within_error": {"tf": 1.4142135623730951}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 12, "i": {"docs": {}, "df": 0, "z": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.pandas.dump_df": {"tf": 1}}, "df": 1}}}}}}}}}, "r": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}, "pyerrors.input.misc.read_pbp": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 10}}, "c": {"docs": {}, "df": 0, "k": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "x": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.dirac.epsilon_tensor": {"tf": 1}, "pyerrors.dirac.epsilon_tensor_rank4": {"tf": 1}}, "df": 2}}}}}}}}}}, "b": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}}, "df": 1}}}}, "d": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}, "r": {"docs": {"pyerrors.correlators.Corr.deriv": {"tf": 1}, "pyerrors.correlators.Corr.second_deriv": {"tf": 1}, "pyerrors.correlators.Corr.m_eff": {"tf": 1}, "pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.dump": {"tf": 1.7320508075688772}, "pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1.7320508075688772}, "pyerrors.input.bdio.read_dSdm": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.create_pobs_string": {"tf": 2}, "pyerrors.input.dobs.write_pobs": {"tf": 2.23606797749979}, "pyerrors.input.dobs.read_pobs": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.read_dobs": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.create_dobs_string": {"tf": 2}, "pyerrors.input.dobs.write_dobs": {"tf": 2.23606797749979}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 2}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1.7320508075688772}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1.7320508075688772}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1.7320508075688772}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1.4142135623730951}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1.7320508075688772}, "pyerrors.input.json.load_json_dict": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_rwms": {"tf": 2}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop": {"tf": 2}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 2.23606797749979}, "pyerrors.input.pandas.to_sql": {"tf": 1.7320508075688772}, "pyerrors.input.pandas.read_sql": {"tf": 1.4142135623730951}, "pyerrors.input.pandas.dump_df": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 2.8284271247461903}, "pyerrors.linalg.einsum": {"tf": 1}, "pyerrors.misc.dump_object": {"tf": 1.4142135623730951}, "pyerrors.misc.load_object": {"tf": 1}, "pyerrors.misc.pseudo_Obs": {"tf": 1}, "pyerrors.misc.gen_correlated_data": {"tf": 1}, "pyerrors.obs.Obs.plot_tauint": {"tf": 1}, "pyerrors.obs.Obs.plot_rho": {"tf": 1}, "pyerrors.obs.Obs.plot_piechart": {"tf": 1}, "pyerrors.obs.Obs.dump": {"tf": 2}, "pyerrors.obs.import_jackknife": {"tf": 1}, "pyerrors.obs.cov_Obs": {"tf": 1}}, "df": 50, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors": {"tf": 3.4641016151377544}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1.4142135623730951}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.import_json_string": {"tf": 1.7320508075688772}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 14, "s": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.pandas.to_sql": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1}, "pyerrors.obs.Obs.__init__": {"tf": 1}}, "df": 7}}}, "d": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 2.23606797749979}}, "df": 1}}}, "c": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.obs.correlate": {"tf": 1}}, "df": 1}}}}}, "u": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}}, "df": 1, "u": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 3}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1.4142135623730951}, "pyerrors.input.json.load_json_dict": {"tf": 1}, "pyerrors.input.misc.read_pbp": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 13, "s": {"docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.import_json_string": {"tf": 1.4142135623730951}, "pyerrors.input.json.load_json": {"tf": 1.4142135623730951}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1.4142135623730951}, "pyerrors.input.json.load_json_dict": {"tf": 1.4142135623730951}}, "df": 11}}}}}}}, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}, "o": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 1}}}}}}, "i": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 2}}}, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1, "d": {"docs": {"pyerrors": {"tf": 2.449489742783178}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 2}}, "a": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}}, "df": 3}}}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}, "p": {"docs": {"pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}, "pyerrors.input.misc.read_pbp": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 9}}, "y": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}}, "df": 1}}}}, "c": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "f": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}, "e": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "a": {"docs": {"pyerrors": {"tf": 1.7320508075688772}}, "df": 1}}}}, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.spaghetti_plot": {"tf": 1}}, "df": 2}}}, "i": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}}, "df": 3}}}}, "a": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.fits.error_band": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}}, "df": 4, "s": {"1": {"docs": {"pyerrors": {"tf": 2.23606797749979}}, "df": 1}, "2": {"docs": {"pyerrors": {"tf": 2.23606797749979}}, "df": 1}, "3": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {"pyerrors": {"tf": 3.872983346207417}, "pyerrors.input": {"tf": 1.7320508075688772}, "pyerrors.misc.pseudo_Obs": {"tf": 1.4142135623730951}, "pyerrors.misc.gen_correlated_data": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs": {"tf": 1}, "pyerrors.obs.Obs.__init__": {"tf": 2}, "pyerrors.obs.Obs.export_jackknife": {"tf": 2.23606797749979}, "pyerrors.obs.import_jackknife": {"tf": 1.7320508075688772}}, "df": 8}}}}, "e": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.correlators.Corr.correlate": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1}, "pyerrors.obs.merge_obs": {"tf": 1}}, "df": 12}}, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "f": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}}, "v": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.obs.Obs.plot_tauint": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.plot_rho": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.plot_piechart": {"tf": 1.4142135623730951}}, "df": 4, "d": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.correlators.Corr.dump": {"tf": 1}, "pyerrors.misc.dump_object": {"tf": 1}, "pyerrors.obs.Obs.dump": {"tf": 1}}, "df": 4}, "s": {"docs": {"pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.obs.Obs.plot_tauint": {"tf": 1}, "pyerrors.obs.Obs.plot_rho": {"tf": 1}, "pyerrors.obs.Obs.plot_piechart": {"tf": 1}}, "df": 6}}}, "r": {"docs": {}, "df": 0, "k": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 1}}}}}, "e": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.correlators.Corr.__init__": {"tf": 1}, "pyerrors.correlators.Corr.item": {"tf": 1}, "pyerrors.correlators.Corr.second_deriv": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.obs.correlate": {"tf": 1}}, "df": 7}}}, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}}, "df": 1}}}, "o": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.input.openQCD.qtop_projection": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.4142135623730951}}, "df": 2}}}}, "t": {"docs": {"pyerrors": {"tf": 2.449489742783178}, "pyerrors.correlators.Corr.GEVP": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 13, "s": {"docs": {"pyerrors.correlators.Corr.set_prange": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1}}, "df": 2}}, "e": {"docs": {"pyerrors": {"tf": 3}, "pyerrors.correlators.Corr.m_eff": {"tf": 1}, "pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.input": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 2.23606797749979}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.linalg.einsum": {"tf": 1}, "pyerrors.obs.Obs.is_zero": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 10}, "l": {"docs": {}, "df": 0, "f": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.fit": {"tf": 1}}, "df": 3}}, "q": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr": {"tf": 1}}, "df": 1}}}}}}, "r": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}}, "df": 1}}}, "i": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "z": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.pandas.dump_df": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1}}, "df": 2}}}}}}}}, "p": {"docs": {"pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}}, "df": 1, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 2}, "pyerrors.input.dobs.import_dobs_string": {"tf": 2.449489742783178}, "pyerrors.input.dobs.read_dobs": {"tf": 2.449489742783178}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 9}}, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}}, "df": 3}}}, "e": {"docs": {"pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}}, "df": 1, "s": {"docs": {"pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 1}}}}}}}, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "h": {"docs": {"pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}}, "df": 1}}}}, "v": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 1}}}}}, "m": {"docs": {}, "df": 0, "i": {"docs": {"pyerrors.obs.covariance": {"tf": 1.4142135623730951}}, "df": 1}}}, "p": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "c": {"docs": {"pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}}, "df": 4, "i": {"docs": {}, "df": 0, "f": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 2}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.4142135623730951}}, "df": 14}, "s": {"docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.correlators.Corr.dump": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.qtop_projection": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.4142135623730951}, "pyerrors.misc.dump_object": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}, "pyerrors.obs.Obs.dump": {"tf": 1}}, "df": 10}}, "c": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}}, "y": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1.4142135623730951}, "pyerrors.input.json.load_json_dict": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 12}}, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.input.openQCD.read_rwms": {"tf": 1}}, "df": 1}}}}}, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 4}}}}, "c": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.correlators.Corr.thin": {"tf": 1.7320508075688772}}, "df": 1}}}, "e": {"docs": {"pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}}, "df": 3}}, "g": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {"pyerrors.correlators.Corr.spaghetti_plot": {"tf": 1}}, "df": 1}}}}}}}}, "y": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "c": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.is_matrix_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.deriv": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.second_deriv": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.m_eff": {"tf": 1}}, "df": 5}, "z": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.symmetric": {"tf": 1}, "pyerrors.correlators.Corr.anti_symmetric": {"tf": 1}}, "df": 2, "s": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.matrix_symmetric": {"tf": 1}}, "df": 2}, "d": {"docs": {"pyerrors.correlators.Corr.m_eff": {"tf": 1}}, "df": 1}}}}, "y": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.T_symmetry": {"tf": 1.4142135623730951}}, "df": 2}}}}}, "b": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}}, "df": 4, "s": {"docs": {"pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}}, "df": 4}}}}}, "n": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "x": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}, "h": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "f": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.roll": {"tf": 1}}, "df": 1, "s": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}}}}, "o": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 6}}}, "w": {"docs": {"pyerrors.obs.Obs.plot_history": {"tf": 1}}, "df": 1, "n": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 1}, "s": {"docs": {"pyerrors.obs.Obs.plot_piechart": {"tf": 1}}, "df": 1}}}, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.bdio.read_ADerrors": {"tf": 1.7320508075688772}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1.7320508075688772}, "pyerrors.input.bdio.read_mesons": {"tf": 1.7320508075688772}, "pyerrors.input.bdio.read_dSdm": {"tf": 1.7320508075688772}}, "df": 4}}}, "p": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 2.6457513110645907}}, "df": 1}}}}, "q": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.covobs.Covobs.errsq": {"tf": 1}}, "df": 2, "s": {"docs": {"pyerrors": {"tf": 3}, "pyerrors.correlators.Corr.fit": {"tf": 1}}, "df": 2}, "d": {"docs": {"pyerrors.fits.Fit_result": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 2}}}}}, "r": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 3, "{": {"docs": {}, "df": 0, "j": {"docs": {}, "df": 0, "^": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}, "\\": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "u": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}}}}}, "l": {"docs": {"pyerrors.input.pandas.read_sql": {"tf": 1.7320508075688772}}, "df": 1, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.pandas.to_sql": {"tf": 1.4142135623730951}, "pyerrors.input.pandas.read_sql": {"tf": 1.4142135623730951}}, "df": 2}}}}}, "m": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 3, "e": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}}}, "e": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 3}}}}}}, "o": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "h": {"docs": {"pyerrors.obs.covariance": {"tf": 1.4142135623730951}}, "df": 1, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}}}}}, "v": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "k": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "p": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1, "p": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 1}}}}}}, "f": {"docs": {}, "df": 0, "q": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.openQCD.read_qtop": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 2}}, "df": 3}}}, "c": {"docs": {}, "df": 0, "f": {"docs": {"pyerrors.input.sfcf.read_sfcf": {"tf": 2.23606797749979}}, "df": 1}}}}, "n": {"docs": {"pyerrors": {"tf": 2}, "pyerrors.correlators.Corr.Hankel": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.prune": {"tf": 2.8284271247461903}, "pyerrors.fits.fit_lin": {"tf": 1.4142135623730951}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 2}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 2}, "pyerrors.obs.covariance": {"tf": 1.4142135623730951}, "pyerrors.obs.import_jackknife": {"tf": 1}, "pyerrors.obs.cov_Obs": {"tf": 1}}, "df": 12, "o": {"docs": {"pyerrors.fits.least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.fit_lin": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.read_dobs": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 11, "n": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.fits.least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.total_least_squares": {"tf": 1.4142135623730951}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 4, "e": {"docs": {"pyerrors.correlators.Corr.GEVP": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 2.23606797749979}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 15}}, "t": {"docs": {"pyerrors": {"tf": 2.6457513110645907}, "pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.correlators.Corr.reweight": {"tf": 1}, "pyerrors.correlators.Corr.m_eff": {"tf": 1}, "pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.read_dobs": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.write_dobs": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_rwms": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.7320508075688772}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.obs.Obs.reweight": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}, "pyerrors.obs.reweight": {"tf": 1}, "pyerrors.obs.correlate": {"tf": 1.4142135623730951}, "pyerrors.obs.covariance": {"tf": 1}, "pyerrors.obs.merge_obs": {"tf": 1}}, "df": 28, "e": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}}, "df": 2, "s": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}, "pyerrors.obs.derived_observable": {"tf": 1}, "pyerrors.obs.correlate": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1.4142135623730951}, "pyerrors.obs.merge_obs": {"tf": 1}}, "df": 7}}}, "w": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "r": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "z": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.reweight": {"tf": 1}, "pyerrors.obs.Obs.reweight": {"tf": 1}, "pyerrors.obs.Obs.plot_rho": {"tf": 1}, "pyerrors.obs.reweight": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 5}}, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 1}}}}}}}}}}}, "i": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1.4142135623730951}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1.4142135623730951}}, "df": 2}}}, "e": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}}, "df": 2}}}}}, "f": {"docs": {}, "df": 0, "f": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 1}}}}}}, "e": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "f": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}}, "w": {"docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.7320508075688772}, "pyerrors.obs.derived_observable": {"tf": 1}, "pyerrors.obs.merge_obs": {"tf": 1}, "pyerrors.obs.cov_Obs": {"tf": 1}}, "df": 6}, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.4142135623730951}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 5, "s": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 2}, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 4}}}}, "c": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.obs.correlate": {"tf": 1}}, "df": 2}}}}}}}, "l": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}}, "df": 1}}}}, "x": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.input.openQCD.read_qtop": {"tf": 1}}, "df": 1}}}}}}, "u": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "m": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}}, "df": 3, "p": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 3.4641016151377544}, "pyerrors.fits.least_squares": {"tf": 1.7320508075688772}, "pyerrors.fits.total_least_squares": {"tf": 1.7320508075688772}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 2.23606797749979}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}, "pyerrors.linalg.matmul": {"tf": 1.4142135623730951}, "pyerrors.linalg.jack_matmul": {"tf": 1.4142135623730951}, "pyerrors.linalg.einsum": {"tf": 2}, "pyerrors.misc.gen_correlated_data": {"tf": 1}, "pyerrors.obs.Obs.__init__": {"tf": 1}, "pyerrors.obs.Obs.is_zero": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1.4142135623730951}, "pyerrors.obs.derived_observable": {"tf": 1.4142135623730951}, "pyerrors.obs.covariance": {"tf": 1}, "pyerrors.obs.import_jackknife": {"tf": 1.4142135623730951}, "pyerrors.roots.find_root": {"tf": 1.7320508075688772}}, "df": 21}}, "b": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.roll": {"tf": 1}, "pyerrors.correlators.Corr.T_symmetry": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 2}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.linalg.matmul": {"tf": 1}, "pyerrors.linalg.jack_matmul": {"tf": 1}, "pyerrors.linalg.einsum": {"tf": 1}, "pyerrors.misc.pseudo_Obs": {"tf": 1}, "pyerrors.misc.gen_correlated_data": {"tf": 1.4142135623730951}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}, "pyerrors.obs.Obs.is_zero_within_error": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1}}, "df": 20, "s": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 2}}}}, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.input": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1.4142135623730951}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 6, "l": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}}, "df": 1}}}}}}}}, "d": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "f": {"docs": {}, "df": 0, "f": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.obs.derived_observable": {"tf": 1}}, "df": 1}}}}}}}}}}}, "p": {"docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 2}, "pyerrors.linalg.eigh": {"tf": 1}, "pyerrors.linalg.eig": {"tf": 1}}, "df": 4}, "a": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 3.872983346207417}, "pyerrors.correlators.Corr.dump": {"tf": 1}, "pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.read_dSdm": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.write_pobs": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.read_dobs": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.create_dobs_string": {"tf": 2}, "pyerrors.input.dobs.write_dobs": {"tf": 2}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}, "pyerrors.input.pandas.to_sql": {"tf": 1.4142135623730951}, "pyerrors.input.sfcf.read_sfcf": {"tf": 2}, "pyerrors.misc.dump_object": {"tf": 1.4142135623730951}, "pyerrors.misc.pseudo_Obs": {"tf": 1.7320508075688772}, "pyerrors.misc.gen_correlated_data": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.dump": {"tf": 1.4142135623730951}, "pyerrors.obs.import_jackknife": {"tf": 1.4142135623730951}, "pyerrors.obs.cov_Obs": {"tf": 1}}, "df": 24, "s": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.read_dobs": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1.7320508075688772}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1}, "pyerrors.obs.Obs.__init__": {"tf": 1}}, "df": 17, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "m": {"docs": {"pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}}, "df": 4}}}}, "d": {"docs": {"pyerrors.obs.Obs.plot_tauint": {"tf": 1}, "pyerrors.obs.Obs.plot_rho": {"tf": 1}, "pyerrors.obs.Obs.plot_piechart": {"tf": 1}}, "df": 3}}}, "n": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}, "d": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 3.605551275463989}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}, "pyerrors.linalg.matmul": {"tf": 1}, "pyerrors.linalg.jack_matmul": {"tf": 1}, "pyerrors.linalg.einsum": {"tf": 1}, "pyerrors.misc.gen_correlated_data": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}, "pyerrors.obs.import_jackknife": {"tf": 1}}, "df": 15}}}}}, "i": {"docs": {}, "df": 0, "m": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}}, "t": {"docs": {}, "df": 0, "h": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.obs.import_jackknife": {"tf": 1}}, "df": 2}, "r": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "c": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1.7320508075688772}}, "df": 1}}}}}, "x": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.correlators.Corr.Hankel": {"tf": 1}, "pyerrors.obs.cov_Obs": {"tf": 1}}, "df": 2}}, "b": {"docs": {}, "df": 0, "y": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}}}}, "r": {"docs": {}, "df": 0, "w": {"docs": {"pyerrors.input.misc.read_pbp": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}}, "df": 2}}}, "x": {"0": {"docs": {"pyerrors": {"tf": 2}, "pyerrors.correlators.Corr.symmetric": {"tf": 1}, "pyerrors.correlators.Corr.anti_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.deriv": {"tf": 1}, "pyerrors.correlators.Corr.second_deriv": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 6}, "1": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.fits.least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.total_least_squares": {"tf": 1.4142135623730951}}, "df": 3}, "2": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.fits.least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.total_least_squares": {"tf": 1.4142135623730951}}, "df": 3}, "docs": {"pyerrors": {"tf": 2.8284271247461903}, "pyerrors.correlators.Corr.show": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.prune": {"tf": 1.4142135623730951}, "pyerrors.fits.least_squares": {"tf": 2.8284271247461903}, "pyerrors.fits.total_least_squares": {"tf": 3}, "pyerrors.fits.fit_lin": {"tf": 1.4142135623730951}, "pyerrors.fits.error_band": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}, "pyerrors.roots.find_root": {"tf": 1.7320508075688772}}, "df": 10, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.fits.fit_lin": {"tf": 1.4142135623730951}}, "df": 1}}}}}, "m": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1.7320508075688772}}, "df": 7}, "i": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1.4142135623730951}}, "df": 1}}}, "[": {"0": {"docs": {"pyerrors.obs.derived_observable": {"tf": 1}}, "df": 1}, "1": {"docs": {"pyerrors.obs.derived_observable": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}}, "y": {"docs": {"pyerrors": {"tf": 2}, "pyerrors.correlators.Corr.show": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.spaghetti_plot": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.total_least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.fit_lin": {"tf": 1.4142135623730951}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 7, "o": {"docs": {}, "df": 0, "u": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 3}}, "e": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}, "r": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.fits.fit_lin": {"tf": 1}}, "df": 2}}}}, "t": {"docs": {"pyerrors.obs.correlate": {"tf": 1}}, "df": 1}}, "l": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1}}, "df": 1}}}}}}, "r": {"0": {"1": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "2": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {"pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 3}, "docs": {"pyerrors": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.misc.read_pbp": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_rwms": {"tf": 2}, "pyerrors.input.openQCD.extract_t0": {"tf": 2.6457513110645907}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.4142135623730951}}, "df": 8, "e": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.linalg.matmul": {"tf": 1}, "pyerrors.linalg.jack_matmul": {"tf": 1}, "pyerrors.linalg.einsum": {"tf": 1}, "pyerrors.misc.gen_correlated_data": {"tf": 1}, "pyerrors.obs.CObs.gamma_method": {"tf": 1}, "pyerrors.obs.CObs.is_zero": {"tf": 1}}, "df": 9}, "d": {"docs": {"pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1.7320508075688772}, "pyerrors.input.bdio.read_dSdm": {"tf": 1.7320508075688772}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 2}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1.7320508075688772}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 2}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 2}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 2}, "pyerrors.input.misc.read_pbp": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_rwms": {"tf": 2.23606797749979}, "pyerrors.input.openQCD.extract_t0": {"tf": 2}, "pyerrors.input.openQCD.read_qtop": {"tf": 2.23606797749979}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 2.23606797749979}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1.7320508075688772}, "pyerrors.input.sfcf.read_sfcf": {"tf": 2.6457513110645907}}, "df": 17, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}}, "df": 3}}}}}, "p": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}}, "l": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "a": {"docs": {"pyerrors": {"tf": 2.6457513110645907}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.read_dobs": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.details": {"tf": 1}, "pyerrors.obs.Obs.plot_rep_dist": {"tf": 1}}, "df": 11, "s": {"docs": {"pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}}, "df": 1, "/": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 4}}}}}}}}}}}}, "u": {"docs": {}, "df": 0, "m": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.input.misc.read_pbp": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_rwms": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.4142135623730951}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.input.utils.check_idl": {"tf": 1}, "pyerrors.obs.Obs.plot_rep_dist": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1.4142135623730951}, "pyerrors.obs.merge_obs": {"tf": 1}}, "df": 12}}}, "k": {"docs": {}, "df": 0, "a": {"docs": {"pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}}, "df": 1}}}, "a": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.pandas.to_sql": {"tf": 1}}, "df": 4, "s": {"docs": {"pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 1}}}}}, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}}, "s": {"docs": {"pyerrors.fits.Fit_result": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 2}}}}}}}, "s": {"docs": {"pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}}, "df": 2, "[": {"0": {"docs": {"pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}}, "df": 2}, "docs": {}, "df": 0}}}, "s": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1}}, "df": 1, "e": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "h": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}, "u": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 4}, "pyerrors.correlators.Corr.show": {"tf": 1.4142135623730951}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.qqplot": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1}}, "df": 5, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.4142135623730951}}, "df": 3}}}, "s": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.fits.Fit_result": {"tf": 1.4142135623730951}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.ks_test": {"tf": 1}}, "df": 4}}}}, "p": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 2.449489742783178}, "pyerrors.correlators.Corr.deriv": {"tf": 1}, "pyerrors.correlators.Corr.second_deriv": {"tf": 1}}, "df": 3, "i": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}}}}}}, "l": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}}, "df": 2}}}}, "i": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.qqplot": {"tf": 1}, "pyerrors.fits.residual_plot": {"tf": 1}}, "df": 3}}}}}}, "c": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}}}}}, "q": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 2}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 8}, "s": {"docs": {"pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}}, "df": 6}, "m": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}}, "df": 2}}}}}}}}}, "g": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1, "l": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}}, "df": 2}}}}}}}, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.correlators.Corr.__init__": {"tf": 1}, "pyerrors.correlators.Corr.plateau": {"tf": 1.4142135623730951}}, "df": 2}}}}, "t": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.correlators.Corr.T_symmetry": {"tf": 1}, "pyerrors.correlators.Corr.deriv": {"tf": 1}, "pyerrors.correlators.Corr.second_deriv": {"tf": 1}, "pyerrors.covobs.Covobs.errsq": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.total_least_squares": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}, "pyerrors.roots.find_root": {"tf": 1}}, "df": 12, "s": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.correlators.Corr.item": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.m_eff": {"tf": 1}, "pyerrors.dirac.Grid_gamma": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.fits.fit_lin": {"tf": 1}, "pyerrors.fits.error_band": {"tf": 1}, "pyerrors.input.misc.read_pbp": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.openQCD.qtop_projection": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}, "pyerrors.obs.Obs.plot_piechart": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1.4142135623730951}, "pyerrors.obs.import_jackknife": {"tf": 1}, "pyerrors.roots.find_root": {"tf": 1}}, "df": 18}, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.GEVP": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.read_pobs": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.read_dobs": {"tf": 1.4142135623730951}, "pyerrors.input.json.import_json_string": {"tf": 1.4142135623730951}, "pyerrors.input.json.load_json": {"tf": 1.4142135623730951}, "pyerrors.input.json.load_json_dict": {"tf": 1.4142135623730951}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 8}}}}}}, "v": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.reverse": {"tf": 1}}, "df": 2, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}, "w": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.reweight": {"tf": 1}, "pyerrors.obs.Obs.reweight": {"tf": 1}, "pyerrors.obs.reweight": {"tf": 1}}, "df": 4, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.reweight": {"tf": 1}, "pyerrors.input.openQCD.qtop_projection": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.obs.Obs.reweight": {"tf": 1}, "pyerrors.obs.reweight": {"tf": 1}, "pyerrors.obs.correlate": {"tf": 1}}, "df": 7}}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.correlators.Corr.reweight": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.obs.Obs.reweight": {"tf": 1.4142135623730951}, "pyerrors.obs.reweight": {"tf": 1.4142135623730951}, "pyerrors.obs.correlate": {"tf": 1}}, "df": 5}}}}}}}}, "i": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.obs.Obs.reweight": {"tf": 1}}, "df": 1}}}}}}}}, "n": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "z": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}}}}}}}}, "f": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 2}}, "df": 4, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.GEVP": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.show": {"tf": 1}}, "df": 2, "s": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1}}, "df": 2}}}}}}}}, "c": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.m_eff": {"tf": 1}}, "df": 1}}}}}}}, "v": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}}, "df": 7}}}}}, "n": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.input.json.import_json_string": {"tf": 1}}, "df": 1}}}}}}}, "r": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 1}}, "g": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "z": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}}, "df": 1}}}}}}}}, "l": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.m_eff": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.dump": {"tf": 1}}, "df": 5}}}}}, "i": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}}, "df": 1}}}}}, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}}, "df": 3}}}}}, "d": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1.4142135623730951}}, "df": 1, "d": {"docs": {"pyerrors.fits.Fit_result": {"tf": 1}}, "df": 1}}}}}, "m": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 1, "d": {"docs": {"pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}}, "df": 4}}}}, "a": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.dobs.read_pobs": {"tf": 1}}, "df": 1}}}}}, "a": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}, "s": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}}, "df": 1}}}}, "n": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 2}, "pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.input.utils.check_idl": {"tf": 1}}, "df": 12, "s": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.obs.Obs.__init__": {"tf": 1}}, "df": 2}}}, "k": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}, "pyerrors.dirac.epsilon_tensor": {"tf": 1}, "pyerrors.dirac.epsilon_tensor_rank4": {"tf": 1}}, "df": 3}, "d": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "m": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}}}, "p": {"docs": {"pyerrors": {"tf": 1.7320508075688772}}, "df": 1}, "d": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 3}}}}, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {"pyerrors.obs.derived_observable": {"tf": 1}}, "df": 1}}}}, "h": {"docs": {}, "df": 0, "o": {"docs": {"pyerrors": {"tf": 2}}, "df": 1}}, "o": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "o": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.m_eff": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.roots.find_root": {"tf": 1.7320508075688772}}, "df": 4, "s": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}}}, "u": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.input": {"tf": 1}}, "df": 2}}}}}, "n": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.openQCD.read_qtop": {"tf": 1}}, "df": 1}}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}}}}, "w": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.7320508075688772}}, "df": 2}}, "i": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 2}}}}, "w": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.openQCD.read_rwms": {"tf": 1}}, "df": 1}}}, "u": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.obs.Obs.is_zero_within_error": {"tf": 1}}, "df": 1}}}, "v": {"1": {"docs": {}, "df": 0, "@": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "@": {"docs": {}, "df": 0, "v": {"2": {"docs": {"pyerrors.correlators.Corr.projected": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}}}}}, "docs": {"pyerrors.correlators.Corr.GEVP": {"tf": 2.23606797749979}, "pyerrors.correlators.Corr.prune": {"tf": 2.23606797749979}, "pyerrors.obs.covariance": {"tf": 1.4142135623730951}}, "df": 3, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 3.1622776601683795}, "pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.correlators.Corr.plottable": {"tf": 1}, "pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.prune": {"tf": 1.4142135623730951}, "pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.fits.Fit_result": {"tf": 2}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.linalg.svd": {"tf": 1}, "pyerrors.misc.pseudo_Obs": {"tf": 1.7320508075688772}, "pyerrors.misc.gen_correlated_data": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 2.449489742783178}, "pyerrors.obs.Obs.gamma_method": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.gm": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1}, "pyerrors.obs.import_jackknife": {"tf": 1}, "pyerrors.obs.cov_Obs": {"tf": 1}}, "df": 20, "s": {"docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.correlators.Corr.show": {"tf": 1.7320508075688772}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.fits.error_band": {"tf": 1}, "pyerrors.fits.ks_test": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1.7320508075688772}}, "df": 7}, "d": {"docs": {"pyerrors": {"tf": 2}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1.4142135623730951}, "pyerrors.input.pandas.to_sql": {"tf": 1}, "pyerrors.input.pandas.read_sql": {"tf": 1}, "pyerrors.input.pandas.dump_df": {"tf": 1}, "pyerrors.linalg.matmul": {"tf": 1}, "pyerrors.linalg.jack_matmul": {"tf": 1}, "pyerrors.linalg.einsum": {"tf": 1}, "pyerrors.linalg.inv": {"tf": 1}, "pyerrors.linalg.cholesky": {"tf": 1}, "pyerrors.linalg.det": {"tf": 1}, "pyerrors.obs.CObs": {"tf": 1}, "pyerrors.roots.find_root": {"tf": 1}}, "df": 13}}}, "i": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}}, "df": 2}}}, "c": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.input": {"tf": 1}}, "df": 2}}}, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.deriv": {"tf": 1}, "pyerrors.correlators.Corr.second_deriv": {"tf": 1}, "pyerrors.correlators.Corr.m_eff": {"tf": 1.4142135623730951}}, "df": 3, "s": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}}, "c": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.covobs.Covobs.errsq": {"tf": 1}}, "df": 2}}}, "b": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1.7320508075688772}}, "df": 1}}}}}}}, "n": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.dobs.import_dobs_string": {"tf": 1}}, "df": 1}}}}}}}}, "i": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "a": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}, "a": {"docs": {"pyerrors": {"tf": 3.7416573867739413}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1}, "pyerrors.fits.Fit_result": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.pandas.dump_df": {"tf": 1}, "pyerrors.linalg.matmul": {"tf": 1}}, "df": 8}, "s": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "z": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.obs.Obs": {"tf": 1}}, "df": 1, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}, "e": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1, "d": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1.4142135623730951}}, "df": 1}}}}}}}}, "e": {"docs": {}, "df": 0, "w": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}}, "df": 1}}}, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}}, "df": 1}}}}, "e": {"docs": {}, "df": 0, "x": {"docs": {"pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}}, "df": 1}}}, "y": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 3}, "s": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 2.23606797749979}, "pyerrors.input.sfcf.read_sfcf": {"tf": 2.449489742783178}}, "df": 6}}}}, "i": {"docs": {}, "df": 0, "f": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}, "b": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}}, "df": 3}}}}}, "c": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1.7320508075688772}}, "df": 2, "s": {"docs": {"pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1.4142135623730951}}, "df": 2}}}}}}, "\\": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "\\": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "{": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "}": {"docs": {}, "df": 0, "^": {"docs": {}, "df": 0, "m": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}}}}}}}}}}}}}}, "u": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.m_eff": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.prune": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1.4142135623730951}, "pyerrors.input.pandas.dump_df": {"tf": 1}, "pyerrors.linalg.jack_matmul": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 2}, "pyerrors.roots.find_root": {"tf": 1}}, "df": 15, "d": {"docs": {"pyerrors": {"tf": 2.8284271247461903}, "pyerrors.correlators.Corr.deriv": {"tf": 1}, "pyerrors.correlators.Corr.second_deriv": {"tf": 1}, "pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1.4142135623730951}, "pyerrors.fits.least_squares": {"tf": 2}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.fits.fit_lin": {"tf": 1.4142135623730951}, "pyerrors.fits.qqplot": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_rwms": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.7320508075688772}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}, "pyerrors.obs.Obs.is_zero_within_error": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1.4142135623730951}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 25}, "r": {"docs": {"pyerrors": {"tf": 2}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 2}, "s": {"docs": {"pyerrors.correlators.Corr.m_eff": {"tf": 1.7320508075688772}, "pyerrors.input.hadrons.Npr_matrix.g5H": {"tf": 1}}, "df": 2}, "f": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.input.openQCD.read_rwms": {"tf": 1}}, "df": 1}}}}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors": {"tf": 2}, "pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}, "pyerrors.obs.derived_observable": {"tf": 1}}, "df": 5}}}, "u": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.correlators.Corr.projected": {"tf": 1}}, "df": 1}}}}}}, "l": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "i": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}, "n": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "q": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 2}}, "df": 1}}}}}, "t": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.4142135623730951}}, "df": 5}}}, "s": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.projected": {"tf": 1.4142135623730951}}, "df": 1}}}}}}}, "c": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}}, "df": 1}}}}}}}}}, "h": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 3}}}}}}}, "l": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}}}, "p": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "k": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}}, "df": 2}}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}}, "df": 1}}}}}}}, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}}, "df": 2}}}}, "p": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 3}, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.utils": {"tf": 1}}, "df": 1}}}}}}}}}, "j": {"docs": {"pyerrors": {"tf": 2}, "pyerrors.correlators.Corr.item": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.prune": {"tf": 1.4142135623730951}, "pyerrors.obs.covariance": {"tf": 1.7320508075688772}}, "df": 4, "o": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "w": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}, "u": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix.g5H": {"tf": 1}}, "df": 3, "u": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}, "l": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "a": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}}}, "m": {"docs": {}, "df": 0, "p": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1, "s": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}}}, "a": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "c": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "i": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}}, "df": 1, "a": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.obs.derived_observable": {"tf": 1}}, "df": 2}}}}}, "k": {"docs": {}, "df": 0, "k": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "f": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input": {"tf": 2.23606797749979}, "pyerrors.linalg.jack_matmul": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 2}, "pyerrors.obs.import_jackknife": {"tf": 1.4142135623730951}}, "df": 4}}}}}, "s": {"docs": {"pyerrors.obs.import_jackknife": {"tf": 1}}, "df": 1}}}}, "s": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 3.7416573867739413}, "pyerrors.correlators.Corr.dump": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1.7320508075688772}, "pyerrors.input.json.dump_to_json": {"tf": 2.23606797749979}, "pyerrors.input.json.import_json_string": {"tf": 1.7320508075688772}, "pyerrors.input.json.load_json": {"tf": 1.4142135623730951}, "pyerrors.input.json.dump_dict_to_json": {"tf": 2.449489742783178}, "pyerrors.input.json.load_json_dict": {"tf": 1.4142135623730951}, "pyerrors.input.pandas.to_sql": {"tf": 1}, "pyerrors.input.pandas.dump_df": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.dump": {"tf": 1.4142135623730951}}, "df": 12}}}, "l": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "}": {"docs": {}, "df": 0, "v": {"docs": {"pyerrors.obs.covariance": {"tf": 1.4142135623730951}}, "df": 1}}, "^": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}, "k": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1.7320508075688772}}, "df": 2, "u": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "k": {"docs": {}, "df": 0, "i": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}, "h": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}}, "e": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "p": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.thin": {"tf": 1}, "pyerrors.obs.correlate": {"tf": 1}}, "df": 3}}, "y": {"docs": {}, "df": 0, "w": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1.4142135623730951}}, "df": 4, "s": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}}}}}}, "o": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, "\u2013": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "v": {"docs": {"pyerrors.fits.ks_test": {"tf": 1}}, "df": 1}}}}}}}}}}}}}}}}}, "a": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "a": {"1": {"docs": {"pyerrors.input.bdio.read_mesons": {"tf": 1}}, "df": 1}, "2": {"docs": {"pyerrors.input.bdio.read_mesons": {"tf": 1}}, "df": 1}, "docs": {"pyerrors.input.bdio.read_dSdm": {"tf": 1}}, "df": 1}}}}, "w": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1.7320508075688772}}, "df": 2}}}}}}, "h": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "p": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 2}, "l": {"docs": {}, "df": 0, "p": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "r": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix.g5H": {"tf": 1.4142135623730951}}, "df": 2}}}, "i": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.linalg.eigh": {"tf": 1}}, "df": 1}}}}}}, "e": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 3}}, "a": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 2}}}}}, "a": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}, "e": {"docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.correlators.Corr": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1.4142135623730951}, "pyerrors.obs.correlate": {"tf": 1}, "pyerrors.roots.find_root": {"tf": 1}}, "df": 19}}, "n": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 2, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1, "d": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 1}}}}}, "k": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.correlators.Corr.Hankel": {"tf": 1.4142135623730951}}, "df": 1}}}}, "s": {"docs": {"pyerrors": {"tf": 2.6457513110645907}, "pyerrors.correlators.Corr.reweight": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.total_least_squares": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.4142135623730951}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.obs.Obs.reweight": {"tf": 1}, "pyerrors.obs.reweight": {"tf": 1}, "pyerrors.obs.correlate": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 18}, "d": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}}, "df": 5}}}}}}, "i": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.plot_history": {"tf": 1.4142135623730951}}, "df": 2}, "i": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}}, "d": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1.4142135623730951}}, "df": 1, "s": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1}}, "df": 1}}}}, "o": {"docs": {}, "df": 0, "w": {"docs": {"pyerrors": {"tf": 2}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.pandas.to_sql": {"tf": 1}}, "df": 5}, "s": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1}}, "df": 1, "n": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "z": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1}}, "df": 1}}}}}}}}, "l": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1, "s": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 1}}}, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.fits.Fit_result": {"tf": 1}}, "df": 1}}}}}}}, "m": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}}}}}}}}}, "t": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, ":": {"docs": {}, "df": 0, "/": {"docs": {}, "df": 0, "/": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "f": {"docs": {"pyerrors.dirac.epsilon_tensor": {"tf": 1}, "pyerrors.dirac.epsilon_tensor_rank4": {"tf": 1}}, "df": 2}}}}}}}}}}}}}}}, "d": {"docs": {}, "df": 0, "f": {"5": {"docs": {"pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}}, "df": 5}, "docs": {}, "df": 0}}, "u": {"docs": {}, "df": 0, "a": {"docs": {"pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 1}}}, "q": {"docs": {}, "df": 0, "c": {"docs": {"pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}}, "df": 1, "d": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "u": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "n": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 2.23606797749979}}, "df": 1}, "i": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 2.449489742783178}}, "df": 1}}}}, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.fits.least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.qqplot": {"tf": 1.4142135623730951}}, "df": 2}}}, "u": {"docs": {}, "df": 0, "m": {"docs": {"pyerrors.correlators.Corr.T_symmetry": {"tf": 1}}, "df": 1}}}}, "r": {"docs": {}, "df": 0, "k": {"docs": {"pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1.7320508075688772}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.4142135623730951}}, "df": 2, "s": {"docs": {"pyerrors.input.sfcf.read_sfcf": {"tf": 1.4142135623730951}}, "df": 1}}}}, "o": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 4}}, "df": 1, "e": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.input.pandas.read_sql": {"tf": 1.4142135623730951}}, "df": 1}}}}, "q": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}}, "df": 2}}}}}, "t": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "p": {"docs": {"pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 2}}}}, "z": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "o": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}, "pyerrors.obs.Obs.is_zero_within_error": {"tf": 1}, "pyerrors.obs.Obs.is_zero": {"tf": 1}, "pyerrors.obs.CObs.is_zero": {"tf": 1}}, "df": 12, "t": {"docs": {}, "df": 0, "h": {"docs": {"pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1}, "pyerrors.obs.import_jackknife": {"tf": 1}}, "df": 3}}, "s": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}}, "df": 1}}}, "u": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.4142135623730951}}, "df": 10}}}}}}}}}}, "pipeline": ["trimmer"], "_isPrebuiltIndex": true};
+ /** pdoc search index */const docs = {"version": "0.9.5", "fields": ["qualname", "fullname", "annotation", "default_value", "signature", "bases", "doc"], "ref": "fullname", "documentStore": {"docs": {"pyerrors": {"fullname": "pyerrors", "modulename": "pyerrors", "kind": "module", "doc": "What is pyerrors?
\n\npyerrors
is a python package for error computation and propagation of Markov chain Monte Carlo data.\nIt is based on the gamma method arXiv:hep-lat/0306017. Some of its features are:
\n\n\n- automatic differentiation for exact linear error propagation as suggested in arXiv:1809.01289 (partly based on the autograd package).
\n- treatment of slow modes in the simulation as suggested in arXiv:1009.5228.
\n- coherent error propagation for data from different Markov chains.
\n- non-linear fits with x- and y-errors and exact linear error propagation based on automatic differentiation as introduced in arXiv:1809.01289.
\n- real and complex matrix operations and their error propagation based on automatic differentiation (Matrix inverse, Cholesky decomposition, calculation of eigenvalues and eigenvectors, singular value decomposition...).
\n
\n\nMore detailed examples can found in the GitHub repository
.
\n\nIf you use pyerrors
for research that leads to a publication please consider citing:
\n\n\n- Fabian Joswig, Simon Kuberski, Justus T. Kuhlmann, Jan Neuendorf, pyerrors: a python framework for error analysis of Monte Carlo data. [arXiv:2209.14371 [hep-lat]].
\n- Ulli Wolff, Monte Carlo errors with less errors. Comput.Phys.Commun. 156 (2004) 143-153, Comput.Phys.Commun. 176 (2007) 383 (erratum).
\n- Alberto Ramos, Automatic differentiation for error analysis of Monte Carlo data. Comput.Phys.Commun. 238 (2019) 19-35.
\n
\n\nand
\n\n\n- Stefan Schaefer, Rainer Sommer, Francesco Virotta, Critical slowing down and error analysis in lattice QCD simulations. Nucl.Phys.B 845 (2011) 93-119.
\n
\n\nwhere applicable.
\n\nThere exist similar publicly available implementations of gamma method error analysis suites in Fortran, Julia and Python.
\n\nBasic example
\n\n\n
import numpy as np\nimport pyerrors as pe\n\nmy_obs = pe.Obs([samples], ['ensemble_name']) # Initialize an Obs object\nmy_new_obs = 2 * np.log(my_obs) / my_obs ** 2 # Construct derived Obs object\nmy_new_obs.gamma_method() # Estimate the statistical error\nprint(my_new_obs) # Print the result to stdout\n> 0.31498(72)\n
\n
\n\nThe Obs
class
\n\npyerrors
introduces a new datatype, Obs
, which simplifies error propagation and estimation for auto- and cross-correlated data.\nAn Obs
object can be initialized with two arguments, the first is a list containing the samples for an observable from a Monte Carlo chain.\nThe samples can either be provided as python list or as numpy array.\nThe second argument is a list containing the names of the respective Monte Carlo chains as strings. These strings uniquely identify a Monte Carlo chain/ensemble.
\n\n\n
import pyerrors as pe\n\nmy_obs = pe.Obs([samples], ['ensemble_name'])\n
\n
\n\nError propagation
\n\nWhen performing mathematical operations on Obs
objects the correct error propagation is intrinsically taken care of using a first order Taylor expansion\n$$\\delta_f^i=\\sum_\\alpha \\bar{f}_\\alpha \\delta_\\alpha^i\\,,\\quad \\delta_\\alpha^i=a_\\alpha^i-\\bar{a}_\\alpha\\,,$$\nas introduced in arXiv:hep-lat/0306017.\nThe required derivatives $\\bar{f}_\\alpha$ are evaluated up to machine precision via automatic differentiation as suggested in arXiv:1809.01289.
\n\nThe Obs
class is designed such that mathematical numpy functions can be used on Obs
just as for regular floats.
\n\n\n
import numpy as np\nimport pyerrors as pe\n\nmy_obs1 = pe.Obs([samples1], ['ensemble_name'])\nmy_obs2 = pe.Obs([samples2], ['ensemble_name'])\n\nmy_sum = my_obs1 + my_obs2\n\nmy_m_eff = np.log(my_obs1 / my_obs2)\n\niamzero = my_m_eff - my_m_eff\n# Check that value and fluctuations are zero within machine precision\nprint(iamzero == 0.0)\n> True\n
\n
\n\nError estimation
\n\nThe error estimation within pyerrors
is based on the gamma method introduced in arXiv:hep-lat/0306017.\nAfter having arrived at the derived quantity of interest the gamma_method
can be called as detailed in the following example.
\n\n\n
my_sum.gamma_method()\nprint(my_sum)\n> 1.70(57)\nmy_sum.details()\n> Result 1.70000000e+00 +/- 5.72046658e-01 +/- 7.56746598e-02 (33.650%)\n> t_int 2.71422900e+00 +/- 6.40320983e-01 S = 2.00\n> 1000 samples in 1 ensemble:\n> \u00b7 Ensemble 'ensemble_name' : 1000 configurations (from 1 to 1000)\n
\n
\n\nWe use the following definition of the integrated autocorrelation time established in Madras & Sokal 1988\n$$\\tau_\\mathrm{int}=\\frac{1}{2}+\\sum_{t=1}^{W}\\rho(t)\\geq \\frac{1}{2}\\,.$$\nThe window $W$ is determined via the automatic windowing procedure described in arXiv:hep-lat/0306017.\nThe standard value for the parameter $S$ of this automatic windowing procedure is $S=2$. Other values for $S$ can be passed to the gamma_method
as parameter.
\n\n\n
my_sum.gamma_method(S=3.0)\nmy_sum.details()\n> Result 1.70000000e+00 +/- 6.30675201e-01 +/- 1.04585650e-01 (37.099%)\n> t_int 3.29909703e+00 +/- 9.77310102e-01 S = 3.00\n> 1000 samples in 1 ensemble:\n> \u00b7 Ensemble 'ensemble_name' : 1000 configurations (from 1 to 1000)\n
\n
\n\nThe integrated autocorrelation time $\\tau_\\mathrm{int}$ and the autocorrelation function $\\rho(W)$ can be monitored via the methods pyerrors.obs.Obs.plot_tauint
and pyerrors.obs.Obs.plot_rho
.
\n\nIf the parameter $S$ is set to zero it is assumed that the dataset does not exhibit any autocorrelation and the window size is chosen to be zero.\nIn this case the error estimate is identical to the sample standard error.
\n\nExponential tails
\n\nSlow modes in the Monte Carlo history can be accounted for by attaching an exponential tail to the autocorrelation function $\\rho$ as suggested in arXiv:1009.5228. The longest autocorrelation time in the history, $\\tau_\\mathrm{exp}$, can be passed to the gamma_method
as parameter. In this case the automatic windowing procedure is vacated and the parameter $S$ does not affect the error estimate.
\n\n\n
my_sum.gamma_method(tau_exp=7.2)\nmy_sum.details()\n> Result 1.70000000e+00 +/- 6.28097762e-01 +/- 5.79077524e-02 (36.947%)\n> t_int 3.27218667e+00 +/- 7.99583654e-01 tau_exp = 7.20, N_sigma = 1\n> 1000 samples in 1 ensemble:\n> \u00b7 Ensemble 'ensemble_name' : 1000 configurations (from 1 to 1000)\n
\n
\n\nFor the full API see pyerrors.obs.Obs.gamma_method
.
\n\nMultiple ensembles/replica
\n\nError propagation for multiple ensembles (Markov chains with different simulation parameters) is handled automatically. Ensembles are uniquely identified by their name
.
\n\n\n
obs1 = pe.Obs([samples1], ['ensemble1'])\nobs2 = pe.Obs([samples2], ['ensemble2'])\n\nmy_sum = obs1 + obs2\nmy_sum.details()\n> Result 2.00697958e+00\n> 1500 samples in 2 ensembles:\n> \u00b7 Ensemble 'ensemble1' : 1000 configurations (from 1 to 1000)\n> \u00b7 Ensemble 'ensemble2' : 500 configurations (from 1 to 500)\n
\n
\n\npyerrors
identifies multiple replica (independent Markov chains with identical simulation parameters) by the vertical bar |
in the name of the data set.
\n\n\n
obs1 = pe.Obs([samples1], ['ensemble1|r01'])\nobs2 = pe.Obs([samples2], ['ensemble1|r02'])\n\n> my_sum = obs1 + obs2\n> my_sum.details()\n> Result 2.00697958e+00\n> 1500 samples in 1 ensemble:\n> \u00b7 Ensemble 'ensemble1'\n> \u00b7 Replicum 'r01' : 1000 configurations (from 1 to 1000)\n> \u00b7 Replicum 'r02' : 500 configurations (from 1 to 500)\n
\n
\n\nError estimation for multiple ensembles
\n\nIn order to keep track of different error analysis parameters for different ensembles one can make use of global dictionaries as detailed in the following example.
\n\n\n
pe.Obs.S_dict['ensemble1'] = 2.5\npe.Obs.tau_exp_dict['ensemble2'] = 8.0\npe.Obs.tau_exp_dict['ensemble3'] = 2.0\n
\n
\n\nIn case the gamma_method
is called without any parameters it will use the values specified in the dictionaries for the respective ensembles.\nPassing arguments to the gamma_method
still dominates over the dictionaries.
\n\nIrregular Monte Carlo chains
\n\nObs
objects defined on irregular Monte Carlo chains can be initialized with the parameter idl
.
\n\n\n
# Observable defined on configurations 20 to 519\nobs1 = pe.Obs([samples1], ['ensemble1'], idl=[range(20, 520)])\nobs1.details()\n> Result 9.98319881e-01\n> 500 samples in 1 ensemble:\n> \u00b7 Ensemble 'ensemble1' : 500 configurations (from 20 to 519)\n\n# Observable defined on every second configuration between 5 and 1003\nobs2 = pe.Obs([samples2], ['ensemble1'], idl=[range(5, 1005, 2)])\nobs2.details()\n> Result 9.99100712e-01\n> 500 samples in 1 ensemble:\n> \u00b7 Ensemble 'ensemble1' : 500 configurations (from 5 to 1003 in steps of 2)\n\n# Observable defined on configurations 2, 9, 28, 29 and 501\nobs3 = pe.Obs([samples3], ['ensemble1'], idl=[[2, 9, 28, 29, 501]])\nobs3.details()\n> Result 1.01718064e+00\n> 5 samples in 1 ensemble:\n> \u00b7 Ensemble 'ensemble1' : 5 configurations (irregular range)\n
\n
\n\nObs
objects defined on regular and irregular histories of the same ensemble can be combined with each other and the correct error propagation and estimation is automatically taken care of.
\n\nWarning: Irregular Monte Carlo chains can result in odd patterns in the autocorrelation functions.\nMake sure to check the autocorrelation time with e.g. pyerrors.obs.Obs.plot_rho
or pyerrors.obs.Obs.plot_tauint
.
\n\nFor the full API see pyerrors.obs.Obs
.
\n\nCorrelators
\n\nWhen one is not interested in single observables but correlation functions, pyerrors
offers the Corr
class which simplifies the corresponding error propagation and provides the user with a set of standard methods. In order to initialize a Corr
objects one needs to arrange the data as a list of Obs
\n\n\n
my_corr = pe.Corr([obs_0, obs_1, obs_2, obs_3])\nprint(my_corr)\n> x0/a Corr(x0/a)\n> ------------------\n> 0 0.7957(80)\n> 1 0.5156(51)\n> 2 0.3227(33)\n> 3 0.2041(21)\n
\n
\n\nIn case the correlation functions are not defined on the outermost timeslices, for example because of fixed boundary conditions, a padding can be introduced.
\n\n\n
my_corr = pe.Corr([obs_0, obs_1, obs_2, obs_3], padding=[1, 1])\nprint(my_corr)\n> x0/a Corr(x0/a)\n> ------------------\n> 0\n> 1 0.7957(80)\n> 2 0.5156(51)\n> 3 0.3227(33)\n> 4 0.2041(21)\n> 5\n
\n
\n\nThe individual entries of a correlator can be accessed via slicing
\n\n\n
print(my_corr[3])\n> 0.3227(33)\n
\n
\n\nError propagation with the Corr
class works very similar to Obs
objects. Mathematical operations are overloaded and Corr
objects can be computed together with other Corr
objects, Obs
objects or real numbers and integers.
\n\n\n
my_new_corr = 0.3 * my_corr[2] * my_corr * my_corr + 12 / my_corr\n
\n
\n\npyerrors
provides the user with a set of regularly used methods for the manipulation of correlator objects:
\n\n\nCorr.gamma_method
applies the gamma method to all entries of the correlator. \nCorr.m_eff
to construct effective masses. Various variants for periodic and fixed temporal boundary conditions are available. \nCorr.deriv
returns the first derivative of the correlator as Corr
. Different discretizations of the numerical derivative are available. \nCorr.second_deriv
returns the second derivative of the correlator as Corr
. Different discretizations of the numerical derivative are available. \nCorr.symmetric
symmetrizes parity even correlations functions, assuming periodic boundary conditions. \nCorr.anti_symmetric
anti-symmetrizes parity odd correlations functions, assuming periodic boundary conditions. \nCorr.T_symmetry
averages a correlator with its time symmetry partner, assuming fixed boundary conditions. \nCorr.plateau
extracts a plateau value from the correlator in a given range. \nCorr.roll
periodically shifts the correlator. \nCorr.reverse
reverses the time ordering of the correlator. \nCorr.correlate
constructs a disconnected correlation function from the correlator and another Corr
or Obs
object. \nCorr.reweight
reweights the correlator. \n
\n\npyerrors
can also handle matrices of correlation functions and extract energy states from these matrices via a generalized eigenvalue problem (see pyerrors.correlators.Corr.GEVP
).
\n\nFor the full API see pyerrors.correlators.Corr
.
\n\nComplex valued observables
\n\npyerrors
can handle complex valued observables via the class pyerrors.obs.CObs
.\nCObs
are initialized with a real and an imaginary part which both can be Obs
valued.
\n\n\n
my_real_part = pe.Obs([samples1], ['ensemble1'])\nmy_imag_part = pe.Obs([samples2], ['ensemble1'])\n\nmy_cobs = pe.CObs(my_real_part, my_imag_part)\nmy_cobs.gamma_method()\nprint(my_cobs)\n> (0.9959(91)+0.659(28)j)\n
\n
\n\nElementary mathematical operations are overloaded and samples are properly propagated as for the Obs
class.
\n\n\n
my_derived_cobs = (my_cobs + my_cobs.conjugate()) / np.abs(my_cobs)\nmy_derived_cobs.gamma_method()\nprint(my_derived_cobs)\n> (1.668(23)+0.0j)\n
\n
\n\nThe Covobs
class
\n\nIn many projects, auxiliary data that is not based on Monte Carlo chains enters. Examples are experimentally determined mesons masses which are used to set the scale or renormalization constants. These numbers come with an error that has to be propagated through the analysis. The Covobs
class allows to define such quantities in pyerrors
. Furthermore, external input might consist of correlated quantities. An example are the parameters of an interpolation formula, which are defined via mean values and a covariance matrix between all parameters. The contribution of the interpolation formula to the error of a derived quantity therefore might depend on the complete covariance matrix.
\n\nThis concept is built into the definition of Covobs
. In pyerrors
, external input is defined by $M$ mean values, a $M\\times M$ covariance matrix, where $M=1$ is permissible, and a name that uniquely identifies the covariance matrix. Below, we define the pion mass, based on its mean value and error, 134.9768(5). Note, that the square of the error enters cov_Obs
, since the second argument of this function is the covariance matrix of the Covobs
.
\n\n\n
import pyerrors.obs as pe\n\nmpi = pe.cov_Obs(134.9768, 0.0005**2, 'pi^0 mass')\nmpi.gamma_method()\nmpi.details()\n> Result 1.34976800e+02 +/- 5.00000000e-04 +/- 0.00000000e+00 (0.000%)\n> pi^0 mass 5.00000000e-04\n> 0 samples in 1 ensemble:\n> \u00b7 Covobs 'pi^0 mass'\n
\n
\n\nThe resulting object mpi
is an Obs
that contains a Covobs
. In the following, it may be handled as any other Obs
. The contribution of the covariance matrix to the error of an Obs
is determined from the $M \\times M$ covariance matrix $\\Sigma$ and the gradient of the Obs
with respect to the external quantities, which is the $1\\times M$ Jacobian matrix $J$, via\n$$s = \\sqrt{J^T \\Sigma J}\\,,$$\nwhere the Jacobian is computed for each derived quantity via automatic differentiation.
\n\nCorrelated auxiliary data is defined similarly to above, e.g., via
\n\n\n
RAP = pe.cov_Obs([16.7457, -19.0475], [[3.49591, -6.07560], [-6.07560, 10.5834]], 'R_AP, 1906.03445, (5.3a)')\nprint(RAP)\n> [Obs[16.7(1.9)], Obs[-19.0(3.3)]]\n
\n
\n\nwhere RAP
now is a list of two Obs
that contains the two correlated parameters.
\n\nSince the gradient of a derived observable with respect to an external covariance matrix is propagated through the entire analysis, the Covobs
class allows to quote the derivative of a result with respect to the external quantities. If these derivatives are published together with the result, small shifts in the definition of external quantities, e.g., the definition of the physical point, can be performed a posteriori based on the published information. This may help to compare results of different groups. The gradient of an Obs
o
with respect to a covariance matrix with the identifying string k
may be accessed via
\n\n\n\nError propagation in iterative algorithms
\n\npyerrors
supports exact linear error propagation for iterative algorithms like various variants of non-linear least squares fits or root finding. The derivatives required for the error propagation are calculated as described in arXiv:1809.01289.
\n\nLeast squares fits
\n\nStandard non-linear least square fits with errors on the dependent but not the independent variables can be performed with pyerrors.fits.least_squares
. As default solver the Levenberg-Marquardt algorithm implemented in scipy is used.
\n\nFit functions have to be of the following form
\n\n\n
import autograd.numpy as anp\n\ndef func(a, x):\n return a[1] * anp.exp(-a[0] * x)\n
\n
\n\nIt is important that numerical functions refer to autograd.numpy
instead of numpy
for the automatic differentiation in iterative algorithms to work properly.
\n\nFits can then be performed via
\n\n\n
fit_result = pe.fits.least_squares(x, y, func)\nprint("\\n", fit_result)\n> Fit with 2 parameters\n> Method: Levenberg-Marquardt\n> `ftol` termination condition is satisfied.\n> chisquare/d.o.f.: 0.9593035785160936\n\n> Goodness of fit:\n> \u03c7\u00b2/d.o.f. = 0.959304\n> p-value = 0.5673\n> Fit parameters:\n> 0 0.0548(28)\n> 1 1.933(64)\n
\n
\n\nwhere x is a list
or numpy.array
of floats
and y is a list
or numpy.array
of Obs
.
\n\nData stored in Corr
objects can be fitted directly using the Corr.fit
method.
\n\n\n
my_corr = pe.Corr(y)\nfit_result = my_corr.fit(func, fitrange=[12, 25])\n
\n
\n\nthis can simplify working with absolute fit ranges and takes care of gaps in the data automatically.
\n\nFor fit functions with multiple independent variables the fit function can be of the form
\n\n\n
def func(a, x):\n (x1, x2) = x\n return a[0] * x1 ** 2 + a[1] * x2\n
\n
\n\npyerrors
also supports correlated fits which can be triggered via the parameter correlated_fit=True
.\nDetails about how the required covariance matrix is estimated can be found in pyerrors.obs.covariance
.
\n\nDirect visualizations of the performed fits can be triggered via resplot=True
or qqplot=True
. For all available options see pyerrors.fits.least_squares
.
\n\nTotal least squares fits
\n\npyerrors
can also fit data with errors on both the dependent and independent variables using the total least squares method also referred to orthogonal distance regression as implemented in scipy, see pyerrors.fits.least_squares
. The syntax is identical to the standard least squares case, the only difference being that x
also has to be a list
or numpy.array
of Obs
.
\n\nFor the full API see pyerrors.fits
for fits and pyerrors.roots
for finding roots of functions.
\n\nMatrix operations
\n\npyerrors
provides wrappers for Obs
- and CObs
-valued matrix operations based on numpy.linalg
. The supported functions include:
\n\n\ninv
for the matrix inverse. \ncholseky
for the Cholesky decomposition. \ndet
for the matrix determinant. \neigh
for eigenvalues and eigenvectors of hermitean matrices. \neig
for eigenvalues of general matrices. \npinv
for the Moore-Penrose pseudoinverse. \nsvd
for the singular-value-decomposition. \n
\n\nFor the full API see pyerrors.linalg
.
\n\nExport data
\n\nThe preferred exported file format within pyerrors
is json.gz. Files written to this format are valid JSON files that have been compressed using gzip. The structure of the content is inspired by the dobs format of the ALPHA collaboration. The aim of the format is to facilitate the storage of data in a self-contained way such that, even years after the creation of the file, it is possible to extract all necessary information:
\n\n\n- What observables are stored? Possibly: How exactly are they defined.
\n- How does each single ensemble or external quantity contribute to the error of the observable?
\n- Who did write the file when and on which machine?
\n
\n\nThis can be achieved by storing all information in one single file. The export routines of pyerrors
are written such that as much information as possible is written automatically as described in the following example
\n\n\n
my_obs = pe.Obs([samples], ["test_ensemble"])\nmy_obs.tag = "My observable"\n\npe.input.json.dump_to_json(my_obs, "test_output_file", description="This file contains a test observable")\n# For a single observable one can equivalently use the class method dump\nmy_obs.dump("test_output_file", description="This file contains a test observable")\n\ncheck = pe.input.json.load_json("test_output_file")\n\nprint(my_obs == check)\n> True\n
\n
\n\nThe format also allows to directly write out the content of Corr
objects or lists and arrays of Obs
objects by passing the desired data to pyerrors.input.json.dump_to_json
.
\n\n\n\nThe first entries of the file provide optional auxiliary information:
\n\n\nprogram
is a string that indicates which program was used to write the file. \nversion
is a string that specifies the version of the format. \nwho
is a string that specifies the user name of the creator of the file. \ndate
is a string and contains the creation date of the file. \nhost
is a string and contains the hostname of the machine where the file has been written. \ndescription
contains information on the content of the file. This field is not filled automatically in pyerrors
. The user is advised to provide as detailed information as possible in this field. Examples are: Input files of measurements or simulations, LaTeX formulae or references to publications to specify how the observables have been computed, details on the analysis strategy, ... This field may be any valid JSON type. Strings, arrays or objects (equivalent to dicts in python) are well suited to provide information. \n
\n\nThe only necessary entry of the file is the field\n-obsdata
, an array that contains the actual data.
\n\nEach entry of the array belongs to a single structure of observables. Currently, these structures can be either of Obs
, list
, numpy.ndarray
, Corr
. All Obs
inside a structure (with dimension > 0) have to be defined on the same set of configurations. Different structures, that are represented by entries of the array obsdata
, are treated independently. Each entry of the array obsdata
has the following required entries:
\n\n\ntype
is a string that specifies the type of the structure. This allows to parse the content to the correct form after reading the file. It is always possible to interpret the content as list of Obs. \nvalue
is an array that contains the mean values of the Obs inside the structure.\nThe following entries are optional: \nlayout
is a string that specifies the layout of multi-dimensional structures. Examples are \"2, 2\" for a 2x2 dimensional matrix or \"64, 4, 4\" for a Corr with $T=64$ and 4x4 matrices on each time slices. \"1\" denotes a single Obs. Multi-dimensional structures are stored in row-major format (see below). \ntag
is any JSON type. It contains additional information concerning the structure. The tag
of an Obs
in pyerrors
is written here. \nreweighted
is a Bool that may be used to specify, whether the Obs
in the structure have been reweighted. \ndata
is an array that contains the data from MC chains. We will define it below. \ncdata
is an array that contains the data from external quantities with an error (Covobs
in pyerrors
). We will define it below. \n
\n\nThe array data
contains the data from MC chains. Each entry of the array corresponds to one ensemble and contains:
\n\n\nid
, a string that contains the name of the ensemble \nreplica
, an array that contains an entry per replica of the ensemble. \n
\n\nEach entry of replica
contains\nname
, a string that contains the name of the replica\ndeltas
, an array that contains the actual data.
\n\nEach entry in deltas
corresponds to one configuration of the replica and has $1+N$ many entries. The first entry is an integer that specifies the configuration number that, together with ensemble and replica name, may be used to uniquely identify the configuration on which the data has been obtained. The following N entries specify the deltas, i.e., the deviation of the observable from the mean value on this configuration, of each Obs
inside the structure. Multi-dimensional structures are stored in a row-major format. For primary observables, such as correlation functions, $value + delta_i$ matches the primary data obtained on the configuration.
\n\nThe array cdata
contains information about the contribution of auxiliary observables, represented by Covobs
in pyerrors
, to the total error of the observables. Each entry of the array belongs to one auxiliary covariance matrix and contains:
\n\n\nid
, a string that identifies the covariance matrix \nlayout
, a string that defines the dimensions of the $M\\times M$ covariance matrix (has to be \"M, M\" or \"1\"). \ncov
, an array that contains the $M\\times M$ many entries of the covariance matrix, stored in row-major format. \ngrad
, an array that contains N entries, one for each Obs
inside the structure. Each entry itself is an array, that contains the M gradients of the Nth observable with respect to the quantity that corresponds to the Mth diagonal entry of the covariance matrix. \n
\n\nA JSON schema that may be used to verify the correctness of a file with respect to the format definition is stored in ./examples/json_schema.json. The schema is a self-descriptive format definition and contains an exemplary file.
\n\nJulia I/O routines for the json.gz format, compatible with ADerrors.jl, can be found here.
\n"}, "pyerrors.correlators": {"fullname": "pyerrors.correlators", "modulename": "pyerrors.correlators", "kind": "module", "doc": "\n"}, "pyerrors.correlators.Corr": {"fullname": "pyerrors.correlators.Corr", "modulename": "pyerrors.correlators", "qualname": "Corr", "kind": "class", "doc": "The class for a correlator (time dependent sequence of pe.Obs).
\n\nEverything, this class does, can be achieved using lists or arrays of Obs.\nBut it is simply more convenient to have a dedicated object for correlators.\nOne often wants to add or multiply correlators of the same length at every timeslice and it is inconvenient\nto iterate over all timeslices for every operation. This is especially true, when dealing with matrices.
\n\nThe correlator can have two types of content: An Obs at every timeslice OR a GEVP\nmatrix at every timeslice. Other dependency (eg. spatial) are not supported.
\n"}, "pyerrors.correlators.Corr.__init__": {"fullname": "pyerrors.correlators.Corr.__init__", "modulename": "pyerrors.correlators", "qualname": "Corr.__init__", "kind": "function", "doc": "Initialize a Corr object.
\n\nParameters
\n\n\n- data_input (list or array):\nlist of Obs or list of arrays of Obs or array of Corrs
\n- padding (list, optional):\nList with two entries where the first labels the padding\nat the front of the correlator and the second the padding\nat the back.
\n- prange (list, optional):\nList containing the first and last timeslice of the plateau\nregion indentified for this correlator.
\n
\n", "signature": "(data_input, padding=[0, 0], prange=None)"}, "pyerrors.correlators.Corr.gamma_method": {"fullname": "pyerrors.correlators.Corr.gamma_method", "modulename": "pyerrors.correlators", "qualname": "Corr.gamma_method", "kind": "function", "doc": "Apply the gamma method to the content of the Corr.
\n", "signature": "(self, **kwargs):", "funcdef": "def"}, "pyerrors.correlators.Corr.gm": {"fullname": "pyerrors.correlators.Corr.gm", "modulename": "pyerrors.correlators", "qualname": "Corr.gm", "kind": "function", "doc": "Apply the gamma method to the content of the Corr.
\n", "signature": "(self, **kwargs):", "funcdef": "def"}, "pyerrors.correlators.Corr.projected": {"fullname": "pyerrors.correlators.Corr.projected", "modulename": "pyerrors.correlators", "qualname": "Corr.projected", "kind": "function", "doc": "We need to project the Correlator with a Vector to get a single value at each timeslice.
\n\nThe method can use one or two vectors.\nIf two are specified it returns v1@G@v2 (the order might be very important.)\nBy default it will return the lowest source, which usually means unsmeared-unsmeared (0,0), but it does not have to
\n", "signature": "(self, vector_l=None, vector_r=None, normalize=False):", "funcdef": "def"}, "pyerrors.correlators.Corr.item": {"fullname": "pyerrors.correlators.Corr.item", "modulename": "pyerrors.correlators", "qualname": "Corr.item", "kind": "function", "doc": "Picks the element [i,j] from every matrix and returns a correlator containing one Obs per timeslice.
\n\nParameters
\n\n\n- i (int):\nFirst index to be picked.
\n- j (int):\nSecond index to be picked.
\n
\n", "signature": "(self, i, j):", "funcdef": "def"}, "pyerrors.correlators.Corr.plottable": {"fullname": "pyerrors.correlators.Corr.plottable", "modulename": "pyerrors.correlators", "qualname": "Corr.plottable", "kind": "function", "doc": "Outputs the correlator in a plotable format.
\n\nOutputs three lists containing the timeslice index, the value on each\ntimeslice and the error on each timeslice.
\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.correlators.Corr.symmetric": {"fullname": "pyerrors.correlators.Corr.symmetric", "modulename": "pyerrors.correlators", "qualname": "Corr.symmetric", "kind": "function", "doc": "Symmetrize the correlator around x0=0.
\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.correlators.Corr.anti_symmetric": {"fullname": "pyerrors.correlators.Corr.anti_symmetric", "modulename": "pyerrors.correlators", "qualname": "Corr.anti_symmetric", "kind": "function", "doc": "Anti-symmetrize the correlator around x0=0.
\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.correlators.Corr.is_matrix_symmetric": {"fullname": "pyerrors.correlators.Corr.is_matrix_symmetric", "modulename": "pyerrors.correlators", "qualname": "Corr.is_matrix_symmetric", "kind": "function", "doc": "Checks whether a correlator matrices is symmetric on every timeslice.
\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.correlators.Corr.matrix_symmetric": {"fullname": "pyerrors.correlators.Corr.matrix_symmetric", "modulename": "pyerrors.correlators", "qualname": "Corr.matrix_symmetric", "kind": "function", "doc": "Symmetrizes the correlator matrices on every timeslice.
\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.correlators.Corr.GEVP": {"fullname": "pyerrors.correlators.Corr.GEVP", "modulename": "pyerrors.correlators", "qualname": "Corr.GEVP", "kind": "function", "doc": "Solve the generalized eigenvalue problem on the correlator matrix and returns the corresponding eigenvectors.
\n\nThe eigenvectors are sorted according to the descending eigenvalues, the zeroth eigenvector(s) correspond to the\nlargest eigenvalue(s). The eigenvector(s) for the individual states can be accessed via slicing
\n\n\n
C.GEVP(t0=2)[0] # Ground state vector(s)\nC.GEVP(t0=2)[:3] # Vectors for the lowest three states\n
\n
\n\nParameters
\n\n\n- t0 (int):\nThe time t0 for the right hand side of the GEVP according to $G(t)v_i=\\lambda_i G(t_0)v_i$
\n- ts (int):\nfixed time $G(t_s)v_i=\\lambda_i G(t_0)v_i$ if sort=None.\nIf sort=\"Eigenvector\" it gives a reference point for the sorting method.
\n- sort (string):\nIf this argument is set, a list of self.T vectors per state is returned. If it is set to None, only one vector is returned.\n
\n- \"Eigenvalue\": The eigenvector is chosen according to which eigenvalue it belongs individually on every timeslice.
\n- \"Eigenvector\": Use the method described in arXiv:2004.10472 to find the set of v(t) belonging to the state.\nThe reference state is identified by its eigenvalue at $t=t_s$.
\n
\n
\n\nOther Parameters
\n\n\n- state (int):\nReturns only the vector(s) for a specified state. The lowest state is zero.
\n
\n", "signature": "(self, t0, ts=None, sort='Eigenvalue', **kwargs):", "funcdef": "def"}, "pyerrors.correlators.Corr.Eigenvalue": {"fullname": "pyerrors.correlators.Corr.Eigenvalue", "modulename": "pyerrors.correlators", "qualname": "Corr.Eigenvalue", "kind": "function", "doc": "Determines the eigenvalue of the GEVP by solving and projecting the correlator
\n\nParameters
\n\n\n- state (int):\nThe state one is interested in ordered by energy. The lowest state is zero.
\n- All other parameters are identical to the ones of Corr.GEVP.
\n
\n", "signature": "(self, t0, ts=None, state=0, sort='Eigenvalue'):", "funcdef": "def"}, "pyerrors.correlators.Corr.Hankel": {"fullname": "pyerrors.correlators.Corr.Hankel", "modulename": "pyerrors.correlators", "qualname": "Corr.Hankel", "kind": "function", "doc": "Constructs an NxN Hankel matrix
\n\nC(t) c(t+1) ... c(t+n-1)\nC(t+1) c(t+2) ... c(t+n)\n.................\nC(t+(n-1)) c(t+n) ... c(t+2(n-1))
\n\nParameters
\n\n\n- N (int):\nDimension of the Hankel matrix
\n- periodic (bool, optional):\ndetermines whether the matrix is extended periodically
\n
\n", "signature": "(self, N, periodic=False):", "funcdef": "def"}, "pyerrors.correlators.Corr.roll": {"fullname": "pyerrors.correlators.Corr.roll", "modulename": "pyerrors.correlators", "qualname": "Corr.roll", "kind": "function", "doc": "Periodically shift the correlator by dt timeslices
\n\nParameters
\n\n\n- dt (int):\nnumber of timeslices
\n
\n", "signature": "(self, dt):", "funcdef": "def"}, "pyerrors.correlators.Corr.reverse": {"fullname": "pyerrors.correlators.Corr.reverse", "modulename": "pyerrors.correlators", "qualname": "Corr.reverse", "kind": "function", "doc": "Reverse the time ordering of the Corr
\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.correlators.Corr.thin": {"fullname": "pyerrors.correlators.Corr.thin", "modulename": "pyerrors.correlators", "qualname": "Corr.thin", "kind": "function", "doc": "Thin out a correlator to suppress correlations
\n\nParameters
\n\n\n- spacing (int):\nKeep only every 'spacing'th entry of the correlator
\n- offset (int):\nOffset the equal spacing
\n
\n", "signature": "(self, spacing=2, offset=0):", "funcdef": "def"}, "pyerrors.correlators.Corr.correlate": {"fullname": "pyerrors.correlators.Corr.correlate", "modulename": "pyerrors.correlators", "qualname": "Corr.correlate", "kind": "function", "doc": "Correlate the correlator with another correlator or Obs
\n\nParameters
\n\n\n- partner (Obs or Corr):\npartner to correlate the correlator with.\nCan either be an Obs which is correlated with all entries of the\ncorrelator or a Corr of same length.
\n
\n", "signature": "(self, partner):", "funcdef": "def"}, "pyerrors.correlators.Corr.reweight": {"fullname": "pyerrors.correlators.Corr.reweight", "modulename": "pyerrors.correlators", "qualname": "Corr.reweight", "kind": "function", "doc": "Reweight the correlator.
\n\nParameters
\n\n\n- weight (Obs):\nReweighting factor. An Observable that has to be defined on a superset of the\nconfigurations in obs[i].idl for all i.
\n- all_configs (bool):\nif True, the reweighted observables are normalized by the average of\nthe reweighting factor on all configurations in weight.idl and not\non the configurations in obs[i].idl.
\n
\n", "signature": "(self, weight, **kwargs):", "funcdef": "def"}, "pyerrors.correlators.Corr.T_symmetry": {"fullname": "pyerrors.correlators.Corr.T_symmetry", "modulename": "pyerrors.correlators", "qualname": "Corr.T_symmetry", "kind": "function", "doc": "Return the time symmetry average of the correlator and its partner
\n\nParameters
\n\n\n- partner (Corr):\nTime symmetry partner of the Corr
\n- partity (int):\nParity quantum number of the correlator, can be +1 or -1
\n
\n", "signature": "(self, partner, parity=1):", "funcdef": "def"}, "pyerrors.correlators.Corr.deriv": {"fullname": "pyerrors.correlators.Corr.deriv", "modulename": "pyerrors.correlators", "qualname": "Corr.deriv", "kind": "function", "doc": "Return the first derivative of the correlator with respect to x0.
\n\nParameters
\n\n\n- variant (str):\ndecides which definition of the finite differences derivative is used.\nAvailable choice: symmetric, forward, backward, improved, log, default: symmetric
\n
\n", "signature": "(self, variant='symmetric'):", "funcdef": "def"}, "pyerrors.correlators.Corr.second_deriv": {"fullname": "pyerrors.correlators.Corr.second_deriv", "modulename": "pyerrors.correlators", "qualname": "Corr.second_deriv", "kind": "function", "doc": "Return the second derivative of the correlator with respect to x0.
\n\nParameters
\n\n\n- variant (str):\ndecides which definition of the finite differences derivative is used.\nAvailable choice: symmetric, improved, log, default: symmetric
\n
\n", "signature": "(self, variant='symmetric'):", "funcdef": "def"}, "pyerrors.correlators.Corr.m_eff": {"fullname": "pyerrors.correlators.Corr.m_eff", "modulename": "pyerrors.correlators", "qualname": "Corr.m_eff", "kind": "function", "doc": "Returns the effective mass of the correlator as correlator object
\n\nParameters
\n\n\n- variant (str):\nlog : uses the standard effective mass log(C(t) / C(t+1))\ncosh, periodic : Use periodicitiy of the correlator by solving C(t) / C(t+1) = cosh(m * (t - T/2)) / cosh(m * (t + 1 - T/2)) for m.\nsinh : Use anti-periodicitiy of the correlator by solving C(t) / C(t+1) = sinh(m * (t - T/2)) / sinh(m * (t + 1 - T/2)) for m.\nSee, e.g., arXiv:1205.5380\narccosh : Uses the explicit form of the symmetrized correlator (not recommended)\nlogsym: uses the symmetric effective mass log(C(t-1) / C(t+1))/2
\n- guess (float):\nguess for the root finder, only relevant for the root variant
\n
\n", "signature": "(self, variant='log', guess=1.0):", "funcdef": "def"}, "pyerrors.correlators.Corr.fit": {"fullname": "pyerrors.correlators.Corr.fit", "modulename": "pyerrors.correlators", "qualname": "Corr.fit", "kind": "function", "doc": "Fits function to the data
\n\nParameters
\n\n\n- function (obj):\nfunction to fit to the data. See fits.least_squares for details.
\n- fitrange (list):\nTwo element list containing the timeslices on which the fit is supposed to start and stop.\nCaution: This range is inclusive as opposed to standard python indexing.\n
fitrange=[4, 6]
corresponds to the three entries 4, 5 and 6.\nIf not specified, self.prange or all timeslices are used. \n- silent (bool):\nDecides whether output is printed to the standard output.
\n
\n", "signature": "(self, function, fitrange=None, silent=False, **kwargs):", "funcdef": "def"}, "pyerrors.correlators.Corr.plateau": {"fullname": "pyerrors.correlators.Corr.plateau", "modulename": "pyerrors.correlators", "qualname": "Corr.plateau", "kind": "function", "doc": "Extract a plateau value from a Corr object
\n\nParameters
\n\n\n- plateau_range (list):\nlist with two entries, indicating the first and the last timeslice\nof the plateau region.
\n- method (str):\nmethod to extract the plateau.\n 'fit' fits a constant to the plateau region\n 'avg', 'average' or 'mean' just average over the given timeslices.
\n- auto_gamma (bool):\napply gamma_method with default parameters to the Corr. Defaults to None
\n
\n", "signature": "(self, plateau_range=None, method='fit', auto_gamma=False):", "funcdef": "def"}, "pyerrors.correlators.Corr.set_prange": {"fullname": "pyerrors.correlators.Corr.set_prange", "modulename": "pyerrors.correlators", "qualname": "Corr.set_prange", "kind": "function", "doc": "Sets the attribute prange of the Corr object.
\n", "signature": "(self, prange):", "funcdef": "def"}, "pyerrors.correlators.Corr.show": {"fullname": "pyerrors.correlators.Corr.show", "modulename": "pyerrors.correlators", "qualname": "Corr.show", "kind": "function", "doc": "Plots the correlator using the tag of the correlator as label if available.
\n\nParameters
\n\n\n- x_range (list):\nlist of two values, determining the range of the x-axis e.g. [4, 8].
\n- comp (Corr or list of Corr):\nCorrelator or list of correlators which are plotted for comparison.\nThe tags of these correlators are used as labels if available.
\n- logscale (bool):\nSets y-axis to logscale.
\n- plateau (Obs):\nPlateau value to be visualized in the figure.
\n- fit_res (Fit_result):\nFit_result object to be visualized.
\n- ylabel (str):\nLabel for the y-axis.
\n- save (str):\npath to file in which the figure should be saved.
\n- auto_gamma (bool):\nApply the gamma method with standard parameters to all correlators and plateau values before plotting.
\n- hide_sigma (float):\nHides data points from the first value on which is consistent with zero within 'hide_sigma' standard errors.
\n- references (list):\nList of floating point values that are displayed as horizontal lines for reference.
\n- title (string):\nOptional title of the figure.
\n
\n", "signature": "(\tself,\tx_range=None,\tcomp=None,\ty_range=None,\tlogscale=False,\tplateau=None,\tfit_res=None,\tylabel=None,\tsave=None,\tauto_gamma=False,\thide_sigma=None,\treferences=None,\ttitle=None):", "funcdef": "def"}, "pyerrors.correlators.Corr.spaghetti_plot": {"fullname": "pyerrors.correlators.Corr.spaghetti_plot", "modulename": "pyerrors.correlators", "qualname": "Corr.spaghetti_plot", "kind": "function", "doc": "Produces a spaghetti plot of the correlator suited to monitor exceptional configurations.
\n\nParameters
\n\n\n- logscale (bool):\nDetermines whether the scale of the y-axis is logarithmic or standard.
\n
\n", "signature": "(self, logscale=True):", "funcdef": "def"}, "pyerrors.correlators.Corr.dump": {"fullname": "pyerrors.correlators.Corr.dump", "modulename": "pyerrors.correlators", "qualname": "Corr.dump", "kind": "function", "doc": "Dumps the Corr into a file of chosen type
\n\nParameters
\n\n\n- filename (str):\nName of the file to be saved.
\n- datatype (str):\nFormat of the exported file. Supported formats include\n\"json.gz\" and \"pickle\"
\n- path (str):\nspecifies a custom path for the file (default '.')
\n
\n", "signature": "(self, filename, datatype='json.gz', **kwargs):", "funcdef": "def"}, "pyerrors.correlators.Corr.print": {"fullname": "pyerrors.correlators.Corr.print", "modulename": "pyerrors.correlators", "qualname": "Corr.print", "kind": "function", "doc": "\n", "signature": "(self, print_range=None):", "funcdef": "def"}, "pyerrors.correlators.Corr.sqrt": {"fullname": "pyerrors.correlators.Corr.sqrt", "modulename": "pyerrors.correlators", "qualname": "Corr.sqrt", "kind": "function", "doc": "\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.correlators.Corr.log": {"fullname": "pyerrors.correlators.Corr.log", "modulename": "pyerrors.correlators", "qualname": "Corr.log", "kind": "function", "doc": "\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.correlators.Corr.exp": {"fullname": "pyerrors.correlators.Corr.exp", "modulename": "pyerrors.correlators", "qualname": "Corr.exp", "kind": "function", "doc": "\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.correlators.Corr.sin": {"fullname": "pyerrors.correlators.Corr.sin", "modulename": "pyerrors.correlators", "qualname": "Corr.sin", "kind": "function", "doc": "\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.correlators.Corr.cos": {"fullname": "pyerrors.correlators.Corr.cos", "modulename": "pyerrors.correlators", "qualname": "Corr.cos", "kind": "function", "doc": "\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.correlators.Corr.tan": {"fullname": "pyerrors.correlators.Corr.tan", "modulename": "pyerrors.correlators", "qualname": "Corr.tan", "kind": "function", "doc": "\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.correlators.Corr.sinh": {"fullname": "pyerrors.correlators.Corr.sinh", "modulename": "pyerrors.correlators", "qualname": "Corr.sinh", "kind": "function", "doc": "\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.correlators.Corr.cosh": {"fullname": "pyerrors.correlators.Corr.cosh", "modulename": "pyerrors.correlators", "qualname": "Corr.cosh", "kind": "function", "doc": "\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.correlators.Corr.tanh": {"fullname": "pyerrors.correlators.Corr.tanh", "modulename": "pyerrors.correlators", "qualname": "Corr.tanh", "kind": "function", "doc": "\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.correlators.Corr.arcsin": {"fullname": "pyerrors.correlators.Corr.arcsin", "modulename": "pyerrors.correlators", "qualname": "Corr.arcsin", "kind": "function", "doc": "\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.correlators.Corr.arccos": {"fullname": "pyerrors.correlators.Corr.arccos", "modulename": "pyerrors.correlators", "qualname": "Corr.arccos", "kind": "function", "doc": "\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.correlators.Corr.arctan": {"fullname": "pyerrors.correlators.Corr.arctan", "modulename": "pyerrors.correlators", "qualname": "Corr.arctan", "kind": "function", "doc": "\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.correlators.Corr.arcsinh": {"fullname": "pyerrors.correlators.Corr.arcsinh", "modulename": "pyerrors.correlators", "qualname": "Corr.arcsinh", "kind": "function", "doc": "\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.correlators.Corr.arccosh": {"fullname": "pyerrors.correlators.Corr.arccosh", "modulename": "pyerrors.correlators", "qualname": "Corr.arccosh", "kind": "function", "doc": "\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.correlators.Corr.arctanh": {"fullname": "pyerrors.correlators.Corr.arctanh", "modulename": "pyerrors.correlators", "qualname": "Corr.arctanh", "kind": "function", "doc": "\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.correlators.Corr.prune": {"fullname": "pyerrors.correlators.Corr.prune", "modulename": "pyerrors.correlators", "qualname": "Corr.prune", "kind": "function", "doc": "Project large correlation matrix to lowest states
\n\nThis method can be used to reduce the size of an (N x N) correlation matrix\nto (Ntrunc x Ntrunc) by solving a GEVP at very early times where the noise\nis still small.
\n\nParameters
\n\n\n- Ntrunc (int):\nRank of the target matrix.
\n- tproj (int):\nTime where the eigenvectors are evaluated, corresponds to ts in the GEVP method.\nThe default value is 3.
\n- t0proj (int):\nTime where the correlation matrix is inverted. Choosing t0proj=1 is strongly\ndiscouraged for O(a) improved theories, since the correctness of the procedure\ncannot be granted in this case. The default value is 2.
\n- basematrix (Corr):\nCorrelation matrix that is used to determine the eigenvectors of the\nlowest states based on a GEVP. basematrix is taken to be the Corr itself if\nis is not specified.
\n
\n\nNotes
\n\nWe have the basematrix $C(t)$ and the target matrix $G(t)$. We start by solving\nthe GEVP $$C(t) v_n(t, t_0) = \\lambda_n(t, t_0) C(t_0) v_n(t, t_0)$$ where $t \\equiv t_\\mathrm{proj}$\nand $t_0 \\equiv t_{0, \\mathrm{proj}}$. The target matrix is projected onto the subspace of the\nresulting eigenvectors $v_n, n=1,\\dots,N_\\mathrm{trunc}$ via\n$$G^\\prime_{i, j}(t) = (v_i, G(t) v_j)$$. This allows to reduce the size of a large\ncorrelation matrix and to remove some noise that is added by irrelevant operators.\nThis may allow to use the GEVP on $G(t)$ at late times such that the theoretically motivated\nbound $t_0 \\leq t/2$ holds, since the condition number of $G(t)$ is decreased, compared to $C(t)$.
\n", "signature": "(self, Ntrunc, tproj=3, t0proj=2, basematrix=None):", "funcdef": "def"}, "pyerrors.covobs": {"fullname": "pyerrors.covobs", "modulename": "pyerrors.covobs", "kind": "module", "doc": "\n"}, "pyerrors.covobs.Covobs": {"fullname": "pyerrors.covobs.Covobs", "modulename": "pyerrors.covobs", "qualname": "Covobs", "kind": "class", "doc": "\n"}, "pyerrors.covobs.Covobs.__init__": {"fullname": "pyerrors.covobs.Covobs.__init__", "modulename": "pyerrors.covobs", "qualname": "Covobs.__init__", "kind": "function", "doc": "Initialize Covobs object.
\n\nParameters
\n\n\n- mean (float):\nMean value of the new Obs
\n- cov (list or array):\n2d Covariance matrix or 1d diagonal entries
\n- name (str):\nidentifier for the covariance matrix
\n- pos (int):\nPosition of the variance belonging to mean in cov.\nIs taken to be 1 if cov is 0-dimensional
\n- grad (list or array):\nGradient of the Covobs wrt. the means belonging to cov.
\n
\n", "signature": "(mean, cov, name, pos=None, grad=None)"}, "pyerrors.covobs.Covobs.errsq": {"fullname": "pyerrors.covobs.Covobs.errsq", "modulename": "pyerrors.covobs", "qualname": "Covobs.errsq", "kind": "function", "doc": "Return the variance (= square of the error) of the Covobs
\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.dirac": {"fullname": "pyerrors.dirac", "modulename": "pyerrors.dirac", "kind": "module", "doc": "\n"}, "pyerrors.dirac.epsilon_tensor": {"fullname": "pyerrors.dirac.epsilon_tensor", "modulename": "pyerrors.dirac", "qualname": "epsilon_tensor", "kind": "function", "doc": "Rank-3 epsilon tensor
\n\nBased on https://codegolf.stackexchange.com/a/160375
\n\nReturns
\n\n\n- elem (int):\nElement (i,j,k) of the epsilon tensor of rank 3
\n
\n", "signature": "(i, j, k):", "funcdef": "def"}, "pyerrors.dirac.epsilon_tensor_rank4": {"fullname": "pyerrors.dirac.epsilon_tensor_rank4", "modulename": "pyerrors.dirac", "qualname": "epsilon_tensor_rank4", "kind": "function", "doc": "Rank-4 epsilon tensor
\n\nExtension of https://codegolf.stackexchange.com/a/160375
\n\nReturns
\n\n\n- elem (int):\nElement (i,j,k,o) of the epsilon tensor of rank 4
\n
\n", "signature": "(i, j, k, o):", "funcdef": "def"}, "pyerrors.dirac.Grid_gamma": {"fullname": "pyerrors.dirac.Grid_gamma", "modulename": "pyerrors.dirac", "qualname": "Grid_gamma", "kind": "function", "doc": "Returns gamma matrix in Grid labeling.
\n", "signature": "(gamma_tag):", "funcdef": "def"}, "pyerrors.fits": {"fullname": "pyerrors.fits", "modulename": "pyerrors.fits", "kind": "module", "doc": "\n"}, "pyerrors.fits.Fit_result": {"fullname": "pyerrors.fits.Fit_result", "modulename": "pyerrors.fits", "qualname": "Fit_result", "kind": "class", "doc": "Represents fit results.
\n\nAttributes
\n\n\n- fit_parameters (list):\nresults for the individual fit parameters,\nalso accessible via indices.
\n- chisquare_by_dof (float):\nreduced chisquare.
\n- p_value (float):\np-value of the fit
\n- t2_p_value (float):\nHotelling t-squared p-value for correlated fits.
\n
\n", "bases": "collections.abc.Sequence"}, "pyerrors.fits.Fit_result.__init__": {"fullname": "pyerrors.fits.Fit_result.__init__", "modulename": "pyerrors.fits", "qualname": "Fit_result.__init__", "kind": "function", "doc": "\n", "signature": "()"}, "pyerrors.fits.Fit_result.gamma_method": {"fullname": "pyerrors.fits.Fit_result.gamma_method", "modulename": "pyerrors.fits", "qualname": "Fit_result.gamma_method", "kind": "function", "doc": "Apply the gamma method to all fit parameters
\n", "signature": "(self, **kwargs):", "funcdef": "def"}, "pyerrors.fits.Fit_result.gm": {"fullname": "pyerrors.fits.Fit_result.gm", "modulename": "pyerrors.fits", "qualname": "Fit_result.gm", "kind": "function", "doc": "Apply the gamma method to all fit parameters
\n", "signature": "(self, **kwargs):", "funcdef": "def"}, "pyerrors.fits.least_squares": {"fullname": "pyerrors.fits.least_squares", "modulename": "pyerrors.fits", "qualname": "least_squares", "kind": "function", "doc": "Performs a non-linear fit to y = func(x).
\n\nParameters
\n\n\n- x (list):\nlist of floats.
\n- y (list):\nlist of Obs.
\nfunc (object):\nfit function, has to be of the form
\n\n\n
import autograd.numpy as anp\n\ndef func(a, x):\n return a[0] + a[1] * x + a[2] * anp.sinh(x)\n
\n
\n\nFor multiple x values func can be of the form
\n\n\n
def func(a, x):\n (x1, x2) = x\n return a[0] * x1 ** 2 + a[1] * x2\n
\n
\n\nIt is important that all numpy functions refer to autograd.numpy, otherwise the differentiation\nwill not work.
\n- priors (list, optional):\npriors has to be a list with an entry for every parameter in the fit. The entries can either be\nObs (e.g. results from a previous fit) or strings containing a value and an error formatted like\n0.548(23), 500(40) or 0.5(0.4)
\n- silent (bool, optional):\nIf true all output to the console is omitted (default False).
\n- initial_guess (list):\ncan provide an initial guess for the input parameters. Relevant for\nnon-linear fits with many parameters. In case of correlated fits the guess is used to perform\nan uncorrelated fit which then serves as guess for the correlated fit.
\n- method (str, optional):\ncan be used to choose an alternative method for the minimization of chisquare.\nThe possible methods are the ones which can be used for scipy.optimize.minimize and\nmigrad of iminuit. If no method is specified, Levenberg-Marquard is used.\nReliable alternatives are migrad, Powell and Nelder-Mead.
\n- correlated_fit (bool):\nIf True, use the full inverse covariance matrix in the definition of the chisquare cost function.\nFor details about how the covariance matrix is estimated see
pyerrors.obs.covariance
.\nIn practice the correlation matrix is Cholesky decomposed and inverted (instead of the covariance matrix).\nThis procedure should be numerically more stable as the correlation matrix is typically better conditioned (Jacobi preconditioning).\nAt the moment this option only works for prior==None
and when no method
is given. \n- expected_chisquare (bool):\nIf True estimates the expected chisquare which is\ncorrected by effects caused by correlated input data (default False).
\n- resplot (bool):\nIf True, a plot which displays fit, data and residuals is generated (default False).
\n- qqplot (bool):\nIf True, a quantile-quantile plot of the fit result is generated (default False).
\n- num_grad (bool):\nUse numerical differentation instead of automatic differentiation to perform the error propagation (default False).
\n
\n\nReturns
\n\n\n- output (Fit_result):\nParameters and information on the fitted result.
\n
\n", "signature": "(x, y, func, priors=None, silent=False, **kwargs):", "funcdef": "def"}, "pyerrors.fits.total_least_squares": {"fullname": "pyerrors.fits.total_least_squares", "modulename": "pyerrors.fits", "qualname": "total_least_squares", "kind": "function", "doc": "Performs a non-linear fit to y = func(x) and returns a list of Obs corresponding to the fit parameters.
\n\nParameters
\n\n\n- x (list):\nlist of Obs, or a tuple of lists of Obs
\n- y (list):\nlist of Obs. The dvalues of the Obs are used as x- and yerror for the fit.
\nfunc (object):\nfunc has to be of the form
\n\n\n
import autograd.numpy as anp\n\ndef func(a, x):\n return a[0] + a[1] * x + a[2] * anp.sinh(x)\n
\n
\n\nFor multiple x values func can be of the form
\n\n\n
def func(a, x):\n (x1, x2) = x\n return a[0] * x1 ** 2 + a[1] * x2\n
\n
\n\nIt is important that all numpy functions refer to autograd.numpy, otherwise the differentiation\nwill not work.
\n- silent (bool, optional):\nIf true all output to the console is omitted (default False).
\n- initial_guess (list):\ncan provide an initial guess for the input parameters. Relevant for non-linear\nfits with many parameters.
\n- expected_chisquare (bool):\nIf true prints the expected chisquare which is\ncorrected by effects caused by correlated input data.\nThis can take a while as the full correlation matrix\nhas to be calculated (default False).
\n- num_grad (bool):\nUse numerical differentation instead of automatic differentiation to perform the error propagation (default False).
\n
\n\nNotes
\n\nBased on the orthogonal distance regression module of scipy.
\n\nReturns
\n\n\n- output (Fit_result):\nParameters and information on the fitted result.
\n
\n", "signature": "(x, y, func, silent=False, **kwargs):", "funcdef": "def"}, "pyerrors.fits.fit_lin": {"fullname": "pyerrors.fits.fit_lin", "modulename": "pyerrors.fits", "qualname": "fit_lin", "kind": "function", "doc": "Performs a linear fit to y = n + m * x and returns two Obs n, m.
\n\nParameters
\n\n\n- x (list):\nCan either be a list of floats in which case no xerror is assumed, or\na list of Obs, where the dvalues of the Obs are used as xerror for the fit.
\n- y (list):\nList of Obs, the dvalues of the Obs are used as yerror for the fit.
\n
\n\nReturns
\n\n\n- fit_parameters (list[Obs]):\nLIist of fitted observables.
\n
\n", "signature": "(x, y, **kwargs):", "funcdef": "def"}, "pyerrors.fits.qqplot": {"fullname": "pyerrors.fits.qqplot", "modulename": "pyerrors.fits", "qualname": "qqplot", "kind": "function", "doc": "Generates a quantile-quantile plot of the fit result which can be used to\n check if the residuals of the fit are gaussian distributed.
\n\nReturns
\n\n\n", "signature": "(x, o_y, func, p):", "funcdef": "def"}, "pyerrors.fits.residual_plot": {"fullname": "pyerrors.fits.residual_plot", "modulename": "pyerrors.fits", "qualname": "residual_plot", "kind": "function", "doc": "Generates a plot which compares the fit to the data and displays the corresponding residuals
\n\nReturns
\n\n\n", "signature": "(x, y, func, fit_res):", "funcdef": "def"}, "pyerrors.fits.error_band": {"fullname": "pyerrors.fits.error_band", "modulename": "pyerrors.fits", "qualname": "error_band", "kind": "function", "doc": "Calculate the error band for an array of sample values x, for given fit function func with optimized parameters beta.
\n\nReturns
\n\n\n- err (np.array(Obs)):\nError band for an array of sample values x
\n
\n", "signature": "(x, func, beta):", "funcdef": "def"}, "pyerrors.fits.ks_test": {"fullname": "pyerrors.fits.ks_test", "modulename": "pyerrors.fits", "qualname": "ks_test", "kind": "function", "doc": "Performs a Kolmogorov\u2013Smirnov test for the p-values of all fit object.
\n\nParameters
\n\n\n- objects (list):\nList of fit results to include in the analysis (optional).
\n
\n\nReturns
\n\n\n", "signature": "(objects=None):", "funcdef": "def"}, "pyerrors.input": {"fullname": "pyerrors.input", "modulename": "pyerrors.input", "kind": "module", "doc": "pyerrors
includes an input
submodule in which input routines and parsers for the output of various numerical programs are contained.
\n\nJackknife samples
\n\nFor comparison with other analysis workflows pyerrors
can also generate jackknife samples from an Obs
object or import jackknife samples into an Obs
object.\nSee pyerrors.obs.Obs.export_jackknife
and pyerrors.obs.import_jackknife
for details.
\n"}, "pyerrors.input.bdio": {"fullname": "pyerrors.input.bdio", "modulename": "pyerrors.input.bdio", "kind": "module", "doc": "\n"}, "pyerrors.input.bdio.read_ADerrors": {"fullname": "pyerrors.input.bdio.read_ADerrors", "modulename": "pyerrors.input.bdio", "qualname": "read_ADerrors", "kind": "function", "doc": "Extract generic MCMC data from a bdio file
\n\nread_ADerrors requires bdio to be compiled into a shared library. This can be achieved by\nadding the flag -fPIC to CC and changing the all target to
\n\nall: bdio.o $(LIBDIR)\n gcc -shared -Wl,-soname,libbdio.so -o $(BUILDDIR)/libbdio.so $(BUILDDIR)/bdio.o\n cp $(BUILDDIR)/libbdio.so $(LIBDIR)/
\n\nParameters
\n\n\n- file_path -- path to the bdio file
\n- bdio_path -- path to the shared bdio library libbdio.so (default ./libbdio.so)
\n
\n\nReturns
\n\n\n- data (List[Obs]):\nExtracted data
\n
\n", "signature": "(file_path, bdio_path='./libbdio.so', **kwargs):", "funcdef": "def"}, "pyerrors.input.bdio.write_ADerrors": {"fullname": "pyerrors.input.bdio.write_ADerrors", "modulename": "pyerrors.input.bdio", "qualname": "write_ADerrors", "kind": "function", "doc": "Write Obs to a bdio file according to ADerrors conventions
\n\nread_mesons requires bdio to be compiled into a shared library. This can be achieved by\nadding the flag -fPIC to CC and changing the all target to
\n\nall: bdio.o $(LIBDIR)\n gcc -shared -Wl,-soname,libbdio.so -o $(BUILDDIR)/libbdio.so $(BUILDDIR)/bdio.o\n cp $(BUILDDIR)/libbdio.so $(LIBDIR)/
\n\nParameters
\n\n\n- file_path -- path to the bdio file
\n- bdio_path -- path to the shared bdio library libbdio.so (default ./libbdio.so)
\n
\n\nReturns
\n\n\n- success (int):\nreturns 0 is successful
\n
\n", "signature": "(obs_list, file_path, bdio_path='./libbdio.so', **kwargs):", "funcdef": "def"}, "pyerrors.input.bdio.read_mesons": {"fullname": "pyerrors.input.bdio.read_mesons", "modulename": "pyerrors.input.bdio", "qualname": "read_mesons", "kind": "function", "doc": "Extract mesons data from a bdio file and return it as a dictionary
\n\nThe dictionary can be accessed with a tuple consisting of (type, source_position, kappa1, kappa2)
\n\nread_mesons requires bdio to be compiled into a shared library. This can be achieved by\nadding the flag -fPIC to CC and changing the all target to
\n\nall: bdio.o $(LIBDIR)\n gcc -shared -Wl,-soname,libbdio.so -o $(BUILDDIR)/libbdio.so $(BUILDDIR)/bdio.o\n cp $(BUILDDIR)/libbdio.so $(LIBDIR)/
\n\nParameters
\n\n\n- file_path (str):\npath to the bdio file
\n- bdio_path (str):\npath to the shared bdio library libbdio.so (default ./libbdio.so)
\n- start (int):\nThe first configuration to be read (default 1)
\n- stop (int):\nThe last configuration to be read (default None)
\n- step (int):\nFixed step size between two measurements (default 1)
\n- alternative_ensemble_name (str):\nManually overwrite ensemble name
\n
\n\nReturns
\n\n\n- data (dict):\nExtracted meson data
\n
\n", "signature": "(file_path, bdio_path='./libbdio.so', **kwargs):", "funcdef": "def"}, "pyerrors.input.bdio.read_dSdm": {"fullname": "pyerrors.input.bdio.read_dSdm", "modulename": "pyerrors.input.bdio", "qualname": "read_dSdm", "kind": "function", "doc": "Extract dSdm data from a bdio file and return it as a dictionary
\n\nThe dictionary can be accessed with a tuple consisting of (type, kappa)
\n\nread_dSdm requires bdio to be compiled into a shared library. This can be achieved by\nadding the flag -fPIC to CC and changing the all target to
\n\nall: bdio.o $(LIBDIR)\n gcc -shared -Wl,-soname,libbdio.so -o $(BUILDDIR)/libbdio.so $(BUILDDIR)/bdio.o\n cp $(BUILDDIR)/libbdio.so $(LIBDIR)/
\n\nParameters
\n\n\n- file_path (str):\npath to the bdio file
\n- bdio_path (str):\npath to the shared bdio library libbdio.so (default ./libbdio.so)
\n- start (int):\nThe first configuration to be read (default 1)
\n- stop (int):\nThe last configuration to be read (default None)
\n- step (int):\nFixed step size between two measurements (default 1)
\n- alternative_ensemble_name (str):\nManually overwrite ensemble name
\n
\n", "signature": "(file_path, bdio_path='./libbdio.so', **kwargs):", "funcdef": "def"}, "pyerrors.input.dobs": {"fullname": "pyerrors.input.dobs", "modulename": "pyerrors.input.dobs", "kind": "module", "doc": "\n"}, "pyerrors.input.dobs.create_pobs_string": {"fullname": "pyerrors.input.dobs.create_pobs_string", "modulename": "pyerrors.input.dobs", "qualname": "create_pobs_string", "kind": "function", "doc": "Export a list of Obs or structures containing Obs to an xml string\naccording to the Zeuthen pobs format.
\n\nTags are not written or recovered automatically. The separator | is removed from the replica names.
\n\nParameters
\n\n\n- obsl (list):\nList of Obs that will be exported.\nThe Obs inside a structure have to be defined on the same ensemble.
\n- name (str):\nThe name of the observable.
\n- spec (str):\nOptional string that describes the contents of the file.
\n- origin (str):\nSpecify where the data has its origin.
\n- symbol (list):\nA list of symbols that describe the observables to be written. May be empty.
\n- enstag (str):\nEnstag that is written to pobs. If None, the ensemble name is used.
\n
\n\nReturns
\n\n\n- xml_str (str):\nXML formatted string of the input data
\n
\n", "signature": "(obsl, name, spec='', origin='', symbol=[], enstag=None):", "funcdef": "def"}, "pyerrors.input.dobs.write_pobs": {"fullname": "pyerrors.input.dobs.write_pobs", "modulename": "pyerrors.input.dobs", "qualname": "write_pobs", "kind": "function", "doc": "Export a list of Obs or structures containing Obs to a .xml.gz file\naccording to the Zeuthen pobs format.
\n\nTags are not written or recovered automatically. The separator | is removed from the replica names.
\n\nParameters
\n\n\n- obsl (list):\nList of Obs that will be exported.\nThe Obs inside a structure have to be defined on the same ensemble.
\n- fname (str):\nFilename of the output file.
\n- name (str):\nThe name of the observable.
\n- spec (str):\nOptional string that describes the contents of the file.
\n- origin (str):\nSpecify where the data has its origin.
\n- symbol (list):\nA list of symbols that describe the observables to be written. May be empty.
\n- enstag (str):\nEnstag that is written to pobs. If None, the ensemble name is used.
\n- gz (bool):\nIf True, the output is a gzipped xml. If False, the output is an xml file.
\n
\n\nReturns
\n\n\n", "signature": "(\tobsl,\tfname,\tname,\tspec='',\torigin='',\tsymbol=[],\tenstag=None,\tgz=True):", "funcdef": "def"}, "pyerrors.input.dobs.read_pobs": {"fullname": "pyerrors.input.dobs.read_pobs", "modulename": "pyerrors.input.dobs", "qualname": "read_pobs", "kind": "function", "doc": "Import a list of Obs from an xml.gz file in the Zeuthen pobs format.
\n\nTags are not written or recovered automatically.
\n\nParameters
\n\n\n- fname (str):\nFilename of the input file.
\n- full_output (bool):\nIf True, a dict containing auxiliary information and the data is returned.\nIf False, only the data is returned as list.
\n- separatior_insertion (str or int):\nstr: replace all occurences of \"separator_insertion\" within the replica names\nby \"|%s\" % (separator_insertion) when constructing the names of the replica.\nint: Insert the separator \"|\" at the position given by separator_insertion.\nNone (default): Replica names remain unchanged.
\n
\n\nReturns
\n\n\n- res (list[Obs]):\nImported data
\n- or
\n- res (dict):\nImported data and meta-data
\n
\n", "signature": "(fname, full_output=False, gz=True, separator_insertion=None):", "funcdef": "def"}, "pyerrors.input.dobs.import_dobs_string": {"fullname": "pyerrors.input.dobs.import_dobs_string", "modulename": "pyerrors.input.dobs", "qualname": "import_dobs_string", "kind": "function", "doc": "Import a list of Obs from a string in the Zeuthen dobs format.
\n\nTags are not written or recovered automatically.
\n\nParameters
\n\n\n- content (str):\nXML string containing the data
\n- noemtpy (bool):\nIf True, ensembles with no contribution to the Obs are not included.\nIf False, ensembles are included as written in the file, possibly with vanishing entries.
\n- full_output (bool):\nIf True, a dict containing auxiliary information and the data is returned.\nIf False, only the data is returned as list.
\n- separatior_insertion (str, int or bool):\nstr: replace all occurences of \"separator_insertion\" within the replica names\nby \"|%s\" % (separator_insertion) when constructing the names of the replica.\nint: Insert the separator \"|\" at the position given by separator_insertion.\nTrue (default): separator \"|\" is inserted after len(ensname), assuming that the\nensemble name is a prefix to the replica name.\nNone or False: No separator is inserted.
\n
\n\nReturns
\n\n\n- res (list[Obs]):\nImported data
\n- or
\n- res (dict):\nImported data and meta-data
\n
\n", "signature": "(content, noempty=False, full_output=False, separator_insertion=True):", "funcdef": "def"}, "pyerrors.input.dobs.read_dobs": {"fullname": "pyerrors.input.dobs.read_dobs", "modulename": "pyerrors.input.dobs", "qualname": "read_dobs", "kind": "function", "doc": "Import a list of Obs from an xml.gz file in the Zeuthen dobs format.
\n\nTags are not written or recovered automatically.
\n\nParameters
\n\n\n- fname (str):\nFilename of the input file.
\n- noemtpy (bool):\nIf True, ensembles with no contribution to the Obs are not included.\nIf False, ensembles are included as written in the file.
\n- full_output (bool):\nIf True, a dict containing auxiliary information and the data is returned.\nIf False, only the data is returned as list.
\n- gz (bool):\nIf True, assumes that data is gzipped. If False, assumes XML file.
\n- separatior_insertion (str, int or bool):\nstr: replace all occurences of \"separator_insertion\" within the replica names\nby \"|%s\" % (separator_insertion) when constructing the names of the replica.\nint: Insert the separator \"|\" at the position given by separator_insertion.\nTrue (default): separator \"|\" is inserted after len(ensname), assuming that the\nensemble name is a prefix to the replica name.\nNone or False: No separator is inserted.
\n
\n\nReturns
\n\n\n- res (list[Obs]):\nImported data
\n- or
\n- res (dict):\nImported data and meta-data
\n
\n", "signature": "(\tfname,\tnoempty=False,\tfull_output=False,\tgz=True,\tseparator_insertion=True):", "funcdef": "def"}, "pyerrors.input.dobs.create_dobs_string": {"fullname": "pyerrors.input.dobs.create_dobs_string", "modulename": "pyerrors.input.dobs", "qualname": "create_dobs_string", "kind": "function", "doc": "Generate the string for the export of a list of Obs or structures containing Obs\nto a .xml.gz file according to the Zeuthen dobs format.
\n\nTags are not written or recovered automatically. The separator |is removed from the replica names.
\n\nParameters
\n\n\n- obsl (list):\nList of Obs that will be exported.\nThe Obs inside a structure do not have to be defined on the same set of configurations,\nbut the storage requirement is increased, if this is not the case.
\n- name (str):\nThe name of the observable.
\n- spec (str):\nOptional string that describes the contents of the file.
\n- origin (str):\nSpecify where the data has its origin.
\n- symbol (list):\nA list of symbols that describe the observables to be written. May be empty.
\n- who (str):\nProvide the name of the person that exports the data.
\n- enstags (dict):\nProvide alternative enstag for ensembles in the form enstags = {ename: enstag}\nOtherwise, the ensemble name is used.
\n
\n\nReturns
\n\n\n- xml_str (str):\nXML string generated from the data
\n
\n", "signature": "(\tobsl,\tname,\tspec='dobs v1.0',\torigin='',\tsymbol=[],\twho=None,\tenstags=None):", "funcdef": "def"}, "pyerrors.input.dobs.write_dobs": {"fullname": "pyerrors.input.dobs.write_dobs", "modulename": "pyerrors.input.dobs", "qualname": "write_dobs", "kind": "function", "doc": "Export a list of Obs or structures containing Obs to a .xml.gz file\naccording to the Zeuthen dobs format.
\n\nTags are not written or recovered automatically. The separator | is removed from the replica names.
\n\nParameters
\n\n\n- obsl (list):\nList of Obs that will be exported.\nThe Obs inside a structure do not have to be defined on the same set of configurations,\nbut the storage requirement is increased, if this is not the case.
\n- fname (str):\nFilename of the output file.
\n- name (str):\nThe name of the observable.
\n- spec (str):\nOptional string that describes the contents of the file.
\n- origin (str):\nSpecify where the data has its origin.
\n- symbol (list):\nA list of symbols that describe the observables to be written. May be empty.
\n- who (str):\nProvide the name of the person that exports the data.
\n- enstags (dict):\nProvide alternative enstag for ensembles in the form enstags = {ename: enstag}\nOtherwise, the ensemble name is used.
\n- gz (bool):\nIf True, the output is a gzipped XML. If False, the output is a XML file.
\n
\n\nReturns
\n\n\n", "signature": "(\tobsl,\tfname,\tname,\tspec='dobs v1.0',\torigin='',\tsymbol=[],\twho=None,\tenstags=None,\tgz=True):", "funcdef": "def"}, "pyerrors.input.hadrons": {"fullname": "pyerrors.input.hadrons", "modulename": "pyerrors.input.hadrons", "kind": "module", "doc": "\n"}, "pyerrors.input.hadrons.read_meson_hd5": {"fullname": "pyerrors.input.hadrons.read_meson_hd5", "modulename": "pyerrors.input.hadrons", "qualname": "read_meson_hd5", "kind": "function", "doc": "Read hadrons meson hdf5 file and extract the meson labeled 'meson'
\n\nParameters
\n\n\n- path (str):\npath to the files to read
\n- filestem (str):\nnamestem of the files to read
\n- ens_id (str):\nname of the ensemble, required for internal bookkeeping
\n- meson (str):\nlabel of the meson to be extracted, standard value meson_0 which\ncorresponds to the pseudoscalar pseudoscalar two-point function.
\n- gammas (tuple of strings):\nInstrad of a meson label one can also provide a tuple of two strings\nindicating the gamma matrices at source and sink.\n(\"Gamma5\", \"Gamma5\") corresponds to the pseudoscalar pseudoscalar\ntwo-point function. The gammas argument dominateds over meson.
\n- idl (range):\nIf specified only configurations in the given range are read in.
\n
\n\nReturns
\n\n\n- corr (Corr):\nCorrelator of the source sink combination in question.
\n
\n", "signature": "(path, filestem, ens_id, meson='meson_0', idl=None, gammas=None):", "funcdef": "def"}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"fullname": "pyerrors.input.hadrons.read_DistillationContraction_hd5", "modulename": "pyerrors.input.hadrons", "qualname": "read_DistillationContraction_hd5", "kind": "function", "doc": "Read hadrons DistillationContraction hdf5 files in given directory structure
\n\nParameters
\n\n\n- path (str):\npath to the directories to read
\n- ens_id (str):\nname of the ensemble, required for internal bookkeeping
\n- diagrams (list):\nList of strings of the diagrams to extract, e.g. [\"direct\", \"box\", \"cross\"].
\n- idl (range):\nIf specified only configurations in the given range are read in.
\n
\n\nReturns
\n\n\n- result (dict):\nextracted DistillationContration data
\n
\n", "signature": "(path, ens_id, diagrams=['direct'], idl=None):", "funcdef": "def"}, "pyerrors.input.hadrons.Npr_matrix": {"fullname": "pyerrors.input.hadrons.Npr_matrix", "modulename": "pyerrors.input.hadrons", "qualname": "Npr_matrix", "kind": "class", "doc": "ndarray(shape, dtype=float, buffer=None, offset=0,\n strides=None, order=None)
\n\nAn array object represents a multidimensional, homogeneous array\nof fixed-size items. An associated data-type object describes the\nformat of each element in the array (its byte-order, how many bytes it\noccupies in memory, whether it is an integer, a floating point number,\nor something else, etc.)
\n\nArrays should be constructed using array
, zeros
or empty
(refer\nto the See Also section below). The parameters given here refer to\na low-level method (ndarray(...)
) for instantiating an array.
\n\nFor more information, refer to the numpy
module and examine the\nmethods and attributes of an array.
\n\nParameters
\n\n\n- (for the __new__ method; see Notes below)
\n- shape (tuple of ints):\nShape of created array.
\n- dtype (data-type, optional):\nAny object that can be interpreted as a numpy data type.
\n- buffer (object exposing buffer interface, optional):\nUsed to fill the array with data.
\n- offset (int, optional):\nOffset of array data in buffer.
\n- strides (tuple of ints, optional):\nStrides of data in memory.
\n- order ({'C', 'F'}, optional):\nRow-major (C-style) or column-major (Fortran-style) order.
\n
\n\nAttributes
\n\n\n- T (ndarray):\nTranspose of the array.
\n- data (buffer):\nThe array's elements, in memory.
\n- dtype (dtype object):\nDescribes the format of the elements in the array.
\n- flags (dict):\nDictionary containing information related to memory use, e.g.,\n'C_CONTIGUOUS', 'OWNDATA', 'WRITEABLE', etc.
\n- flat (numpy.flatiter object):\nFlattened version of the array as an iterator. The iterator\nallows assignments, e.g.,
x.flat = 3
(See ndarray.flat
for\nassignment examples; TODO). \n- imag (ndarray):\nImaginary part of the array.
\n- real (ndarray):\nReal part of the array.
\n- size (int):\nNumber of elements in the array.
\n- itemsize (int):\nThe memory use of each array element in bytes.
\n- nbytes (int):\nThe total number of bytes required to store the array data,\ni.e.,
itemsize * size
. \n- ndim (int):\nThe array's number of dimensions.
\n- shape (tuple of ints):\nShape of the array.
\n- strides (tuple of ints):\nThe step-size required to move from one element to the next in\nmemory. For example, a contiguous
(3, 4)
array of type\nint16
in C-order has strides (8, 2)
. This implies that\nto move from element to element in memory requires jumps of 2 bytes.\nTo move from row-to-row, one needs to jump 8 bytes at a time\n(2 * 4
). \n- ctypes (ctypes object):\nClass containing properties of the array needed for interaction\nwith ctypes.
\n- base (ndarray):\nIf the array is a view into another array, that array is its
base
\n(unless that array is also a view). The base
array is where the\narray data is actually stored. \n
\n\nSee Also
\n\narray
: Construct an array.
\nzeros
: Create an array, each element of which is zero.
\nempty
: Create an array, but leave its allocated memory unchanged (i.e.,\nit contains \"garbage\").
\ndtype
: Create a data-type.
\nnumpy.typing.NDArray
: An ndarray alias :term:generic <generic type>
\nw.r.t. its dtype.type <numpy.dtype.type>
.
\n\nNotes
\n\nThere are two modes of creating an array using __new__
:
\n\n\n- If
buffer
is None, then only shape
, dtype
, and order
\nare used. \n- If
buffer
is an object exposing the buffer interface, then\nall keywords are interpreted. \n
\n\nNo __init__
method is needed because the array is fully initialized\nafter the __new__
method.
\n\nExamples
\n\nThese examples illustrate the low-level ndarray
constructor. Refer\nto the See Also
section above for easier ways of constructing an\nndarray.
\n\nFirst mode, buffer
is None:
\n\n\n
>>> np.ndarray(shape=(2,2), dtype=float, order='F')\narray([[0.0e+000, 0.0e+000], # random\n [ nan, 2.5e-323]])\n
\n
\n\nSecond mode:
\n\n\n
>>> np.ndarray((2,), buffer=np.array([1,2,3]),\n... offset=np.int_().itemsize,\n... dtype=int) # offset = 1*itemsize, i.e. skip first element\narray([2, 3])\n
\n
\n", "bases": "numpy.ndarray"}, "pyerrors.input.hadrons.Npr_matrix.__init__": {"fullname": "pyerrors.input.hadrons.Npr_matrix.__init__", "modulename": "pyerrors.input.hadrons", "qualname": "Npr_matrix.__init__", "kind": "function", "doc": "\n", "signature": "()"}, "pyerrors.input.hadrons.Npr_matrix.g5H": {"fullname": "pyerrors.input.hadrons.Npr_matrix.g5H", "modulename": "pyerrors.input.hadrons", "qualname": "Npr_matrix.g5H", "kind": "variable", "doc": "Gamma_5 hermitean conjugate
\n\nUses the fact that the propagator is gamma5 hermitean, so just the\nin and out momenta of the propagator are exchanged.
\n"}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"fullname": "pyerrors.input.hadrons.read_ExternalLeg_hd5", "modulename": "pyerrors.input.hadrons", "qualname": "read_ExternalLeg_hd5", "kind": "function", "doc": "Read hadrons ExternalLeg hdf5 file and output an array of CObs
\n\nParameters
\n\n\n- path (str):\npath to the files to read
\n- filestem (str):\nnamestem of the files to read
\n- ens_id (str):\nname of the ensemble, required for internal bookkeeping
\n- idl (range):\nIf specified only configurations in the given range are read in.
\n
\n\nReturns
\n\n\n- result (Npr_matrix):\nread Cobs-matrix
\n
\n", "signature": "(path, filestem, ens_id, idl=None):", "funcdef": "def"}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"fullname": "pyerrors.input.hadrons.read_Bilinear_hd5", "modulename": "pyerrors.input.hadrons", "qualname": "read_Bilinear_hd5", "kind": "function", "doc": "Read hadrons Bilinear hdf5 file and output an array of CObs
\n\nParameters
\n\n\n- path (str):\npath to the files to read
\n- filestem (str):\nnamestem of the files to read
\n- ens_id (str):\nname of the ensemble, required for internal bookkeeping
\n- idl (range):\nIf specified only configurations in the given range are read in.
\n
\n\nReturns
\n\n\n- result_dict (dict[Npr_matrix]):\nextracted Bilinears
\n
\n", "signature": "(path, filestem, ens_id, idl=None):", "funcdef": "def"}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"fullname": "pyerrors.input.hadrons.read_Fourquark_hd5", "modulename": "pyerrors.input.hadrons", "qualname": "read_Fourquark_hd5", "kind": "function", "doc": "Read hadrons FourquarkFullyConnected hdf5 file and output an array of CObs
\n\nParameters
\n\n\n- path (str):\npath to the files to read
\n- filestem (str):\nnamestem of the files to read
\n- ens_id (str):\nname of the ensemble, required for internal bookkeeping
\n- idl (range):\nIf specified only configurations in the given range are read in.
\n- vertices (list):\nVertex functions to be extracted.
\n
\n\nReturns
\n\n\n- result_dict (dict):\nextracted fourquark matrizes
\n
\n", "signature": "(path, filestem, ens_id, idl=None, vertices=['VA', 'AV']):", "funcdef": "def"}, "pyerrors.input.json": {"fullname": "pyerrors.input.json", "modulename": "pyerrors.input.json", "kind": "module", "doc": "\n"}, "pyerrors.input.json.create_json_string": {"fullname": "pyerrors.input.json.create_json_string", "modulename": "pyerrors.input.json", "qualname": "create_json_string", "kind": "function", "doc": "Generate the string for the export of a list of Obs or structures containing Obs\nto a .json(.gz) file
\n\nParameters
\n\n\n- ol (list):\nList of objects that will be exported. At the moment, these objects can be\neither of: Obs, list, numpy.ndarray, Corr.\nAll Obs inside a structure have to be defined on the same set of configurations.
\n- description (str):\nOptional string that describes the contents of the json file.
\n- indent (int):\nSpecify the indentation level of the json file. None or 0 is permissible and\nsaves disk space.
\n
\n\nReturns
\n\n\n- json_string (str):\nString for export to .json(.gz) file
\n
\n", "signature": "(ol, description='', indent=1):", "funcdef": "def"}, "pyerrors.input.json.dump_to_json": {"fullname": "pyerrors.input.json.dump_to_json", "modulename": "pyerrors.input.json", "qualname": "dump_to_json", "kind": "function", "doc": "Export a list of Obs or structures containing Obs to a .json(.gz) file
\n\nParameters
\n\n\n- ol (list):\nList of objects that will be exported. At the moment, these objects can be\neither of: Obs, list, numpy.ndarray, Corr.\nAll Obs inside a structure have to be defined on the same set of configurations.
\n- fname (str):\nFilename of the output file.
\n- description (str):\nOptional string that describes the contents of the json file.
\n- indent (int):\nSpecify the indentation level of the json file. None or 0 is permissible and\nsaves disk space.
\n- gz (bool):\nIf True, the output is a gzipped json. If False, the output is a json file.
\n
\n\nReturns
\n\n\n", "signature": "(ol, fname, description='', indent=1, gz=True):", "funcdef": "def"}, "pyerrors.input.json.import_json_string": {"fullname": "pyerrors.input.json.import_json_string", "modulename": "pyerrors.input.json", "qualname": "import_json_string", "kind": "function", "doc": "Reconstruct a list of Obs or structures containing Obs from a json string.
\n\nThe following structures are supported: Obs, list, numpy.ndarray, Corr\nIf the list contains only one element, it is unpacked from the list.
\n\nParameters
\n\n\n- json_string (str):\njson string containing the data.
\n- verbose (bool):\nPrint additional information that was written to the file.
\n- full_output (bool):\nIf True, a dict containing auxiliary information and the data is returned.\nIf False, only the data is returned.
\n
\n\nReturns
\n\n\n- result (list[Obs]):\nreconstructed list of observables from the json string
\n- or
\n- result (Obs):\nonly one observable if the list only has one entry
\n- or
\n- result (dict):\nif full_output=True
\n
\n", "signature": "(json_string, verbose=True, full_output=False):", "funcdef": "def"}, "pyerrors.input.json.load_json": {"fullname": "pyerrors.input.json.load_json", "modulename": "pyerrors.input.json", "qualname": "load_json", "kind": "function", "doc": "Import a list of Obs or structures containing Obs from a .json(.gz) file.
\n\nThe following structures are supported: Obs, list, numpy.ndarray, Corr\nIf the list contains only one element, it is unpacked from the list.
\n\nParameters
\n\n\n- fname (str):\nFilename of the input file.
\n- verbose (bool):\nPrint additional information that was written to the file.
\n- gz (bool):\nIf True, assumes that data is gzipped. If False, assumes JSON file.
\n- full_output (bool):\nIf True, a dict containing auxiliary information and the data is returned.\nIf False, only the data is returned.
\n
\n\nReturns
\n\n\n- result (list[Obs]):\nreconstructed list of observables from the json string
\n- or
\n- result (Obs):\nonly one observable if the list only has one entry
\n- or
\n- result (dict):\nif full_output=True
\n
\n", "signature": "(fname, verbose=True, gz=True, full_output=False):", "funcdef": "def"}, "pyerrors.input.json.dump_dict_to_json": {"fullname": "pyerrors.input.json.dump_dict_to_json", "modulename": "pyerrors.input.json", "qualname": "dump_dict_to_json", "kind": "function", "doc": "Export a dict of Obs or structures containing Obs to a .json(.gz) file
\n\nParameters
\n\n\n- od (dict):\nDict of JSON valid structures and objects that will be exported.\nAt the moment, these objects can be either of: Obs, list, numpy.ndarray, Corr.\nAll Obs inside a structure have to be defined on the same set of configurations.
\n- fname (str):\nFilename of the output file.
\n- description (str):\nOptional string that describes the contents of the json file.
\n- indent (int):\nSpecify the indentation level of the json file. None or 0 is permissible and\nsaves disk space.
\n- reps (str):\nSpecify the structure of the placeholder in exported dict to be reps[0-9]+.
\n- gz (bool):\nIf True, the output is a gzipped json. If False, the output is a json file.
\n
\n\nReturns
\n\n\n", "signature": "(od, fname, description='', indent=1, reps='DICTOBS', gz=True):", "funcdef": "def"}, "pyerrors.input.json.load_json_dict": {"fullname": "pyerrors.input.json.load_json_dict", "modulename": "pyerrors.input.json", "qualname": "load_json_dict", "kind": "function", "doc": "Import a dict of Obs or structures containing Obs from a .json(.gz) file.
\n\nThe following structures are supported: Obs, list, numpy.ndarray, Corr
\n\nParameters
\n\n\n- fname (str):\nFilename of the input file.
\n- verbose (bool):\nPrint additional information that was written to the file.
\n- gz (bool):\nIf True, assumes that data is gzipped. If False, assumes JSON file.
\n- full_output (bool):\nIf True, a dict containing auxiliary information and the data is returned.\nIf False, only the data is returned.
\n- reps (str):\nSpecify the structure of the placeholder in imported dict to be reps[0-9]+.
\n
\n\nReturns
\n\n\n- data (Obs / list / Corr):\nRead data
\n- or
\n- data (dict):\nRead data and meta-data
\n
\n", "signature": "(fname, verbose=True, gz=True, full_output=False, reps='DICTOBS'):", "funcdef": "def"}, "pyerrors.input.misc": {"fullname": "pyerrors.input.misc", "modulename": "pyerrors.input.misc", "kind": "module", "doc": "\n"}, "pyerrors.input.misc.read_pbp": {"fullname": "pyerrors.input.misc.read_pbp", "modulename": "pyerrors.input.misc", "qualname": "read_pbp", "kind": "function", "doc": "Read pbp format from given folder structure.
\n\nParameters
\n\n\n- r_start (list):\nlist which contains the first config to be read for each replicum
\n- r_stop (list):\nlist which contains the last config to be read for each replicum
\n
\n\nReturns
\n\n\n- result (list[Obs]):\nlist of observables read
\n
\n", "signature": "(path, prefix, **kwargs):", "funcdef": "def"}, "pyerrors.input.openQCD": {"fullname": "pyerrors.input.openQCD", "modulename": "pyerrors.input.openQCD", "kind": "module", "doc": "\n"}, "pyerrors.input.openQCD.read_rwms": {"fullname": "pyerrors.input.openQCD.read_rwms", "modulename": "pyerrors.input.openQCD", "qualname": "read_rwms", "kind": "function", "doc": "Read rwms format from given folder structure. Returns a list of length nrw
\n\nParameters
\n\n\n- path (str):\npath that contains the data files
\n- prefix (str):\nall files in path that start with prefix are considered as input files.\nMay be used together postfix to consider only special file endings.\nPrefix is ignored, if the keyword 'files' is used.
\n- version (str):\nversion of openQCD, default 2.0
\n- names (list):\nlist of names that is assigned to the data according according\nto the order in the file list. Use careful, if you do not provide file names!
\n- r_start (list):\nlist which contains the first config to be read for each replicum
\n- r_stop (list):\nlist which contains the last config to be read for each replicum
\n- r_step (int):\ninteger that defines a fixed step size between two measurements (in units of configs)\nIf not given, r_step=1 is assumed.
\n- postfix (str):\npostfix of the file to read, e.g. '.ms1' for openQCD-files
\n- files (list):\nlist which contains the filenames to be read. No automatic detection of\nfiles performed if given.
\n- print_err (bool):\nPrint additional information that is useful for debugging.
\n
\n\nReturns
\n\n\n- rwms (Obs):\nReweighting factors read
\n
\n", "signature": "(path, prefix, version='2.0', names=None, **kwargs):", "funcdef": "def"}, "pyerrors.input.openQCD.extract_t0": {"fullname": "pyerrors.input.openQCD.extract_t0", "modulename": "pyerrors.input.openQCD", "qualname": "extract_t0", "kind": "function", "doc": "Extract t0 from given .ms.dat files. Returns t0 as Obs.
\n\nIt is assumed that all boundary effects have\nsufficiently decayed at x0=xmin.\nThe data around the zero crossing of t^2 - 0.3\nis fitted with a linear function\nfrom which the exact root is extracted.
\n\nIt is assumed that one measurement is performed for each config.\nIf this is not the case, the resulting idl, as well as the handling\nof r_start, r_stop and r_step is wrong and the user has to correct\nthis in the resulting observable.
\n\nParameters
\n\n\n- path (str):\nPath to .ms.dat files
\n- prefix (str):\nEnsemble prefix
\n- dtr_read (int):\nDetermines how many trajectories should be skipped\nwhen reading the ms.dat files.\nCorresponds to dtr_cnfg / dtr_ms in the openQCD input file.
\n- xmin (int):\nFirst timeslice where the boundary\neffects have sufficiently decayed.
\n- spatial_extent (int):\nspatial extent of the lattice, required for normalization.
\n- fit_range (int):\nNumber of data points left and right of the zero\ncrossing to be included in the linear fit. (Default: 5)
\n- r_start (list):\nlist which contains the first config to be read for each replicum.
\n- r_stop (list):\nlist which contains the last config to be read for each replicum.
\n- r_step (int):\ninteger that defines a fixed step size between two measurements (in units of configs)\nIf not given, r_step=1 is assumed.
\n- plaquette (bool):\nIf true extract the plaquette estimate of t0 instead.
\n- names (list):\nlist of names that is assigned to the data according according\nto the order in the file list. Use careful, if you do not provide file names!
\n- files (list):\nlist which contains the filenames to be read. No automatic detection of\nfiles performed if given.
\n- plot_fit (bool):\nIf true, the fit for the extraction of t0 is shown together with the data.
\n- assume_thermalization (bool):\nIf True: If the first record divided by the distance between two measurements is larger than\n1, it is assumed that this is due to thermalization and the first measurement belongs\nto the first config (default).\nIf False: The config numbers are assumed to be traj_number // difference
\n
\n\nReturns
\n\n\n- t0 (Obs):\nExtracted t0
\n
\n", "signature": "(path, prefix, dtr_read, xmin, spatial_extent, fit_range=5, **kwargs):", "funcdef": "def"}, "pyerrors.input.openQCD.read_qtop": {"fullname": "pyerrors.input.openQCD.read_qtop", "modulename": "pyerrors.input.openQCD", "qualname": "read_qtop", "kind": "function", "doc": "Read the topologial charge based on openQCD gradient flow measurements.
\n\nParameters
\n\n\n- path (str):\npath of the measurement files
\n- prefix (str):\nprefix of the measurement files, e.g. _id0_r0.ms.dat.\nIgnored if file names are passed explicitly via keyword files.
\n- c (double):\nSmearing radius in units of the lattice extent, c = sqrt(8 t0) / L.
\n- dtr_cnfg (int):\n(optional) parameter that specifies the number of measurements\nbetween two configs.\nIf it is not set, the distance between two measurements\nin the file is assumed to be the distance between two configurations.
\n- steps (int):\n(optional) Distance between two configurations in units of trajectories /\n cycles. Assumed to be the distance between two measurements * dtr_cnfg if not given
\n- version (str):\nEither openQCD or sfqcd, depending on the data.
\n- L (int):\nspatial length of the lattice in L/a.\nHAS to be set if version != sfqcd, since openQCD does not provide\nthis in the header
\n- r_start (list):\nlist which contains the first config to be read for each replicum.
\n- r_stop (list):\nlist which contains the last config to be read for each replicum.
\n- files (list):\nspecify the exact files that need to be read\nfrom path, practical if e.g. only one replicum is needed
\n- postfix (str):\npostfix of the file to read, e.g. '.gfms.dat' for openQCD-files
\n- names (list):\nAlternative labeling for replicas/ensembles.\nHas to have the appropriate length.
\n- Zeuthen_flow (bool):\n(optional) If True, the Zeuthen flow is used for Qtop. Only possible\nfor version=='sfqcd' If False, the Wilson flow is used.
\n- integer_charge (bool):\nIf True, the charge is rounded towards the nearest integer on each config.
\n
\n\nReturns
\n\n\n- result (Obs):\nRead topological charge
\n
\n", "signature": "(path, prefix, c, dtr_cnfg=1, version='openQCD', **kwargs):", "funcdef": "def"}, "pyerrors.input.openQCD.read_gf_coupling": {"fullname": "pyerrors.input.openQCD.read_gf_coupling", "modulename": "pyerrors.input.openQCD", "qualname": "read_gf_coupling", "kind": "function", "doc": "Read the gradient flow coupling based on sfqcd gradient flow measurements. See 1607.06423 for details.
\n\nNote: The current implementation only works for c=0.3 and T=L. The definition of the coupling in 1607.06423 requires projection to topological charge zero which is not done within this function but has to be performed in a separate step.
\n\nParameters
\n\n\n- path (str):\npath of the measurement files
\n- prefix (str):\nprefix of the measurement files, e.g. _id0_r0.ms.dat.\nIgnored if file names are passed explicitly via keyword files.
\n- c (double):\nSmearing radius in units of the lattice extent, c = sqrt(8 t0) / L.
\n- dtr_cnfg (int):\n(optional) parameter that specifies the number of measurements\nbetween two configs.\nIf it is not set, the distance between two measurements\nin the file is assumed to be the distance between two configurations.
\n- steps (int):\n(optional) Distance between two configurations in units of trajectories /\n cycles. Assumed to be the distance between two measurements * dtr_cnfg if not given
\n- r_start (list):\nlist which contains the first config to be read for each replicum.
\n- r_stop (list):\nlist which contains the last config to be read for each replicum.
\n- files (list):\nspecify the exact files that need to be read\nfrom path, practical if e.g. only one replicum is needed
\n- names (list):\nAlternative labeling for replicas/ensembles.\nHas to have the appropriate length.
\n- postfix (str):\npostfix of the file to read, e.g. '.gfms.dat' for openQCD-files
\n- Zeuthen_flow (bool):\n(optional) If True, the Zeuthen flow is used for the coupling. If False, the Wilson flow is used.
\n
\n", "signature": "(path, prefix, c, dtr_cnfg=1, Zeuthen_flow=True, **kwargs):", "funcdef": "def"}, "pyerrors.input.openQCD.qtop_projection": {"fullname": "pyerrors.input.openQCD.qtop_projection", "modulename": "pyerrors.input.openQCD", "qualname": "qtop_projection", "kind": "function", "doc": "Returns the projection to the topological charge sector defined by target.
\n\nParameters
\n\n\n- path (Obs):\nTopological charge.
\n- target (int):\nSpecifies the topological sector to be reweighted to (default 0)
\n
\n\nReturns
\n\n\n- reto (Obs):\nprojection to the topological charge sector defined by target
\n
\n", "signature": "(qtop, target=0):", "funcdef": "def"}, "pyerrors.input.openQCD.read_qtop_sector": {"fullname": "pyerrors.input.openQCD.read_qtop_sector", "modulename": "pyerrors.input.openQCD", "qualname": "read_qtop_sector", "kind": "function", "doc": "Constructs reweighting factors to a specified topological sector.
\n\nParameters
\n\n\n- path (str):\npath of the measurement files
\n- prefix (str):\nprefix of the measurement files, e.g. _id0_r0.ms.dat
\n- c (double):\nSmearing radius in units of the lattice extent, c = sqrt(8 t0) / L
\n- target (int):\nSpecifies the topological sector to be reweighted to (default 0)
\n- dtr_cnfg (int):\n(optional) parameter that specifies the number of trajectories\nbetween two configs.\nif it is not set, the distance between two measurements\nin the file is assumed to be the distance between two configurations.
\n- steps (int):\n(optional) Distance between two configurations in units of trajectories /\n cycles. Assumed to be the distance between two measurements * dtr_cnfg if not given
\n- version (str):\nversion string of the openQCD (sfqcd) version used to create\nthe ensemble. Default is 2.0. May also be set to sfqcd.
\n- L (int):\nspatial length of the lattice in L/a.\nHAS to be set if version != sfqcd, since openQCD does not provide\nthis in the header
\n- r_start (list):\noffset of the first ensemble, making it easier to match\nlater on with other Obs
\n- r_stop (list):\nlast configurations that need to be read (per replicum)
\n- files (list):\nspecify the exact files that need to be read\nfrom path, practical if e.g. only one replicum is needed
\n- names (list):\nAlternative labeling for replicas/ensembles.\nHas to have the appropriate length
\n- Zeuthen_flow (bool):\n(optional) If True, the Zeuthen flow is used for Qtop. Only possible\nfor version=='sfqcd' If False, the Wilson flow is used.
\n
\n\nReturns
\n\n\n- reto (Obs):\nprojection to the topological charge sector defined by target
\n
\n", "signature": "(path, prefix, c, target=0, **kwargs):", "funcdef": "def"}, "pyerrors.input.openQCD.read_ms5_xsf": {"fullname": "pyerrors.input.openQCD.read_ms5_xsf", "modulename": "pyerrors.input.openQCD", "qualname": "read_ms5_xsf", "kind": "function", "doc": "Read data from files in the specified directory with the specified prefix and quark combination extension, and return a Corr
object containing the data.
\n\nParameters
\n\n\n\nReturns
\n\n\n- Corr: A complex valued
Corr
object containing the data read from the files. In case of boudary to bulk correlators. \n- or
\n- CObs: A complex valued
CObs
object containing the data read from the files. In case of boudary to boundary correlators. \n
\n\nRaises
\n\n\n- FileNotFoundError: If no files matching the specified prefix and quark combination extension are found in the specified directory.
\n- IOError: If there is an error reading a file.
\n- struct.error: If there is an error unpacking binary data.
\n
\n", "signature": "(path, prefix, qc, corr, sep='r', **kwargs):", "funcdef": "def"}, "pyerrors.input.pandas": {"fullname": "pyerrors.input.pandas", "modulename": "pyerrors.input.pandas", "kind": "module", "doc": "\n"}, "pyerrors.input.pandas.to_sql": {"fullname": "pyerrors.input.pandas.to_sql", "modulename": "pyerrors.input.pandas", "qualname": "to_sql", "kind": "function", "doc": "Write DataFrame including Obs or Corr valued columns to sqlite database.
\n\nParameters
\n\n\n- df (pandas.DataFrame):\nDataframe to be written to the database.
\n- table_name (str):\nName of the table in the database.
\n- db (str):\nPath to the sqlite database.
\n- if exists (str):\nHow to behave if table already exists. Options 'fail', 'replace', 'append'.
\n- gz (bool):\nIf True the json strings are gzipped.
\n
\n\nReturns
\n\n\n", "signature": "(df, table_name, db, if_exists='fail', gz=True, **kwargs):", "funcdef": "def"}, "pyerrors.input.pandas.read_sql": {"fullname": "pyerrors.input.pandas.read_sql", "modulename": "pyerrors.input.pandas", "qualname": "read_sql", "kind": "function", "doc": "Execute SQL query on sqlite database and obtain DataFrame including Obs or Corr valued columns.
\n\nParameters
\n\n\n- sql (str):\nSQL query to be executed.
\n- db (str):\nPath to the sqlite database.
\n- auto_gamma (bool):\nIf True applies the gamma_method to all imported Obs objects with the default parameters for\nthe error analysis. Default False.
\n
\n\nReturns
\n\n\n- data (pandas.DataFrame):\nDataframe with the content of the sqlite database.
\n
\n", "signature": "(sql, db, auto_gamma=False, **kwargs):", "funcdef": "def"}, "pyerrors.input.pandas.dump_df": {"fullname": "pyerrors.input.pandas.dump_df", "modulename": "pyerrors.input.pandas", "qualname": "dump_df", "kind": "function", "doc": "Exports a pandas DataFrame containing Obs valued columns to a (gzipped) csv file.
\n\nBefore making use of pandas to_csv functionality Obs objects are serialized via the standardized\njson format of pyerrors.
\n\nParameters
\n\n\n- df (pandas.DataFrame):\nDataframe to be dumped to a file.
\n- fname (str):\nFilename of the output file.
\n- gz (bool):\nIf True, the output is a gzipped csv file. If False, the output is a csv file.
\n
\n\nReturns
\n\n\n", "signature": "(df, fname, gz=True):", "funcdef": "def"}, "pyerrors.input.pandas.load_df": {"fullname": "pyerrors.input.pandas.load_df", "modulename": "pyerrors.input.pandas", "qualname": "load_df", "kind": "function", "doc": "Imports a pandas DataFrame from a csv.(gz) file in which Obs objects are serialized as json strings.
\n\nParameters
\n\n\n- fname (str):\nFilename of the input file.
\n- auto_gamma (bool):\nIf True applies the gamma_method to all imported Obs objects with the default parameters for\nthe error analysis. Default False.
\n- gz (bool):\nIf True, assumes that data is gzipped. If False, assumes JSON file.
\n
\n\nReturns
\n\n\n- data (pandas.DataFrame):\nDataframe with the content of the sqlite database.
\n
\n", "signature": "(fname, auto_gamma=False, gz=True):", "funcdef": "def"}, "pyerrors.input.sfcf": {"fullname": "pyerrors.input.sfcf", "modulename": "pyerrors.input.sfcf", "kind": "module", "doc": "\n"}, "pyerrors.input.sfcf.read_sfcf": {"fullname": "pyerrors.input.sfcf.read_sfcf", "modulename": "pyerrors.input.sfcf", "qualname": "read_sfcf", "kind": "function", "doc": "Read sfcf c format from given folder structure.
\n\nParameters
\n\n\n- path (str):\nPath to the sfcf files.
\n- prefix (str):\nPrefix of the sfcf files.
\n- name (str):\nName of the correlation function to read.
\n- quarks (str):\nLabel of the quarks used in the sfcf input file. e.g. \"quark quark\"\nfor version 0.0 this does NOT need to be given with the typical \" - \"\nthat is present in the output file,\nthis is done automatically for this version
\n- corr_type (str):\nType of correlation function to read. Can be\n
\n- 'bi' for boundary-inner
\n- 'bb' for boundary-boundary
\n- 'bib' for boundary-inner-boundary
\n
\n- noffset (int):\nOffset of the source (only relevant when wavefunctions are used)
\n- wf (int):\nID of wave function
\n- wf2 (int):\nID of the second wavefunction\n(only relevant for boundary-to-boundary correlation functions)
\n- im (bool):\nif True, read imaginary instead of real part\nof the correlation function.
\n- names (list):\nAlternative labeling for replicas/ensembles.\nHas to have the appropriate length
\n- ens_name (str):\nreplaces the name of the ensemble
\n- version (str):\nversion of SFCF, with which the measurement was done.\nif the compact output option (-c) was specified,\nappend a \"c\" to the version (e.g. \"1.0c\")\nif the append output option (-a) was specified,\nappend an \"a\" to the version
\n- cfg_separator (str):\nString that separates the ensemble identifier from the configuration number (default 'n').
\n- replica (list):\nlist of replica to be read, default is all
\n- files (list):\nlist of files to be read per replica, default is all.\nfor non-compact output format, hand the folders to be read here.
\n- check_configs (list[list[int]]):\nlist of list of supposed configs, eg. [range(1,1000)]\nfor one replicum with 1000 configs
\n
\n\nReturns
\n\n\n- result (list[Obs]):\nlist of Observables with length T, observable per timeslice.\nbb-type correlators have length 1.
\n
\n", "signature": "(\tpath,\tprefix,\tname,\tquarks='.*',\tcorr_type='bi',\tnoffset=0,\twf=0,\twf2=0,\tversion='1.0c',\tcfg_separator='n',\t**kwargs):", "funcdef": "def"}, "pyerrors.input.utils": {"fullname": "pyerrors.input.utils", "modulename": "pyerrors.input.utils", "kind": "module", "doc": "Utilities for the input
\n"}, "pyerrors.input.utils.check_idl": {"fullname": "pyerrors.input.utils.check_idl", "modulename": "pyerrors.input.utils", "qualname": "check_idl", "kind": "function", "doc": "Checks if list of configurations is contained in an idl
\n\nParameters
\n\n\n- idl (range or list):\nidl of the current replicum
\n- che (list):\nlist of configurations to be checked against
\n
\n\nReturns
\n\n\n- miss_str (str):\nstring with integers of which idls are missing
\n
\n", "signature": "(idl, che):", "funcdef": "def"}, "pyerrors.linalg": {"fullname": "pyerrors.linalg", "modulename": "pyerrors.linalg", "kind": "module", "doc": "\n"}, "pyerrors.linalg.matmul": {"fullname": "pyerrors.linalg.matmul", "modulename": "pyerrors.linalg", "qualname": "matmul", "kind": "function", "doc": "Matrix multiply all operands.
\n\nParameters
\n\n\n- operands (numpy.ndarray):\nArbitrary number of 2d-numpy arrays which can be real or complex\nObs valued.
\n- This implementation is faster compared to standard multiplication via the @ operator.
\n
\n", "signature": "(*operands):", "funcdef": "def"}, "pyerrors.linalg.jack_matmul": {"fullname": "pyerrors.linalg.jack_matmul", "modulename": "pyerrors.linalg", "qualname": "jack_matmul", "kind": "function", "doc": "Matrix multiply both operands making use of the jackknife approximation.
\n\nParameters
\n\n\n- operands (numpy.ndarray):\nArbitrary number of 2d-numpy arrays which can be real or complex\nObs valued.
\n- For large matrices this is considerably faster compared to matmul.
\n
\n", "signature": "(*operands):", "funcdef": "def"}, "pyerrors.linalg.einsum": {"fullname": "pyerrors.linalg.einsum", "modulename": "pyerrors.linalg", "qualname": "einsum", "kind": "function", "doc": "Wrapper for numpy.einsum
\n\nParameters
\n\n\n- subscripts (str):\nSubscripts for summation (see numpy documentation for details)
\n- operands (numpy.ndarray):\nArbitrary number of 2d-numpy arrays which can be real or complex\nObs valued.
\n
\n", "signature": "(subscripts, *operands):", "funcdef": "def"}, "pyerrors.linalg.inv": {"fullname": "pyerrors.linalg.inv", "modulename": "pyerrors.linalg", "qualname": "inv", "kind": "function", "doc": "Inverse of Obs or CObs valued matrices.
\n", "signature": "(x):", "funcdef": "def"}, "pyerrors.linalg.cholesky": {"fullname": "pyerrors.linalg.cholesky", "modulename": "pyerrors.linalg", "qualname": "cholesky", "kind": "function", "doc": "Cholesky decomposition of Obs valued matrices.
\n", "signature": "(x):", "funcdef": "def"}, "pyerrors.linalg.det": {"fullname": "pyerrors.linalg.det", "modulename": "pyerrors.linalg", "qualname": "det", "kind": "function", "doc": "Determinant of Obs valued matrices.
\n", "signature": "(x):", "funcdef": "def"}, "pyerrors.linalg.eigh": {"fullname": "pyerrors.linalg.eigh", "modulename": "pyerrors.linalg", "qualname": "eigh", "kind": "function", "doc": "Computes the eigenvalues and eigenvectors of a given hermitian matrix of Obs according to np.linalg.eigh.
\n", "signature": "(obs, **kwargs):", "funcdef": "def"}, "pyerrors.linalg.eig": {"fullname": "pyerrors.linalg.eig", "modulename": "pyerrors.linalg", "qualname": "eig", "kind": "function", "doc": "Computes the eigenvalues of a given matrix of Obs according to np.linalg.eig.
\n", "signature": "(obs, **kwargs):", "funcdef": "def"}, "pyerrors.linalg.pinv": {"fullname": "pyerrors.linalg.pinv", "modulename": "pyerrors.linalg", "qualname": "pinv", "kind": "function", "doc": "Computes the Moore-Penrose pseudoinverse of a matrix of Obs.
\n", "signature": "(obs, **kwargs):", "funcdef": "def"}, "pyerrors.linalg.svd": {"fullname": "pyerrors.linalg.svd", "modulename": "pyerrors.linalg", "qualname": "svd", "kind": "function", "doc": "Computes the singular value decomposition of a matrix of Obs.
\n", "signature": "(obs, **kwargs):", "funcdef": "def"}, "pyerrors.misc": {"fullname": "pyerrors.misc", "modulename": "pyerrors.misc", "kind": "module", "doc": "\n"}, "pyerrors.misc.dump_object": {"fullname": "pyerrors.misc.dump_object", "modulename": "pyerrors.misc", "qualname": "dump_object", "kind": "function", "doc": "Dump object into pickle file.
\n\nParameters
\n\n\n- obj (object):\nobject to be saved in the pickle file
\n- name (str):\nname of the file
\n- path (str):\nspecifies a custom path for the file (default '.')
\n
\n\nReturns
\n\n\n", "signature": "(obj, name, **kwargs):", "funcdef": "def"}, "pyerrors.misc.load_object": {"fullname": "pyerrors.misc.load_object", "modulename": "pyerrors.misc", "qualname": "load_object", "kind": "function", "doc": "Load object from pickle file.
\n\nParameters
\n\n\n- path (str):\npath to the file
\n
\n\nReturns
\n\n\n- object (Obs):\nLoaded Object
\n
\n", "signature": "(path):", "funcdef": "def"}, "pyerrors.misc.pseudo_Obs": {"fullname": "pyerrors.misc.pseudo_Obs", "modulename": "pyerrors.misc", "qualname": "pseudo_Obs", "kind": "function", "doc": "Generate an Obs object with given value, dvalue and name for test purposes
\n\nParameters
\n\n\n- value (float):\ncentral value of the Obs to be generated.
\n- dvalue (float):\nerror of the Obs to be generated.
\n- name (str):\nname of the ensemble for which the Obs is to be generated.
\n- samples (int):\nnumber of samples for the Obs (default 1000).
\n
\n\nReturns
\n\n\n- res (Obs):\nGenerated Observable
\n
\n", "signature": "(value, dvalue, name, samples=1000):", "funcdef": "def"}, "pyerrors.misc.gen_correlated_data": {"fullname": "pyerrors.misc.gen_correlated_data", "modulename": "pyerrors.misc", "qualname": "gen_correlated_data", "kind": "function", "doc": "Generate observables with given covariance and autocorrelation times.
\n\nParameters
\n\n\n- means (list):\nlist containing the mean value of each observable.
\n- cov (numpy.ndarray):\ncovariance matrix for the data to be generated.
\n- name (str):\nensemble name for the data to be geneated.
\n- tau (float or list):\ncan either be a real number or a list with an entry for\nevery dataset.
\n- samples (int):\nnumber of samples to be generated for each observable.
\n
\n\nReturns
\n\n\n- corr_obs (list[Obs]):\nGenerated observable list
\n
\n", "signature": "(means, cov, name, tau=0.5, samples=1000):", "funcdef": "def"}, "pyerrors.mpm": {"fullname": "pyerrors.mpm", "modulename": "pyerrors.mpm", "kind": "module", "doc": "\n"}, "pyerrors.mpm.matrix_pencil_method": {"fullname": "pyerrors.mpm.matrix_pencil_method", "modulename": "pyerrors.mpm", "qualname": "matrix_pencil_method", "kind": "function", "doc": "Matrix pencil method to extract k energy levels from data
\n\nImplementation of the matrix pencil method based on\neq. (2.17) of Y. Hua, T. K. Sarkar, IEEE Trans. Acoust. 38, 814-824 (1990)
\n\nParameters
\n\n\n- data (list):\ncan be a list of Obs for the analysis of a single correlator, or a list of lists\nof Obs if several correlators are to analyzed at once.
\n- k (int):\nNumber of states to extract (default 1).
\n- p (int):\nmatrix pencil parameter which filters noise. The optimal value is expected between\nlen(data)/3 and 2*len(data)/3. The computation is more expensive the closer p is\nto len(data)/2 but could possibly suppress more noise (default len(data)//2).
\n
\n\nReturns
\n\n\n- energy_levels (list[Obs]):\nExtracted energy levels
\n
\n", "signature": "(corrs, k=1, p=None, **kwargs):", "funcdef": "def"}, "pyerrors.obs": {"fullname": "pyerrors.obs", "modulename": "pyerrors.obs", "kind": "module", "doc": "\n"}, "pyerrors.obs.Obs": {"fullname": "pyerrors.obs.Obs", "modulename": "pyerrors.obs", "qualname": "Obs", "kind": "class", "doc": "Class for a general observable.
\n\nInstances of Obs are the basic objects of a pyerrors error analysis.\nThey are initialized with a list which contains arrays of samples for\ndifferent ensembles/replica and another list of same length which contains\nthe names of the ensembles/replica. Mathematical operations can be\nperformed on instances. The result is another instance of Obs. The error of\nan instance can be computed with the gamma_method. Also contains additional\nmethods for output and visualization of the error calculation.
\n\nAttributes
\n\n\n- S_global (float):\nStandard value for S (default 2.0)
\n- S_dict (dict):\nDictionary for S values. If an entry for a given ensemble\nexists this overwrites the standard value for that ensemble.
\n- tau_exp_global (float):\nStandard value for tau_exp (default 0.0)
\n- tau_exp_dict (dict):\nDictionary for tau_exp values. If an entry for a given ensemble exists\nthis overwrites the standard value for that ensemble.
\n- N_sigma_global (float):\nStandard value for N_sigma (default 1.0)
\n- N_sigma_dict (dict):\nDictionary for N_sigma values. If an entry for a given ensemble exists\nthis overwrites the standard value for that ensemble.
\n
\n"}, "pyerrors.obs.Obs.__init__": {"fullname": "pyerrors.obs.Obs.__init__", "modulename": "pyerrors.obs", "qualname": "Obs.__init__", "kind": "function", "doc": "Initialize Obs object.
\n\nParameters
\n\n\n- samples (list):\nlist of numpy arrays containing the Monte Carlo samples
\n- names (list):\nlist of strings labeling the individual samples
\n- idl (list, optional):\nlist of ranges or lists on which the samples are defined
\n
\n", "signature": "(samples, names, idl=None, **kwargs)"}, "pyerrors.obs.Obs.gamma_method": {"fullname": "pyerrors.obs.Obs.gamma_method", "modulename": "pyerrors.obs", "qualname": "Obs.gamma_method", "kind": "function", "doc": "Estimate the error and related properties of the Obs.
\n\nParameters
\n\n\n- S (float):\nspecifies a custom value for the parameter S (default 2.0).\nIf set to 0 it is assumed that the data exhibits no\nautocorrelation. In this case the error estimates coincides\nwith the sample standard error.
\n- tau_exp (float):\npositive value triggers the critical slowing down analysis\n(default 0.0).
\n- N_sigma (float):\nnumber of standard deviations from zero until the tail is\nattached to the autocorrelation function (default 1).
\n- fft (bool):\ndetermines whether the fft algorithm is used for the computation\nof the autocorrelation function (default True)
\n
\n", "signature": "(self, **kwargs):", "funcdef": "def"}, "pyerrors.obs.Obs.gm": {"fullname": "pyerrors.obs.Obs.gm", "modulename": "pyerrors.obs", "qualname": "Obs.gm", "kind": "function", "doc": "Estimate the error and related properties of the Obs.
\n\nParameters
\n\n\n- S (float):\nspecifies a custom value for the parameter S (default 2.0).\nIf set to 0 it is assumed that the data exhibits no\nautocorrelation. In this case the error estimates coincides\nwith the sample standard error.
\n- tau_exp (float):\npositive value triggers the critical slowing down analysis\n(default 0.0).
\n- N_sigma (float):\nnumber of standard deviations from zero until the tail is\nattached to the autocorrelation function (default 1).
\n- fft (bool):\ndetermines whether the fft algorithm is used for the computation\nof the autocorrelation function (default True)
\n
\n", "signature": "(self, **kwargs):", "funcdef": "def"}, "pyerrors.obs.Obs.details": {"fullname": "pyerrors.obs.Obs.details", "modulename": "pyerrors.obs", "qualname": "Obs.details", "kind": "function", "doc": "Output detailed properties of the Obs.
\n\nParameters
\n\n\n- ens_content (bool):\nprint details about the ensembles and replica if true.
\n
\n", "signature": "(self, ens_content=True):", "funcdef": "def"}, "pyerrors.obs.Obs.reweight": {"fullname": "pyerrors.obs.Obs.reweight", "modulename": "pyerrors.obs", "qualname": "Obs.reweight", "kind": "function", "doc": "Reweight the obs with given rewighting factors.
\n\nParameters
\n\n\n- weight (Obs):\nReweighting factor. An Observable that has to be defined on a superset of the\nconfigurations in obs[i].idl for all i.
\n- all_configs (bool):\nif True, the reweighted observables are normalized by the average of\nthe reweighting factor on all configurations in weight.idl and not\non the configurations in obs[i].idl. Default False.
\n
\n", "signature": "(self, weight):", "funcdef": "def"}, "pyerrors.obs.Obs.is_zero_within_error": {"fullname": "pyerrors.obs.Obs.is_zero_within_error", "modulename": "pyerrors.obs", "qualname": "Obs.is_zero_within_error", "kind": "function", "doc": "Checks whether the observable is zero within 'sigma' standard errors.
\n\nParameters
\n\n\n- sigma (int):\nNumber of standard errors used for the check.
\n- Works only properly when the gamma method was run.
\n
\n", "signature": "(self, sigma=1):", "funcdef": "def"}, "pyerrors.obs.Obs.is_zero": {"fullname": "pyerrors.obs.Obs.is_zero", "modulename": "pyerrors.obs", "qualname": "Obs.is_zero", "kind": "function", "doc": "Checks whether the observable is zero within a given tolerance.
\n\nParameters
\n\n\n- atol (float):\nAbsolute tolerance (for details see numpy documentation).
\n
\n", "signature": "(self, atol=1e-10):", "funcdef": "def"}, "pyerrors.obs.Obs.plot_tauint": {"fullname": "pyerrors.obs.Obs.plot_tauint", "modulename": "pyerrors.obs", "qualname": "Obs.plot_tauint", "kind": "function", "doc": "Plot integrated autocorrelation time for each ensemble.
\n\nParameters
\n\n\n- save (str):\nsaves the figure to a file named 'save' if.
\n
\n", "signature": "(self, save=None):", "funcdef": "def"}, "pyerrors.obs.Obs.plot_rho": {"fullname": "pyerrors.obs.Obs.plot_rho", "modulename": "pyerrors.obs", "qualname": "Obs.plot_rho", "kind": "function", "doc": "Plot normalized autocorrelation function time for each ensemble.
\n\nParameters
\n\n\n- save (str):\nsaves the figure to a file named 'save' if.
\n
\n", "signature": "(self, save=None):", "funcdef": "def"}, "pyerrors.obs.Obs.plot_rep_dist": {"fullname": "pyerrors.obs.Obs.plot_rep_dist", "modulename": "pyerrors.obs", "qualname": "Obs.plot_rep_dist", "kind": "function", "doc": "Plot replica distribution for each ensemble with more than one replicum.
\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.obs.Obs.plot_history": {"fullname": "pyerrors.obs.Obs.plot_history", "modulename": "pyerrors.obs", "qualname": "Obs.plot_history", "kind": "function", "doc": "Plot derived Monte Carlo history for each ensemble
\n\nParameters
\n\n\n- expand (bool):\nshow expanded history for irregular Monte Carlo chains (default: True).
\n
\n", "signature": "(self, expand=True):", "funcdef": "def"}, "pyerrors.obs.Obs.plot_piechart": {"fullname": "pyerrors.obs.Obs.plot_piechart", "modulename": "pyerrors.obs", "qualname": "Obs.plot_piechart", "kind": "function", "doc": "Plot piechart which shows the fractional contribution of each\nensemble to the error and returns a dictionary containing the fractions.
\n\nParameters
\n\n\n- save (str):\nsaves the figure to a file named 'save' if.
\n
\n", "signature": "(self, save=None):", "funcdef": "def"}, "pyerrors.obs.Obs.dump": {"fullname": "pyerrors.obs.Obs.dump", "modulename": "pyerrors.obs", "qualname": "Obs.dump", "kind": "function", "doc": "Dump the Obs to a file 'name' of chosen format.
\n\nParameters
\n\n\n- filename (str):\nname of the file to be saved.
\n- datatype (str):\nFormat of the exported file. Supported formats include\n\"json.gz\" and \"pickle\"
\n- description (str):\nDescription for output file, only relevant for json.gz format.
\n- path (str):\nspecifies a custom path for the file (default '.')
\n
\n", "signature": "(self, filename, datatype='json.gz', description='', **kwargs):", "funcdef": "def"}, "pyerrors.obs.Obs.export_jackknife": {"fullname": "pyerrors.obs.Obs.export_jackknife", "modulename": "pyerrors.obs", "qualname": "Obs.export_jackknife", "kind": "function", "doc": "Export jackknife samples from the Obs
\n\nReturns
\n\n\n- numpy.ndarray: Returns a numpy array of length N + 1 where N is the number of samples\nfor the given ensemble and replicum. The zeroth entry of the array contains\nthe mean value of the Obs, entries 1 to N contain the N jackknife samples\nderived from the Obs. The current implementation only works for observables\ndefined on exactly one ensemble and replicum. The derived jackknife samples\nshould agree with samples from a full jackknife analysis up to O(1/N).
\n
\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.obs.Obs.sqrt": {"fullname": "pyerrors.obs.Obs.sqrt", "modulename": "pyerrors.obs", "qualname": "Obs.sqrt", "kind": "function", "doc": "\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.obs.Obs.log": {"fullname": "pyerrors.obs.Obs.log", "modulename": "pyerrors.obs", "qualname": "Obs.log", "kind": "function", "doc": "\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.obs.Obs.exp": {"fullname": "pyerrors.obs.Obs.exp", "modulename": "pyerrors.obs", "qualname": "Obs.exp", "kind": "function", "doc": "\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.obs.Obs.sin": {"fullname": "pyerrors.obs.Obs.sin", "modulename": "pyerrors.obs", "qualname": "Obs.sin", "kind": "function", "doc": "\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.obs.Obs.cos": {"fullname": "pyerrors.obs.Obs.cos", "modulename": "pyerrors.obs", "qualname": "Obs.cos", "kind": "function", "doc": "\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.obs.Obs.tan": {"fullname": "pyerrors.obs.Obs.tan", "modulename": "pyerrors.obs", "qualname": "Obs.tan", "kind": "function", "doc": "\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.obs.Obs.arcsin": {"fullname": "pyerrors.obs.Obs.arcsin", "modulename": "pyerrors.obs", "qualname": "Obs.arcsin", "kind": "function", "doc": "\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.obs.Obs.arccos": {"fullname": "pyerrors.obs.Obs.arccos", "modulename": "pyerrors.obs", "qualname": "Obs.arccos", "kind": "function", "doc": "\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.obs.Obs.arctan": {"fullname": "pyerrors.obs.Obs.arctan", "modulename": "pyerrors.obs", "qualname": "Obs.arctan", "kind": "function", "doc": "\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.obs.Obs.sinh": {"fullname": "pyerrors.obs.Obs.sinh", "modulename": "pyerrors.obs", "qualname": "Obs.sinh", "kind": "function", "doc": "\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.obs.Obs.cosh": {"fullname": "pyerrors.obs.Obs.cosh", "modulename": "pyerrors.obs", "qualname": "Obs.cosh", "kind": "function", "doc": "\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.obs.Obs.tanh": {"fullname": "pyerrors.obs.Obs.tanh", "modulename": "pyerrors.obs", "qualname": "Obs.tanh", "kind": "function", "doc": "\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.obs.Obs.arcsinh": {"fullname": "pyerrors.obs.Obs.arcsinh", "modulename": "pyerrors.obs", "qualname": "Obs.arcsinh", "kind": "function", "doc": "\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.obs.Obs.arccosh": {"fullname": "pyerrors.obs.Obs.arccosh", "modulename": "pyerrors.obs", "qualname": "Obs.arccosh", "kind": "function", "doc": "\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.obs.Obs.arctanh": {"fullname": "pyerrors.obs.Obs.arctanh", "modulename": "pyerrors.obs", "qualname": "Obs.arctanh", "kind": "function", "doc": "\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.obs.CObs": {"fullname": "pyerrors.obs.CObs", "modulename": "pyerrors.obs", "qualname": "CObs", "kind": "class", "doc": "Class for a complex valued observable.
\n"}, "pyerrors.obs.CObs.__init__": {"fullname": "pyerrors.obs.CObs.__init__", "modulename": "pyerrors.obs", "qualname": "CObs.__init__", "kind": "function", "doc": "\n", "signature": "(real, imag=0.0)"}, "pyerrors.obs.CObs.gamma_method": {"fullname": "pyerrors.obs.CObs.gamma_method", "modulename": "pyerrors.obs", "qualname": "CObs.gamma_method", "kind": "function", "doc": "Executes the gamma_method for the real and the imaginary part.
\n", "signature": "(self, **kwargs):", "funcdef": "def"}, "pyerrors.obs.CObs.is_zero": {"fullname": "pyerrors.obs.CObs.is_zero", "modulename": "pyerrors.obs", "qualname": "CObs.is_zero", "kind": "function", "doc": "Checks whether both real and imaginary part are zero within machine precision.
\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.obs.CObs.conjugate": {"fullname": "pyerrors.obs.CObs.conjugate", "modulename": "pyerrors.obs", "qualname": "CObs.conjugate", "kind": "function", "doc": "\n", "signature": "(self):", "funcdef": "def"}, "pyerrors.obs.derived_observable": {"fullname": "pyerrors.obs.derived_observable", "modulename": "pyerrors.obs", "qualname": "derived_observable", "kind": "function", "doc": "Construct a derived Obs according to func(data, **kwargs) using automatic differentiation.
\n\nParameters
\n\n\n- func (object):\narbitrary function of the form func(data, **kwargs). For the\nautomatic differentiation to work, all numpy functions have to have\nthe autograd wrapper (use 'import autograd.numpy as anp').
\n- data (list):\nlist of Obs, e.g. [obs1, obs2, obs3].
\n- num_grad (bool):\nif True, numerical derivatives are used instead of autograd\n(default False). To control the numerical differentiation the\nkwargs of numdifftools.step_generators.MaxStepGenerator\ncan be used.
\n- man_grad (list):\nmanually supply a list or an array which contains the jacobian\nof func. Use cautiously, supplying the wrong derivative will\nnot be intercepted.
\n
\n\nNotes
\n\nFor simple mathematical operations it can be practical to use anonymous\nfunctions. For the ratio of two observables one can e.g. use
\n\nnew_obs = derived_observable(lambda x: x[0] / x[1], [obs1, obs2])
\n", "signature": "(func, data, array_mode=False, **kwargs):", "funcdef": "def"}, "pyerrors.obs.reweight": {"fullname": "pyerrors.obs.reweight", "modulename": "pyerrors.obs", "qualname": "reweight", "kind": "function", "doc": "Reweight a list of observables.
\n\nParameters
\n\n\n- weight (Obs):\nReweighting factor. An Observable that has to be defined on a superset of the\nconfigurations in obs[i].idl for all i.
\n- obs (list):\nlist of Obs, e.g. [obs1, obs2, obs3].
\n- all_configs (bool):\nif True, the reweighted observables are normalized by the average of\nthe reweighting factor on all configurations in weight.idl and not\non the configurations in obs[i].idl. Default False.
\n
\n", "signature": "(weight, obs, **kwargs):", "funcdef": "def"}, "pyerrors.obs.correlate": {"fullname": "pyerrors.obs.correlate", "modulename": "pyerrors.obs", "qualname": "correlate", "kind": "function", "doc": "Correlate two observables.
\n\nParameters
\n\n\n- obs_a (Obs):\nFirst observable
\n- obs_b (Obs):\nSecond observable
\n
\n\nNotes
\n\nKeep in mind to only correlate primary observables which have not been reweighted\nyet. The reweighting has to be applied after correlating the observables.\nCurrently only works if ensembles are identical (this is not strictly necessary).
\n", "signature": "(obs_a, obs_b):", "funcdef": "def"}, "pyerrors.obs.covariance": {"fullname": "pyerrors.obs.covariance", "modulename": "pyerrors.obs", "qualname": "covariance", "kind": "function", "doc": "Calculates the error covariance matrix of a set of observables.
\n\nWARNING: This function should be used with care, especially for observables with support on multiple\n ensembles with differing autocorrelations. See the notes below for details.
\n\nThe gamma method has to be applied first to all observables.
\n\nParameters
\n\n\n- obs (list or numpy.ndarray):\nList or one dimensional array of Obs
\n- visualize (bool):\nIf True plots the corresponding normalized correlation matrix (default False).
\n- correlation (bool):\nIf True the correlation matrix instead of the error covariance matrix is returned (default False).
\n- smooth (None or int):\nIf smooth is an integer 'E' between 2 and the dimension of the matrix minus 1 the eigenvalue\nsmoothing procedure of hep-lat/9412087 is applied to the correlation matrix which leaves the\nlargest E eigenvalues essentially unchanged and smoothes the smaller eigenvalues to avoid extremely\nsmall ones.
\n
\n\nNotes
\n\nThe error covariance is defined such that it agrees with the squared standard error for two identical observables\n$$\\operatorname{cov}(a,a)=\\sum_{s=1}^N\\delta_a^s\\delta_a^s/N^2=\\Gamma_{aa}(0)/N=\\operatorname{var}(a)/N=\\sigma_a^2$$\nin the absence of autocorrelation.\nThe error covariance is estimated by calculating the correlation matrix assuming no autocorrelation and then rescaling the correlation matrix by the full errors including the previous gamma method estimate for the autocorrelation of the observables. The covariance at windowsize 0 is guaranteed to be positive semi-definite\n$$\\sum_{i,j}v_i\\Gamma_{ij}(0)v_j=\\frac{1}{N}\\sum_{s=1}^N\\sum_{i,j}v_i\\delta_i^s\\delta_j^s v_j=\\frac{1}{N}\\sum_{s=1}^N\\sum_{i}|v_i\\delta_i^s|^2\\geq 0\\,,$$ for every $v\\in\\mathbb{R}^M$, while such an identity does not hold for larger windows/lags.\nFor observables defined on a single ensemble our approximation is equivalent to assuming that the integrated autocorrelation time of an off-diagonal element is equal to the geometric mean of the integrated autocorrelation times of the corresponding diagonal elements.\n$$\\tau_{\\mathrm{int}, ij}=\\sqrt{\\tau_{\\mathrm{int}, i}\\times \\tau_{\\mathrm{int}, j}}$$\nThis construction ensures that the estimated covariance matrix is positive semi-definite (up to numerical rounding errors).
\n", "signature": "(obs, visualize=False, correlation=False, smooth=None, **kwargs):", "funcdef": "def"}, "pyerrors.obs.import_jackknife": {"fullname": "pyerrors.obs.import_jackknife", "modulename": "pyerrors.obs", "qualname": "import_jackknife", "kind": "function", "doc": "Imports jackknife samples and returns an Obs
\n\nParameters
\n\n\n- jacks (numpy.ndarray):\nnumpy array containing the mean value as zeroth entry and\nthe N jackknife samples as first to Nth entry.
\n- name (str):\nname of the ensemble the samples are defined on.
\n
\n", "signature": "(jacks, name, idl=None):", "funcdef": "def"}, "pyerrors.obs.merge_obs": {"fullname": "pyerrors.obs.merge_obs", "modulename": "pyerrors.obs", "qualname": "merge_obs", "kind": "function", "doc": "Combine all observables in list_of_obs into one new observable
\n\nParameters
\n\n\n- list_of_obs (list):\nlist of the Obs object to be combined
\n
\n\nNotes
\n\nIt is not possible to combine obs which are based on the same replicum
\n", "signature": "(list_of_obs):", "funcdef": "def"}, "pyerrors.obs.cov_Obs": {"fullname": "pyerrors.obs.cov_Obs", "modulename": "pyerrors.obs", "qualname": "cov_Obs", "kind": "function", "doc": "Create an Obs based on mean(s) and a covariance matrix
\n\nParameters
\n\n\n- mean (list of floats or float):\nN mean value(s) of the new Obs
\n- cov (list or array):\n2d (NxN) Covariance matrix, 1d diagonal entries or 0d covariance
\n- name (str):\nidentifier for the covariance matrix
\n- grad (list or array):\nGradient of the Covobs wrt. the means belonging to cov.
\n
\n", "signature": "(means, cov, name, grad=None):", "funcdef": "def"}, "pyerrors.roots": {"fullname": "pyerrors.roots", "modulename": "pyerrors.roots", "kind": "module", "doc": "\n"}, "pyerrors.roots.find_root": {"fullname": "pyerrors.roots.find_root", "modulename": "pyerrors.roots", "qualname": "find_root", "kind": "function", "doc": "Finds the root of the function func(x, d) where d is an Obs
.
\n\nParameters
\n\n\n\nReturns
\n\n\n- res (Obs):\n
Obs
valued root of the function. \n
\n", "signature": "(d, func, guess=1.0, **kwargs):", "funcdef": "def"}, "pyerrors.version": {"fullname": "pyerrors.version", "modulename": "pyerrors.version", "kind": "module", "doc": "\n"}}, "docInfo": {"pyerrors": {"qualname": 0, "fullname": 1, "annotation": 0, "default_value": 0, "signature": 0, "bases": 0, "doc": 8007}, "pyerrors.correlators": {"qualname": 0, "fullname": 2, "annotation": 0, "default_value": 0, "signature": 0, "bases": 0, "doc": 3}, "pyerrors.correlators.Corr": {"qualname": 1, "fullname": 3, "annotation": 0, "default_value": 0, "signature": 0, "bases": 0, "doc": 108}, "pyerrors.correlators.Corr.__init__": {"qualname": 3, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 40, "bases": 0, "doc": 94}, "pyerrors.correlators.Corr.gamma_method": {"qualname": 3, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 18, "bases": 0, "doc": 13}, "pyerrors.correlators.Corr.gm": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 18, "bases": 0, "doc": 13}, "pyerrors.correlators.Corr.projected": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 43, "bases": 0, "doc": 64}, "pyerrors.correlators.Corr.item": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 21, "bases": 0, "doc": 53}, "pyerrors.correlators.Corr.plottable": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 31}, "pyerrors.correlators.Corr.symmetric": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 9}, "pyerrors.correlators.Corr.anti_symmetric": {"qualname": 3, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 10}, "pyerrors.correlators.Corr.is_matrix_symmetric": {"qualname": 4, "fullname": 6, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 13}, "pyerrors.correlators.Corr.matrix_symmetric": {"qualname": 3, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 10}, "pyerrors.correlators.Corr.GEVP": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 47, "bases": 0, "doc": 326}, "pyerrors.correlators.Corr.Eigenvalue": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 50, "bases": 0, "doc": 59}, "pyerrors.correlators.Corr.Hankel": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 26, "bases": 0, "doc": 67}, "pyerrors.correlators.Corr.roll": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 16, "bases": 0, "doc": 26}, "pyerrors.correlators.Corr.reverse": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 9}, "pyerrors.correlators.Corr.thin": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 31, "bases": 0, "doc": 43}, "pyerrors.correlators.Corr.correlate": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 16, "bases": 0, "doc": 53}, "pyerrors.correlators.Corr.reweight": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 23, "bases": 0, "doc": 79}, "pyerrors.correlators.Corr.T_symmetry": {"qualname": 3, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 26, "bases": 0, "doc": 51}, "pyerrors.correlators.Corr.deriv": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 25, "bases": 0, "doc": 47}, "pyerrors.correlators.Corr.second_deriv": {"qualname": 3, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 25, "bases": 0, "doc": 45}, "pyerrors.correlators.Corr.m_eff": {"qualname": 3, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 36, "bases": 0, "doc": 148}, "pyerrors.correlators.Corr.fit": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 43, "bases": 0, "doc": 110}, "pyerrors.correlators.Corr.plateau": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 47, "bases": 0, "doc": 92}, "pyerrors.correlators.Corr.set_prange": {"qualname": 3, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 16, "bases": 0, "doc": 11}, "pyerrors.correlators.Corr.show": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 149, "bases": 0, "doc": 241}, "pyerrors.correlators.Corr.spaghetti_plot": {"qualname": 3, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 21, "bases": 0, "doc": 42}, "pyerrors.correlators.Corr.dump": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 38, "bases": 0, "doc": 69}, "pyerrors.correlators.Corr.print": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 22, "bases": 0, "doc": 3}, "pyerrors.correlators.Corr.sqrt": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 3}, "pyerrors.correlators.Corr.log": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 3}, "pyerrors.correlators.Corr.exp": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 3}, "pyerrors.correlators.Corr.sin": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 3}, "pyerrors.correlators.Corr.cos": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 3}, "pyerrors.correlators.Corr.tan": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 3}, "pyerrors.correlators.Corr.sinh": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 3}, "pyerrors.correlators.Corr.cosh": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 3}, "pyerrors.correlators.Corr.tanh": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 3}, "pyerrors.correlators.Corr.arcsin": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 3}, "pyerrors.correlators.Corr.arccos": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 3}, "pyerrors.correlators.Corr.arctan": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 3}, "pyerrors.correlators.Corr.arcsinh": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 3}, "pyerrors.correlators.Corr.arccosh": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 3}, "pyerrors.correlators.Corr.arctanh": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 3}, "pyerrors.correlators.Corr.prune": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 46, "bases": 0, "doc": 325}, "pyerrors.covobs": {"qualname": 0, "fullname": 2, "annotation": 0, "default_value": 0, "signature": 0, "bases": 0, "doc": 3}, "pyerrors.covobs.Covobs": {"qualname": 1, "fullname": 3, "annotation": 0, "default_value": 0, "signature": 0, "bases": 0, "doc": 3}, "pyerrors.covobs.Covobs.__init__": {"qualname": 3, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 39, "bases": 0, "doc": 100}, "pyerrors.covobs.Covobs.errsq": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 12}, "pyerrors.dirac": {"qualname": 0, "fullname": 2, "annotation": 0, "default_value": 0, "signature": 0, "bases": 0, "doc": 3}, "pyerrors.dirac.epsilon_tensor": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 21, "bases": 0, "doc": 40}, "pyerrors.dirac.epsilon_tensor_rank4": {"qualname": 3, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 26, "bases": 0, "doc": 41}, "pyerrors.dirac.Grid_gamma": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 12, "bases": 0, "doc": 9}, "pyerrors.fits": {"qualname": 0, "fullname": 2, "annotation": 0, "default_value": 0, "signature": 0, "bases": 0, "doc": 3}, "pyerrors.fits.Fit_result": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 0, "bases": 3, "doc": 75}, "pyerrors.fits.Fit_result.__init__": {"qualname": 4, "fullname": 6, "annotation": 0, "default_value": 0, "signature": 4, "bases": 0, "doc": 3}, "pyerrors.fits.Fit_result.gamma_method": {"qualname": 4, "fullname": 6, "annotation": 0, "default_value": 0, "signature": 18, "bases": 0, "doc": 10}, "pyerrors.fits.Fit_result.gm": {"qualname": 3, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 18, "bases": 0, "doc": 10}, "pyerrors.fits.least_squares": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 48, "bases": 0, "doc": 682}, "pyerrors.fits.total_least_squares": {"qualname": 3, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 38, "bases": 0, "doc": 468}, "pyerrors.fits.fit_lin": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 23, "bases": 0, "doc": 110}, "pyerrors.fits.qqplot": {"qualname": 1, "fullname": 3, "annotation": 0, "default_value": 0, "signature": 27, "bases": 0, "doc": 39}, "pyerrors.fits.residual_plot": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 27, "bases": 0, "doc": 29}, "pyerrors.fits.error_band": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 21, "bases": 0, "doc": 48}, "pyerrors.fits.ks_test": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 16, "bases": 0, "doc": 52}, "pyerrors.input": {"qualname": 0, "fullname": 2, "annotation": 0, "default_value": 0, "signature": 0, "bases": 0, "doc": 81}, "pyerrors.input.bdio": {"qualname": 0, "fullname": 3, "annotation": 0, "default_value": 0, "signature": 0, "bases": 0, "doc": 3}, "pyerrors.input.bdio.read_ADerrors": {"qualname": 2, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 35, "bases": 0, "doc": 122}, "pyerrors.input.bdio.write_ADerrors": {"qualname": 2, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 41, "bases": 0, "doc": 126}, "pyerrors.input.bdio.read_mesons": {"qualname": 2, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 35, "bases": 0, "doc": 211}, "pyerrors.input.bdio.read_dSdm": {"qualname": 2, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 35, "bases": 0, "doc": 191}, "pyerrors.input.dobs": {"qualname": 0, "fullname": 3, "annotation": 0, "default_value": 0, "signature": 0, "bases": 0, "doc": 3}, "pyerrors.input.dobs.create_pobs_string": {"qualname": 3, "fullname": 6, "annotation": 0, "default_value": 0, "signature": 62, "bases": 0, "doc": 186}, "pyerrors.input.dobs.write_pobs": {"qualname": 2, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 85, "bases": 0, "doc": 214}, "pyerrors.input.dobs.read_pobs": {"qualname": 2, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 43, "bases": 0, "doc": 164}, "pyerrors.input.dobs.import_dobs_string": {"qualname": 3, "fullname": 6, "annotation": 0, "default_value": 0, "signature": 43, "bases": 0, "doc": 218}, "pyerrors.input.dobs.read_dobs": {"qualname": 2, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 58, "bases": 0, "doc": 237}, "pyerrors.input.dobs.create_dobs_string": {"qualname": 3, "fullname": 6, "annotation": 0, "default_value": 0, "signature": 82, "bases": 0, "doc": 229}, "pyerrors.input.dobs.write_dobs": {"qualname": 2, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 99, "bases": 0, "doc": 252}, "pyerrors.input.hadrons": {"qualname": 0, "fullname": 3, "annotation": 0, "default_value": 0, "signature": 0, "bases": 0, "doc": 3}, "pyerrors.input.hadrons.read_meson_hd5": {"qualname": 3, "fullname": 6, "annotation": 0, "default_value": 0, "signature": 57, "bases": 0, "doc": 181}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"qualname": 3, "fullname": 6, "annotation": 0, "default_value": 0, "signature": 45, "bases": 0, "doc": 106}, "pyerrors.input.hadrons.Npr_matrix": {"qualname": 2, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 0, "bases": 2, "doc": 1069}, "pyerrors.input.hadrons.Npr_matrix.__init__": {"qualname": 4, "fullname": 7, "annotation": 0, "default_value": 0, "signature": 4, "bases": 0, "doc": 3}, "pyerrors.input.hadrons.Npr_matrix.g5H": {"qualname": 3, "fullname": 6, "annotation": 0, "default_value": 0, "signature": 0, "bases": 0, "doc": 30}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"qualname": 3, "fullname": 6, "annotation": 0, "default_value": 0, "signature": 32, "bases": 0, "doc": 99}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"qualname": 3, "fullname": 6, "annotation": 0, "default_value": 0, "signature": 32, "bases": 0, "doc": 99}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"qualname": 3, "fullname": 6, "annotation": 0, "default_value": 0, "signature": 60, "bases": 0, "doc": 112}, "pyerrors.input.json": {"qualname": 0, "fullname": 3, "annotation": 0, "default_value": 0, "signature": 0, "bases": 0, "doc": 3}, "pyerrors.input.json.create_json_string": {"qualname": 3, "fullname": 6, "annotation": 0, "default_value": 0, "signature": 34, "bases": 0, "doc": 138}, "pyerrors.input.json.dump_to_json": {"qualname": 3, "fullname": 6, "annotation": 0, "default_value": 0, "signature": 49, "bases": 0, "doc": 159}, "pyerrors.input.json.import_json_string": {"qualname": 3, "fullname": 6, "annotation": 0, "default_value": 0, "signature": 33, "bases": 0, "doc": 168}, "pyerrors.input.json.load_json": {"qualname": 2, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 42, "bases": 0, "doc": 188}, "pyerrors.input.json.dump_dict_to_json": {"qualname": 4, "fullname": 7, "annotation": 0, "default_value": 0, "signature": 63, "bases": 0, "doc": 184}, "pyerrors.input.json.load_json_dict": {"qualname": 3, "fullname": 6, "annotation": 0, "default_value": 0, "signature": 56, "bases": 0, "doc": 172}, "pyerrors.input.misc": {"qualname": 0, "fullname": 3, "annotation": 0, "default_value": 0, "signature": 0, "bases": 0, "doc": 3}, "pyerrors.input.misc.read_pbp": {"qualname": 2, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 23, "bases": 0, "doc": 75}, "pyerrors.input.openQCD": {"qualname": 0, "fullname": 3, "annotation": 0, "default_value": 0, "signature": 0, "bases": 0, "doc": 3}, "pyerrors.input.openQCD.read_rwms": {"qualname": 2, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 48, "bases": 0, "doc": 271}, "pyerrors.input.openQCD.extract_t0": {"qualname": 2, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 51, "bases": 0, "doc": 473}, "pyerrors.input.openQCD.read_qtop": {"qualname": 2, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 53, "bases": 0, "doc": 383}, "pyerrors.input.openQCD.read_gf_coupling": {"qualname": 3, "fullname": 6, "annotation": 0, "default_value": 0, "signature": 50, "bases": 0, "doc": 345}, "pyerrors.input.openQCD.qtop_projection": {"qualname": 2, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 21, "bases": 0, "doc": 72}, "pyerrors.input.openQCD.read_qtop_sector": {"qualname": 3, "fullname": 6, "annotation": 0, "default_value": 0, "signature": 38, "bases": 0, "doc": 363}, "pyerrors.input.openQCD.read_ms5_xsf": {"qualname": 3, "fullname": 6, "annotation": 0, "default_value": 0, "signature": 47, "bases": 0, "doc": 276}, "pyerrors.input.pandas": {"qualname": 0, "fullname": 3, "annotation": 0, "default_value": 0, "signature": 0, "bases": 0, "doc": 3}, "pyerrors.input.pandas.to_sql": {"qualname": 2, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 54, "bases": 0, "doc": 113}, "pyerrors.input.pandas.read_sql": {"qualname": 2, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 34, "bases": 0, "doc": 105}, "pyerrors.input.pandas.dump_df": {"qualname": 2, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 26, "bases": 0, "doc": 111}, "pyerrors.input.pandas.load_df": {"qualname": 2, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 32, "bases": 0, "doc": 115}, "pyerrors.input.sfcf": {"qualname": 0, "fullname": 3, "annotation": 0, "default_value": 0, "signature": 0, "bases": 0, "doc": 3}, "pyerrors.input.sfcf.read_sfcf": {"qualname": 2, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 128, "bases": 0, "doc": 422}, "pyerrors.input.utils": {"qualname": 0, "fullname": 3, "annotation": 0, "default_value": 0, "signature": 0, "bases": 0, "doc": 6}, "pyerrors.input.utils.check_idl": {"qualname": 2, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 16, "bases": 0, "doc": 70}, "pyerrors.linalg": {"qualname": 0, "fullname": 2, "annotation": 0, "default_value": 0, "signature": 0, "bases": 0, "doc": 3}, "pyerrors.linalg.matmul": {"qualname": 1, "fullname": 3, "annotation": 0, "default_value": 0, "signature": 13, "bases": 0, "doc": 54}, "pyerrors.linalg.jack_matmul": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 13, "bases": 0, "doc": 58}, "pyerrors.linalg.einsum": {"qualname": 1, "fullname": 3, "annotation": 0, "default_value": 0, "signature": 18, "bases": 0, "doc": 52}, "pyerrors.linalg.inv": {"qualname": 1, "fullname": 3, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 10}, "pyerrors.linalg.cholesky": {"qualname": 1, "fullname": 3, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 9}, "pyerrors.linalg.det": {"qualname": 1, "fullname": 3, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 8}, "pyerrors.linalg.eigh": {"qualname": 1, "fullname": 3, "annotation": 0, "default_value": 0, "signature": 18, "bases": 0, "doc": 20}, "pyerrors.linalg.eig": {"qualname": 1, "fullname": 3, "annotation": 0, "default_value": 0, "signature": 18, "bases": 0, "doc": 17}, "pyerrors.linalg.pinv": {"qualname": 1, "fullname": 3, "annotation": 0, "default_value": 0, "signature": 18, "bases": 0, "doc": 13}, "pyerrors.linalg.svd": {"qualname": 1, "fullname": 3, "annotation": 0, "default_value": 0, "signature": 18, "bases": 0, "doc": 13}, "pyerrors.misc": {"qualname": 0, "fullname": 2, "annotation": 0, "default_value": 0, "signature": 0, "bases": 0, "doc": 3}, "pyerrors.misc.dump_object": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 23, "bases": 0, "doc": 69}, "pyerrors.misc.load_object": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 42}, "pyerrors.misc.pseudo_Obs": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 31, "bases": 0, "doc": 105}, "pyerrors.misc.gen_correlated_data": {"qualname": 3, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 42, "bases": 0, "doc": 127}, "pyerrors.mpm": {"qualname": 0, "fullname": 2, "annotation": 0, "default_value": 0, "signature": 0, "bases": 0, "doc": 3}, "pyerrors.mpm.matrix_pencil_method": {"qualname": 3, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 38, "bases": 0, "doc": 165}, "pyerrors.obs": {"qualname": 0, "fullname": 2, "annotation": 0, "default_value": 0, "signature": 0, "bases": 0, "doc": 3}, "pyerrors.obs.Obs": {"qualname": 1, "fullname": 3, "annotation": 0, "default_value": 0, "signature": 0, "bases": 0, "doc": 238}, "pyerrors.obs.Obs.__init__": {"qualname": 3, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 31, "bases": 0, "doc": 62}, "pyerrors.obs.Obs.gamma_method": {"qualname": 3, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 18, "bases": 0, "doc": 133}, "pyerrors.obs.Obs.gm": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 18, "bases": 0, "doc": 133}, "pyerrors.obs.Obs.details": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 22, "bases": 0, "doc": 34}, "pyerrors.obs.Obs.reweight": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 16, "bases": 0, "doc": 85}, "pyerrors.obs.Obs.is_zero_within_error": {"qualname": 5, "fullname": 7, "annotation": 0, "default_value": 0, "signature": 21, "bases": 0, "doc": 50}, "pyerrors.obs.Obs.is_zero": {"qualname": 3, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 22, "bases": 0, "doc": 35}, "pyerrors.obs.Obs.plot_tauint": {"qualname": 3, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 21, "bases": 0, "doc": 34}, "pyerrors.obs.Obs.plot_rho": {"qualname": 3, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 21, "bases": 0, "doc": 35}, "pyerrors.obs.Obs.plot_rep_dist": {"qualname": 4, "fullname": 6, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 14}, "pyerrors.obs.Obs.plot_history": {"qualname": 3, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 21, "bases": 0, "doc": 35}, "pyerrors.obs.Obs.plot_piechart": {"qualname": 3, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 21, "bases": 0, "doc": 47}, "pyerrors.obs.Obs.dump": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 51, "bases": 0, "doc": 89}, "pyerrors.obs.Obs.export_jackknife": {"qualname": 3, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 101}, "pyerrors.obs.Obs.sqrt": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 3}, "pyerrors.obs.Obs.log": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 3}, "pyerrors.obs.Obs.exp": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 3}, "pyerrors.obs.Obs.sin": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 3}, "pyerrors.obs.Obs.cos": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 3}, "pyerrors.obs.Obs.tan": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 3}, "pyerrors.obs.Obs.arcsin": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 3}, "pyerrors.obs.Obs.arccos": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 3}, "pyerrors.obs.Obs.arctan": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 3}, "pyerrors.obs.Obs.sinh": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 3}, "pyerrors.obs.Obs.cosh": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 3}, "pyerrors.obs.Obs.tanh": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 3}, "pyerrors.obs.Obs.arcsinh": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 3}, "pyerrors.obs.Obs.arccosh": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 3}, "pyerrors.obs.Obs.arctanh": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 3}, "pyerrors.obs.CObs": {"qualname": 1, "fullname": 3, "annotation": 0, "default_value": 0, "signature": 0, "bases": 0, "doc": 9}, "pyerrors.obs.CObs.__init__": {"qualname": 3, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 20, "bases": 0, "doc": 3}, "pyerrors.obs.CObs.gamma_method": {"qualname": 3, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 18, "bases": 0, "doc": 14}, "pyerrors.obs.CObs.is_zero": {"qualname": 3, "fullname": 5, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 15}, "pyerrors.obs.CObs.conjugate": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 11, "bases": 0, "doc": 3}, "pyerrors.obs.derived_observable": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 34, "bases": 0, "doc": 184}, "pyerrors.obs.reweight": {"qualname": 1, "fullname": 3, "annotation": 0, "default_value": 0, "signature": 23, "bases": 0, "doc": 99}, "pyerrors.obs.correlate": {"qualname": 1, "fullname": 3, "annotation": 0, "default_value": 0, "signature": 18, "bases": 0, "doc": 75}, "pyerrors.obs.covariance": {"qualname": 1, "fullname": 3, "annotation": 0, "default_value": 0, "signature": 48, "bases": 0, "doc": 374}, "pyerrors.obs.import_jackknife": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 26, "bases": 0, "doc": 61}, "pyerrors.obs.merge_obs": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 13, "bases": 0, "doc": 56}, "pyerrors.obs.cov_Obs": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 31, "bases": 0, "doc": 90}, "pyerrors.roots": {"qualname": 0, "fullname": 2, "annotation": 0, "default_value": 0, "signature": 0, "bases": 0, "doc": 3}, "pyerrors.roots.find_root": {"qualname": 2, "fullname": 4, "annotation": 0, "default_value": 0, "signature": 34, "bases": 0, "doc": 181}, "pyerrors.version": {"qualname": 0, "fullname": 2, "annotation": 0, "default_value": 0, "signature": 0, "bases": 0, "doc": 3}}, "length": 181, "save": true}, "index": {"qualname": {"root": {"docs": {"pyerrors.correlators.Corr.__init__": {"tf": 1}, "pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.fits.Fit_result.__init__": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix.__init__": {"tf": 1}, "pyerrors.obs.Obs.__init__": {"tf": 1}, "pyerrors.obs.CObs.__init__": {"tf": 1}}, "df": 6, "c": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.correlators.Corr.__init__": {"tf": 1}, "pyerrors.correlators.Corr.gamma_method": {"tf": 1}, "pyerrors.correlators.Corr.gm": {"tf": 1}, "pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.correlators.Corr.item": {"tf": 1}, "pyerrors.correlators.Corr.plottable": {"tf": 1}, "pyerrors.correlators.Corr.symmetric": {"tf": 1}, "pyerrors.correlators.Corr.anti_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.is_matrix_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.matrix_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 1}, "pyerrors.correlators.Corr.Hankel": {"tf": 1}, "pyerrors.correlators.Corr.roll": {"tf": 1}, "pyerrors.correlators.Corr.reverse": {"tf": 1}, "pyerrors.correlators.Corr.thin": {"tf": 1}, "pyerrors.correlators.Corr.correlate": {"tf": 1}, "pyerrors.correlators.Corr.reweight": {"tf": 1}, "pyerrors.correlators.Corr.T_symmetry": {"tf": 1}, "pyerrors.correlators.Corr.deriv": {"tf": 1}, "pyerrors.correlators.Corr.second_deriv": {"tf": 1}, "pyerrors.correlators.Corr.m_eff": {"tf": 1}, "pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.correlators.Corr.set_prange": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.correlators.Corr.spaghetti_plot": {"tf": 1}, "pyerrors.correlators.Corr.dump": {"tf": 1}, "pyerrors.correlators.Corr.print": {"tf": 1}, "pyerrors.correlators.Corr.sqrt": {"tf": 1}, "pyerrors.correlators.Corr.log": {"tf": 1}, "pyerrors.correlators.Corr.exp": {"tf": 1}, "pyerrors.correlators.Corr.sin": {"tf": 1}, "pyerrors.correlators.Corr.cos": {"tf": 1}, "pyerrors.correlators.Corr.tan": {"tf": 1}, "pyerrors.correlators.Corr.sinh": {"tf": 1}, "pyerrors.correlators.Corr.cosh": {"tf": 1}, "pyerrors.correlators.Corr.tanh": {"tf": 1}, "pyerrors.correlators.Corr.arcsin": {"tf": 1}, "pyerrors.correlators.Corr.arccos": {"tf": 1}, "pyerrors.correlators.Corr.arctan": {"tf": 1}, "pyerrors.correlators.Corr.arcsinh": {"tf": 1}, "pyerrors.correlators.Corr.arccosh": {"tf": 1}, "pyerrors.correlators.Corr.arctanh": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 46, "e": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.correlate": {"tf": 1}, "pyerrors.obs.correlate": {"tf": 1}}, "df": 2, "d": {"docs": {"pyerrors.misc.gen_correlated_data": {"tf": 1}}, "df": 1}}}}}}}}, "s": {"docs": {"pyerrors.correlators.Corr.cos": {"tf": 1}, "pyerrors.obs.Obs.cos": {"tf": 1}}, "df": 2, "h": {"docs": {"pyerrors.correlators.Corr.cosh": {"tf": 1}, "pyerrors.obs.Obs.cosh": {"tf": 1}}, "df": 2}}, "v": {"docs": {"pyerrors.obs.cov_Obs": {"tf": 1}}, "df": 1, "o": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.covobs.Covobs": {"tf": 1}, "pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.covobs.Covobs.errsq": {"tf": 1}}, "df": 3}}}, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}}}}}}, "u": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}}, "df": 1}}}}}}, "b": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.obs.CObs": {"tf": 1}, "pyerrors.obs.CObs.__init__": {"tf": 1}, "pyerrors.obs.CObs.gamma_method": {"tf": 1}, "pyerrors.obs.CObs.is_zero": {"tf": 1}, "pyerrors.obs.CObs.conjugate": {"tf": 1}}, "df": 5}}, "n": {"docs": {}, "df": 0, "j": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.obs.CObs.conjugate": {"tf": 1}}, "df": 1}}}}}}}}, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1}}, "df": 3}}}}}, "h": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "k": {"docs": {"pyerrors.input.utils.check_idl": {"tf": 1}}, "df": 1}}}, "o": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "k": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.linalg.cholesky": {"tf": 1}}, "df": 1}}}}}}}}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.__init__": {"tf": 1}, "pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.fits.Fit_result.__init__": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix.__init__": {"tf": 1}, "pyerrors.obs.Obs.__init__": {"tf": 1}, "pyerrors.obs.CObs.__init__": {"tf": 1}}, "df": 6}}, "v": {"docs": {"pyerrors.linalg.inv": {"tf": 1}}, "df": 1}}, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "m": {"docs": {"pyerrors.correlators.Corr.item": {"tf": 1}}, "df": 1}}}, "s": {"docs": {"pyerrors.correlators.Corr.is_matrix_symmetric": {"tf": 1}, "pyerrors.obs.Obs.is_zero_within_error": {"tf": 1}, "pyerrors.obs.Obs.is_zero": {"tf": 1}, "pyerrors.obs.CObs.is_zero": {"tf": 1}}, "df": 4}, "m": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.obs.import_jackknife": {"tf": 1}}, "df": 3}}}}}, "d": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.input.utils.check_idl": {"tf": 1}}, "df": 1}}}, "g": {"5": {"docs": {}, "df": 0, "h": {"docs": {"pyerrors.input.hadrons.Npr_matrix.g5H": {"tf": 1}}, "df": 1}}, "docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "a": {"docs": {"pyerrors.correlators.Corr.gamma_method": {"tf": 1}, "pyerrors.dirac.Grid_gamma": {"tf": 1}, "pyerrors.fits.Fit_result.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.CObs.gamma_method": {"tf": 1}}, "df": 5}}}}, "m": {"docs": {"pyerrors.correlators.Corr.gm": {"tf": 1}, "pyerrors.fits.Fit_result.gm": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}}, "df": 3}, "e": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, "p": {"docs": {"pyerrors.correlators.Corr.GEVP": {"tf": 1}}, "df": 1}}, "n": {"docs": {"pyerrors.misc.gen_correlated_data": {"tf": 1}}, "df": 1}}, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.dirac.Grid_gamma": {"tf": 1}}, "df": 1}}}, "f": {"docs": {"pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}}, "df": 1}}, "m": {"docs": {"pyerrors.correlators.Corr.m_eff": {"tf": 1}}, "df": 1, "e": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.gamma_method": {"tf": 1}, "pyerrors.fits.Fit_result.gamma_method": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.CObs.gamma_method": {"tf": 1}}, "df": 5}}}}, "s": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}}, "df": 1, "s": {"docs": {"pyerrors.input.bdio.read_mesons": {"tf": 1}}, "df": 1}}}}, "r": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.obs.merge_obs": {"tf": 1}}, "df": 1}}}}, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "x": {"docs": {"pyerrors.correlators.Corr.is_matrix_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.matrix_symmetric": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix.__init__": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix.g5H": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 6}}}, "m": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.linalg.matmul": {"tf": 1}, "pyerrors.linalg.jack_matmul": {"tf": 1}}, "df": 2}}}}}, "s": {"5": {"docs": {"pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}}, "p": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "j": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.projected": {"tf": 1}}, "df": 1}}, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.openQCD.qtop_projection": {"tf": 1}}, "df": 1}}}}}}}}, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.set_prange": {"tf": 1}}, "df": 1}}}}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.print": {"tf": 1}}, "df": 1}}}, "u": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 1}}}}, "l": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.spaghetti_plot": {"tf": 1}, "pyerrors.fits.residual_plot": {"tf": 1}, "pyerrors.obs.Obs.plot_tauint": {"tf": 1}, "pyerrors.obs.Obs.plot_rho": {"tf": 1}, "pyerrors.obs.Obs.plot_rep_dist": {"tf": 1}, "pyerrors.obs.Obs.plot_history": {"tf": 1}, "pyerrors.obs.Obs.plot_piechart": {"tf": 1}}, "df": 7, "t": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.plottable": {"tf": 1}}, "df": 1}}}}}}}, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "u": {"docs": {"pyerrors.correlators.Corr.plateau": {"tf": 1}}, "df": 1}}}}}}, "o": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}}, "df": 3}}}, "b": {"docs": {}, "df": 0, "p": {"docs": {"pyerrors.input.misc.read_pbp": {"tf": 1}}, "df": 1}}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "v": {"docs": {"pyerrors.linalg.pinv": {"tf": 1}}, "df": 1}}, "e": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.obs.Obs.plot_piechart": {"tf": 1}}, "df": 1}}}}}}}, "s": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "o": {"docs": {"pyerrors.misc.pseudo_Obs": {"tf": 1}}, "df": 1}}}}}, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 1}}}}}}, "s": {"docs": {}, "df": 0, "y": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "c": {"docs": {"pyerrors.correlators.Corr.symmetric": {"tf": 1}, "pyerrors.correlators.Corr.anti_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.is_matrix_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.matrix_symmetric": {"tf": 1}}, "df": 4}}, "y": {"docs": {"pyerrors.correlators.Corr.T_symmetry": {"tf": 1}}, "df": 1}}}}}}}, "e": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.second_deriv": {"tf": 1}}, "df": 1}}}, "t": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 1}}}}, "t": {"docs": {"pyerrors.correlators.Corr.set_prange": {"tf": 1}}, "df": 1}}, "h": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "w": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1}}, "df": 1}}}, "p": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {"pyerrors.correlators.Corr.spaghetti_plot": {"tf": 1}}, "df": 1}}}}}}}}, "q": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.sqrt": {"tf": 1}, "pyerrors.obs.Obs.sqrt": {"tf": 1}}, "df": 2}}, "u": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}}, "df": 2}}}}}, "l": {"docs": {"pyerrors.input.pandas.to_sql": {"tf": 1}, "pyerrors.input.pandas.read_sql": {"tf": 1}}, "df": 2}}, "i": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.correlators.Corr.sin": {"tf": 1}, "pyerrors.obs.Obs.sin": {"tf": 1}}, "df": 2, "h": {"docs": {"pyerrors.correlators.Corr.sinh": {"tf": 1}, "pyerrors.obs.Obs.sinh": {"tf": 1}}, "df": 2}}}, "t": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.import_json_string": {"tf": 1}}, "df": 5}}}}}, "f": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "f": {"docs": {"pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 1}}}, "v": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.linalg.svd": {"tf": 1}}, "df": 1}}}, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {"pyerrors.correlators.Corr.anti_symmetric": {"tf": 1}}, "df": 1}}}, "r": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.correlators.Corr.arcsin": {"tf": 1}, "pyerrors.obs.Obs.arcsin": {"tf": 1}}, "df": 2, "h": {"docs": {"pyerrors.correlators.Corr.arcsinh": {"tf": 1}, "pyerrors.obs.Obs.arcsinh": {"tf": 1}}, "df": 2}}}}, "c": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.correlators.Corr.arccos": {"tf": 1}, "pyerrors.obs.Obs.arccos": {"tf": 1}}, "df": 2, "h": {"docs": {"pyerrors.correlators.Corr.arccosh": {"tf": 1}, "pyerrors.obs.Obs.arccosh": {"tf": 1}}, "df": 2}}}}, "t": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.correlators.Corr.arctan": {"tf": 1}, "pyerrors.obs.Obs.arctan": {"tf": 1}}, "df": 2, "h": {"docs": {"pyerrors.correlators.Corr.arctanh": {"tf": 1}, "pyerrors.obs.Obs.arctanh": {"tf": 1}}, "df": 2}}}}}}, "d": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}}, "df": 2}}}}}}}}, "e": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.linalg.eig": {"tf": 1}}, "df": 1, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.Eigenvalue": {"tf": 1}}, "df": 1}}}}}}}, "h": {"docs": {"pyerrors.linalg.eigh": {"tf": 1}}, "df": 1}}, "n": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "m": {"docs": {"pyerrors.linalg.einsum": {"tf": 1}}, "df": 1}}}}}, "f": {"docs": {}, "df": 0, "f": {"docs": {"pyerrors.correlators.Corr.m_eff": {"tf": 1}}, "df": 1}}, "x": {"docs": {}, "df": 0, "p": {"docs": {"pyerrors.correlators.Corr.exp": {"tf": 1}, "pyerrors.obs.Obs.exp": {"tf": 1}}, "df": 2, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.obs.Obs.export_jackknife": {"tf": 1}}, "df": 1}}}}, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}}, "df": 1}}}}}}}}, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 1}}}}}}, "r": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "q": {"docs": {"pyerrors.covobs.Covobs.errsq": {"tf": 1}}, "df": 1}}, "o": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.fits.error_band": {"tf": 1}, "pyerrors.obs.Obs.is_zero_within_error": {"tf": 1}}, "df": 2}}}}, "p": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.dirac.epsilon_tensor": {"tf": 1}, "pyerrors.dirac.epsilon_tensor_rank4": {"tf": 1}}, "df": 2}}}}}}}, "h": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "k": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.correlators.Corr.Hankel": {"tf": 1}}, "df": 1}}}}}, "d": {"5": {"docs": {"pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}}, "df": 5}, "docs": {}, "df": 0}, "i": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.obs.Obs.plot_history": {"tf": 1}}, "df": 1}}}}}}}, "r": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.correlators.Corr.roll": {"tf": 1}}, "df": 1}}, "o": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.roots.find_root": {"tf": 1}}, "df": 1}}}, "e": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.reverse": {"tf": 1}}, "df": 1}}}}}, "w": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.reweight": {"tf": 1}, "pyerrors.obs.Obs.reweight": {"tf": 1}, "pyerrors.obs.reweight": {"tf": 1}}, "df": 3}}}}}}, "s": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.fits.Fit_result": {"tf": 1}, "pyerrors.fits.Fit_result.__init__": {"tf": 1}, "pyerrors.fits.Fit_result.gamma_method": {"tf": 1}, "pyerrors.fits.Fit_result.gm": {"tf": 1}}, "df": 4}}}, "i": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.fits.residual_plot": {"tf": 1}}, "df": 1}}}}}}, "a": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}, "pyerrors.input.misc.read_pbp": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}, "pyerrors.input.pandas.read_sql": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 18}}, "p": {"docs": {"pyerrors.obs.Obs.plot_rep_dist": {"tf": 1}}, "df": 1}}, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "k": {"4": {"docs": {"pyerrors.dirac.epsilon_tensor_rank4": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}}}, "w": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.openQCD.read_rwms": {"tf": 1}}, "df": 1}}}, "h": {"docs": {}, "df": 0, "o": {"docs": {"pyerrors.obs.Obs.plot_rho": {"tf": 1}}, "df": 1}}}, "t": {"0": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 1}, "docs": {"pyerrors.correlators.Corr.T_symmetry": {"tf": 1}}, "df": 1, "h": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.correlators.Corr.thin": {"tf": 1}}, "df": 1}}}, "a": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.correlators.Corr.tan": {"tf": 1}, "pyerrors.obs.Obs.tan": {"tf": 1}}, "df": 2, "h": {"docs": {"pyerrors.correlators.Corr.tanh": {"tf": 1}, "pyerrors.obs.Obs.tanh": {"tf": 1}}, "df": 2}}, "u": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.obs.Obs.plot_tauint": {"tf": 1}}, "df": 1}}}}}, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.dirac.epsilon_tensor": {"tf": 1}, "pyerrors.dirac.epsilon_tensor_rank4": {"tf": 1}}, "df": 2}}}}, "s": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.fits.ks_test": {"tf": 1}}, "df": 1}}}, "o": {"docs": {"pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.pandas.to_sql": {"tf": 1}}, "df": 3, "t": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.fits.total_least_squares": {"tf": 1}}, "df": 1}}}}}, "d": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "v": {"docs": {"pyerrors.correlators.Corr.deriv": {"tf": 1}, "pyerrors.correlators.Corr.second_deriv": {"tf": 1}}, "df": 2, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.obs.derived_observable": {"tf": 1}}, "df": 1}}}}}, "t": {"docs": {"pyerrors.linalg.det": {"tf": 1}}, "df": 1, "a": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.obs.Obs.details": {"tf": 1}}, "df": 1}}}}}}, "u": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "p": {"docs": {"pyerrors.correlators.Corr.dump": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.pandas.dump_df": {"tf": 1}, "pyerrors.misc.dump_object": {"tf": 1}, "pyerrors.obs.Obs.dump": {"tf": 1}}, "df": 6}}}, "s": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "m": {"docs": {"pyerrors.input.bdio.read_dSdm": {"tf": 1}}, "df": 1}}}, "o": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}}, "df": 4}}}, "i": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.obs.Obs.plot_rep_dist": {"tf": 1}}, "df": 1, "i": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}}, "df": 1}}}}}}}}}}}}}}}}}}}}}, "c": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}}, "df": 2}}}, "f": {"docs": {"pyerrors.input.pandas.dump_df": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1}}, "df": 2}, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "a": {"docs": {"pyerrors.misc.gen_correlated_data": {"tf": 1}}, "df": 1}}}}, "f": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.fits.Fit_result": {"tf": 1}, "pyerrors.fits.Fit_result.__init__": {"tf": 1}, "pyerrors.fits.Fit_result.gamma_method": {"tf": 1}, "pyerrors.fits.Fit_result.gm": {"tf": 1}, "pyerrors.fits.fit_lin": {"tf": 1}}, "df": 6}, "n": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.roots.find_root": {"tf": 1}}, "df": 1}}}, "o": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "q": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "k": {"docs": {"pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}}, "df": 1}}}}}}}}}, "l": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.correlators.Corr.log": {"tf": 1}, "pyerrors.obs.Obs.log": {"tf": 1}}, "df": 2}, "a": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1}, "pyerrors.misc.load_object": {"tf": 1}}, "df": 4}}}, "e": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}}, "df": 2}}}}, "i": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.fits.fit_lin": {"tf": 1}}, "df": 1}}}, "q": {"docs": {}, "df": 0, "q": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.fits.qqplot": {"tf": 1}}, "df": 1}}}}}, "t": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "p": {"docs": {"pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.qtop_projection": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 3}}}}, "b": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.fits.error_band": {"tf": 1}}, "df": 1}}}, "i": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}}, "df": 1}}}}}}}}, "k": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.fits.ks_test": {"tf": 1}}, "df": 1}}, "w": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}}, "df": 3}}}}, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.obs.Obs.is_zero_within_error": {"tf": 1}}, "df": 1}}}}}}, "n": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix.__init__": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix.g5H": {"tf": 1}}, "df": 3}}}, "j": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}}, "df": 6}}}, "a": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "k": {"docs": {"pyerrors.linalg.jack_matmul": {"tf": 1}}, "df": 1, "k": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "f": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.obs.Obs.export_jackknife": {"tf": 1}, "pyerrors.obs.import_jackknife": {"tf": 1}}, "df": 2}}}}}}}}}, "x": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "f": {"docs": {"pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}}, "df": 1}}}, "o": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "j": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.misc.dump_object": {"tf": 1}, "pyerrors.misc.load_object": {"tf": 1}}, "df": 2}}}}, "s": {"docs": {"pyerrors.misc.pseudo_Obs": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1}, "pyerrors.obs.Obs.__init__": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}, "pyerrors.obs.Obs.details": {"tf": 1}, "pyerrors.obs.Obs.reweight": {"tf": 1}, "pyerrors.obs.Obs.is_zero_within_error": {"tf": 1}, "pyerrors.obs.Obs.is_zero": {"tf": 1}, "pyerrors.obs.Obs.plot_tauint": {"tf": 1}, "pyerrors.obs.Obs.plot_rho": {"tf": 1}, "pyerrors.obs.Obs.plot_rep_dist": {"tf": 1}, "pyerrors.obs.Obs.plot_history": {"tf": 1}, "pyerrors.obs.Obs.plot_piechart": {"tf": 1}, "pyerrors.obs.Obs.dump": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1}, "pyerrors.obs.Obs.sqrt": {"tf": 1}, "pyerrors.obs.Obs.log": {"tf": 1}, "pyerrors.obs.Obs.exp": {"tf": 1}, "pyerrors.obs.Obs.sin": {"tf": 1}, "pyerrors.obs.Obs.cos": {"tf": 1}, "pyerrors.obs.Obs.tan": {"tf": 1}, "pyerrors.obs.Obs.arcsin": {"tf": 1}, "pyerrors.obs.Obs.arccos": {"tf": 1}, "pyerrors.obs.Obs.arctan": {"tf": 1}, "pyerrors.obs.Obs.sinh": {"tf": 1}, "pyerrors.obs.Obs.cosh": {"tf": 1}, "pyerrors.obs.Obs.tanh": {"tf": 1}, "pyerrors.obs.Obs.arcsinh": {"tf": 1}, "pyerrors.obs.Obs.arccosh": {"tf": 1}, "pyerrors.obs.Obs.arctanh": {"tf": 1}, "pyerrors.obs.merge_obs": {"tf": 1}, "pyerrors.obs.cov_Obs": {"tf": 1}}, "df": 33, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.obs.derived_observable": {"tf": 1}}, "df": 1}}}}}}}}}}, "z": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "o": {"docs": {"pyerrors.obs.Obs.is_zero_within_error": {"tf": 1}, "pyerrors.obs.Obs.is_zero": {"tf": 1}, "pyerrors.obs.CObs.is_zero": {"tf": 1}}, "df": 3}}}}}}, "fullname": {"root": {"docs": {"pyerrors.correlators.Corr.__init__": {"tf": 1}, "pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.fits.Fit_result.__init__": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix.__init__": {"tf": 1}, "pyerrors.obs.Obs.__init__": {"tf": 1}, "pyerrors.obs.CObs.__init__": {"tf": 1}}, "df": 6, "p": {"docs": {}, "df": 0, "y": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators": {"tf": 1}, "pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.correlators.Corr.__init__": {"tf": 1}, "pyerrors.correlators.Corr.gamma_method": {"tf": 1}, "pyerrors.correlators.Corr.gm": {"tf": 1}, "pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.correlators.Corr.item": {"tf": 1}, "pyerrors.correlators.Corr.plottable": {"tf": 1}, "pyerrors.correlators.Corr.symmetric": {"tf": 1}, "pyerrors.correlators.Corr.anti_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.is_matrix_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.matrix_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 1}, "pyerrors.correlators.Corr.Hankel": {"tf": 1}, "pyerrors.correlators.Corr.roll": {"tf": 1}, "pyerrors.correlators.Corr.reverse": {"tf": 1}, "pyerrors.correlators.Corr.thin": {"tf": 1}, "pyerrors.correlators.Corr.correlate": {"tf": 1}, "pyerrors.correlators.Corr.reweight": {"tf": 1}, "pyerrors.correlators.Corr.T_symmetry": {"tf": 1}, "pyerrors.correlators.Corr.deriv": {"tf": 1}, "pyerrors.correlators.Corr.second_deriv": {"tf": 1}, "pyerrors.correlators.Corr.m_eff": {"tf": 1}, "pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.correlators.Corr.set_prange": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.correlators.Corr.spaghetti_plot": {"tf": 1}, "pyerrors.correlators.Corr.dump": {"tf": 1}, "pyerrors.correlators.Corr.print": {"tf": 1}, "pyerrors.correlators.Corr.sqrt": {"tf": 1}, "pyerrors.correlators.Corr.log": {"tf": 1}, "pyerrors.correlators.Corr.exp": {"tf": 1}, "pyerrors.correlators.Corr.sin": {"tf": 1}, "pyerrors.correlators.Corr.cos": {"tf": 1}, "pyerrors.correlators.Corr.tan": {"tf": 1}, "pyerrors.correlators.Corr.sinh": {"tf": 1}, "pyerrors.correlators.Corr.cosh": {"tf": 1}, "pyerrors.correlators.Corr.tanh": {"tf": 1}, "pyerrors.correlators.Corr.arcsin": {"tf": 1}, "pyerrors.correlators.Corr.arccos": {"tf": 1}, "pyerrors.correlators.Corr.arctan": {"tf": 1}, "pyerrors.correlators.Corr.arcsinh": {"tf": 1}, "pyerrors.correlators.Corr.arccosh": {"tf": 1}, "pyerrors.correlators.Corr.arctanh": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1}, "pyerrors.covobs": {"tf": 1}, "pyerrors.covobs.Covobs": {"tf": 1}, "pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.covobs.Covobs.errsq": {"tf": 1}, "pyerrors.dirac": {"tf": 1}, "pyerrors.dirac.epsilon_tensor": {"tf": 1}, "pyerrors.dirac.epsilon_tensor_rank4": {"tf": 1}, "pyerrors.dirac.Grid_gamma": {"tf": 1}, "pyerrors.fits": {"tf": 1}, "pyerrors.fits.Fit_result": {"tf": 1}, "pyerrors.fits.Fit_result.__init__": {"tf": 1}, "pyerrors.fits.Fit_result.gamma_method": {"tf": 1}, "pyerrors.fits.Fit_result.gm": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.fits.fit_lin": {"tf": 1}, "pyerrors.fits.qqplot": {"tf": 1}, "pyerrors.fits.residual_plot": {"tf": 1}, "pyerrors.fits.error_band": {"tf": 1}, "pyerrors.fits.ks_test": {"tf": 1}, "pyerrors.input": {"tf": 1}, "pyerrors.input.bdio": {"tf": 1}, "pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}, "pyerrors.input.dobs": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.hadrons": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix.__init__": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix.g5H": {"tf": 1}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}, "pyerrors.input.json": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}, "pyerrors.input.misc": {"tf": 1}, "pyerrors.input.misc.read_pbp": {"tf": 1}, "pyerrors.input.openQCD": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.qtop_projection": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}, "pyerrors.input.pandas": {"tf": 1}, "pyerrors.input.pandas.to_sql": {"tf": 1}, "pyerrors.input.pandas.read_sql": {"tf": 1}, "pyerrors.input.pandas.dump_df": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1}, "pyerrors.input.sfcf": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.input.utils": {"tf": 1}, "pyerrors.input.utils.check_idl": {"tf": 1}, "pyerrors.linalg": {"tf": 1}, "pyerrors.linalg.matmul": {"tf": 1}, "pyerrors.linalg.jack_matmul": {"tf": 1}, "pyerrors.linalg.einsum": {"tf": 1}, "pyerrors.linalg.inv": {"tf": 1}, "pyerrors.linalg.cholesky": {"tf": 1}, "pyerrors.linalg.det": {"tf": 1}, "pyerrors.linalg.eigh": {"tf": 1}, "pyerrors.linalg.eig": {"tf": 1}, "pyerrors.linalg.pinv": {"tf": 1}, "pyerrors.linalg.svd": {"tf": 1}, "pyerrors.misc": {"tf": 1}, "pyerrors.misc.dump_object": {"tf": 1}, "pyerrors.misc.load_object": {"tf": 1}, "pyerrors.misc.pseudo_Obs": {"tf": 1}, "pyerrors.misc.gen_correlated_data": {"tf": 1}, "pyerrors.mpm": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}, "pyerrors.obs": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1}, "pyerrors.obs.Obs.__init__": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}, "pyerrors.obs.Obs.details": {"tf": 1}, "pyerrors.obs.Obs.reweight": {"tf": 1}, "pyerrors.obs.Obs.is_zero_within_error": {"tf": 1}, "pyerrors.obs.Obs.is_zero": {"tf": 1}, "pyerrors.obs.Obs.plot_tauint": {"tf": 1}, "pyerrors.obs.Obs.plot_rho": {"tf": 1}, "pyerrors.obs.Obs.plot_rep_dist": {"tf": 1}, "pyerrors.obs.Obs.plot_history": {"tf": 1}, "pyerrors.obs.Obs.plot_piechart": {"tf": 1}, "pyerrors.obs.Obs.dump": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1}, "pyerrors.obs.Obs.sqrt": {"tf": 1}, "pyerrors.obs.Obs.log": {"tf": 1}, "pyerrors.obs.Obs.exp": {"tf": 1}, "pyerrors.obs.Obs.sin": {"tf": 1}, "pyerrors.obs.Obs.cos": {"tf": 1}, "pyerrors.obs.Obs.tan": {"tf": 1}, "pyerrors.obs.Obs.arcsin": {"tf": 1}, "pyerrors.obs.Obs.arccos": {"tf": 1}, "pyerrors.obs.Obs.arctan": {"tf": 1}, "pyerrors.obs.Obs.sinh": {"tf": 1}, "pyerrors.obs.Obs.cosh": {"tf": 1}, "pyerrors.obs.Obs.tanh": {"tf": 1}, "pyerrors.obs.Obs.arcsinh": {"tf": 1}, "pyerrors.obs.Obs.arccosh": {"tf": 1}, "pyerrors.obs.Obs.arctanh": {"tf": 1}, "pyerrors.obs.CObs": {"tf": 1}, "pyerrors.obs.CObs.__init__": {"tf": 1}, "pyerrors.obs.CObs.gamma_method": {"tf": 1}, "pyerrors.obs.CObs.is_zero": {"tf": 1}, "pyerrors.obs.CObs.conjugate": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}, "pyerrors.obs.reweight": {"tf": 1}, "pyerrors.obs.correlate": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}, "pyerrors.obs.import_jackknife": {"tf": 1}, "pyerrors.obs.merge_obs": {"tf": 1}, "pyerrors.obs.cov_Obs": {"tf": 1}, "pyerrors.roots": {"tf": 1}, "pyerrors.roots.find_root": {"tf": 1}, "pyerrors.version": {"tf": 1}}, "df": 181}}}}}}}, "r": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "j": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.projected": {"tf": 1}}, "df": 1}}, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.openQCD.qtop_projection": {"tf": 1}}, "df": 1}}}}}}}}, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.set_prange": {"tf": 1}}, "df": 1}}}}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.print": {"tf": 1}}, "df": 1}}}, "u": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 1}}}}, "l": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.spaghetti_plot": {"tf": 1}, "pyerrors.fits.residual_plot": {"tf": 1}, "pyerrors.obs.Obs.plot_tauint": {"tf": 1}, "pyerrors.obs.Obs.plot_rho": {"tf": 1}, "pyerrors.obs.Obs.plot_rep_dist": {"tf": 1}, "pyerrors.obs.Obs.plot_history": {"tf": 1}, "pyerrors.obs.Obs.plot_piechart": {"tf": 1}}, "df": 7, "t": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.plottable": {"tf": 1}}, "df": 1}}}}}}}, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "u": {"docs": {"pyerrors.correlators.Corr.plateau": {"tf": 1}}, "df": 1}}}}}}, "o": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}}, "df": 3}}}, "b": {"docs": {}, "df": 0, "p": {"docs": {"pyerrors.input.misc.read_pbp": {"tf": 1}}, "df": 1}}, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.pandas": {"tf": 1}, "pyerrors.input.pandas.to_sql": {"tf": 1}, "pyerrors.input.pandas.read_sql": {"tf": 1}, "pyerrors.input.pandas.dump_df": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1}}, "df": 5}}}}}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "v": {"docs": {"pyerrors.linalg.pinv": {"tf": 1}}, "df": 1}}, "e": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.obs.Obs.plot_piechart": {"tf": 1}}, "df": 1}}}}}}}, "s": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "o": {"docs": {"pyerrors.misc.pseudo_Obs": {"tf": 1}}, "df": 1}}}}}, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 1}}}}}}, "c": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.correlators.Corr.__init__": {"tf": 1}, "pyerrors.correlators.Corr.gamma_method": {"tf": 1}, "pyerrors.correlators.Corr.gm": {"tf": 1}, "pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.correlators.Corr.item": {"tf": 1}, "pyerrors.correlators.Corr.plottable": {"tf": 1}, "pyerrors.correlators.Corr.symmetric": {"tf": 1}, "pyerrors.correlators.Corr.anti_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.is_matrix_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.matrix_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 1}, "pyerrors.correlators.Corr.Hankel": {"tf": 1}, "pyerrors.correlators.Corr.roll": {"tf": 1}, "pyerrors.correlators.Corr.reverse": {"tf": 1}, "pyerrors.correlators.Corr.thin": {"tf": 1}, "pyerrors.correlators.Corr.correlate": {"tf": 1}, "pyerrors.correlators.Corr.reweight": {"tf": 1}, "pyerrors.correlators.Corr.T_symmetry": {"tf": 1}, "pyerrors.correlators.Corr.deriv": {"tf": 1}, "pyerrors.correlators.Corr.second_deriv": {"tf": 1}, "pyerrors.correlators.Corr.m_eff": {"tf": 1}, "pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.correlators.Corr.set_prange": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.correlators.Corr.spaghetti_plot": {"tf": 1}, "pyerrors.correlators.Corr.dump": {"tf": 1}, "pyerrors.correlators.Corr.print": {"tf": 1}, "pyerrors.correlators.Corr.sqrt": {"tf": 1}, "pyerrors.correlators.Corr.log": {"tf": 1}, "pyerrors.correlators.Corr.exp": {"tf": 1}, "pyerrors.correlators.Corr.sin": {"tf": 1}, "pyerrors.correlators.Corr.cos": {"tf": 1}, "pyerrors.correlators.Corr.tan": {"tf": 1}, "pyerrors.correlators.Corr.sinh": {"tf": 1}, "pyerrors.correlators.Corr.cosh": {"tf": 1}, "pyerrors.correlators.Corr.tanh": {"tf": 1}, "pyerrors.correlators.Corr.arcsin": {"tf": 1}, "pyerrors.correlators.Corr.arccos": {"tf": 1}, "pyerrors.correlators.Corr.arctan": {"tf": 1}, "pyerrors.correlators.Corr.arcsinh": {"tf": 1}, "pyerrors.correlators.Corr.arccosh": {"tf": 1}, "pyerrors.correlators.Corr.arctanh": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 46, "e": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.correlators": {"tf": 1}, "pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.correlators.Corr.__init__": {"tf": 1}, "pyerrors.correlators.Corr.gamma_method": {"tf": 1}, "pyerrors.correlators.Corr.gm": {"tf": 1}, "pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.correlators.Corr.item": {"tf": 1}, "pyerrors.correlators.Corr.plottable": {"tf": 1}, "pyerrors.correlators.Corr.symmetric": {"tf": 1}, "pyerrors.correlators.Corr.anti_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.is_matrix_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.matrix_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 1}, "pyerrors.correlators.Corr.Hankel": {"tf": 1}, "pyerrors.correlators.Corr.roll": {"tf": 1}, "pyerrors.correlators.Corr.reverse": {"tf": 1}, "pyerrors.correlators.Corr.thin": {"tf": 1}, "pyerrors.correlators.Corr.correlate": {"tf": 1}, "pyerrors.correlators.Corr.reweight": {"tf": 1}, "pyerrors.correlators.Corr.T_symmetry": {"tf": 1}, "pyerrors.correlators.Corr.deriv": {"tf": 1}, "pyerrors.correlators.Corr.second_deriv": {"tf": 1}, "pyerrors.correlators.Corr.m_eff": {"tf": 1}, "pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.correlators.Corr.set_prange": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.correlators.Corr.spaghetti_plot": {"tf": 1}, "pyerrors.correlators.Corr.dump": {"tf": 1}, "pyerrors.correlators.Corr.print": {"tf": 1}, "pyerrors.correlators.Corr.sqrt": {"tf": 1}, "pyerrors.correlators.Corr.log": {"tf": 1}, "pyerrors.correlators.Corr.exp": {"tf": 1}, "pyerrors.correlators.Corr.sin": {"tf": 1}, "pyerrors.correlators.Corr.cos": {"tf": 1}, "pyerrors.correlators.Corr.tan": {"tf": 1}, "pyerrors.correlators.Corr.sinh": {"tf": 1}, "pyerrors.correlators.Corr.cosh": {"tf": 1}, "pyerrors.correlators.Corr.tanh": {"tf": 1}, "pyerrors.correlators.Corr.arcsin": {"tf": 1}, "pyerrors.correlators.Corr.arccos": {"tf": 1}, "pyerrors.correlators.Corr.arctan": {"tf": 1}, "pyerrors.correlators.Corr.arcsinh": {"tf": 1}, "pyerrors.correlators.Corr.arccosh": {"tf": 1}, "pyerrors.correlators.Corr.arctanh": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 47}}}, "e": {"docs": {"pyerrors.correlators.Corr.correlate": {"tf": 1}, "pyerrors.obs.correlate": {"tf": 1}}, "df": 2, "d": {"docs": {"pyerrors.misc.gen_correlated_data": {"tf": 1}}, "df": 1}}}}}}}}, "s": {"docs": {"pyerrors.correlators.Corr.cos": {"tf": 1}, "pyerrors.obs.Obs.cos": {"tf": 1}}, "df": 2, "h": {"docs": {"pyerrors.correlators.Corr.cosh": {"tf": 1}, "pyerrors.obs.Obs.cosh": {"tf": 1}}, "df": 2}}, "v": {"docs": {"pyerrors.obs.cov_Obs": {"tf": 1}}, "df": 1, "o": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.covobs": {"tf": 1}, "pyerrors.covobs.Covobs": {"tf": 1.4142135623730951}, "pyerrors.covobs.Covobs.__init__": {"tf": 1.4142135623730951}, "pyerrors.covobs.Covobs.errsq": {"tf": 1.4142135623730951}}, "df": 4}}}, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}}}}}}, "u": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}}, "df": 1}}}}}}, "b": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.obs.CObs": {"tf": 1}, "pyerrors.obs.CObs.__init__": {"tf": 1}, "pyerrors.obs.CObs.gamma_method": {"tf": 1}, "pyerrors.obs.CObs.is_zero": {"tf": 1}, "pyerrors.obs.CObs.conjugate": {"tf": 1}}, "df": 5}}, "n": {"docs": {}, "df": 0, "j": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.obs.CObs.conjugate": {"tf": 1}}, "df": 1}}}}}}}}, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1}}, "df": 3}}}}}, "h": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "k": {"docs": {"pyerrors.input.utils.check_idl": {"tf": 1}}, "df": 1}}}, "o": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "k": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.linalg.cholesky": {"tf": 1}}, "df": 1}}}}}}}}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.__init__": {"tf": 1}, "pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.fits.Fit_result.__init__": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix.__init__": {"tf": 1}, "pyerrors.obs.Obs.__init__": {"tf": 1}, "pyerrors.obs.CObs.__init__": {"tf": 1}}, "df": 6}}, "p": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.input": {"tf": 1}, "pyerrors.input.bdio": {"tf": 1}, "pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}, "pyerrors.input.dobs": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.hadrons": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix.__init__": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix.g5H": {"tf": 1}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}, "pyerrors.input.json": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}, "pyerrors.input.misc": {"tf": 1}, "pyerrors.input.misc.read_pbp": {"tf": 1}, "pyerrors.input.openQCD": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.qtop_projection": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}, "pyerrors.input.pandas": {"tf": 1}, "pyerrors.input.pandas.to_sql": {"tf": 1}, "pyerrors.input.pandas.read_sql": {"tf": 1}, "pyerrors.input.pandas.dump_df": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1}, "pyerrors.input.sfcf": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.input.utils": {"tf": 1}, "pyerrors.input.utils.check_idl": {"tf": 1}}, "df": 49}}}, "v": {"docs": {"pyerrors.linalg.inv": {"tf": 1}}, "df": 1}}, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "m": {"docs": {"pyerrors.correlators.Corr.item": {"tf": 1}}, "df": 1}}}, "s": {"docs": {"pyerrors.correlators.Corr.is_matrix_symmetric": {"tf": 1}, "pyerrors.obs.Obs.is_zero_within_error": {"tf": 1}, "pyerrors.obs.Obs.is_zero": {"tf": 1}, "pyerrors.obs.CObs.is_zero": {"tf": 1}}, "df": 4}, "m": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.obs.import_jackknife": {"tf": 1}}, "df": 3}}}}}, "d": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.input.utils.check_idl": {"tf": 1}}, "df": 1}}}, "g": {"5": {"docs": {}, "df": 0, "h": {"docs": {"pyerrors.input.hadrons.Npr_matrix.g5H": {"tf": 1}}, "df": 1}}, "docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "a": {"docs": {"pyerrors.correlators.Corr.gamma_method": {"tf": 1}, "pyerrors.dirac.Grid_gamma": {"tf": 1}, "pyerrors.fits.Fit_result.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.CObs.gamma_method": {"tf": 1}}, "df": 5}}}}, "m": {"docs": {"pyerrors.correlators.Corr.gm": {"tf": 1}, "pyerrors.fits.Fit_result.gm": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}}, "df": 3}, "e": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, "p": {"docs": {"pyerrors.correlators.Corr.GEVP": {"tf": 1}}, "df": 1}}, "n": {"docs": {"pyerrors.misc.gen_correlated_data": {"tf": 1}}, "df": 1}}, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.dirac.Grid_gamma": {"tf": 1}}, "df": 1}}}, "f": {"docs": {"pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}}, "df": 1}}, "m": {"docs": {"pyerrors.correlators.Corr.m_eff": {"tf": 1}}, "df": 1, "e": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.gamma_method": {"tf": 1}, "pyerrors.fits.Fit_result.gamma_method": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.CObs.gamma_method": {"tf": 1}}, "df": 5}}}}, "s": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}}, "df": 1, "s": {"docs": {"pyerrors.input.bdio.read_mesons": {"tf": 1}}, "df": 1}}}}, "r": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.obs.merge_obs": {"tf": 1}}, "df": 1}}}}, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "x": {"docs": {"pyerrors.correlators.Corr.is_matrix_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.matrix_symmetric": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix.__init__": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix.g5H": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 6}}}, "m": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.linalg.matmul": {"tf": 1}, "pyerrors.linalg.jack_matmul": {"tf": 1}}, "df": 2}}}}}, "i": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "c": {"docs": {"pyerrors.input.misc": {"tf": 1}, "pyerrors.input.misc.read_pbp": {"tf": 1}, "pyerrors.misc": {"tf": 1}, "pyerrors.misc.dump_object": {"tf": 1}, "pyerrors.misc.load_object": {"tf": 1}, "pyerrors.misc.pseudo_Obs": {"tf": 1}, "pyerrors.misc.gen_correlated_data": {"tf": 1}}, "df": 7}}}, "s": {"5": {"docs": {"pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "p": {"docs": {}, "df": 0, "m": {"docs": {"pyerrors.mpm": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 2}}}, "s": {"docs": {}, "df": 0, "y": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "c": {"docs": {"pyerrors.correlators.Corr.symmetric": {"tf": 1}, "pyerrors.correlators.Corr.anti_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.is_matrix_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.matrix_symmetric": {"tf": 1}}, "df": 4}}, "y": {"docs": {"pyerrors.correlators.Corr.T_symmetry": {"tf": 1}}, "df": 1}}}}}}}, "e": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.second_deriv": {"tf": 1}}, "df": 1}}}, "t": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 1}}}}, "t": {"docs": {"pyerrors.correlators.Corr.set_prange": {"tf": 1}}, "df": 1}}, "h": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "w": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1}}, "df": 1}}}, "p": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {"pyerrors.correlators.Corr.spaghetti_plot": {"tf": 1}}, "df": 1}}}}}}}}, "q": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.sqrt": {"tf": 1}, "pyerrors.obs.Obs.sqrt": {"tf": 1}}, "df": 2}}, "u": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}}, "df": 2}}}}}, "l": {"docs": {"pyerrors.input.pandas.to_sql": {"tf": 1}, "pyerrors.input.pandas.read_sql": {"tf": 1}}, "df": 2}}, "i": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.correlators.Corr.sin": {"tf": 1}, "pyerrors.obs.Obs.sin": {"tf": 1}}, "df": 2, "h": {"docs": {"pyerrors.correlators.Corr.sinh": {"tf": 1}, "pyerrors.obs.Obs.sinh": {"tf": 1}}, "df": 2}}}, "t": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.import_json_string": {"tf": 1}}, "df": 5}}}}}, "f": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "f": {"docs": {"pyerrors.input.sfcf": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.4142135623730951}}, "df": 2}}}, "v": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.linalg.svd": {"tf": 1}}, "df": 1}}}, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {"pyerrors.correlators.Corr.anti_symmetric": {"tf": 1}}, "df": 1}}}, "r": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.correlators.Corr.arcsin": {"tf": 1}, "pyerrors.obs.Obs.arcsin": {"tf": 1}}, "df": 2, "h": {"docs": {"pyerrors.correlators.Corr.arcsinh": {"tf": 1}, "pyerrors.obs.Obs.arcsinh": {"tf": 1}}, "df": 2}}}}, "c": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.correlators.Corr.arccos": {"tf": 1}, "pyerrors.obs.Obs.arccos": {"tf": 1}}, "df": 2, "h": {"docs": {"pyerrors.correlators.Corr.arccosh": {"tf": 1}, "pyerrors.obs.Obs.arccosh": {"tf": 1}}, "df": 2}}}}, "t": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.correlators.Corr.arctan": {"tf": 1}, "pyerrors.obs.Obs.arctan": {"tf": 1}}, "df": 2, "h": {"docs": {"pyerrors.correlators.Corr.arctanh": {"tf": 1}, "pyerrors.obs.Obs.arctanh": {"tf": 1}}, "df": 2}}}}}}, "d": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}}, "df": 2}}}}}}}}, "e": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.linalg.eig": {"tf": 1}}, "df": 1, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.Eigenvalue": {"tf": 1}}, "df": 1}}}}}}}, "h": {"docs": {"pyerrors.linalg.eigh": {"tf": 1}}, "df": 1}}, "n": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "m": {"docs": {"pyerrors.linalg.einsum": {"tf": 1}}, "df": 1}}}}}, "f": {"docs": {}, "df": 0, "f": {"docs": {"pyerrors.correlators.Corr.m_eff": {"tf": 1}}, "df": 1}}, "x": {"docs": {}, "df": 0, "p": {"docs": {"pyerrors.correlators.Corr.exp": {"tf": 1}, "pyerrors.obs.Obs.exp": {"tf": 1}}, "df": 2, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.obs.Obs.export_jackknife": {"tf": 1}}, "df": 1}}}}, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}}, "df": 1}}}}}}}}, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 1}}}}}}, "r": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "q": {"docs": {"pyerrors.covobs.Covobs.errsq": {"tf": 1}}, "df": 1}}, "o": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.fits.error_band": {"tf": 1}, "pyerrors.obs.Obs.is_zero_within_error": {"tf": 1}}, "df": 2}}}}, "p": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.dirac.epsilon_tensor": {"tf": 1}, "pyerrors.dirac.epsilon_tensor_rank4": {"tf": 1}}, "df": 2}}}}}}}, "h": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "k": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.correlators.Corr.Hankel": {"tf": 1}}, "df": 1}}}}, "d": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.hadrons": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix.__init__": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix.g5H": {"tf": 1}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}}, "df": 9}}}}}}, "d": {"5": {"docs": {"pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}}, "df": 5}, "docs": {}, "df": 0}, "i": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.obs.Obs.plot_history": {"tf": 1}}, "df": 1}}}}}}}, "r": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.correlators.Corr.roll": {"tf": 1}}, "df": 1}}, "o": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.roots.find_root": {"tf": 1}}, "df": 1, "s": {"docs": {"pyerrors.roots": {"tf": 1}, "pyerrors.roots.find_root": {"tf": 1}}, "df": 2}}}}, "e": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.reverse": {"tf": 1}}, "df": 1}}}}}, "w": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.reweight": {"tf": 1}, "pyerrors.obs.Obs.reweight": {"tf": 1}, "pyerrors.obs.reweight": {"tf": 1}}, "df": 3}}}}}}, "s": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.fits.Fit_result": {"tf": 1}, "pyerrors.fits.Fit_result.__init__": {"tf": 1}, "pyerrors.fits.Fit_result.gamma_method": {"tf": 1}, "pyerrors.fits.Fit_result.gm": {"tf": 1}}, "df": 4}}}, "i": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.fits.residual_plot": {"tf": 1}}, "df": 1}}}}}}, "a": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}, "pyerrors.input.misc.read_pbp": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}, "pyerrors.input.pandas.read_sql": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 18}}, "p": {"docs": {"pyerrors.obs.Obs.plot_rep_dist": {"tf": 1}}, "df": 1}}, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "k": {"4": {"docs": {"pyerrors.dirac.epsilon_tensor_rank4": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}}}, "w": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.openQCD.read_rwms": {"tf": 1}}, "df": 1}}}, "h": {"docs": {}, "df": 0, "o": {"docs": {"pyerrors.obs.Obs.plot_rho": {"tf": 1}}, "df": 1}}}, "t": {"0": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 1}, "docs": {"pyerrors.correlators.Corr.T_symmetry": {"tf": 1}}, "df": 1, "h": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.correlators.Corr.thin": {"tf": 1}}, "df": 1}}}, "a": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.correlators.Corr.tan": {"tf": 1}, "pyerrors.obs.Obs.tan": {"tf": 1}}, "df": 2, "h": {"docs": {"pyerrors.correlators.Corr.tanh": {"tf": 1}, "pyerrors.obs.Obs.tanh": {"tf": 1}}, "df": 2}}, "u": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.obs.Obs.plot_tauint": {"tf": 1}}, "df": 1}}}}}, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.dirac.epsilon_tensor": {"tf": 1}, "pyerrors.dirac.epsilon_tensor_rank4": {"tf": 1}}, "df": 2}}}}, "s": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.fits.ks_test": {"tf": 1}}, "df": 1}}}, "o": {"docs": {"pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.pandas.to_sql": {"tf": 1}}, "df": 3, "t": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.fits.total_least_squares": {"tf": 1}}, "df": 1}}}}}, "d": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "v": {"docs": {"pyerrors.correlators.Corr.deriv": {"tf": 1}, "pyerrors.correlators.Corr.second_deriv": {"tf": 1}}, "df": 2, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.obs.derived_observable": {"tf": 1}}, "df": 1}}}}}, "t": {"docs": {"pyerrors.linalg.det": {"tf": 1}}, "df": 1, "a": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.obs.Obs.details": {"tf": 1}}, "df": 1}}}}}}, "u": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "p": {"docs": {"pyerrors.correlators.Corr.dump": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.pandas.dump_df": {"tf": 1}, "pyerrors.misc.dump_object": {"tf": 1}, "pyerrors.obs.Obs.dump": {"tf": 1}}, "df": 6}}}, "i": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "c": {"docs": {"pyerrors.dirac": {"tf": 1}, "pyerrors.dirac.epsilon_tensor": {"tf": 1}, "pyerrors.dirac.epsilon_tensor_rank4": {"tf": 1}, "pyerrors.dirac.Grid_gamma": {"tf": 1}}, "df": 4}}}, "s": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.obs.Obs.plot_rep_dist": {"tf": 1}}, "df": 1, "i": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}}, "df": 1}}}}}}}}}}}}}}}}}}}}}, "c": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}}, "df": 2}}}, "s": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "m": {"docs": {"pyerrors.input.bdio.read_dSdm": {"tf": 1}}, "df": 1}}}, "o": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.dobs": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.read_dobs": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.write_dobs": {"tf": 1.4142135623730951}}, "df": 8}}}, "f": {"docs": {"pyerrors.input.pandas.dump_df": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1}}, "df": 2}, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "a": {"docs": {"pyerrors.misc.gen_correlated_data": {"tf": 1}}, "df": 1}}}}, "f": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.fits.Fit_result": {"tf": 1}, "pyerrors.fits.Fit_result.__init__": {"tf": 1}, "pyerrors.fits.Fit_result.gamma_method": {"tf": 1}, "pyerrors.fits.Fit_result.gm": {"tf": 1}, "pyerrors.fits.fit_lin": {"tf": 1}}, "df": 6, "s": {"docs": {"pyerrors.fits": {"tf": 1}, "pyerrors.fits.Fit_result": {"tf": 1}, "pyerrors.fits.Fit_result.__init__": {"tf": 1}, "pyerrors.fits.Fit_result.gamma_method": {"tf": 1}, "pyerrors.fits.Fit_result.gm": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.fits.fit_lin": {"tf": 1}, "pyerrors.fits.qqplot": {"tf": 1}, "pyerrors.fits.residual_plot": {"tf": 1}, "pyerrors.fits.error_band": {"tf": 1}, "pyerrors.fits.ks_test": {"tf": 1}}, "df": 12}}, "n": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.roots.find_root": {"tf": 1}}, "df": 1}}}, "o": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "q": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "k": {"docs": {"pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}}, "df": 1}}}}}}}}}, "l": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.correlators.Corr.log": {"tf": 1}, "pyerrors.obs.Obs.log": {"tf": 1}}, "df": 2}, "a": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1}, "pyerrors.misc.load_object": {"tf": 1}}, "df": 4}}}, "e": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}}, "df": 2}}}}, "i": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.fits.fit_lin": {"tf": 1}}, "df": 1, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.linalg": {"tf": 1}, "pyerrors.linalg.matmul": {"tf": 1}, "pyerrors.linalg.jack_matmul": {"tf": 1}, "pyerrors.linalg.einsum": {"tf": 1}, "pyerrors.linalg.inv": {"tf": 1}, "pyerrors.linalg.cholesky": {"tf": 1}, "pyerrors.linalg.det": {"tf": 1}, "pyerrors.linalg.eigh": {"tf": 1}, "pyerrors.linalg.eig": {"tf": 1}, "pyerrors.linalg.pinv": {"tf": 1}, "pyerrors.linalg.svd": {"tf": 1}}, "df": 11}}}}}}, "q": {"docs": {}, "df": 0, "q": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.fits.qqplot": {"tf": 1}}, "df": 1}}}}}, "t": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "p": {"docs": {"pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.qtop_projection": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 3}}}}, "b": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.fits.error_band": {"tf": 1}}, "df": 1}}}, "d": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {"pyerrors.input.bdio": {"tf": 1}, "pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}}, "df": 5}}}, "i": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}}, "df": 1}}}}}}}}, "k": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.fits.ks_test": {"tf": 1}}, "df": 1}}, "w": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}}, "df": 3}}}}, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.obs.Obs.is_zero_within_error": {"tf": 1}}, "df": 1}}}}}}, "n": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix.__init__": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix.g5H": {"tf": 1}}, "df": 3}}}, "j": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.json": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1.4142135623730951}, "pyerrors.input.json.dump_to_json": {"tf": 1.4142135623730951}, "pyerrors.input.json.import_json_string": {"tf": 1.4142135623730951}, "pyerrors.input.json.load_json": {"tf": 1.4142135623730951}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1.4142135623730951}, "pyerrors.input.json.load_json_dict": {"tf": 1.4142135623730951}}, "df": 7}}}, "a": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "k": {"docs": {"pyerrors.linalg.jack_matmul": {"tf": 1}}, "df": 1, "k": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "f": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.obs.Obs.export_jackknife": {"tf": 1}, "pyerrors.obs.import_jackknife": {"tf": 1}}, "df": 2}}}}}}}}}, "o": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "q": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.openQCD": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.qtop_projection": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}}, "df": 8}}}}}}, "b": {"docs": {}, "df": 0, "j": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.misc.dump_object": {"tf": 1}, "pyerrors.misc.load_object": {"tf": 1}}, "df": 2}}}}, "s": {"docs": {"pyerrors.misc.pseudo_Obs": {"tf": 1}, "pyerrors.obs": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.__init__": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.gamma_method": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.gm": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.details": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.reweight": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.is_zero_within_error": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.is_zero": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.plot_tauint": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.plot_rho": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.plot_rep_dist": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.plot_history": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.plot_piechart": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.dump": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.sqrt": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.log": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.exp": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.sin": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.cos": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.tan": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.arcsin": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.arccos": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.arctan": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.sinh": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.cosh": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.tanh": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.arcsinh": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.arccosh": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.arctanh": {"tf": 1.4142135623730951}, "pyerrors.obs.CObs": {"tf": 1}, "pyerrors.obs.CObs.__init__": {"tf": 1}, "pyerrors.obs.CObs.gamma_method": {"tf": 1}, "pyerrors.obs.CObs.is_zero": {"tf": 1}, "pyerrors.obs.CObs.conjugate": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}, "pyerrors.obs.reweight": {"tf": 1}, "pyerrors.obs.correlate": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}, "pyerrors.obs.import_jackknife": {"tf": 1}, "pyerrors.obs.merge_obs": {"tf": 1.4142135623730951}, "pyerrors.obs.cov_Obs": {"tf": 1.4142135623730951}}, "df": 44, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.obs.derived_observable": {"tf": 1}}, "df": 1}}}}}}}}}}, "x": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "f": {"docs": {"pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}}, "df": 1}}}, "u": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.utils": {"tf": 1}, "pyerrors.input.utils.check_idl": {"tf": 1}}, "df": 2}}}}}, "z": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "o": {"docs": {"pyerrors.obs.Obs.is_zero_within_error": {"tf": 1}, "pyerrors.obs.Obs.is_zero": {"tf": 1}, "pyerrors.obs.CObs.is_zero": {"tf": 1}}, "df": 3}}}}, "v": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.version": {"tf": 1}}, "df": 1}}}}}}}}}, "annotation": {"root": {"docs": {}, "df": 0}}, "default_value": {"root": {"docs": {}, "df": 0}}, "signature": {"root": {"0": {"docs": {"pyerrors.correlators.Corr.__init__": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 1}, "pyerrors.correlators.Corr.thin": {"tf": 1}, "pyerrors.correlators.Corr.m_eff": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.qtop_projection": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.7320508075688772}, "pyerrors.misc.gen_correlated_data": {"tf": 1}, "pyerrors.obs.CObs.__init__": {"tf": 1.4142135623730951}, "pyerrors.roots.find_root": {"tf": 1}}, "df": 14, "c": {"docs": {"pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 1}}, "1": {"0": {"0": {"0": {"docs": {"pyerrors.misc.pseudo_Obs": {"tf": 1}, "pyerrors.misc.gen_correlated_data": {"tf": 1}}, "df": 2}, "docs": {}, "df": 0}, "docs": {"pyerrors.obs.Obs.is_zero": {"tf": 1}}, "df": 1}, "docs": {"pyerrors.correlators.Corr.T_symmetry": {"tf": 1}, "pyerrors.correlators.Corr.m_eff": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}, "pyerrors.obs.Obs.is_zero_within_error": {"tf": 1}, "pyerrors.roots.find_root": {"tf": 1}}, "df": 11, "e": {"docs": {"pyerrors.obs.Obs.is_zero": {"tf": 1}}, "df": 1}}, "2": {"docs": {"pyerrors.correlators.Corr.thin": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}}, "df": 3}, "3": {"9": {"docs": {"pyerrors.correlators.Corr.GEVP": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.deriv": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.second_deriv": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.m_eff": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.plateau": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.dump": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.read_ADerrors": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.read_mesons": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.read_dSdm": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.create_pobs_string": {"tf": 2}, "pyerrors.input.dobs.write_pobs": {"tf": 2}, "pyerrors.input.dobs.create_dobs_string": {"tf": 2}, "pyerrors.input.dobs.write_dobs": {"tf": 2}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 2}, "pyerrors.input.json.create_json_string": {"tf": 1.4142135623730951}, "pyerrors.input.json.dump_to_json": {"tf": 1.4142135623730951}, "pyerrors.input.json.dump_dict_to_json": {"tf": 2}, "pyerrors.input.json.load_json_dict": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_rwms": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1.4142135623730951}, "pyerrors.input.pandas.to_sql": {"tf": 1.4142135623730951}, "pyerrors.input.sfcf.read_sfcf": {"tf": 2.8284271247461903}, "pyerrors.obs.Obs.dump": {"tf": 2}}, "df": 28}, "docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 1}, "5": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.misc.gen_correlated_data": {"tf": 1}}, "df": 2}, "docs": {"pyerrors.correlators.Corr.__init__": {"tf": 5.744562646538029}, "pyerrors.correlators.Corr.gamma_method": {"tf": 4}, "pyerrors.correlators.Corr.gm": {"tf": 4}, "pyerrors.correlators.Corr.projected": {"tf": 5.830951894845301}, "pyerrors.correlators.Corr.item": {"tf": 4.242640687119285}, "pyerrors.correlators.Corr.plottable": {"tf": 3.1622776601683795}, "pyerrors.correlators.Corr.symmetric": {"tf": 3.1622776601683795}, "pyerrors.correlators.Corr.anti_symmetric": {"tf": 3.1622776601683795}, "pyerrors.correlators.Corr.is_matrix_symmetric": {"tf": 3.1622776601683795}, "pyerrors.correlators.Corr.matrix_symmetric": {"tf": 3.1622776601683795}, "pyerrors.correlators.Corr.GEVP": {"tf": 6.164414002968976}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 6.324555320336759}, "pyerrors.correlators.Corr.Hankel": {"tf": 4.69041575982343}, "pyerrors.correlators.Corr.roll": {"tf": 3.7416573867739413}, "pyerrors.correlators.Corr.reverse": {"tf": 3.1622776601683795}, "pyerrors.correlators.Corr.thin": {"tf": 5.0990195135927845}, "pyerrors.correlators.Corr.correlate": {"tf": 3.7416573867739413}, "pyerrors.correlators.Corr.reweight": {"tf": 4.47213595499958}, "pyerrors.correlators.Corr.T_symmetry": {"tf": 4.69041575982343}, "pyerrors.correlators.Corr.deriv": {"tf": 4.47213595499958}, "pyerrors.correlators.Corr.second_deriv": {"tf": 4.47213595499958}, "pyerrors.correlators.Corr.m_eff": {"tf": 5.291502622129181}, "pyerrors.correlators.Corr.fit": {"tf": 6}, "pyerrors.correlators.Corr.plateau": {"tf": 6}, "pyerrors.correlators.Corr.set_prange": {"tf": 3.7416573867739413}, "pyerrors.correlators.Corr.show": {"tf": 10.908712114635714}, "pyerrors.correlators.Corr.spaghetti_plot": {"tf": 4.242640687119285}, "pyerrors.correlators.Corr.dump": {"tf": 5.477225575051661}, "pyerrors.correlators.Corr.print": {"tf": 4.242640687119285}, "pyerrors.correlators.Corr.sqrt": {"tf": 3.1622776601683795}, "pyerrors.correlators.Corr.log": {"tf": 3.1622776601683795}, "pyerrors.correlators.Corr.exp": {"tf": 3.1622776601683795}, "pyerrors.correlators.Corr.sin": {"tf": 3.1622776601683795}, "pyerrors.correlators.Corr.cos": {"tf": 3.1622776601683795}, "pyerrors.correlators.Corr.tan": {"tf": 3.1622776601683795}, "pyerrors.correlators.Corr.sinh": {"tf": 3.1622776601683795}, "pyerrors.correlators.Corr.cosh": {"tf": 3.1622776601683795}, "pyerrors.correlators.Corr.tanh": {"tf": 3.1622776601683795}, "pyerrors.correlators.Corr.arcsin": {"tf": 3.1622776601683795}, "pyerrors.correlators.Corr.arccos": {"tf": 3.1622776601683795}, "pyerrors.correlators.Corr.arctan": {"tf": 3.1622776601683795}, "pyerrors.correlators.Corr.arcsinh": {"tf": 3.1622776601683795}, "pyerrors.correlators.Corr.arccosh": {"tf": 3.1622776601683795}, "pyerrors.correlators.Corr.arctanh": {"tf": 3.1622776601683795}, "pyerrors.correlators.Corr.prune": {"tf": 6.164414002968976}, "pyerrors.covobs.Covobs.__init__": {"tf": 5.656854249492381}, "pyerrors.covobs.Covobs.errsq": {"tf": 3.1622776601683795}, "pyerrors.dirac.epsilon_tensor": {"tf": 4.242640687119285}, "pyerrors.dirac.epsilon_tensor_rank4": {"tf": 4.69041575982343}, "pyerrors.dirac.Grid_gamma": {"tf": 3.1622776601683795}, "pyerrors.fits.Fit_result.__init__": {"tf": 2}, "pyerrors.fits.Fit_result.gamma_method": {"tf": 4}, "pyerrors.fits.Fit_result.gm": {"tf": 4}, "pyerrors.fits.least_squares": {"tf": 6.324555320336759}, "pyerrors.fits.total_least_squares": {"tf": 5.656854249492381}, "pyerrors.fits.fit_lin": {"tf": 4.47213595499958}, "pyerrors.fits.qqplot": {"tf": 4.69041575982343}, "pyerrors.fits.residual_plot": {"tf": 4.69041575982343}, "pyerrors.fits.error_band": {"tf": 4.242640687119285}, "pyerrors.fits.ks_test": {"tf": 3.7416573867739413}, "pyerrors.input.bdio.read_ADerrors": {"tf": 5.0990195135927845}, "pyerrors.input.bdio.write_ADerrors": {"tf": 5.477225575051661}, "pyerrors.input.bdio.read_mesons": {"tf": 5.0990195135927845}, "pyerrors.input.bdio.read_dSdm": {"tf": 5.0990195135927845}, "pyerrors.input.dobs.create_pobs_string": {"tf": 7.14142842854285}, "pyerrors.input.dobs.write_pobs": {"tf": 8.426149773176359}, "pyerrors.input.dobs.read_pobs": {"tf": 5.830951894845301}, "pyerrors.input.dobs.import_dobs_string": {"tf": 5.830951894845301}, "pyerrors.input.dobs.read_dobs": {"tf": 6.855654600401044}, "pyerrors.input.dobs.create_dobs_string": {"tf": 8.12403840463596}, "pyerrors.input.dobs.write_dobs": {"tf": 8.94427190999916}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 6.6332495807108}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 6}, "pyerrors.input.hadrons.Npr_matrix.__init__": {"tf": 2}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 5.0990195135927845}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 5.0990195135927845}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 6.855654600401044}, "pyerrors.input.json.create_json_string": {"tf": 5.291502622129181}, "pyerrors.input.json.dump_to_json": {"tf": 6.324555320336759}, "pyerrors.input.json.import_json_string": {"tf": 5.0990195135927845}, "pyerrors.input.json.load_json": {"tf": 5.830951894845301}, "pyerrors.input.json.dump_dict_to_json": {"tf": 7.0710678118654755}, "pyerrors.input.json.load_json_dict": {"tf": 6.6332495807108}, "pyerrors.input.misc.read_pbp": {"tf": 4.47213595499958}, "pyerrors.input.openQCD.read_rwms": {"tf": 6.164414002968976}, "pyerrors.input.openQCD.extract_t0": {"tf": 6.324555320336759}, "pyerrors.input.openQCD.read_qtop": {"tf": 6.48074069840786}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 6.324555320336759}, "pyerrors.input.openQCD.qtop_projection": {"tf": 4.242640687119285}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 5.656854249492381}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 6.164414002968976}, "pyerrors.input.pandas.to_sql": {"tf": 6.48074069840786}, "pyerrors.input.pandas.read_sql": {"tf": 5.291502622129181}, "pyerrors.input.pandas.dump_df": {"tf": 4.69041575982343}, "pyerrors.input.pandas.load_df": {"tf": 5.0990195135927845}, "pyerrors.input.sfcf.read_sfcf": {"tf": 10}, "pyerrors.input.utils.check_idl": {"tf": 3.7416573867739413}, "pyerrors.linalg.matmul": {"tf": 3.4641016151377544}, "pyerrors.linalg.jack_matmul": {"tf": 3.4641016151377544}, "pyerrors.linalg.einsum": {"tf": 4}, "pyerrors.linalg.inv": {"tf": 3.1622776601683795}, "pyerrors.linalg.cholesky": {"tf": 3.1622776601683795}, "pyerrors.linalg.det": {"tf": 3.1622776601683795}, "pyerrors.linalg.eigh": {"tf": 4}, "pyerrors.linalg.eig": {"tf": 4}, "pyerrors.linalg.pinv": {"tf": 4}, "pyerrors.linalg.svd": {"tf": 4}, "pyerrors.misc.dump_object": {"tf": 4.47213595499958}, "pyerrors.misc.load_object": {"tf": 3.1622776601683795}, "pyerrors.misc.pseudo_Obs": {"tf": 5.0990195135927845}, "pyerrors.misc.gen_correlated_data": {"tf": 5.830951894845301}, "pyerrors.mpm.matrix_pencil_method": {"tf": 5.656854249492381}, "pyerrors.obs.Obs.__init__": {"tf": 5.0990195135927845}, "pyerrors.obs.Obs.gamma_method": {"tf": 4}, "pyerrors.obs.Obs.gm": {"tf": 4}, "pyerrors.obs.Obs.details": {"tf": 4.242640687119285}, "pyerrors.obs.Obs.reweight": {"tf": 3.7416573867739413}, "pyerrors.obs.Obs.is_zero_within_error": {"tf": 4.242640687119285}, "pyerrors.obs.Obs.is_zero": {"tf": 4.242640687119285}, "pyerrors.obs.Obs.plot_tauint": {"tf": 4.242640687119285}, "pyerrors.obs.Obs.plot_rho": {"tf": 4.242640687119285}, "pyerrors.obs.Obs.plot_rep_dist": {"tf": 3.1622776601683795}, "pyerrors.obs.Obs.plot_history": {"tf": 4.242640687119285}, "pyerrors.obs.Obs.plot_piechart": {"tf": 4.242640687119285}, "pyerrors.obs.Obs.dump": {"tf": 6.324555320336759}, "pyerrors.obs.Obs.export_jackknife": {"tf": 3.1622776601683795}, "pyerrors.obs.Obs.sqrt": {"tf": 3.1622776601683795}, "pyerrors.obs.Obs.log": {"tf": 3.1622776601683795}, "pyerrors.obs.Obs.exp": {"tf": 3.1622776601683795}, "pyerrors.obs.Obs.sin": {"tf": 3.1622776601683795}, "pyerrors.obs.Obs.cos": {"tf": 3.1622776601683795}, "pyerrors.obs.Obs.tan": {"tf": 3.1622776601683795}, "pyerrors.obs.Obs.arcsin": {"tf": 3.1622776601683795}, "pyerrors.obs.Obs.arccos": {"tf": 3.1622776601683795}, "pyerrors.obs.Obs.arctan": {"tf": 3.1622776601683795}, "pyerrors.obs.Obs.sinh": {"tf": 3.1622776601683795}, "pyerrors.obs.Obs.cosh": {"tf": 3.1622776601683795}, "pyerrors.obs.Obs.tanh": {"tf": 3.1622776601683795}, "pyerrors.obs.Obs.arcsinh": {"tf": 3.1622776601683795}, "pyerrors.obs.Obs.arccosh": {"tf": 3.1622776601683795}, "pyerrors.obs.Obs.arctanh": {"tf": 3.1622776601683795}, "pyerrors.obs.CObs.__init__": {"tf": 4}, "pyerrors.obs.CObs.gamma_method": {"tf": 4}, "pyerrors.obs.CObs.is_zero": {"tf": 3.1622776601683795}, "pyerrors.obs.CObs.conjugate": {"tf": 3.1622776601683795}, "pyerrors.obs.derived_observable": {"tf": 5.291502622129181}, "pyerrors.obs.reweight": {"tf": 4.47213595499958}, "pyerrors.obs.correlate": {"tf": 3.7416573867739413}, "pyerrors.obs.covariance": {"tf": 6.324555320336759}, "pyerrors.obs.import_jackknife": {"tf": 4.69041575982343}, "pyerrors.obs.merge_obs": {"tf": 3.1622776601683795}, "pyerrors.obs.cov_Obs": {"tf": 5.0990195135927845}, "pyerrors.roots.find_root": {"tf": 5.291502622129181}}, "df": 153, "d": {"docs": {"pyerrors.roots.find_root": {"tf": 1}}, "df": 1, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "a": {"docs": {"pyerrors.correlators.Corr.__init__": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}}, "df": 2, "t": {"docs": {}, "df": 0, "y": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.dump": {"tf": 1}, "pyerrors.obs.Obs.dump": {"tf": 1}}, "df": 2}}}}}}}, "t": {"docs": {"pyerrors.correlators.Corr.roll": {"tf": 1}}, "df": 1, "r": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}}, "df": 3}}, "o": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}}, "df": 2}}}, "i": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}}, "df": 1}}}}}}, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}}, "df": 1}}}}, "c": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}}, "df": 2}}}}}}, "e": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.obs.Obs.dump": {"tf": 1}}, "df": 4}}}}}}}}}}, "f": {"docs": {"pyerrors.input.pandas.to_sql": {"tf": 1}, "pyerrors.input.pandas.dump_df": {"tf": 1}}, "df": 2}, "b": {"docs": {"pyerrors.input.pandas.to_sql": {"tf": 1}, "pyerrors.input.pandas.read_sql": {"tf": 1}}, "df": 2}, "v": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.misc.pseudo_Obs": {"tf": 1}}, "df": 1}}}}}}, "i": {"docs": {"pyerrors.correlators.Corr.item": {"tf": 1}, "pyerrors.dirac.epsilon_tensor": {"tf": 1}, "pyerrors.dirac.epsilon_tensor_rank4": {"tf": 1}}, "df": 3, "n": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.__init__": {"tf": 1}}, "df": 1}}}, "s": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}}, "df": 3}}}}}}}, "d": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}}, "df": 3}}}}}, "d": {"docs": {"pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}}, "df": 5, "l": {"docs": {"pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}, "pyerrors.input.utils.check_idl": {"tf": 1}, "pyerrors.obs.Obs.__init__": {"tf": 1}, "pyerrors.obs.import_jackknife": {"tf": 1}}, "df": 8}}, "f": {"docs": {"pyerrors.input.pandas.to_sql": {"tf": 1}}, "df": 1}, "m": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.obs.CObs.__init__": {"tf": 1}}, "df": 1}}}}, "p": {"docs": {"pyerrors.fits.qqplot": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 2, "a": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.correlators.Corr.__init__": {"tf": 1}}, "df": 1}}}}}, "r": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.correlators.Corr.correlate": {"tf": 1}, "pyerrors.correlators.Corr.T_symmetry": {"tf": 1}}, "df": 2}}}}, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.correlators.Corr.T_symmetry": {"tf": 1}}, "df": 1}}}}, "t": {"docs": {}, "df": 0, "h": {"docs": {"pyerrors.input.bdio.read_ADerrors": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.read_mesons": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.read_dSdm": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}, "pyerrors.input.misc.read_pbp": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.misc.load_object": {"tf": 1}}, "df": 18}}}, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.__init__": {"tf": 1}, "pyerrors.correlators.Corr.set_prange": {"tf": 1}}, "df": 2}}}}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.print": {"tf": 1}}, "df": 1}}, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}}, "df": 1}}}}, "e": {"docs": {}, "df": 0, "f": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "x": {"docs": {"pyerrors.input.misc.read_pbp": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 8}}}}}, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "c": {"docs": {"pyerrors.correlators.Corr.Hankel": {"tf": 1}}, "df": 1}}}}}}}, "l": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "u": {"docs": {"pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1}}, "df": 2}}}}}}, "o": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.covobs.Covobs.__init__": {"tf": 1}}, "df": 1}}}, "n": {"docs": {"pyerrors.correlators.Corr.Hankel": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 2, "o": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.__init__": {"tf": 1}, "pyerrors.correlators.Corr.projected": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 1}, "pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 3.1622776601683795}, "pyerrors.correlators.Corr.print": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1}, "pyerrors.covobs.Covobs.__init__": {"tf": 1.4142135623730951}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.ks_test": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.write_dobs": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}, "pyerrors.obs.Obs.__init__": {"tf": 1}, "pyerrors.obs.Obs.plot_tauint": {"tf": 1}, "pyerrors.obs.Obs.plot_rho": {"tf": 1}, "pyerrors.obs.Obs.plot_piechart": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}, "pyerrors.obs.import_jackknife": {"tf": 1}, "pyerrors.obs.cov_Obs": {"tf": 1}}, "df": 31}}, "r": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "z": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.projected": {"tf": 1}}, "df": 1}}}}}}}, "e": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}}, "df": 2}}}}}, "f": {"docs": {}, "df": 0, "f": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 1}}}}}}, "t": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "c": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 1}}}}}, "a": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.pandas.to_sql": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.misc.dump_object": {"tf": 1}, "pyerrors.misc.pseudo_Obs": {"tf": 1}, "pyerrors.misc.gen_correlated_data": {"tf": 1}, "pyerrors.obs.import_jackknife": {"tf": 1}, "pyerrors.obs.cov_Obs": {"tf": 1}}, "df": 12, "s": {"docs": {"pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.obs.Obs.__init__": {"tf": 1}}, "df": 2}}}}}, "s": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "f": {"docs": {"pyerrors.correlators.Corr.gamma_method": {"tf": 1}, "pyerrors.correlators.Corr.gm": {"tf": 1}, "pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.correlators.Corr.item": {"tf": 1}, "pyerrors.correlators.Corr.plottable": {"tf": 1}, "pyerrors.correlators.Corr.symmetric": {"tf": 1}, "pyerrors.correlators.Corr.anti_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.is_matrix_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.matrix_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 1}, "pyerrors.correlators.Corr.Hankel": {"tf": 1}, "pyerrors.correlators.Corr.roll": {"tf": 1}, "pyerrors.correlators.Corr.reverse": {"tf": 1}, "pyerrors.correlators.Corr.thin": {"tf": 1}, "pyerrors.correlators.Corr.correlate": {"tf": 1}, "pyerrors.correlators.Corr.reweight": {"tf": 1}, "pyerrors.correlators.Corr.T_symmetry": {"tf": 1}, "pyerrors.correlators.Corr.deriv": {"tf": 1}, "pyerrors.correlators.Corr.second_deriv": {"tf": 1}, "pyerrors.correlators.Corr.m_eff": {"tf": 1}, "pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.correlators.Corr.set_prange": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.correlators.Corr.spaghetti_plot": {"tf": 1}, "pyerrors.correlators.Corr.dump": {"tf": 1}, "pyerrors.correlators.Corr.print": {"tf": 1}, "pyerrors.correlators.Corr.sqrt": {"tf": 1}, "pyerrors.correlators.Corr.log": {"tf": 1}, "pyerrors.correlators.Corr.exp": {"tf": 1}, "pyerrors.correlators.Corr.sin": {"tf": 1}, "pyerrors.correlators.Corr.cos": {"tf": 1}, "pyerrors.correlators.Corr.tan": {"tf": 1}, "pyerrors.correlators.Corr.sinh": {"tf": 1}, "pyerrors.correlators.Corr.cosh": {"tf": 1}, "pyerrors.correlators.Corr.tanh": {"tf": 1}, "pyerrors.correlators.Corr.arcsin": {"tf": 1}, "pyerrors.correlators.Corr.arccos": {"tf": 1}, "pyerrors.correlators.Corr.arctan": {"tf": 1}, "pyerrors.correlators.Corr.arcsinh": {"tf": 1}, "pyerrors.correlators.Corr.arccosh": {"tf": 1}, "pyerrors.correlators.Corr.arctanh": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1}, "pyerrors.covobs.Covobs.errsq": {"tf": 1}, "pyerrors.fits.Fit_result.gamma_method": {"tf": 1}, "pyerrors.fits.Fit_result.gm": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}, "pyerrors.obs.Obs.details": {"tf": 1}, "pyerrors.obs.Obs.reweight": {"tf": 1}, "pyerrors.obs.Obs.is_zero_within_error": {"tf": 1}, "pyerrors.obs.Obs.is_zero": {"tf": 1}, "pyerrors.obs.Obs.plot_tauint": {"tf": 1}, "pyerrors.obs.Obs.plot_rho": {"tf": 1}, "pyerrors.obs.Obs.plot_rep_dist": {"tf": 1}, "pyerrors.obs.Obs.plot_history": {"tf": 1}, "pyerrors.obs.Obs.plot_piechart": {"tf": 1}, "pyerrors.obs.Obs.dump": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1}, "pyerrors.obs.Obs.sqrt": {"tf": 1}, "pyerrors.obs.Obs.log": {"tf": 1}, "pyerrors.obs.Obs.exp": {"tf": 1}, "pyerrors.obs.Obs.sin": {"tf": 1}, "pyerrors.obs.Obs.cos": {"tf": 1}, "pyerrors.obs.Obs.tan": {"tf": 1}, "pyerrors.obs.Obs.arcsin": {"tf": 1}, "pyerrors.obs.Obs.arccos": {"tf": 1}, "pyerrors.obs.Obs.arctan": {"tf": 1}, "pyerrors.obs.Obs.sinh": {"tf": 1}, "pyerrors.obs.Obs.cosh": {"tf": 1}, "pyerrors.obs.Obs.tanh": {"tf": 1}, "pyerrors.obs.Obs.arcsinh": {"tf": 1}, "pyerrors.obs.Obs.arccosh": {"tf": 1}, "pyerrors.obs.Obs.arctanh": {"tf": 1}, "pyerrors.obs.CObs.gamma_method": {"tf": 1}, "pyerrors.obs.CObs.is_zero": {"tf": 1}, "pyerrors.obs.CObs.conjugate": {"tf": 1}}, "df": 78}}, "p": {"docs": {"pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}}, "df": 1, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 4}}}}}}}}, "o": {"docs": {"pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}}, "df": 4, "r": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 1}}, "df": 2}}}, "t": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.Eigenvalue": {"tf": 1}}, "df": 1}}}, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.json.import_json_string": {"tf": 1}}, "df": 1}}}}}, "p": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.correlators.Corr.thin": {"tf": 1}}, "df": 1}}}}, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 1}}}}}, "e": {"docs": {}, "df": 0, "c": {"docs": {"pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}}, "df": 4}}}, "y": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "c": {"docs": {"pyerrors.correlators.Corr.deriv": {"tf": 1}, "pyerrors.correlators.Corr.second_deriv": {"tf": 1}}, "df": 2}}}}}}, "b": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}}, "df": 4}}}}}, "i": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}}, "df": 3}}}}, "g": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "a": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.obs.Obs.is_zero_within_error": {"tf": 1}}, "df": 2}}}}, "a": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.obs.Obs.plot_tauint": {"tf": 1}, "pyerrors.obs.Obs.plot_rho": {"tf": 1}, "pyerrors.obs.Obs.plot_piechart": {"tf": 1}}, "df": 4}}, "m": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.misc.pseudo_Obs": {"tf": 1}, "pyerrors.misc.gen_correlated_data": {"tf": 1}, "pyerrors.obs.Obs.__init__": {"tf": 1}}, "df": 3}}}}}}, "q": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.input.pandas.read_sql": {"tf": 1}}, "df": 1}}, "u": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.linalg.einsum": {"tf": 1}}, "df": 1}}}}}}}}}, "m": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "h": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}}}}, "k": {"docs": {"pyerrors.dirac.epsilon_tensor": {"tf": 1}, "pyerrors.dirac.epsilon_tensor_rank4": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 3, "w": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.correlators.Corr.gamma_method": {"tf": 1}, "pyerrors.correlators.Corr.gm": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.reweight": {"tf": 1}, "pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.correlators.Corr.dump": {"tf": 1}, "pyerrors.fits.Fit_result.gamma_method": {"tf": 1}, "pyerrors.fits.Fit_result.gm": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.fits.fit_lin": {"tf": 1}, "pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}, "pyerrors.input.misc.read_pbp": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}, "pyerrors.input.pandas.to_sql": {"tf": 1}, "pyerrors.input.pandas.read_sql": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.linalg.eigh": {"tf": 1}, "pyerrors.linalg.eig": {"tf": 1}, "pyerrors.linalg.pinv": {"tf": 1}, "pyerrors.linalg.svd": {"tf": 1}, "pyerrors.misc.dump_object": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}, "pyerrors.obs.Obs.__init__": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}, "pyerrors.obs.Obs.dump": {"tf": 1}, "pyerrors.obs.CObs.gamma_method": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}, "pyerrors.obs.reweight": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}, "pyerrors.roots.find_root": {"tf": 1}}, "df": 40}}}}}}, "v": {"1": {"docs": {"pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}}, "df": 2}, "docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.correlators.Corr.projected": {"tf": 1.4142135623730951}}, "df": 1}}}}, "r": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}}, "df": 1}}}}}, "b": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}}, "df": 3}}}}, "s": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 3}}}}}}, "a": {"docs": {"pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}}, "df": 1, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.deriv": {"tf": 1}, "pyerrors.correlators.Corr.second_deriv": {"tf": 1}, "pyerrors.correlators.Corr.m_eff": {"tf": 1}}, "df": 3}}}}}, "l": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.misc.pseudo_Obs": {"tf": 1}}, "df": 1}}}}, "i": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "z": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}}}}}}}, "l": {"docs": {"pyerrors.correlators.Corr.projected": {"tf": 1}}, "df": 1, "o": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.correlators.Corr.m_eff": {"tf": 1}}, "df": 1, "s": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.correlators.Corr.spaghetti_plot": {"tf": 1}}, "df": 2}}}}}}}, "i": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {"pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}}, "df": 4}}}}}, "s": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.obs.merge_obs": {"tf": 1}}, "df": 2}}}}, "r": {"docs": {"pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}}, "df": 2, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.print": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 4}}}}, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.fits.residual_plot": {"tf": 1}}, "df": 2}, "f": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1}}, "df": 1}}}}}}}}, "p": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}}, "df": 2}}, "a": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 1}, "l": {"docs": {"pyerrors.obs.CObs.__init__": {"tf": 1}}, "df": 1}}}}, "f": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.correlators.Corr.Hankel": {"tf": 1}, "pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1.4142135623730951}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.read_dobs": {"tf": 1.4142135623730951}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}, "pyerrors.input.pandas.read_sql": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1.4142135623730951}}, "df": 17}}}, "i": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.input.pandas.to_sql": {"tf": 1}}, "df": 1}}}, "u": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "c": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.fits.qqplot": {"tf": 1}, "pyerrors.fits.residual_plot": {"tf": 1}, "pyerrors.fits.error_band": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}, "pyerrors.roots.find_root": {"tf": 1}}, "df": 7, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.correlators.Corr.fit": {"tf": 1}}, "df": 1}}}}}}, "l": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}}, "df": 6}}}, "i": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.fits.residual_plot": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 4, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.fit": {"tf": 1}}, "df": 1}}}}}}, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}}, "df": 4, "n": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.dump": {"tf": 1}, "pyerrors.obs.Obs.dump": {"tf": 1}}, "df": 2}}}}, "s": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "m": {"docs": {"pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}}, "df": 4}}}}}}}, "n": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}, "pyerrors.input.pandas.dump_df": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1}}, "df": 10}}}}, "l": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "w": {"docs": {"pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}}, "df": 1}}}}, "j": {"docs": {"pyerrors.correlators.Corr.item": {"tf": 1}, "pyerrors.dirac.epsilon_tensor": {"tf": 1}, "pyerrors.dirac.epsilon_tensor_rank4": {"tf": 1}}, "df": 3, "s": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.correlators.Corr.dump": {"tf": 1}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.obs.Obs.dump": {"tf": 1}}, "df": 3}}}, "a": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "k": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.obs.import_jackknife": {"tf": 1}}, "df": 1}}}}}, "t": {"0": {"docs": {"pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 1}}, "df": 2, "p": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "j": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 1}}}}}, "docs": {}, "df": 0, "s": {"docs": {"pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 1}}, "df": 2}, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1}}, "df": 1}}}}, "r": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.spaghetti_plot": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1.4142135623730951}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.pandas.to_sql": {"tf": 1}, "pyerrors.input.pandas.dump_df": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1}, "pyerrors.obs.Obs.details": {"tf": 1}, "pyerrors.obs.Obs.plot_history": {"tf": 1}}, "df": 17}}}, "p": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "j": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 1}}}}, "a": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.dirac.Grid_gamma": {"tf": 1}}, "df": 1}, "r": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.input.openQCD.qtop_projection": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 2}}}}, "b": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.pandas.to_sql": {"tf": 1}}, "df": 1}}}, "u": {"docs": {"pyerrors.misc.gen_correlated_data": {"tf": 1}}, "df": 1}}, "y": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 1}}}}, "e": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 1}}, "df": 2}}}}}}}}}, "n": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}, "pyerrors.obs.Obs.details": {"tf": 1}}, "df": 6, "t": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}}, "df": 2, "s": {"docs": {"pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}}, "df": 2}}}}}}, "x": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 1}}}}, "i": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.pandas.to_sql": {"tf": 1}}, "df": 1}}}}, "p": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.obs.Obs.plot_history": {"tf": 1}}, "df": 1}}}}}}, "o": {"docs": {"pyerrors.dirac.epsilon_tensor_rank4": {"tf": 1}, "pyerrors.fits.qqplot": {"tf": 1}}, "df": 2, "f": {"docs": {"pyerrors.obs.merge_obs": {"tf": 1}}, "df": 1, "f": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.thin": {"tf": 1}}, "df": 1}}}}}, "b": {"docs": {}, "df": 0, "j": {"docs": {"pyerrors.misc.dump_object": {"tf": 1}}, "df": 1, "e": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.fits.ks_test": {"tf": 1}}, "df": 1}}}}}, "s": {"docs": {"pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.linalg.eigh": {"tf": 1}, "pyerrors.linalg.eig": {"tf": 1}, "pyerrors.linalg.pinv": {"tf": 1}, "pyerrors.linalg.svd": {"tf": 1}, "pyerrors.obs.reweight": {"tf": 1}, "pyerrors.obs.correlate": {"tf": 1.4142135623730951}, "pyerrors.obs.covariance": {"tf": 1}, "pyerrors.obs.merge_obs": {"tf": 1}}, "df": 9, "l": {"docs": {"pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}}, "df": 4}}}, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}}, "df": 4}}}}}, "u": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}}, "df": 6}}}}}, "l": {"docs": {"pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}}, "df": 2}, "d": {"docs": {"pyerrors.input.json.dump_dict_to_json": {"tf": 1}}, "df": 1}, "p": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "q": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.openQCD.read_qtop": {"tf": 1}}, "df": 1}}}}, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.linalg.matmul": {"tf": 1}, "pyerrors.linalg.jack_matmul": {"tf": 1}, "pyerrors.linalg.einsum": {"tf": 1}}, "df": 3}}}}}}}}, "w": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.reweight": {"tf": 1}, "pyerrors.obs.Obs.reweight": {"tf": 1}, "pyerrors.obs.reweight": {"tf": 1}}, "df": 3}}}}}, "h": {"docs": {}, "df": 0, "o": {"docs": {"pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}}, "df": 2}}, "f": {"2": {"docs": {"pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 1}, "docs": {"pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 1}}, "g": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.correlators.Corr.m_eff": {"tf": 1}, "pyerrors.roots.find_root": {"tf": 1}}, "df": 2}}}}, "a": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "a": {"docs": {"pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.dirac.Grid_gamma": {"tf": 1}, "pyerrors.input.pandas.read_sql": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1}}, "df": 5, "s": {"docs": {"pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}}, "df": 1}}}}}, "z": {"docs": {"pyerrors.correlators.Corr.dump": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}, "pyerrors.input.pandas.to_sql": {"tf": 1}, "pyerrors.input.pandas.dump_df": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1}, "pyerrors.obs.Obs.dump": {"tf": 1}}, "df": 13}, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.obs.cov_Obs": {"tf": 1}}, "df": 2}}}}, "m": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.plateau": {"tf": 1}}, "df": 1}}}}, "a": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.covobs.Covobs.__init__": {"tf": 1}}, "df": 1, "s": {"docs": {"pyerrors.misc.gen_correlated_data": {"tf": 1}, "pyerrors.obs.cov_Obs": {"tf": 1}}, "df": 2}}}, "s": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.hadrons.read_meson_hd5": {"tf": 1.4142135623730951}}, "df": 1}}}}, "o": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.obs.derived_observable": {"tf": 1}}, "df": 1}}}}, "a": {"docs": {"pyerrors.obs.correlate": {"tf": 1}}, "df": 1, "u": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "o": {"docs": {"pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.input.pandas.read_sql": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1}}, "df": 4}}}, "v": {"docs": {"pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}}, "df": 1}, "t": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.obs.Obs.is_zero": {"tf": 1}}, "df": 1}}}, "r": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.obs.derived_observable": {"tf": 1}}, "df": 1}}}}}, "x": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.fits.fit_lin": {"tf": 1}, "pyerrors.fits.qqplot": {"tf": 1}, "pyerrors.fits.residual_plot": {"tf": 1}, "pyerrors.fits.error_band": {"tf": 1}, "pyerrors.linalg.inv": {"tf": 1}, "pyerrors.linalg.cholesky": {"tf": 1}, "pyerrors.linalg.det": {"tf": 1}}, "df": 10, "m": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 1}}}}, "c": {"docs": {"pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 3, "o": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "p": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1}}, "df": 1}}, "v": {"docs": {"pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.misc.gen_correlated_data": {"tf": 1}, "pyerrors.obs.cov_Obs": {"tf": 1}}, "df": 3}, "n": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.obs.Obs.details": {"tf": 1}}, "df": 2}}}}}, "r": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 2, "s": {"docs": {"pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 1}, "e": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}}}}}}}}, "n": {"docs": {}, "df": 0, "f": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}}, "df": 2}}}, "f": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 1}}, "h": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.utils.check_idl": {"tf": 1}}, "df": 1}}}, "y": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.fits.fit_lin": {"tf": 1}, "pyerrors.fits.qqplot": {"tf": 1}, "pyerrors.fits.residual_plot": {"tf": 1}}, "df": 6, "l": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1}}, "df": 1}}}}}}, "h": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1}}, "df": 1}}}}, "b": {"docs": {"pyerrors.obs.correlate": {"tf": 1}}, "df": 1, "a": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "x": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 1}}}}}}}}}, "e": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "a": {"docs": {"pyerrors.fits.error_band": {"tf": 1}}, "df": 1}}}, "d": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {"pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}}, "df": 4}}}, "i": {"docs": {"pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 1}}, "z": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}}, "df": 1}}}}}}}, "q": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "p": {"docs": {"pyerrors.input.openQCD.qtop_projection": {"tf": 1}}, "df": 1}}}, "c": {"docs": {"pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}}, "df": 1}, "u": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "k": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 1}}}}}}}}, "bases": {"root": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.fits.Fit_result": {"tf": 1}}, "df": 1}}}}}}}}}}}, "a": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "c": {"docs": {"pyerrors.fits.Fit_result": {"tf": 1}}, "df": 1}}}, "s": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "q": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.fits.Fit_result": {"tf": 1}}, "df": 1}}}}}}}}, "n": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}}}, "d": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}}}}}}}}, "doc": {"root": {"0": {"0": {"0": {"0": {"0": {"0": {"0": {"0": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1, "+": {"0": {"0": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "docs": {}, "df": 0}}}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "5": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {"pyerrors": {"tf": 1}}, "df": 1}, "6": {"9": {"7": {"9": {"5": {"8": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "+": {"0": {"0": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}, "docs": {}, "df": 0}, "docs": {}, "df": 0}}}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}, "1": {"2": {"8": {"9": {"docs": {"pyerrors": {"tf": 2}}, "df": 1}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "7": {"1": {"8": {"0": {"6": {"4": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "+": {"0": {"0": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "docs": {}, "df": 0}}}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {"pyerrors": {"tf": 3}}, "df": 1}, "2": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}, "3": {"4": {"4": {"5": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "4": {"5": {"8": {"5": {"6": {"5": {"0": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "7": {"5": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}, "5": {"4": {"8": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "6": {"4": {"2": {"3": {"docs": {"pyerrors.input.openQCD.read_gf_coupling": {"tf": 1.4142135623730951}}, "df": 1}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "7": {"5": {"6": {"0": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "9": {"9": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "docs": {"pyerrors": {"tf": 6.164414002968976}, "pyerrors.correlators.Corr.projected": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.symmetric": {"tf": 1}, "pyerrors.correlators.Corr.anti_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.prune": {"tf": 2.6457513110645907}, "pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 2.23606797749979}, "pyerrors.fits.total_least_squares": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.7320508075688772}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.qtop_projection": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.4142135623730951}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs": {"tf": 2}, "pyerrors.obs.Obs.gamma_method": {"tf": 2}, "pyerrors.obs.Obs.gm": {"tf": 2}, "pyerrors.obs.covariance": {"tf": 2}}, "df": 25, "e": {"docs": {}, "df": 0, "+": {"0": {"0": {"0": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}}, "df": 1}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}}, "c": {"docs": {"pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 1}, "d": {"docs": {"pyerrors.obs.cov_Obs": {"tf": 1}}, "df": 1}}, "1": {"0": {"0": {"0": {"docs": {"pyerrors": {"tf": 3.605551275463989}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.4142135623730951}, "pyerrors.misc.pseudo_Obs": {"tf": 1}}, "df": 3}, "3": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}, "5": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "4": {"7": {"2": {"docs": {"pyerrors.correlators.Corr.GEVP": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {"pyerrors": {"tf": 1}}, "df": 1}, "1": {"9": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "2": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}, "3": {"4": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}, "docs": {}, "df": 0}, "4": {"3": {"7": {"1": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "5": {"0": {"0": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}, "docs": {}, "df": 0}, "3": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "6": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "6": {"0": {"7": {"docs": {"pyerrors.input.openQCD.read_gf_coupling": {"tf": 1.4142135623730951}}, "df": 1}, "docs": {}, "df": 0}, "docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}, "7": {"6": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {"pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 1}, "9": {"0": {"6": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "8": {"8": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "9": {"0": {"docs": {"pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "docs": {"pyerrors": {"tf": 1.7320508075688772}}, "df": 1}, "docs": {"pyerrors": {"tf": 6.164414002968976}, "pyerrors.correlators.Corr.Hankel": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.T_symmetry": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.m_eff": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.prune": {"tf": 1.4142135623730951}, "pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.total_least_squares": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.read_mesons": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.read_dSdm": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.4142135623730951}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.7320508075688772}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1.4142135623730951}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 20, "}": {"docs": {}, "df": 0, "^": {"docs": {}, "df": 0, "{": {"docs": {}, "df": 0, "w": {"docs": {}, "df": 0, "}": {"docs": {}, "df": 0, "\\": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "o": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}}, "n": {"docs": {}, "df": 0, "\\": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "a": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}}}, "s": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "m": {"docs": {"pyerrors.obs.covariance": {"tf": 1.4142135623730951}}, "df": 1}}}}}}}, "\\": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}, "+": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "d": {"docs": {"pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.obs.cov_Obs": {"tf": 1}}, "df": 2}, "*": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "z": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}}}}}}}}, "/": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.obs.Obs.export_jackknife": {"tf": 1}}, "df": 1}}}, "2": {"0": {"0": {"4": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "7": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "1": {"1": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "9": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "4": {"1": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}, "docs": {}, "df": 0}, "docs": {"pyerrors": {"tf": 2}}, "df": 1}, "1": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}, "3": {"8": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {"pyerrors": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}}, "df": 2}, "5": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "7": {"2": {"1": {"8": {"6": {"6": {"7": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "+": {"0": {"0": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "docs": {}, "df": 0}}}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "8": {"0": {"9": {"7": {"7": {"6": {"2": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {"pyerrors": {"tf": 2}}, "df": 1}, "9": {"9": {"0": {"9": {"7": {"0": {"3": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "+": {"0": {"0": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "docs": {}, "df": 0}}}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}, "docs": {"pyerrors": {"tf": 5}, "pyerrors.correlators.Corr.GEVP": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.m_eff": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.total_least_squares": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 3}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}, "pyerrors.roots.find_root": {"tf": 1}}, "df": 15, "x": {"2": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "d": {"docs": {"pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.linalg.matmul": {"tf": 1}, "pyerrors.linalg.jack_matmul": {"tf": 1}, "pyerrors.linalg.einsum": {"tf": 1}, "pyerrors.obs.cov_Obs": {"tf": 1}}, "df": 5}, "*": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 1}}}}}, "3": {"0": {"6": {"7": {"5": {"2": {"0": {"1": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "1": {"4": {"9": {"8": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "2": {"2": {"7": {"docs": {"pyerrors": {"tf": 1.7320508075688772}}, "df": 1}, "docs": {}, "df": 0}, "3": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "3": {"docs": {"pyerrors": {"tf": 2}}, "df": 1}, "4": {"9": {"7": {"6": {"8": {"0": {"0": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "+": {"0": {"2": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "docs": {}, "df": 0}}}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "5": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "6": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "7": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "8": {"3": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {"pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 1}, "9": {"docs": {"pyerrors": {"tf": 7.745966692414834}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}}, "df": 2}, "docs": {"pyerrors": {"tf": 3.605551275463989}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1}, "pyerrors.dirac.epsilon_tensor": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 2}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1.4142135623730951}}, "df": 8, "a": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "4": {"0": {"3": {"2": {"0": {"9": {"8": {"3": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {"pyerrors.fits.least_squares": {"tf": 1}}, "df": 1}, "9": {"5": {"9": {"1": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.fit": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.dirac.epsilon_tensor_rank4": {"tf": 1.4142135623730951}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}}, "df": 6, "x": {"4": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}}, "5": {"0": {"0": {"docs": {"pyerrors": {"tf": 2.8284271247461903}, "pyerrors.fits.least_squares": {"tf": 1}}, "df": 2}, "1": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}, "docs": {}, "df": 0}, "1": {"5": {"6": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}, "docs": {}, "df": 0}, "9": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}, "docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}, "2": {"0": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "2": {"8": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "3": {"8": {"0": {"docs": {"pyerrors.correlators.Corr.m_eff": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "4": {"8": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "6": {"7": {"3": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "4": {"6": {"5": {"9": {"8": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "7": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "8": {"3": {"4": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {"pyerrors": {"tf": 3.605551275463989}, "pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix.g5H": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 5, "e": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}, "6": {"4": {"docs": {"pyerrors": {"tf": 1.7320508075688772}}, "df": 1}, "5": {"0": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "9": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "6": {"8": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.correlators.Corr.fit": {"tf": 1.4142135623730951}}, "df": 2}, "7": {"0": {"0": {"0": {"0": {"0": {"0": {"0": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "+": {"0": {"0": {"docs": {"pyerrors": {"tf": 1.7320508075688772}}, "df": 1}, "docs": {}, "df": 0}, "docs": {}, "df": 0}}}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {"pyerrors": {"tf": 1}}, "df": 1}, "1": {"4": {"2": {"2": {"9": {"0": {"0": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "+": {"0": {"0": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "docs": {}, "df": 0}}}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "2": {"0": {"4": {"6": {"6": {"5": {"8": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {"pyerrors": {"tf": 1}}, "df": 1}, "4": {"5": {"7": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "7": {"3": {"1": {"0": {"1": {"0": {"2": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "9": {"0": {"7": {"7": {"5": {"2": {"4": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "5": {"7": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {"pyerrors": {"tf": 2.23606797749979}}, "df": 1}, "8": {"0": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}, "1": {"4": {"docs": {"pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "2": {"4": {"docs": {"pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "4": {"5": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 6}, "9": {"1": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "3": {"3": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {"pyerrors": {"tf": 1}}, "df": 1}, "4": {"7": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "5": {"9": {"3": {"0": {"3": {"5": {"7": {"8": {"5": {"1": {"6": {"0": {"9": {"3": {"6": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "4": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "7": {"6": {"8": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "8": {"3": {"1": {"9": {"8": {"8": {"1": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "9": {"1": {"0": {"0": {"7": {"1": {"2": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "5": {"8": {"3": {"6": {"5": {"4": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "9": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {"pyerrors": {"tf": 2.449489742783178}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}}, "df": 3}, "docs": {"pyerrors": {"tf": 62.928530890209096}, "pyerrors.correlators": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr": {"tf": 3}, "pyerrors.correlators.Corr.__init__": {"tf": 5.0990195135927845}, "pyerrors.correlators.Corr.gamma_method": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.gm": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.projected": {"tf": 2.23606797749979}, "pyerrors.correlators.Corr.item": {"tf": 4.58257569495584}, "pyerrors.correlators.Corr.plottable": {"tf": 2.449489742783178}, "pyerrors.correlators.Corr.symmetric": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.anti_symmetric": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.is_matrix_symmetric": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.matrix_symmetric": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.GEVP": {"tf": 10.535653752852738}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 4.358898943540674}, "pyerrors.correlators.Corr.Hankel": {"tf": 4.58257569495584}, "pyerrors.correlators.Corr.roll": {"tf": 3.605551275463989}, "pyerrors.correlators.Corr.reverse": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.thin": {"tf": 4.242640687119285}, "pyerrors.correlators.Corr.correlate": {"tf": 3.7416573867739413}, "pyerrors.correlators.Corr.reweight": {"tf": 4.58257569495584}, "pyerrors.correlators.Corr.T_symmetry": {"tf": 4.242640687119285}, "pyerrors.correlators.Corr.deriv": {"tf": 3.7416573867739413}, "pyerrors.correlators.Corr.second_deriv": {"tf": 3.7416573867739413}, "pyerrors.correlators.Corr.m_eff": {"tf": 5.830951894845301}, "pyerrors.correlators.Corr.fit": {"tf": 5.291502622129181}, "pyerrors.correlators.Corr.plateau": {"tf": 5}, "pyerrors.correlators.Corr.set_prange": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.show": {"tf": 8.660254037844387}, "pyerrors.correlators.Corr.spaghetti_plot": {"tf": 3.872983346207417}, "pyerrors.correlators.Corr.dump": {"tf": 5.0990195135927845}, "pyerrors.correlators.Corr.print": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.sqrt": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.log": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.exp": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.sin": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.cos": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.tan": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.sinh": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.cosh": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.tanh": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.arcsin": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.arccos": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.arctan": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.arcsinh": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.arccosh": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.arctanh": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.prune": {"tf": 6.855654600401044}, "pyerrors.covobs": {"tf": 1.7320508075688772}, "pyerrors.covobs.Covobs": {"tf": 1.7320508075688772}, "pyerrors.covobs.Covobs.__init__": {"tf": 5.916079783099616}, "pyerrors.covobs.Covobs.errsq": {"tf": 1.4142135623730951}, "pyerrors.dirac": {"tf": 1.7320508075688772}, "pyerrors.dirac.epsilon_tensor": {"tf": 4.123105625617661}, "pyerrors.dirac.epsilon_tensor_rank4": {"tf": 4.123105625617661}, "pyerrors.dirac.Grid_gamma": {"tf": 1.7320508075688772}, "pyerrors.fits": {"tf": 1.7320508075688772}, "pyerrors.fits.Fit_result": {"tf": 5.656854249492381}, "pyerrors.fits.Fit_result.__init__": {"tf": 1.7320508075688772}, "pyerrors.fits.Fit_result.gamma_method": {"tf": 1.4142135623730951}, "pyerrors.fits.Fit_result.gm": {"tf": 1.4142135623730951}, "pyerrors.fits.least_squares": {"tf": 16.431676725154983}, "pyerrors.fits.total_least_squares": {"tf": 15.427248620541512}, "pyerrors.fits.fit_lin": {"tf": 5.916079783099616}, "pyerrors.fits.qqplot": {"tf": 3.605551275463989}, "pyerrors.fits.residual_plot": {"tf": 3.4641016151377544}, "pyerrors.fits.error_band": {"tf": 3.7416573867739413}, "pyerrors.fits.ks_test": {"tf": 5}, "pyerrors.input": {"tf": 4.69041575982343}, "pyerrors.input.bdio": {"tf": 1.7320508075688772}, "pyerrors.input.bdio.read_ADerrors": {"tf": 6.164414002968976}, "pyerrors.input.bdio.write_ADerrors": {"tf": 6.164414002968976}, "pyerrors.input.bdio.read_mesons": {"tf": 8.12403840463596}, "pyerrors.input.bdio.read_dSdm": {"tf": 7.416198487095663}, "pyerrors.input.dobs": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.create_pobs_string": {"tf": 7.745966692414834}, "pyerrors.input.dobs.write_pobs": {"tf": 8.426149773176359}, "pyerrors.input.dobs.read_pobs": {"tf": 7.280109889280518}, "pyerrors.input.dobs.import_dobs_string": {"tf": 7.681145747868608}, "pyerrors.input.dobs.read_dobs": {"tf": 8.12403840463596}, "pyerrors.input.dobs.create_dobs_string": {"tf": 8.06225774829855}, "pyerrors.input.dobs.write_dobs": {"tf": 8.774964387392123}, "pyerrors.input.hadrons": {"tf": 1.7320508075688772}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 7.3484692283495345}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 6.557438524302}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 20.904544960366874}, "pyerrors.input.hadrons.Npr_matrix.__init__": {"tf": 1.7320508075688772}, "pyerrors.input.hadrons.Npr_matrix.g5H": {"tf": 2.23606797749979}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 6.324555320336759}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 6.324555320336759}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 6.782329983125268}, "pyerrors.input.json": {"tf": 1.7320508075688772}, "pyerrors.input.json.create_json_string": {"tf": 6.082762530298219}, "pyerrors.input.json.dump_to_json": {"tf": 6.928203230275509}, "pyerrors.input.json.import_json_string": {"tf": 7.681145747868608}, "pyerrors.input.json.load_json": {"tf": 8.06225774829855}, "pyerrors.input.json.dump_dict_to_json": {"tf": 7.3484692283495345}, "pyerrors.input.json.load_json_dict": {"tf": 7.937253933193772}, "pyerrors.input.misc": {"tf": 1.7320508075688772}, "pyerrors.input.misc.read_pbp": {"tf": 5.477225575051661}, "pyerrors.input.openQCD": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_rwms": {"tf": 8.54400374531753}, "pyerrors.input.openQCD.extract_t0": {"tf": 10.44030650891055}, "pyerrors.input.openQCD.read_qtop": {"tf": 10.246950765959598}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 8.888194417315589}, "pyerrors.input.openQCD.qtop_projection": {"tf": 5.656854249492381}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 9.797958971132712}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 10}, "pyerrors.input.pandas": {"tf": 1.7320508075688772}, "pyerrors.input.pandas.to_sql": {"tf": 7}, "pyerrors.input.pandas.read_sql": {"tf": 6.244997998398398}, "pyerrors.input.pandas.dump_df": {"tf": 6.324555320336759}, "pyerrors.input.pandas.load_df": {"tf": 6.244997998398398}, "pyerrors.input.sfcf": {"tf": 1.7320508075688772}, "pyerrors.input.sfcf.read_sfcf": {"tf": 11.090536506409418}, "pyerrors.input.utils": {"tf": 1.4142135623730951}, "pyerrors.input.utils.check_idl": {"tf": 5.385164807134504}, "pyerrors.linalg": {"tf": 1.7320508075688772}, "pyerrors.linalg.matmul": {"tf": 4.58257569495584}, "pyerrors.linalg.jack_matmul": {"tf": 4.47213595499958}, "pyerrors.linalg.einsum": {"tf": 4.47213595499958}, "pyerrors.linalg.inv": {"tf": 1.7320508075688772}, "pyerrors.linalg.cholesky": {"tf": 1.7320508075688772}, "pyerrors.linalg.det": {"tf": 1.7320508075688772}, "pyerrors.linalg.eigh": {"tf": 1.7320508075688772}, "pyerrors.linalg.eig": {"tf": 1.7320508075688772}, "pyerrors.linalg.pinv": {"tf": 1.7320508075688772}, "pyerrors.linalg.svd": {"tf": 1.7320508075688772}, "pyerrors.misc": {"tf": 1.7320508075688772}, "pyerrors.misc.dump_object": {"tf": 5.916079783099616}, "pyerrors.misc.load_object": {"tf": 5}, "pyerrors.misc.pseudo_Obs": {"tf": 6.557438524302}, "pyerrors.misc.gen_correlated_data": {"tf": 7.0710678118654755}, "pyerrors.mpm": {"tf": 1.7320508075688772}, "pyerrors.mpm.matrix_pencil_method": {"tf": 6.324555320336759}, "pyerrors.obs": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs": {"tf": 6.928203230275509}, "pyerrors.obs.Obs.__init__": {"tf": 4.898979485566356}, "pyerrors.obs.Obs.gamma_method": {"tf": 5.744562646538029}, "pyerrors.obs.Obs.gm": {"tf": 5.744562646538029}, "pyerrors.obs.Obs.details": {"tf": 3.872983346207417}, "pyerrors.obs.Obs.reweight": {"tf": 4.58257569495584}, "pyerrors.obs.Obs.is_zero_within_error": {"tf": 4.47213595499958}, "pyerrors.obs.Obs.is_zero": {"tf": 3.872983346207417}, "pyerrors.obs.Obs.plot_tauint": {"tf": 3.872983346207417}, "pyerrors.obs.Obs.plot_rho": {"tf": 3.872983346207417}, "pyerrors.obs.Obs.plot_rep_dist": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.plot_history": {"tf": 3.7416573867739413}, "pyerrors.obs.Obs.plot_piechart": {"tf": 3.872983346207417}, "pyerrors.obs.Obs.dump": {"tf": 5.744562646538029}, "pyerrors.obs.Obs.export_jackknife": {"tf": 3.7416573867739413}, "pyerrors.obs.Obs.sqrt": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.log": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.exp": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.sin": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.cos": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.tan": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.arcsin": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.arccos": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.arctan": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.sinh": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.cosh": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.tanh": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.arcsinh": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.arccosh": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.arctanh": {"tf": 1.7320508075688772}, "pyerrors.obs.CObs": {"tf": 1.7320508075688772}, "pyerrors.obs.CObs.__init__": {"tf": 1.7320508075688772}, "pyerrors.obs.CObs.gamma_method": {"tf": 1.7320508075688772}, "pyerrors.obs.CObs.is_zero": {"tf": 1.7320508075688772}, "pyerrors.obs.CObs.conjugate": {"tf": 1.7320508075688772}, "pyerrors.obs.derived_observable": {"tf": 6.4031242374328485}, "pyerrors.obs.reweight": {"tf": 5.196152422706632}, "pyerrors.obs.correlate": {"tf": 4.898979485566356}, "pyerrors.obs.covariance": {"tf": 6.6332495807108}, "pyerrors.obs.import_jackknife": {"tf": 4.47213595499958}, "pyerrors.obs.merge_obs": {"tf": 4.123105625617661}, "pyerrors.obs.cov_Obs": {"tf": 5.385164807134504}, "pyerrors.roots": {"tf": 1.7320508075688772}, "pyerrors.roots.find_root": {"tf": 10.488088481701515}, "pyerrors.version": {"tf": 1.7320508075688772}}, "df": 181, "w": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 2, "h": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}}, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 2.449489742783178}, "pyerrors.correlators.Corr.__init__": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 2}, "pyerrors.fits.fit_lin": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1}, "pyerrors.roots.find_root": {"tf": 1}}, "df": 12}}, "n": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.obs.Obs.is_zero_within_error": {"tf": 1}}, "df": 10}, "t": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.is_matrix_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.Hankel": {"tf": 1}, "pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.correlators.Corr.spaghetti_plot": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}, "pyerrors.obs.Obs.is_zero_within_error": {"tf": 1}, "pyerrors.obs.Obs.is_zero": {"tf": 1}, "pyerrors.obs.CObs.is_zero": {"tf": 1}}, "df": 11}}}}}, "i": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "h": {"docs": {"pyerrors": {"tf": 3.1622776601683795}, "pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.correlate": {"tf": 1}, "pyerrors.correlators.Corr.deriv": {"tf": 1}, "pyerrors.correlators.Corr.second_deriv": {"tf": 1}, "pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1.7320508075688772}, "pyerrors.fits.least_squares": {"tf": 2}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.fits.fit_lin": {"tf": 1}, "pyerrors.fits.qqplot": {"tf": 1}, "pyerrors.fits.residual_plot": {"tf": 1}, "pyerrors.input": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.misc.read_pbp": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_rwms": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.extract_t0": {"tf": 2}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1.7320508075688772}, "pyerrors.input.pandas.load_df": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.input.utils.check_idl": {"tf": 1}, "pyerrors.linalg.matmul": {"tf": 1}, "pyerrors.linalg.jack_matmul": {"tf": 1}, "pyerrors.linalg.einsum": {"tf": 1}, "pyerrors.misc.pseudo_Obs": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.__init__": {"tf": 1}, "pyerrors.obs.Obs.plot_piechart": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}, "pyerrors.obs.correlate": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}, "pyerrors.obs.merge_obs": {"tf": 1}}, "df": 36}}, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 2}}}, "o": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}}, "df": 3}}, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "h": {"docs": {"pyerrors": {"tf": 5.830951894845301}, "pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.correlators.Corr.__init__": {"tf": 1}, "pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.correlators.Corr.correlate": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.deriv": {"tf": 1}, "pyerrors.correlators.Corr.second_deriv": {"tf": 1}, "pyerrors.correlators.Corr.plateau": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.show": {"tf": 1.4142135623730951}, "pyerrors.fits.least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.fits.error_band": {"tf": 1}, "pyerrors.input": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}, "pyerrors.input.pandas.read_sql": {"tf": 1.4142135623730951}, "pyerrors.input.pandas.load_df": {"tf": 1.4142135623730951}, "pyerrors.input.sfcf.read_sfcf": {"tf": 2}, "pyerrors.input.utils.check_idl": {"tf": 1}, "pyerrors.misc.pseudo_Obs": {"tf": 1}, "pyerrors.misc.gen_correlated_data": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}, "pyerrors.obs.Obs.reweight": {"tf": 1}, "pyerrors.obs.Obs.plot_rep_dist": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 2}}, "df": 35, "i": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.obs.Obs.is_zero_within_error": {"tf": 1}, "pyerrors.obs.Obs.is_zero": {"tf": 1}, "pyerrors.obs.CObs.is_zero": {"tf": 1}}, "df": 9}}, "o": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}, "n": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "w": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors": {"tf": 1.7320508075688772}}, "df": 1}}}, "s": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "z": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}, "/": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}}}}}}}}, "l": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}}, "df": 12}, "s": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 3}}}}}, "o": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "f": {"docs": {}, "df": 0, "f": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}, "r": {"docs": {}, "df": 0, "k": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}}, "df": 4, "s": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.obs.Obs.is_zero_within_error": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1}, "pyerrors.obs.correlate": {"tf": 1}}, "df": 6}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}, "f": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "w": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input": {"tf": 1}}, "df": 1}}}}}}}}, "e": {"docs": {"pyerrors": {"tf": 2}, "pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1.4142135623730951}}, "df": 3, "l": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 2}}, "i": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.reweight": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.reweight": {"tf": 1.4142135623730951}, "pyerrors.obs.reweight": {"tf": 1.4142135623730951}}, "df": 3}}}}}, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 2}}}}}, "y": {"docs": {"pyerrors": {"tf": 1}}, "df": 1, "s": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}, "s": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.is_zero_within_error": {"tf": 1}}, "df": 6}, "n": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.correlators.Corr": {"tf": 1}}, "df": 1}}}, "v": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 1, "f": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 1, "s": {"docs": {"pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 1}}}}}}}}}}}}, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.linalg.einsum": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}, "pyerrors.roots.find_root": {"tf": 1}}, "df": 3, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.write_pobs": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.read_dobs": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.write_dobs": {"tf": 1.4142135623730951}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}, "pyerrors.input.pandas.to_sql": {"tf": 1}}, "df": 12}}}, "e": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.pandas.to_sql": {"tf": 1}}, "df": 3, "a": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}}}}}}, "t": {"docs": {"pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.obs.cov_Obs": {"tf": 1}}, "df": 2}, "o": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}}, "df": 2}}}}, "l": {"docs": {"pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}}, "df": 4}, "f": {"2": {"docs": {"pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 1}, "docs": {"pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 1}}, "i": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.item": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.GEVP": {"tf": 2.449489742783178}, "pyerrors.correlators.Corr.reweight": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1.4142135623730951}, "pyerrors.dirac.epsilon_tensor": {"tf": 1}, "pyerrors.dirac.epsilon_tensor_rank4": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.reweight": {"tf": 1}, "pyerrors.obs.reweight": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1.4142135623730951}}, "df": 11, "s": {"docs": {"pyerrors": {"tf": 8}, "pyerrors.correlators.Corr": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.is_matrix_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 2.6457513110645907}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.Hankel": {"tf": 1}, "pyerrors.correlators.Corr.correlate": {"tf": 1}, "pyerrors.correlators.Corr.deriv": {"tf": 1}, "pyerrors.correlators.Corr.second_deriv": {"tf": 1}, "pyerrors.correlators.Corr.fit": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.correlators.Corr.spaghetti_plot": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 3.4641016151377544}, "pyerrors.covobs.Covobs.__init__": {"tf": 1.4142135623730951}, "pyerrors.fits.least_squares": {"tf": 3.4641016151377544}, "pyerrors.fits.total_least_squares": {"tf": 1.7320508075688772}, "pyerrors.fits.fit_lin": {"tf": 1}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.write_pobs": {"tf": 2.23606797749979}, "pyerrors.input.dobs.read_pobs": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.import_dobs_string": {"tf": 2.23606797749979}, "pyerrors.input.dobs.read_dobs": {"tf": 2.449489742783178}, "pyerrors.input.dobs.create_dobs_string": {"tf": 2}, "pyerrors.input.dobs.write_dobs": {"tf": 2.449489742783178}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 3.4641016151377544}, "pyerrors.input.hadrons.Npr_matrix.g5H": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1.7320508075688772}, "pyerrors.input.json.import_json_string": {"tf": 1.7320508075688772}, "pyerrors.input.json.load_json": {"tf": 2}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1.7320508075688772}, "pyerrors.input.json.load_json_dict": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_rwms": {"tf": 2.23606797749979}, "pyerrors.input.openQCD.extract_t0": {"tf": 3.605551275463989}, "pyerrors.input.openQCD.read_qtop": {"tf": 2.449489742783178}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 2.449489742783178}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 2.449489742783178}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1.4142135623730951}, "pyerrors.input.pandas.dump_df": {"tf": 1.4142135623730951}, "pyerrors.input.pandas.load_df": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 2}, "pyerrors.input.utils.check_idl": {"tf": 1}, "pyerrors.linalg.matmul": {"tf": 1}, "pyerrors.linalg.jack_matmul": {"tf": 1}, "pyerrors.misc.pseudo_Obs": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.gm": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.is_zero_within_error": {"tf": 1}, "pyerrors.obs.Obs.is_zero": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1}, "pyerrors.obs.correlate": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 3}, "pyerrors.obs.merge_obs": {"tf": 1}, "pyerrors.roots.find_root": {"tf": 1}}, "df": 57}, "t": {"docs": {"pyerrors": {"tf": 3.1622776601683795}, "pyerrors.correlators.Corr": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.projected": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.GEVP": {"tf": 1.7320508075688772}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.7320508075688772}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}, "pyerrors.obs.merge_obs": {"tf": 1}}, "df": 20, "s": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.T_symmetry": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 2}}, "df": 8, "e": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "f": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 2}}}}, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1.7320508075688772}}, "df": 1}}}, "e": {"docs": {"pyerrors.correlators.Corr": {"tf": 1}}, "df": 1}, "o": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}}, "df": 1}}}}}, "m": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1, "i": {"docs": {}, "df": 0, "z": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1.7320508075688772}}, "df": 1}}}}}}}, "n": {"docs": {"pyerrors": {"tf": 8.18535277187245}, "pyerrors.correlators.Corr.plottable": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 1}, "pyerrors.correlators.Corr.reweight": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.show": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.prune": {"tf": 1.4142135623730951}, "pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.dirac.Grid_gamma": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 2}, "pyerrors.fits.fit_lin": {"tf": 1}, "pyerrors.fits.ks_test": {"tf": 1}, "pyerrors.input": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.read_dobs": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1.7320508075688772}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1.7320508075688772}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 3.3166247903554}, "pyerrors.input.hadrons.Npr_matrix.g5H": {"tf": 1}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1.4142135623730951}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.extract_t0": {"tf": 2.23606797749979}, "pyerrors.input.openQCD.read_qtop": {"tf": 2.23606797749979}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 2.23606797749979}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 2.23606797749979}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 2.23606797749979}, "pyerrors.input.pandas.to_sql": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.4142135623730951}, "pyerrors.input.utils.check_idl": {"tf": 1}, "pyerrors.misc.dump_object": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}, "pyerrors.obs.Obs.reweight": {"tf": 1.7320508075688772}, "pyerrors.obs.reweight": {"tf": 1.7320508075688772}, "pyerrors.obs.correlate": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}, "pyerrors.obs.merge_obs": {"tf": 1}}, "df": 45, "t": {"1": {"6": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.item": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.GEVP": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 1}, "pyerrors.correlators.Corr.Hankel": {"tf": 1}, "pyerrors.correlators.Corr.roll": {"tf": 1}, "pyerrors.correlators.Corr.thin": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.T_symmetry": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1.7320508075688772}, "pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.dirac.epsilon_tensor": {"tf": 1}, "pyerrors.dirac.epsilon_tensor_rank4": {"tf": 1}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1.7320508075688772}, "pyerrors.input.bdio.read_dSdm": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.read_pobs": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.read_dobs": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 2.6457513110645907}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 2.23606797749979}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.qtop_projection": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 2}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.7320508075688772}, "pyerrors.misc.pseudo_Obs": {"tf": 1}, "pyerrors.misc.gen_correlated_data": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.is_zero_within_error": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 34, "r": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 2}}, "df": 1}, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}}}}}, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1}}, "df": 1, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 1}}, "df": 2}}}}}, "p": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}}}}}}}, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1}}, "df": 1, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}}, "df": 1}}}}}}, "n": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}}, "df": 5}}}, "f": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}}, "df": 1}}}}, "a": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}}}}}, "c": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.obs.derived_observable": {"tf": 1}}, "df": 1}}}}}}}, "g": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.plot_tauint": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1.4142135623730951}}, "df": 3}}}}}, "e": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.4142135623730951}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 6, "s": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.input.utils.check_idl": {"tf": 1}}, "df": 2}}}}}, "o": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.dump": {"tf": 1}, "pyerrors.input": {"tf": 1}, "pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.misc.dump_object": {"tf": 1}, "pyerrors.obs.merge_obs": {"tf": 1}}, "df": 10}, "s": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 2}}, "df": 1}}, "v": {"docs": {"pyerrors": {"tf": 1}}, "df": 1, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.linalg.inv": {"tf": 1}}, "df": 3}}, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}}, "df": 2}}}}}}, "i": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1, "i": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.fits.least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.total_least_squares": {"tf": 1.4142135623730951}, "pyerrors.roots.find_root": {"tf": 1}}, "df": 3, "i": {"docs": {}, "df": 0, "z": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.__init__": {"tf": 1}, "pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.obs.Obs.__init__": {"tf": 1}}, "df": 4, "d": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1}}, "df": 3}}}}}}}}}, "d": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 2}}, "df": 1, "l": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}}}}, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}}, "df": 3, "i": {"docs": {}, "df": 0, "f": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.__init__": {"tf": 1}}, "df": 1}}}}}, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}}, "df": 3}}}}}}}, "x": {"docs": {"pyerrors.correlators.Corr.item": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.plottable": {"tf": 1}}, "df": 2, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.correlators.Corr.fit": {"tf": 1}}, "df": 1}}}}}, "i": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.fits.Fit_result": {"tf": 1}, "pyerrors.obs.Obs.__init__": {"tf": 1}}, "df": 4, "l": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.correlators.Corr.GEVP": {"tf": 1}}, "df": 1}}}}}}}}, "c": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}}, "df": 2}}}}}, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.fits.Fit_result": {"tf": 1}}, "df": 1}}}}}, "p": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 2.449489742783178}, "pyerrors.correlators.Corr.__init__": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.total_least_squares": {"tf": 1.4142135623730951}, "pyerrors.input": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.input.utils": {"tf": 1}}, "df": 15}}}, "f": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 3.1622776601683795}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}, "pyerrors.input.json.import_json_string": {"tf": 1.4142135623730951}, "pyerrors.input.json.load_json": {"tf": 1.4142135623730951}, "pyerrors.input.json.load_json_dict": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}}, "df": 11}}}}}}}}}, "s": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 7}}}, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}}, "df": 1}}}, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}}}}}}, "c": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.obs.Obs": {"tf": 1.4142135623730951}}, "df": 1, "s": {"docs": {"pyerrors.obs.Obs": {"tf": 1.4142135623730951}}, "df": 1}}}}}}, "p": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}, "i": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 2}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}}, "df": 8}}}, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}}, "df": 3, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.dobs.read_pobs": {"tf": 2}, "pyerrors.input.dobs.import_dobs_string": {"tf": 2}, "pyerrors.input.dobs.read_dobs": {"tf": 2}}, "df": 3}}}, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.dobs.import_dobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.read_dobs": {"tf": 1.4142135623730951}}, "df": 2}}}}}}, "c": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.dump": {"tf": 1}, "pyerrors.fits.ks_test": {"tf": 1}, "pyerrors.obs.Obs.dump": {"tf": 1}}, "df": 4, "s": {"docs": {"pyerrors.input": {"tf": 1}}, "df": 1}, "d": {"docs": {"pyerrors.input.dobs.import_dobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.read_dobs": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 3}}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.pandas.to_sql": {"tf": 1}, "pyerrors.input.pandas.read_sql": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 3}}}}, "s": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.fit": {"tf": 1}}, "df": 1}}}}}}, "o": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr": {"tf": 1}}, "df": 1}}}}}}}}}, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}}, "df": 2}}}}}}}, "n": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.input.sfcf.read_sfcf": {"tf": 1.4142135623730951}}, "df": 1}}}}, "f": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 2}, "pyerrors.correlators.Corr.reweight": {"tf": 1}, "pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.prune": {"tf": 1}, "pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 2.449489742783178}, "pyerrors.fits.total_least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.qqplot": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.read_pobs": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.import_dobs_string": {"tf": 2}, "pyerrors.input.dobs.read_dobs": {"tf": 2.449489742783178}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1.7320508075688772}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.7320508075688772}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1.4142135623730951}, "pyerrors.input.json.import_json_string": {"tf": 2.23606797749979}, "pyerrors.input.json.load_json": {"tf": 2.6457513110645907}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1.4142135623730951}, "pyerrors.input.json.load_json_dict": {"tf": 2}, "pyerrors.input.openQCD.read_rwms": {"tf": 2}, "pyerrors.input.openQCD.extract_t0": {"tf": 3}, "pyerrors.input.openQCD.read_qtop": {"tf": 2.8284271247461903}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 2.449489742783178}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 2.449489742783178}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1.7320508075688772}, "pyerrors.input.pandas.to_sql": {"tf": 1.7320508075688772}, "pyerrors.input.pandas.read_sql": {"tf": 1}, "pyerrors.input.pandas.dump_df": {"tf": 1.4142135623730951}, "pyerrors.input.pandas.load_df": {"tf": 1.7320508075688772}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.7320508075688772}, "pyerrors.input.utils.check_idl": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}, "pyerrors.obs.Obs.details": {"tf": 1}, "pyerrors.obs.Obs.reweight": {"tf": 1}, "pyerrors.obs.Obs.plot_tauint": {"tf": 1}, "pyerrors.obs.Obs.plot_rho": {"tf": 1}, "pyerrors.obs.Obs.plot_piechart": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}, "pyerrors.obs.reweight": {"tf": 1}, "pyerrors.obs.correlate": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1.7320508075688772}}, "df": 54}, "m": {"docs": {"pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 1, "p": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.linalg.matmul": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1}}, "df": 4, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}}}}}}}, "i": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}}}, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 2.6457513110645907}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.input": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}, "pyerrors.roots.find_root": {"tf": 1}}, "df": 11, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}}, "df": 4}}}, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.dobs.read_pobs": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.read_dobs": {"tf": 1.4142135623730951}, "pyerrors.input.json.load_json_dict": {"tf": 1}, "pyerrors.input.pandas.read_sql": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1}}, "df": 6}}, "s": {"docs": {"pyerrors.input.pandas.load_df": {"tf": 1}, "pyerrors.obs.import_jackknife": {"tf": 1}}, "df": 2}}}}, "r": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.deriv": {"tf": 1}, "pyerrors.correlators.Corr.second_deriv": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 3}}}}}}, "a": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 2, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.obs.CObs.gamma_method": {"tf": 1}, "pyerrors.obs.CObs.is_zero": {"tf": 1}}, "df": 5}}}}}}}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}}, "df": 1}}}}}}, "d": {"0": {"docs": {"pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 3}, "docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.4142135623730951}}, "df": 7, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "f": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}, "i": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}}, "df": 2}, "s": {"docs": {"pyerrors": {"tf": 1.7320508075688772}}, "df": 1}, "r": {"docs": {"pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.obs.cov_Obs": {"tf": 1}}, "df": 3}}}}, "c": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 1}, "pyerrors.obs.correlate": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 4}}}, "t": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}}}}, "l": {"docs": {"pyerrors": {"tf": 2}, "pyerrors.correlators.Corr.reweight": {"tf": 1.7320508075688772}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.utils.check_idl": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.__init__": {"tf": 1}, "pyerrors.obs.Obs.reweight": {"tf": 1.7320508075688772}, "pyerrors.obs.reweight": {"tf": 1.7320508075688772}}, "df": 12, "s": {"docs": {"pyerrors.input.utils.check_idl": {"tf": 1}}, "df": 1}}}, "a": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "z": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "o": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}}}}}}, "r": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.obs.Obs.plot_history": {"tf": 1}}, "df": 2}}}}}, "l": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 1}}}}}}}}}, "/": {"docs": {}, "df": 0, "o": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "l": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}}}}}}}}, "g": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}}, "df": 3}}}}}}, "o": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}}, "df": 1}}}}}}, "e": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 1}}}, "\\": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "a": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}}}, "d": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "a": {"docs": {"pyerrors.obs.covariance": {"tf": 1.4142135623730951}}, "df": 1}}}}}}, "j": {"docs": {"pyerrors.obs.covariance": {"tf": 1.4142135623730951}}, "df": 1}, "^": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "\\": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "a": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}}}}, "|": {"docs": {}, "df": 0, "^": {"2": {"docs": {}, "df": 0, "\\": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "q": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}}}, "docs": {}, "df": 0}}}}, "}": {"docs": {}, "df": 0, "|": {"docs": {}, "df": 0, "v": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}, "\\": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}}}}}}, "p": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.fits.Fit_result": {"tf": 2}, "pyerrors.fits.ks_test": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1.4142135623730951}}, "df": 4, "y": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 6.6332495807108}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.input": {"tf": 2}, "pyerrors.input.pandas.dump_df": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1}}, "df": 5}}}}}}, "t": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.correlators.Corr.fit": {"tf": 1}}, "df": 2}}}}}, "a": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "k": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}}}}}, "r": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.obs.CObs.gamma_method": {"tf": 1}, "pyerrors.obs.CObs.is_zero": {"tf": 1}}, "df": 5, "l": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "n": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.correlate": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.T_symmetry": {"tf": 1.7320508075688772}}, "df": 3}}}, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.correlators.Corr.T_symmetry": {"tf": 1}}, "df": 1}}}}, "a": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors": {"tf": 2.6457513110645907}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}}, "df": 8, "s": {"docs": {"pyerrors": {"tf": 3}, "pyerrors.correlators.Corr.__init__": {"tf": 1}, "pyerrors.correlators.Corr.item": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.Hankel": {"tf": 1}, "pyerrors.correlators.Corr.roll": {"tf": 1}, "pyerrors.correlators.Corr.thin": {"tf": 1}, "pyerrors.correlators.Corr.correlate": {"tf": 1}, "pyerrors.correlators.Corr.reweight": {"tf": 1}, "pyerrors.correlators.Corr.T_symmetry": {"tf": 1}, "pyerrors.correlators.Corr.deriv": {"tf": 1}, "pyerrors.correlators.Corr.second_deriv": {"tf": 1}, "pyerrors.correlators.Corr.m_eff": {"tf": 1}, "pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.correlators.Corr.plateau": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.show": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.spaghetti_plot": {"tf": 1}, "pyerrors.correlators.Corr.dump": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1}, "pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.fits.Fit_result": {"tf": 1.4142135623730951}, "pyerrors.fits.Fit_result.gamma_method": {"tf": 1}, "pyerrors.fits.Fit_result.gm": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 2}, "pyerrors.fits.total_least_squares": {"tf": 2.23606797749979}, "pyerrors.fits.fit_lin": {"tf": 1.4142135623730951}, "pyerrors.fits.error_band": {"tf": 1}, "pyerrors.fits.ks_test": {"tf": 1}, "pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}, "pyerrors.input.misc.read_pbp": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.qtop_projection": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}, "pyerrors.input.pandas.to_sql": {"tf": 1}, "pyerrors.input.pandas.read_sql": {"tf": 1.4142135623730951}, "pyerrors.input.pandas.dump_df": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1.4142135623730951}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.input.utils.check_idl": {"tf": 1}, "pyerrors.linalg.matmul": {"tf": 1}, "pyerrors.linalg.jack_matmul": {"tf": 1}, "pyerrors.linalg.einsum": {"tf": 1}, "pyerrors.misc.dump_object": {"tf": 1}, "pyerrors.misc.load_object": {"tf": 1}, "pyerrors.misc.pseudo_Obs": {"tf": 1}, "pyerrors.misc.gen_correlated_data": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}, "pyerrors.obs.Obs.__init__": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}, "pyerrors.obs.Obs.details": {"tf": 1}, "pyerrors.obs.Obs.reweight": {"tf": 1}, "pyerrors.obs.Obs.is_zero_within_error": {"tf": 1}, "pyerrors.obs.Obs.is_zero": {"tf": 1}, "pyerrors.obs.Obs.plot_tauint": {"tf": 1}, "pyerrors.obs.Obs.plot_rho": {"tf": 1}, "pyerrors.obs.Obs.plot_history": {"tf": 1}, "pyerrors.obs.Obs.plot_piechart": {"tf": 1}, "pyerrors.obs.Obs.dump": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}, "pyerrors.obs.reweight": {"tf": 1}, "pyerrors.obs.correlate": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}, "pyerrors.obs.import_jackknife": {"tf": 1}, "pyerrors.obs.merge_obs": {"tf": 1}, "pyerrors.obs.cov_Obs": {"tf": 1}, "pyerrors.roots.find_root": {"tf": 1}}, "df": 94}}}}}}}, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.T_symmetry": {"tf": 1}}, "df": 2}}}, "s": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}}, "df": 1, "r": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input": {"tf": 1}}, "df": 1}}}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}}, "df": 1}}}}}, "s": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.roots.find_root": {"tf": 1}}, "df": 4}}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}}}}}, "t": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}, "h": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.correlators.Corr.dump": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.read_ADerrors": {"tf": 2}, "pyerrors.input.bdio.write_ADerrors": {"tf": 2}, "pyerrors.input.bdio.read_mesons": {"tf": 2}, "pyerrors.input.bdio.read_dSdm": {"tf": 2}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_rwms": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.qtop_projection": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}, "pyerrors.input.pandas.to_sql": {"tf": 1}, "pyerrors.input.pandas.read_sql": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.4142135623730951}, "pyerrors.misc.dump_object": {"tf": 1.4142135623730951}, "pyerrors.misc.load_object": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.dump": {"tf": 1.4142135623730951}}, "df": 24}}, "d": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.__init__": {"tf": 1.7320508075688772}}, "df": 2}}}}}, "n": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.pandas.to_sql": {"tf": 1}, "pyerrors.input.pandas.read_sql": {"tf": 1}, "pyerrors.input.pandas.dump_df": {"tf": 1.7320508075688772}, "pyerrors.input.pandas.load_df": {"tf": 1.4142135623730951}}, "df": 4}}}}}, "r": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 3.872983346207417}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}}, "df": 3}}}, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1.7320508075688772}}, "df": 1}}, "o": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.input.hadrons.Npr_matrix.g5H": {"tf": 1.4142135623730951}}, "df": 1}}}}}}, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.is_zero_within_error": {"tf": 1}}, "df": 2}}, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}, "pyerrors.obs.Obs.details": {"tf": 1}}, "df": 4}}}}}}}, "v": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.write_dobs": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 10, "d": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "s": {"docs": {"pyerrors": {"tf": 1.7320508075688772}}, "df": 1}}}}}, "c": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.prune": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 4}}}}}}, "b": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "m": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}}, "df": 2}}}}, "j": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 2, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.correlators.Corr.Eigenvalue": {"tf": 1}}, "df": 1}}, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.qtop_projection": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 3}}}, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 1}}}}}}, "g": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "m": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1, "s": {"docs": {"pyerrors.input": {"tf": 1}}, "df": 1}}}}}, "d": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.correlators.Corr.spaghetti_plot": {"tf": 1}}, "df": 1}}}}}}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 3.4641016151377544}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.details": {"tf": 1}}, "df": 6, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.fit": {"tf": 1}}, "df": 1}}, "s": {"docs": {"pyerrors.fits.total_least_squares": {"tf": 1}}, "df": 1}}}, "m": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.obs.correlate": {"tf": 1}}, "df": 2}}}}, "o": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}}, "df": 1, "s": {"docs": {"pyerrors.fits.least_squares": {"tf": 1.4142135623730951}}, "df": 1}}}}, "e": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.obs.CObs.is_zero": {"tf": 1}}, "df": 2}}}}}, "o": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}}, "df": 1}}}}}}}}}}}}, "f": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}, "i": {"docs": {}, "df": 0, "x": {"docs": {"pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 2}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.4142135623730951}}, "df": 9}}}, "v": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 2}}}}}, "s": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 1}}}}}, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.__init__": {"tf": 1}, "pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.correlators.Corr.set_prange": {"tf": 1}}, "df": 3}}}, "c": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}}, "df": 1}, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}}, "df": 4}}}}}}}}, "u": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1}}, "df": 1, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}, "l": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}, "s": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}}}}}}}, "r": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.misc.pseudo_Obs": {"tf": 1}}, "df": 1}}}}}}}, "l": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}, "o": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 2}, "pyerrors.correlators.Corr.spaghetti_plot": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.qqplot": {"tf": 1}, "pyerrors.fits.residual_plot": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.obs.Obs.plot_tauint": {"tf": 1}, "pyerrors.obs.Obs.plot_rho": {"tf": 1}, "pyerrors.obs.Obs.plot_rep_dist": {"tf": 1}, "pyerrors.obs.Obs.plot_history": {"tf": 1}, "pyerrors.obs.Obs.plot_piechart": {"tf": 1}}, "df": 11, "a": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.plottable": {"tf": 1}}, "df": 1}}}}, "s": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 2}, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1}}, "df": 1}}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1}}, "df": 1}}}}}}, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "u": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.__init__": {"tf": 1}, "pyerrors.correlators.Corr.plateau": {"tf": 2.23606797749979}, "pyerrors.correlators.Corr.show": {"tf": 1.7320508075688772}}, "df": 4}}}}, "c": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}}, "df": 2}}}}}}}}, "q": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1.4142135623730951}}, "df": 1}}}}}}}}, "h": {"docs": {}, "df": 0, "y": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 2}}, "df": 1, "i": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}}, "e": {"docs": {"pyerrors": {"tf": 5.477225575051661}, "pyerrors.correlators.Corr": {"tf": 1}}, "df": 2, "r": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.item": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.4142135623730951}}, "df": 5, "f": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "m": {"docs": {"pyerrors.fits.least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.total_least_squares": {"tf": 1}}, "df": 2, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 2}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1}}, "df": 5}}, "s": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.fits.fit_lin": {"tf": 1}, "pyerrors.fits.ks_test": {"tf": 1}}, "df": 4}}}}}, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "c": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.Hankel": {"tf": 1}, "pyerrors.correlators.Corr.m_eff": {"tf": 1}}, "df": 3, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.Hankel": {"tf": 1}, "pyerrors.correlators.Corr.roll": {"tf": 1}}, "df": 3}}}}, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.correlators.Corr.m_eff": {"tf": 1.4142135623730951}}, "df": 1}}}}}}}}}, "m": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}}, "df": 4}}}}}}}}, "s": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}}, "df": 2}}}}, "n": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.linalg.pinv": {"tf": 1}}, "df": 2}}}}, "c": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.mpm.matrix_pencil_method": {"tf": 1.7320508075688772}}, "df": 1}}}}}, "i": {"docs": {"pyerrors": {"tf": 1}}, "df": 1, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "^": {"0": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}, "docs": {}, "df": 0}, "n": {"docs": {}, "df": 0, "v": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "c": {"docs": {}, "df": 0, "k": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.correlators.Corr.item": {"tf": 1}}, "df": 1}, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.item": {"tf": 1.4142135623730951}}, "df": 1}}, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.dump": {"tf": 1}, "pyerrors.misc.dump_object": {"tf": 1.4142135623730951}, "pyerrors.misc.load_object": {"tf": 1}, "pyerrors.obs.Obs.dump": {"tf": 1}}, "df": 4}}}}, "e": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.obs.Obs.plot_piechart": {"tf": 1}}, "df": 1}}}}}}}, "o": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 5, "s": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 2}}}}, "s": {"docs": {"pyerrors.covobs.Covobs.__init__": {"tf": 1}}, "df": 1, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}, "f": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "x": {"docs": {"pyerrors.input.openQCD.read_rwms": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1.4142135623730951}}, "df": 3}}}}, "s": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 2}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.obs.merge_obs": {"tf": 1}}, "df": 5}, "y": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 3}}}}}, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}}, "df": 5}}, "v": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1.4142135623730951}}, "df": 3}}}}}}, "w": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}}, "df": 1}}}}, "b": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.dobs.create_pobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.write_pobs": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.read_pobs": {"tf": 1}}, "df": 3}}}, "s": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.linalg.pinv": {"tf": 1}}, "df": 2}}}}}}}, "s": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.input.hadrons.read_meson_hd5": {"tf": 2}}, "df": 1}}}}}}}}}}}, "b": {"docs": {}, "df": 0, "p": {"docs": {"pyerrors.input.misc.read_pbp": {"tf": 1}}, "df": 1}}}, "a": {"docs": {"pyerrors": {"tf": 8.366600265340756}, "pyerrors.correlators.Corr": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.__init__": {"tf": 1}, "pyerrors.correlators.Corr.projected": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.item": {"tf": 1}, "pyerrors.correlators.Corr.plottable": {"tf": 1}, "pyerrors.correlators.Corr.is_matrix_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.thin": {"tf": 1}, "pyerrors.correlators.Corr.correlate": {"tf": 1}, "pyerrors.correlators.Corr.reweight": {"tf": 1}, "pyerrors.correlators.Corr.plateau": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.spaghetti_plot": {"tf": 1}, "pyerrors.correlators.Corr.dump": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.prune": {"tf": 2}, "pyerrors.fits.least_squares": {"tf": 3.605551275463989}, "pyerrors.fits.total_least_squares": {"tf": 3.3166247903554}, "pyerrors.fits.fit_lin": {"tf": 1.7320508075688772}, "pyerrors.fits.qqplot": {"tf": 1}, "pyerrors.fits.residual_plot": {"tf": 1}, "pyerrors.fits.ks_test": {"tf": 1}, "pyerrors.input.bdio.read_ADerrors": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.read_mesons": {"tf": 2}, "pyerrors.input.bdio.read_dSdm": {"tf": 2}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.write_pobs": {"tf": 2.23606797749979}, "pyerrors.input.dobs.read_pobs": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.import_dobs_string": {"tf": 2}, "pyerrors.input.dobs.read_dobs": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.create_dobs_string": {"tf": 2}, "pyerrors.input.dobs.write_dobs": {"tf": 2.449489742783178}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 3}, "pyerrors.input.json.create_json_string": {"tf": 1.7320508075688772}, "pyerrors.input.json.dump_to_json": {"tf": 2.23606797749979}, "pyerrors.input.json.import_json_string": {"tf": 1.7320508075688772}, "pyerrors.input.json.load_json": {"tf": 1.7320508075688772}, "pyerrors.input.json.dump_dict_to_json": {"tf": 2.23606797749979}, "pyerrors.input.json.load_json_dict": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_rwms": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 2.23606797749979}, "pyerrors.input.pandas.dump_df": {"tf": 2.23606797749979}, "pyerrors.input.pandas.load_df": {"tf": 1.4142135623730951}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.7320508075688772}, "pyerrors.linalg.eigh": {"tf": 1}, "pyerrors.linalg.eig": {"tf": 1}, "pyerrors.linalg.pinv": {"tf": 1}, "pyerrors.linalg.svd": {"tf": 1}, "pyerrors.misc.dump_object": {"tf": 1}, "pyerrors.misc.gen_correlated_data": {"tf": 1.4142135623730951}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs": {"tf": 2.449489742783178}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}, "pyerrors.obs.Obs.reweight": {"tf": 1}, "pyerrors.obs.Obs.is_zero": {"tf": 1}, "pyerrors.obs.Obs.plot_tauint": {"tf": 1}, "pyerrors.obs.Obs.plot_rho": {"tf": 1}, "pyerrors.obs.Obs.plot_piechart": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.dump": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1.4142135623730951}, "pyerrors.obs.CObs": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1.4142135623730951}, "pyerrors.obs.reweight": {"tf": 1.4142135623730951}, "pyerrors.obs.correlate": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 2.23606797749979}, "pyerrors.obs.cov_Obs": {"tf": 1}}, "df": 71, "n": {"docs": {"pyerrors": {"tf": 5.0990195135927845}, "pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.correlators.Corr.Hankel": {"tf": 1}, "pyerrors.correlators.Corr.correlate": {"tf": 1}, "pyerrors.correlators.Corr.reweight": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 2.23606797749979}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.fits.error_band": {"tf": 1.4142135623730951}, "pyerrors.input": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 3.605551275463989}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1.4142135623730951}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.input.utils.check_idl": {"tf": 1}, "pyerrors.misc.pseudo_Obs": {"tf": 1}, "pyerrors.misc.gen_correlated_data": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 2}, "pyerrors.obs.Obs.reweight": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}, "pyerrors.obs.reweight": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1.7320508075688772}, "pyerrors.obs.import_jackknife": {"tf": 1}, "pyerrors.obs.cov_Obs": {"tf": 1}, "pyerrors.roots.find_root": {"tf": 1}}, "df": 31, "d": {"docs": {"pyerrors": {"tf": 7}, "pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.correlators.Corr.__init__": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.item": {"tf": 1}, "pyerrors.correlators.Corr.plottable": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 1}, "pyerrors.correlators.Corr.reweight": {"tf": 1}, "pyerrors.correlators.Corr.T_symmetry": {"tf": 1}, "pyerrors.correlators.Corr.fit": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.correlators.Corr.dump": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1.7320508075688772}, "pyerrors.fits.least_squares": {"tf": 2.6457513110645907}, "pyerrors.fits.total_least_squares": {"tf": 1.7320508075688772}, "pyerrors.fits.fit_lin": {"tf": 1}, "pyerrors.fits.residual_plot": {"tf": 1}, "pyerrors.input": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.read_dSdm": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.read_pobs": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.read_dobs": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.7320508075688772}, "pyerrors.input.hadrons.Npr_matrix.g5H": {"tf": 1}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1.4142135623730951}, "pyerrors.input.json.load_json_dict": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.extract_t0": {"tf": 2}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1.7320508075688772}, "pyerrors.input.pandas.read_sql": {"tf": 1}, "pyerrors.linalg.eigh": {"tf": 1}, "pyerrors.misc.pseudo_Obs": {"tf": 1}, "pyerrors.misc.gen_correlated_data": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}, "pyerrors.obs.Obs.details": {"tf": 1}, "pyerrors.obs.Obs.reweight": {"tf": 1}, "pyerrors.obs.Obs.plot_piechart": {"tf": 1}, "pyerrors.obs.Obs.dump": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1.4142135623730951}, "pyerrors.obs.CObs.gamma_method": {"tf": 1}, "pyerrors.obs.CObs.is_zero": {"tf": 1}, "pyerrors.obs.reweight": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1.7320508075688772}, "pyerrors.obs.import_jackknife": {"tf": 1.4142135623730951}, "pyerrors.obs.cov_Obs": {"tf": 1}}, "df": 60}, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "y": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 2.8284271247461903}, "pyerrors.fits.ks_test": {"tf": 1}, "pyerrors.input": {"tf": 1}, "pyerrors.input.pandas.read_sql": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1}}, "df": 10}}}, "z": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 1}}}}}}, "y": {"docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.roots.find_root": {"tf": 1}}, "df": 3}, "t": {"docs": {}, "df": 0, "i": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.anti_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.m_eff": {"tf": 1}}, "df": 3}}, "o": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.correlate": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1.4142135623730951}}, "df": 4}}}}, "n": {"docs": {}, "df": 0, "y": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.obs.derived_observable": {"tf": 1}}, "df": 1}}}}}}}, "p": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.fits.least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.total_least_squares": {"tf": 1.4142135623730951}, "pyerrors.obs.derived_observable": {"tf": 1}, "pyerrors.roots.find_root": {"tf": 1.4142135623730951}}, "df": 5}}, "r": {"docs": {}, "df": 0, "x": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, ":": {"1": {"0": {"0": {"9": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "2": {"0": {"5": {"docs": {"pyerrors.correlators.Corr.m_eff": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "8": {"0": {"9": {"docs": {"pyerrors": {"tf": 2}}, "df": 1}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "2": {"0": {"0": {"4": {"docs": {"pyerrors.correlators.Corr.GEVP": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "2": {"0": {"9": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "p": {"docs": {"pyerrors": {"tf": 2}}, "df": 1}}}}}}}, "e": {"docs": {"pyerrors": {"tf": 5.477225575051661}, "pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 1}, "pyerrors.correlators.Corr.reweight": {"tf": 1}, "pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.prune": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.fits.fit_lin": {"tf": 1.4142135623730951}, "pyerrors.fits.qqplot": {"tf": 1}, "pyerrors.input": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.read_dobs": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.7320508075688772}, "pyerrors.input.hadrons.Npr_matrix.g5H": {"tf": 1}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1.4142135623730951}, "pyerrors.input.pandas.to_sql": {"tf": 1}, "pyerrors.input.pandas.dump_df": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.input.utils.check_idl": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.__init__": {"tf": 1}, "pyerrors.obs.Obs.reweight": {"tf": 1}, "pyerrors.obs.CObs.is_zero": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}, "pyerrors.obs.reweight": {"tf": 1}, "pyerrors.obs.correlate": {"tf": 1}, "pyerrors.obs.import_jackknife": {"tf": 1}, "pyerrors.obs.merge_obs": {"tf": 1}}, "df": 51}, "g": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}}, "df": 3, "s": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1.4142135623730951}}, "df": 2}}}}}}}, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 4.47213595499958}, "pyerrors.correlators.Corr.__init__": {"tf": 1.4142135623730951}, "pyerrors.covobs.Covobs.__init__": {"tf": 1.4142135623730951}, "pyerrors.fits.error_band": {"tf": 1.7320508075688772}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 6.082762530298219}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1.4142135623730951}, "pyerrors.obs.derived_observable": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}, "pyerrors.obs.import_jackknife": {"tf": 1}, "pyerrors.obs.cov_Obs": {"tf": 1.4142135623730951}}, "df": 13, "s": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.correlators.Corr.__init__": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.linalg.matmul": {"tf": 1}, "pyerrors.linalg.jack_matmul": {"tf": 1}, "pyerrors.linalg.einsum": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1}, "pyerrors.obs.Obs.__init__": {"tf": 1}}, "df": 9}}, "n": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}, "i": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}, "o": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.symmetric": {"tf": 1}, "pyerrors.correlators.Corr.anti_symmetric": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 3}}}}, "c": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "h": {"docs": {"pyerrors.correlators.Corr.m_eff": {"tf": 1}}, "df": 1}}}}}, "b": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.linalg.matmul": {"tf": 1}, "pyerrors.linalg.jack_matmul": {"tf": 1}, "pyerrors.linalg.einsum": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}}, "df": 4}}}}}}}}, "u": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "o": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.input.pandas.read_sql": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1}}, "df": 5, "m": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "c": {"docs": {"pyerrors": {"tf": 3.1622776601683795}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1.4142135623730951}}, "df": 6, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 9}}}}}}}}}, "g": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.fits.least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.total_least_squares": {"tf": 1.4142135623730951}, "pyerrors.obs.derived_observable": {"tf": 1.7320508075688772}, "pyerrors.roots.find_root": {"tf": 1.4142135623730951}}, "df": 5}}}}, "c": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 2.8284271247461903}, "pyerrors.misc.gen_correlated_data": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.gm": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.plot_tauint": {"tf": 1}, "pyerrors.obs.Obs.plot_rho": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 2.23606797749979}}, "df": 7, "s": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}}}}}}}}}}}}, "x": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}}, "df": 7}}}}}}}}, "s": {"docs": {"pyerrors": {"tf": 6}, "pyerrors.correlators.Corr.m_eff": {"tf": 1}, "pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1.7320508075688772}, "pyerrors.fits.least_squares": {"tf": 1.7320508075688772}, "pyerrors.fits.total_least_squares": {"tf": 1.7320508075688772}, "pyerrors.fits.fit_lin": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.read_dobs": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.7320508075688772}, "pyerrors.input.pandas.load_df": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}, "pyerrors.obs.import_jackknife": {"tf": 1.4142135623730951}, "pyerrors.roots.find_root": {"tf": 1}}, "df": 19, "s": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 1, "d": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.fits.fit_lin": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 2.23606797749979}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}}, "df": 9}, "s": {"docs": {"pyerrors.input.dobs.read_dobs": {"tf": 1.4142135623730951}, "pyerrors.input.json.load_json": {"tf": 1.4142135623730951}, "pyerrors.input.json.load_json_dict": {"tf": 1.4142135623730951}, "pyerrors.input.pandas.load_df": {"tf": 1.4142135623730951}}, "df": 4}}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1.4142135623730951}}, "df": 4}}}}}, "o": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}}}}}}, "i": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1, "s": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}}}}, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 2}}}}}}}, "l": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "o": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}, "p": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "a": {"docs": {"pyerrors": {"tf": 2.23606797749979}}, "df": 1, "^": {"docs": {}, "df": 0, "i": {"docs": {"pyerrors": {"tf": 1.7320508075688772}}, "df": 1}}}}}, "l": {"docs": {"pyerrors": {"tf": 2.449489742783178}, "pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 1}, "pyerrors.correlators.Corr.correlate": {"tf": 1}, "pyerrors.correlators.Corr.reweight": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.fits.Fit_result.gamma_method": {"tf": 1}, "pyerrors.fits.Fit_result.gm": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.total_least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.ks_test": {"tf": 1}, "pyerrors.input.bdio.read_ADerrors": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.read_mesons": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.read_dSdm": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.pandas.read_sql": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.4142135623730951}, "pyerrors.linalg.matmul": {"tf": 1}, "pyerrors.obs.Obs.reweight": {"tf": 1.7320508075688772}, "pyerrors.obs.derived_observable": {"tf": 1}, "pyerrors.obs.reweight": {"tf": 1.7320508075688772}, "pyerrors.obs.covariance": {"tf": 1}, "pyerrors.obs.merge_obs": {"tf": 1}}, "df": 34, "o": {"docs": {}, "df": 0, "w": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 1, "s": {"docs": {"pyerrors": {"tf": 2}, "pyerrors.correlators.Corr.prune": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 3}}, "c": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}}}}}}, "s": {"docs": {}, "df": 0, "o": {"docs": {"pyerrors": {"tf": 2.449489742783178}, "pyerrors.fits.Fit_result": {"tf": 1}, "pyerrors.input": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 2}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1}}, "df": 7}}, "g": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "m": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}}, "df": 3, "s": {"docs": {"pyerrors": {"tf": 1.7320508075688772}}, "df": 1}}}}}}}}, "w": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "y": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 9, "s": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}}, "df": 1}}}}}}}}}}, "i": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}}, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.input.pandas.to_sql": {"tf": 1}}, "df": 1}}}}}}, "p": {"docs": {"pyerrors": {"tf": 1}}, "df": 1, "p": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.input.pandas.read_sql": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1}}, "df": 3}, "d": {"docs": {"pyerrors.obs.correlate": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1.4142135623730951}}, "df": 2}}}, "y": {"docs": {"pyerrors.correlators.Corr.gamma_method": {"tf": 1}, "pyerrors.correlators.Corr.gm": {"tf": 1}, "pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.fits.Fit_result.gamma_method": {"tf": 1}, "pyerrors.fits.Fit_result.gm": {"tf": 1}}, "df": 6}}, "r": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 4}}}}}}, "x": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.linalg.jack_matmul": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 2}}}}}}}}}}, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.pandas.to_sql": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.7320508075688772}}, "df": 2}}}}, "i": {"docs": {"pyerrors": {"tf": 2.23606797749979}}, "df": 1}}, "v": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.correlators.Corr.deriv": {"tf": 1}, "pyerrors.correlators.Corr.second_deriv": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1.4142135623730951}}, "df": 4}}}}}}}, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.reweight": {"tf": 1}, "pyerrors.correlators.Corr.T_symmetry": {"tf": 1}, "pyerrors.correlators.Corr.plateau": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.reweight": {"tf": 1}, "pyerrors.obs.reweight": {"tf": 1}}, "df": 5, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}, "g": {"docs": {"pyerrors.correlators.Corr.plateau": {"tf": 1}}, "df": 1}, "o": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}}, "f": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.obs.correlate": {"tf": 1}}, "df": 5}}}, "f": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}, "t": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.__init__": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1.4142135623730951}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 18, "t": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}}, "df": 2}}}}}, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.set_prange": {"tf": 1}}, "df": 1, "s": {"docs": {"pyerrors.fits.Fit_result": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs": {"tf": 1}}, "df": 3}}}}}}}}, "o": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.obs.Obs.is_zero": {"tf": 1}}, "df": 1}}}, "m": {"docs": {}, "df": 0, "p": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "c": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}, "r": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.correlators.Corr.GEVP": {"tf": 1.7320508075688772}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.4142135623730951}, "pyerrors.linalg.eigh": {"tf": 1}, "pyerrors.linalg.eig": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}}, "df": 11}}}}}}, "e": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}}, "df": 4}}, "i": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.fits.Fit_result": {"tf": 1}}, "df": 1}}}}}}}}, "h": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}}, "df": 6}}}}}}, "t": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1, "l": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}}}}}, "o": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 1}}}}}, "b": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1, "o": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.obs.Obs.is_zero": {"tf": 1}}, "df": 2}}}}}, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}}}, "o": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 2}}, "u": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.obs.Obs.details": {"tf": 1}}, "df": 3}}}}, "i": {"docs": {}, "df": 0, "m": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "d": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}, "d": {"docs": {"pyerrors.correlators.Corr": {"tf": 1}}, "df": 1, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1}}, "df": 7}}}}}}, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}}, "df": 4}}}, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 1}}}, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}}, "df": 3}}}}}}}, "x": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.spaghetti_plot": {"tf": 1}}, "df": 2}}}, "g": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1.4142135623730951}, "pyerrors.input.utils.check_idl": {"tf": 1}}, "df": 2}}}}}, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.obs.Obs.export_jackknife": {"tf": 1}}, "df": 1, "s": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}}}, "^": {"2": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "\\": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "a": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}}}}, "/": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "^": {"2": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}}}}}, "a": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}, "f": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}}, "df": 2, "o": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors": {"tf": 6.855654600401044}, "pyerrors.correlators.Corr": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.__init__": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 2.23606797749979}, "pyerrors.correlators.Corr.reweight": {"tf": 1}, "pyerrors.correlators.Corr.m_eff": {"tf": 2}, "pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.dump": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1}, "pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.fits.Fit_result": {"tf": 1.4142135623730951}, "pyerrors.fits.least_squares": {"tf": 3}, "pyerrors.fits.total_least_squares": {"tf": 2}, "pyerrors.fits.fit_lin": {"tf": 1.4142135623730951}, "pyerrors.fits.error_band": {"tf": 1.7320508075688772}, "pyerrors.fits.ks_test": {"tf": 1}, "pyerrors.input": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 2.6457513110645907}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1.4142135623730951}, "pyerrors.input.misc.read_pbp": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_rwms": {"tf": 2}, "pyerrors.input.openQCD.extract_t0": {"tf": 2.23606797749979}, "pyerrors.input.openQCD.read_qtop": {"tf": 2.449489742783178}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 2.6457513110645907}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1.7320508075688772}, "pyerrors.input.pandas.read_sql": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 3}, "pyerrors.input.utils": {"tf": 1}, "pyerrors.linalg.jack_matmul": {"tf": 1}, "pyerrors.linalg.einsum": {"tf": 1.7320508075688772}, "pyerrors.misc.dump_object": {"tf": 1}, "pyerrors.misc.pseudo_Obs": {"tf": 1.7320508075688772}, "pyerrors.misc.gen_correlated_data": {"tf": 2}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 3.872983346207417}, "pyerrors.obs.Obs.gamma_method": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.gm": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.reweight": {"tf": 1}, "pyerrors.obs.Obs.is_zero_within_error": {"tf": 1}, "pyerrors.obs.Obs.is_zero": {"tf": 1}, "pyerrors.obs.Obs.plot_tauint": {"tf": 1}, "pyerrors.obs.Obs.plot_rho": {"tf": 1}, "pyerrors.obs.Obs.plot_rep_dist": {"tf": 1}, "pyerrors.obs.Obs.plot_history": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.dump": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1.4142135623730951}, "pyerrors.obs.CObs": {"tf": 1}, "pyerrors.obs.CObs.gamma_method": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1.7320508075688772}, "pyerrors.obs.reweight": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 2.6457513110645907}, "pyerrors.obs.cov_Obs": {"tf": 1}, "pyerrors.roots.find_root": {"tf": 1}}, "df": 63, "t": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 2}}}}, "m": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.m_eff": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.total_least_squares": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}}, "df": 7, "u": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "a": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1, "e": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}, "a": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 3.605551275463989}, "pyerrors.correlators.Corr.plottable": {"tf": 1}, "pyerrors.correlators.Corr.dump": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}, "pyerrors.input.misc.read_pbp": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.pandas.dump_df": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.dump": {"tf": 1.7320508075688772}}, "df": 16, "s": {"docs": {"pyerrors.correlators.Corr.dump": {"tf": 1}, "pyerrors.obs.Obs.dump": {"tf": 1}}, "df": 2}, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}}, "df": 2}}}}}}, "w": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.deriv": {"tf": 1}}, "df": 1}}}}}, "u": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}}, "df": 2}}, "r": {"docs": {}, "df": 0, "q": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "k": {"docs": {"pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}}, "df": 1, "f": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "y": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}}, "df": 1}}}}}}}}}}}}}}}}}}}}}, "l": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "w": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors": {"tf": 3}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}}, "df": 5}}}}}}, "d": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.input.misc.read_pbp": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 3, "s": {"docs": {"pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 1}}}}}}, "e": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}}, "r": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "m": {"docs": {"pyerrors": {"tf": 4.358898943540674}, "pyerrors.correlators.Corr.item": {"tf": 1}, "pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.input": {"tf": 1}, "pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.7320508075688772}, "pyerrors.input.json.import_json_string": {"tf": 1.7320508075688772}, "pyerrors.input.json.load_json": {"tf": 1.7320508075688772}, "pyerrors.input.json.load_json_dict": {"tf": 1}, "pyerrors.input.misc.read_pbp": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1.7320508075688772}, "pyerrors.input.pandas.load_df": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.4142135623730951}, "pyerrors.misc.load_object": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1.7320508075688772}}, "df": 34}, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.__init__": {"tf": 1}}, "df": 1}}}, "a": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "w": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "k": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}, "n": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "o": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}, "c": {"docs": {}, "df": 0, "{": {"1": {"docs": {}, "df": 0, "}": {"docs": {}, "df": 0, "{": {"2": {"docs": {"pyerrors": {"tf": 1}}, "df": 1, "}": {"docs": {}, "df": 0, "+": {"docs": {}, "df": 0, "\\": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "m": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}}, "docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "}": {"docs": {}, "df": 0, "\\": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "m": {"docs": {"pyerrors.obs.covariance": {"tf": 1.4142135623730951}}, "df": 1}}}}}}}}}, "docs": {}, "df": 0}, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.obs.Obs.plot_piechart": {"tf": 1}}, "df": 1}}, "s": {"docs": {"pyerrors.obs.Obs.plot_piechart": {"tf": 1}}, "df": 1}}}}}}}}, "i": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 3.7416573867739413}, "pyerrors.correlators.Corr.fit": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1.7320508075688772}, "pyerrors.fits.Fit_result": {"tf": 2}, "pyerrors.fits.Fit_result.gamma_method": {"tf": 1}, "pyerrors.fits.Fit_result.gm": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 3.1622776601683795}, "pyerrors.fits.total_least_squares": {"tf": 2}, "pyerrors.fits.fit_lin": {"tf": 2}, "pyerrors.fits.qqplot": {"tf": 1.4142135623730951}, "pyerrors.fits.residual_plot": {"tf": 1}, "pyerrors.fits.error_band": {"tf": 1}, "pyerrors.fits.ks_test": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.extract_t0": {"tf": 2}}, "df": 15, "s": {"docs": {"pyerrors": {"tf": 3.7416573867739413}, "pyerrors.correlators.Corr.fit": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.fits.Fit_result": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.total_least_squares": {"tf": 1}}, "df": 6}, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.fits.fit_lin": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 5}}}, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.fit": {"tf": 1.4142135623730951}}, "df": 2}}}}}}, "r": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.correlators.Corr.__init__": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.item": {"tf": 1}, "pyerrors.correlators.Corr.deriv": {"tf": 1}, "pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}, "pyerrors.input.misc.read_pbp": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 2.23606797749979}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.obs.correlate": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}, "pyerrors.obs.import_jackknife": {"tf": 1}}, "df": 18}}}, "x": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 7}}}, "n": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.GEVP": {"tf": 1}}, "df": 1, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}}}, "e": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.correlators.Corr.m_eff": {"tf": 1}}, "df": 1}}, "s": {"docs": {"pyerrors.roots.find_root": {"tf": 1}}, "df": 1}}, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.deriv": {"tf": 1}, "pyerrors.correlators.Corr.second_deriv": {"tf": 1}}, "df": 2}}}}, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 4.358898943540674}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.correlators.Corr.dump": {"tf": 2}, "pyerrors.input.bdio.read_ADerrors": {"tf": 1.7320508075688772}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1.7320508075688772}, "pyerrors.input.bdio.read_mesons": {"tf": 1.7320508075688772}, "pyerrors.input.bdio.read_dSdm": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 2}, "pyerrors.input.dobs.read_pobs": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 2}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.write_dobs": {"tf": 2}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 2}, "pyerrors.input.json.dump_to_json": {"tf": 2.23606797749979}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 2}, "pyerrors.input.json.dump_dict_to_json": {"tf": 2.23606797749979}, "pyerrors.input.json.load_json_dict": {"tf": 2}, "pyerrors.input.openQCD.read_rwms": {"tf": 2}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}, "pyerrors.input.pandas.dump_df": {"tf": 2.23606797749979}, "pyerrors.input.pandas.load_df": {"tf": 1.7320508075688772}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.4142135623730951}, "pyerrors.misc.dump_object": {"tf": 2}, "pyerrors.misc.load_object": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.plot_tauint": {"tf": 1}, "pyerrors.obs.Obs.plot_rho": {"tf": 1}, "pyerrors.obs.Obs.plot_piechart": {"tf": 1}, "pyerrors.obs.Obs.dump": {"tf": 2.23606797749979}}, "df": 39, "s": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_rwms": {"tf": 2.6457513110645907}, "pyerrors.input.openQCD.extract_t0": {"tf": 2.23606797749979}, "pyerrors.input.openQCD.read_qtop": {"tf": 2.449489742783178}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 2.449489742783178}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 2}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 2.6457513110645907}, "pyerrors.input.sfcf.read_sfcf": {"tf": 2}}, "df": 13, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "m": {"docs": {"pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}}, "df": 4}}}}, "n": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.dump": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}, "pyerrors.input.pandas.dump_df": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1}, "pyerrors.obs.Obs.dump": {"tf": 1}}, "df": 12, "s": {"docs": {"pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 2}}}}, "o": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "f": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}}, "df": 1}}}}}}}}}}}}}}, "l": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 1}}}}}, "e": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 2}}, "df": 1}}}, "g": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.plot_tauint": {"tf": 1}, "pyerrors.obs.Obs.plot_rho": {"tf": 1}, "pyerrors.obs.Obs.plot_piechart": {"tf": 1}}, "df": 4}}}}}, "a": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}, "c": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}}, "t": {"docs": {"pyerrors.input.hadrons.Npr_matrix.g5H": {"tf": 1}}, "df": 1, "o": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.correlators.Corr.reweight": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.reweight": {"tf": 1.4142135623730951}, "pyerrors.obs.reweight": {"tf": 1.4142135623730951}}, "df": 3, "s": {"docs": {"pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.obs.Obs.reweight": {"tf": 1}}, "df": 3}}}}}, "l": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.fits.least_squares": {"tf": 2.23606797749979}, "pyerrors.fits.total_least_squares": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.read_dobs": {"tf": 2}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1.4142135623730951}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.pandas.read_sql": {"tf": 1}, "pyerrors.input.pandas.dump_df": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.reweight": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}, "pyerrors.obs.reweight": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1.4142135623730951}}, "df": 23}}}, "i": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.input.pandas.to_sql": {"tf": 1}}, "df": 1}}, "s": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.linalg.matmul": {"tf": 1}, "pyerrors.linalg.jack_matmul": {"tf": 1}}, "df": 2}}}}}, "^": {"docs": {}, "df": 0, "i": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "u": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "c": {"docs": {"pyerrors": {"tf": 2}, "pyerrors.fits.least_squares": {"tf": 2.23606797749979}, "pyerrors.fits.total_least_squares": {"tf": 2.449489742783178}, "pyerrors.fits.error_band": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 2}, "pyerrors.roots.find_root": {"tf": 1.7320508075688772}}, "df": 6, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.correlators.Corr.fit": {"tf": 1.7320508075688772}, "pyerrors.fits.least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.error_band": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 2}, "pyerrors.obs.Obs.gamma_method": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.gm": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.plot_rho": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}, "pyerrors.roots.find_root": {"tf": 2}}, "df": 14, "s": {"docs": {"pyerrors": {"tf": 3.605551275463989}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1.4142135623730951}, "pyerrors.roots.find_root": {"tf": 1}}, "df": 7}, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.input.pandas.dump_df": {"tf": 1}}, "df": 1}}}}}}}}}}}, "l": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.json.import_json_string": {"tf": 1.4142135623730951}, "pyerrors.input.json.load_json": {"tf": 1.4142135623730951}, "pyerrors.input.json.load_json_dict": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 11, "y": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}}, "r": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}}}}}, "l": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.m_eff": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.fits.Fit_result": {"tf": 1.7320508075688772}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}, "pyerrors.misc.pseudo_Obs": {"tf": 1.4142135623730951}, "pyerrors.misc.gen_correlated_data": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.gamma_method": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.gm": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.is_zero": {"tf": 1}, "pyerrors.obs.cov_Obs": {"tf": 1}, "pyerrors.roots.find_root": {"tf": 1}}, "df": 13, "s": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.fit_lin": {"tf": 1}, "pyerrors.obs.cov_Obs": {"tf": 1}}, "df": 4}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 2}}}}}, "w": {"docs": {"pyerrors.input.openQCD.read_qtop": {"tf": 2}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 2.23606797749979}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.7320508075688772}}, "df": 3}}, "u": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}}}}}, "a": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}}, "df": 4, "s": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}, "t": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1.7320508075688772}}, "df": 1, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}}}, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}}}}}}}, "t": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}, "p": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "c": {"docs": {"pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}}, "df": 4}}}, "n": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}, "pyerrors.input.pandas.dump_df": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1}}, "df": 10}}}}, "f": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.obs.Obs.gamma_method": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.gm": {"tf": 1.4142135623730951}}, "df": 2}}}, "e": {"docs": {"pyerrors": {"tf": 2}, "pyerrors.correlators.Corr.m_eff": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 2.23606797749979}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.4142135623730951}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.4142135623730951}, "pyerrors.obs.derived_observable": {"tf": 1.4142135623730951}, "pyerrors.obs.reweight": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1.4142135623730951}}, "df": 14, "r": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.fits.error_band": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}}, "df": 2, "o": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors": {"tf": 5.916079783099616}, "pyerrors.correlators.Corr.plottable": {"tf": 1}, "pyerrors.covobs.Covobs.errsq": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.fits.error_band": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1.7320508075688772}, "pyerrors.input.pandas.read_sql": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1}, "pyerrors.misc.pseudo_Obs": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.gamma_method": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.gm": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.plot_piechart": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 2.23606797749979}}, "df": 15, "s": {"docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.obs.Obs.is_zero_within_error": {"tf": 1.4142135623730951}, "pyerrors.obs.covariance": {"tf": 1.4142135623730951}}, "df": 4}}}, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "m": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}, "x": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 5, "l": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1}}, "df": 2}}}}, "m": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 2.449489742783178}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.roots.find_root": {"tf": 1}}, "df": 3, "s": {"docs": {"pyerrors": {"tf": 2}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.7320508075688772}}, "df": 2, "/": {"docs": {}, "df": 0, "j": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}}}}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}}}}, "i": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1}}, "df": 1, "s": {"docs": {"pyerrors.input.pandas.to_sql": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs": {"tf": 1.7320508075688772}}, "df": 2}}}}, "p": {"docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.obs.Obs": {"tf": 2}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}, "pyerrors.roots.find_root": {"tf": 1}}, "df": 5, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}, "d": {"docs": {"pyerrors.obs.Obs.plot_history": {"tf": 1}}, "df": 1, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.obs.Obs.plot_history": {"tf": 1}}, "df": 1}}}}}, "o": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}}}}}}}, "r": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.input": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1.4142135623730951}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1}}, "df": 10, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.dump": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.dump": {"tf": 1}}, "df": 10}}, "s": {"docs": {"pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.pandas.dump_df": {"tf": 1}}, "df": 3}}}, "s": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}}, "df": 1}}}}}, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}}}}}, "c": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.fits.least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.total_least_squares": {"tf": 1.4142135623730951}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 3}}}}, "n": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 1}}}}}}, "l": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.m_eff": {"tf": 1}}, "df": 1, "l": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}}, "df": 2}}}}}}}}, "h": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1}}, "df": 1, "s": {"docs": {"pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}}, "df": 2}}}}}}, "t": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.plateau": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1.4142135623730951}}, "df": 10, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.4142135623730951}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 8}}, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 1}}}}}}, "e": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}}}}, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors": {"tf": 2.8284271247461903}}, "df": 1, "l": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}}, "df": 1}}}}}}}, "n": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.Hankel": {"tf": 1}}, "df": 1}}}, "s": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.dirac.epsilon_tensor_rank4": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1.7320508075688772}}, "df": 2}}}}, "t": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 4}}}}, "e": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}, "c": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.pandas.read_sql": {"tf": 1}}, "df": 1, "d": {"docs": {"pyerrors.input.pandas.read_sql": {"tf": 1}}, "df": 1}, "s": {"docs": {"pyerrors.obs.CObs.gamma_method": {"tf": 1}}, "df": 1}}}}}}, "c": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.correlators.Corr.spaghetti_plot": {"tf": 1}}, "df": 1}}}}}}}}, "h": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.hadrons.Npr_matrix.g5H": {"tf": 1}}, "df": 1}}}}}}}}, "i": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.linalg.eig": {"tf": 1}}, "df": 2, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 2.23606797749979}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 4, "s": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.linalg.eigh": {"tf": 1}, "pyerrors.linalg.eig": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1.4142135623730951}}, "df": 5}}}}}, "e": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.correlators.Corr.GEVP": {"tf": 2.23606797749979}}, "df": 1, "s": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.GEVP": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.prune": {"tf": 1.7320508075688772}, "pyerrors.linalg.eigh": {"tf": 1}}, "df": 4}}}}}}}}}, "h": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.linalg.eigh": {"tf": 1}}, "df": 2}}, "t": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.correlate": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.fit_lin": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.misc.gen_correlated_data": {"tf": 1}}, "df": 9}}}}, "n": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "m": {"docs": {"pyerrors.linalg.einsum": {"tf": 1}}, "df": 1}}}}}, "n": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.obs.Obs.details": {"tf": 1}}, "df": 7, "e": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"1": {"docs": {"pyerrors": {"tf": 3.4641016151377544}}, "df": 1, "|": {"docs": {}, "df": 0, "r": {"0": {"1": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "2": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "docs": {}, "df": 0}}}, "2": {"docs": {"pyerrors": {"tf": 1.7320508075688772}}, "df": 1}, "3": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {"pyerrors": {"tf": 5.5677643628300215}, "pyerrors.input.bdio.read_mesons": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.read_dSdm": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.write_pobs": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.4142135623730951}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.4142135623730951}, "pyerrors.misc.pseudo_Obs": {"tf": 1}, "pyerrors.misc.gen_correlated_data": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 2.449489742783178}, "pyerrors.obs.Obs.plot_tauint": {"tf": 1}, "pyerrors.obs.Obs.plot_rho": {"tf": 1}, "pyerrors.obs.Obs.plot_rep_dist": {"tf": 1}, "pyerrors.obs.Obs.plot_history": {"tf": 1}, "pyerrors.obs.Obs.plot_piechart": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1.4142135623730951}, "pyerrors.obs.covariance": {"tf": 1}, "pyerrors.obs.import_jackknife": {"tf": 1}}, "df": 28, "s": {"docs": {"pyerrors": {"tf": 2.449489742783178}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.read_dobs": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.obs.Obs.details": {"tf": 1}, "pyerrors.obs.correlate": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 8, "/": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "a": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1.4142135623730951}}, "df": 2}}}}}}}}}}}}}}, "t": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.dobs.create_pobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.write_pobs": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.write_dobs": {"tf": 1.4142135623730951}}, "df": 4, "s": {"docs": {"pyerrors.input.dobs.create_dobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.write_dobs": {"tf": 1.4142135623730951}}, "df": 2}}}}, "n": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}}, "df": 2}}}}, "u": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}}}, "t": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 3.1622776601683795}, "pyerrors.correlators.Corr.__init__": {"tf": 1}, "pyerrors.correlators.Corr.correlate": {"tf": 1}, "pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1}, "pyerrors.obs.cov_Obs": {"tf": 1}}, "df": 10}}}, "y": {"docs": {"pyerrors": {"tf": 3.3166247903554}, "pyerrors.correlators.Corr.thin": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.misc.gen_correlated_data": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1}, "pyerrors.obs.import_jackknife": {"tf": 1.4142135623730951}}, "df": 9}}, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}}}, "i": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1.7320508075688772}}, "df": 3}}}}, "a": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}}, "df": 2}}}, "d": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.openQCD.read_rwms": {"tf": 1}}, "df": 1}}}}}}, "s": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 5, "d": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1.4142135623730951}}, "df": 3}, "s": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}}, "df": 3}}, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 2.23606797749979}}, "df": 1}}}}}}}, "a": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}}}}, "p": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 2}}}}}}}}, "s": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}}}}}}}}, "v": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 2}}}}}}}, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr": {"tf": 2}, "pyerrors.correlators.Corr.item": {"tf": 1}, "pyerrors.correlators.Corr.is_matrix_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.matrix_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.thin": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.misc.gen_correlated_data": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 10, "t": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.correlators.Corr": {"tf": 1}}, "df": 1}}}}}}}, "n": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}}}, "f": {"docs": {}, "df": 0, "f": {"docs": {"pyerrors": {"tf": 2}}, "df": 1, "e": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.m_eff": {"tf": 1.7320508075688772}}, "df": 2}}}, "s": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.4142135623730951}}, "df": 3}}}}}}, "a": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "h": {"docs": {"pyerrors": {"tf": 3.605551275463989}, "pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.correlators.Corr.plottable": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.7320508075688772}, "pyerrors.input.misc.read_pbp": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_rwms": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1.4142135623730951}, "pyerrors.misc.gen_correlated_data": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.plot_tauint": {"tf": 1}, "pyerrors.obs.Obs.plot_rho": {"tf": 1}, "pyerrors.obs.Obs.plot_rep_dist": {"tf": 1}, "pyerrors.obs.Obs.plot_history": {"tf": 1}, "pyerrors.obs.Obs.plot_piechart": {"tf": 1}}, "df": 15}}, "r": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 1}}}, "s": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 2}}}}}, "l": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "m": {"docs": {"pyerrors.dirac.epsilon_tensor": {"tf": 1}, "pyerrors.dirac.epsilon_tensor_rank4": {"tf": 1}}, "df": 2, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.item": {"tf": 1}, "pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.dirac.epsilon_tensor": {"tf": 1}, "pyerrors.dirac.epsilon_tensor_rank4": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 2.6457513110645907}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 8, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}, "s": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1.7320508075688772}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 2}}}}}}, "s": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}}, "q": {"docs": {"pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 1, "u": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "v": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1.4142135623730951}}, "df": 1, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 2, "l": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}}}}, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.correlators.Corr.thin": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 2}}}}, "g": {"docs": {"pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 2}, "p": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.dirac.epsilon_tensor": {"tf": 1.4142135623730951}, "pyerrors.dirac.epsilon_tensor_rank4": {"tf": 1.4142135623730951}}, "df": 2}}}}}}, "m": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}}, "df": 5}}}}, "t": {"docs": {}, "df": 0, "c": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}}, "df": 1}}}, "c": {"docs": {"pyerrors.correlators.Corr.GEVP": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.Hankel": {"tf": 3}, "pyerrors.correlators.Corr.m_eff": {"tf": 2.8284271247461903}, "pyerrors.correlators.Corr.prune": {"tf": 2}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 2}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.4142135623730951}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.7320508075688772}}, "df": 9, "o": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "p": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1}}, "df": 1, "u": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1.7320508075688772}}, "df": 1, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}}, "df": 4}}}}}, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs": {"tf": 1}}, "df": 2}, "s": {"docs": {"pyerrors.linalg.eigh": {"tf": 1}, "pyerrors.linalg.eig": {"tf": 1}, "pyerrors.linalg.pinv": {"tf": 1}, "pyerrors.linalg.svd": {"tf": 1}}, "df": 4}}}}, "l": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "x": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1.4142135623730951}, "pyerrors.linalg.matmul": {"tf": 1}, "pyerrors.linalg.jack_matmul": {"tf": 1}, "pyerrors.linalg.einsum": {"tf": 1}, "pyerrors.obs.CObs": {"tf": 1}}, "df": 6}, "t": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}}, "df": 1, "d": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}, "pyerrors.linalg.matmul": {"tf": 1}, "pyerrors.linalg.jack_matmul": {"tf": 1}}, "df": 3}, "s": {"docs": {"pyerrors.fits.residual_plot": {"tf": 1}}, "df": 1}}, "i": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.input": {"tf": 1}}, "df": 2}}}}}, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}, "c": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.input.sfcf.read_sfcf": {"tf": 1.4142135623730951}}, "df": 1}}}, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}, "i": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}}, "df": 4}}}}}, "m": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1.7320508075688772}}, "df": 1}}}, "b": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.obs.merge_obs": {"tf": 1.4142135623730951}}, "df": 1, "d": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.obs.merge_obs": {"tf": 1}}, "df": 2}}, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1.7320508075688772}}, "df": 2}}}}}}}}, "e": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "/": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "/": {"1": {"6": {"0": {"3": {"7": {"5": {"docs": {"pyerrors.dirac.epsilon_tensor": {"tf": 1}, "pyerrors.dirac.epsilon_tensor_rank4": {"tf": 1}}, "df": 2}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}}}}, "h": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}, "n": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}}, "df": 2, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.openQCD.read_rwms": {"tf": 1}}, "df": 1}}, "a": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.linalg.jack_matmul": {"tf": 1}}, "df": 1}}}}}}}, "s": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1}}, "df": 1, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1}}, "df": 1}}}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}}, "df": 2}}}}}}, "t": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}}, "df": 3, "s": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.Hankel": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 3}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 4}}, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}, "o": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": null}}, "df": 1}}}}}}, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.plateau": {"tf": 1}}, "df": 1, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}, "o": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}}, "df": 2}}}}, "t": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.obs.Obs.export_jackknife": {"tf": 1}}, "df": 1, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.__init__": {"tf": 1}, "pyerrors.correlators.Corr.item": {"tf": 1}, "pyerrors.correlators.Corr.plottable": {"tf": 1}, "pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.import_json_string": {"tf": 1.7320508075688772}, "pyerrors.input.json.load_json": {"tf": 1.4142135623730951}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1.7320508075688772}, "pyerrors.input.pandas.dump_df": {"tf": 1}, "pyerrors.misc.gen_correlated_data": {"tf": 1}, "pyerrors.obs.Obs.__init__": {"tf": 1}, "pyerrors.obs.Obs.plot_piechart": {"tf": 1}, "pyerrors.obs.import_jackknife": {"tf": 1}}, "df": 26}}}, "s": {"docs": {"pyerrors": {"tf": 5}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.misc.read_pbp": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_rwms": {"tf": 2}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}}, "df": 12}, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.input": {"tf": 1}, "pyerrors.input.utils.check_idl": {"tf": 1}}, "df": 3}}}}}, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.obs.Obs.plot_piechart": {"tf": 1}}, "df": 4}}}, "e": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}, "o": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.obs.derived_observable": {"tf": 1}}, "df": 1}}}, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.correlators.Corr.gamma_method": {"tf": 1}, "pyerrors.correlators.Corr.gm": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.pandas.read_sql": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1}, "pyerrors.obs.Obs.details": {"tf": 1}}, "df": 8, "s": {"docs": {"pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}}, "df": 7}}}}, "i": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}}, "df": 1}}}}}}}, "f": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.misc.read_pbp": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_rwms": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.extract_t0": {"tf": 2.23606797749979}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1.4142135623730951}}, "df": 5, "u": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 2.449489742783178}, "pyerrors.input.bdio.read_mesons": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.read_dSdm": {"tf": 1.4142135623730951}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 4, "s": {"docs": {"pyerrors": {"tf": 3.605551275463989}, "pyerrors.correlators.Corr.reweight": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.spaghetti_plot": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.7320508075688772}, "pyerrors.input.utils.check_idl": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.reweight": {"tf": 1.7320508075688772}, "pyerrors.obs.reweight": {"tf": 1.7320508075688772}}, "df": 19}}}}}}}}, "s": {"docs": {"pyerrors.correlators.Corr.reweight": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.reweight": {"tf": 1}, "pyerrors.obs.reweight": {"tf": 1}}, "df": 9}}}}, "d": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 2, "s": {"docs": {"pyerrors": {"tf": 2.23606797749979}}, "df": 1}, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}}, "df": 1}}}}}}}}, "j": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix.g5H": {"tf": 1}}, "df": 2}}}}}}, "c": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "r": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}}, "v": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr": {"tf": 1}}, "df": 1}}}}, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.bdio.write_ADerrors": {"tf": 1}}, "df": 1}}}}}}}}}, "r": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors": {"tf": 6.6332495807108}, "pyerrors.correlators.Corr.__init__": {"tf": 1}, "pyerrors.correlators.Corr.gamma_method": {"tf": 1}, "pyerrors.correlators.Corr.gm": {"tf": 1}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 1}, "pyerrors.correlators.Corr.reverse": {"tf": 1}, "pyerrors.correlators.Corr.correlate": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.T_symmetry": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.plateau": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.set_prange": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.dump": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1.4142135623730951}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 2}, "pyerrors.input.pandas.to_sql": {"tf": 1}, "pyerrors.input.pandas.read_sql": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.misc.gen_correlated_data": {"tf": 1}}, "df": 25, "e": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.correlate": {"tf": 1.4142135623730951}, "pyerrors.obs.correlate": {"tf": 1.4142135623730951}}, "df": 3, "d": {"docs": {"pyerrors": {"tf": 2.449489742783178}, "pyerrors.correlators.Corr.correlate": {"tf": 1}, "pyerrors.fits.Fit_result": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 2}, "pyerrors.fits.total_least_squares": {"tf": 1}}, "df": 5}}, "o": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors": {"tf": 3.3166247903554}, "pyerrors.correlators.Corr": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.__init__": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.correlators.Corr.item": {"tf": 1}, "pyerrors.correlators.Corr.plottable": {"tf": 1}, "pyerrors.correlators.Corr.symmetric": {"tf": 1}, "pyerrors.correlators.Corr.anti_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.is_matrix_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.matrix_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 1}, "pyerrors.correlators.Corr.roll": {"tf": 1}, "pyerrors.correlators.Corr.thin": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.correlate": {"tf": 2}, "pyerrors.correlators.Corr.reweight": {"tf": 1}, "pyerrors.correlators.Corr.T_symmetry": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.deriv": {"tf": 1}, "pyerrors.correlators.Corr.second_deriv": {"tf": 1}, "pyerrors.correlators.Corr.m_eff": {"tf": 2.23606797749979}, "pyerrors.correlators.Corr.show": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.spaghetti_plot": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 25, "s": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.show": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1.4142135623730951}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 6}}}, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.correlators.Corr.prune": {"tf": 2.23606797749979}, "pyerrors.fits.least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 2}, "pyerrors.obs.covariance": {"tf": 2.449489742783178}}, "df": 6, "s": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.thin": {"tf": 1}}, "df": 2}}}, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.obs.correlate": {"tf": 1}}, "df": 1}}}}}}, "c": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 2, "n": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 2}}}}, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}}, "df": 2}}}}, "s": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.GEVP": {"tf": 1}}, "df": 1, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.fits.residual_plot": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1.4142135623730951}}, "df": 5}}}, "s": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 5}}}}}}}, "s": {"docs": {"pyerrors.correlators.Corr.__init__": {"tf": 1}}, "df": 1}}}, "b": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 3.605551275463989}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1.4142135623730951}, "pyerrors.linalg.inv": {"tf": 1}}, "df": 6}}, "v": {"docs": {"pyerrors": {"tf": 2}, "pyerrors.covobs.Covobs.__init__": {"tf": 2}, "pyerrors.misc.gen_correlated_data": {"tf": 1}, "pyerrors.obs.cov_Obs": {"tf": 1.4142135623730951}}, "df": 4, "o": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 3.1622776601683795}, "pyerrors.covobs.Covobs.__init__": {"tf": 1.4142135623730951}, "pyerrors.covobs.Covobs.errsq": {"tf": 1}, "pyerrors.obs.cov_Obs": {"tf": 1}}, "df": 4}}}, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 4}, "pyerrors.covobs.Covobs.__init__": {"tf": 1.4142135623730951}, "pyerrors.fits.least_squares": {"tf": 2}, "pyerrors.misc.gen_correlated_data": {"tf": 1.4142135623730951}, "pyerrors.obs.covariance": {"tf": 2.449489742783178}, "pyerrors.obs.cov_Obs": {"tf": 2}}, "df": 6}}}}}}}}, "l": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}}}}}, "u": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1, "s": {"docs": {"pyerrors.input.pandas.to_sql": {"tf": 1}, "pyerrors.input.pandas.read_sql": {"tf": 1}, "pyerrors.input.pandas.dump_df": {"tf": 1}}, "df": 3}}}}}, "s": {"docs": {}, "df": 0, "h": {"docs": {"pyerrors.correlators.Corr.m_eff": {"tf": 1.7320508075688772}}, "df": 1}, "t": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}}, "df": 1}}, "u": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.openQCD.read_gf_coupling": {"tf": 1.7320508075688772}}, "df": 1}}}}}, "l": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 1}}}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}}, "df": 2}}}}}}}}, "h": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1, "s": {"docs": {"pyerrors": {"tf": 3.1622776601683795}, "pyerrors.obs.Obs.plot_history": {"tf": 1}}, "df": 2}, "/": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}}}}}}, "n": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}}, "df": 4}}}}}, "r": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.openQCD.read_qtop": {"tf": 2}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.qtop_projection": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 4}}}}, "o": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "k": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.linalg.cholesky": {"tf": 1}}, "df": 3}}}}, "s": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "k": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}, "s": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.dump": {"tf": 1}, "pyerrors.obs.Obs.dump": {"tf": 1}}, "df": 4}}}, "i": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.deriv": {"tf": 1}, "pyerrors.correlators.Corr.second_deriv": {"tf": 1}}, "df": 2}}}, "o": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 1}}}, "e": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}}, "df": 1}}}}, "e": {"docs": {"pyerrors.input.utils.check_idl": {"tf": 1}}, "df": 1, "c": {"docs": {}, "df": 0, "k": {"docs": {"pyerrors": {"tf": 2}, "pyerrors.fits.qqplot": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.obs.Obs.is_zero_within_error": {"tf": 1}}, "df": 4, "s": {"docs": {"pyerrors.correlators.Corr.is_matrix_symmetric": {"tf": 1}, "pyerrors.input.utils.check_idl": {"tf": 1}, "pyerrors.obs.Obs.is_zero_within_error": {"tf": 1}, "pyerrors.obs.Obs.is_zero": {"tf": 1}, "pyerrors.obs.CObs.is_zero": {"tf": 1}}, "df": 5}, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.utils.check_idl": {"tf": 1}}, "df": 1}}}}}, "i": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "q": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.fits.Fit_result": {"tf": 1.4142135623730951}, "pyerrors.fits.least_squares": {"tf": 2}, "pyerrors.fits.total_least_squares": {"tf": 1.4142135623730951}}, "df": 4}}}}}}}}, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "o": {"docs": {"pyerrors": {"tf": 3.4641016151377544}, "pyerrors.obs.Obs.__init__": {"tf": 1}, "pyerrors.obs.Obs.plot_history": {"tf": 1.4142135623730951}}, "df": 3}}, "e": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 2, "f": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 2}}}}}, "l": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1}}, "df": 2}}, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}, "e": {"docs": {"pyerrors.fits.error_band": {"tf": 1}}, "df": 1, "d": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}}, "df": 2}, "s": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}}}}}, "l": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}}}}, "n": {"docs": {"pyerrors": {"tf": 5.744562646538029}, "pyerrors.correlators.Corr": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.correlate": {"tf": 1}, "pyerrors.correlators.Corr.T_symmetry": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 2.23606797749979}, "pyerrors.fits.total_least_squares": {"tf": 1.7320508075688772}, "pyerrors.fits.fit_lin": {"tf": 1}, "pyerrors.fits.qqplot": {"tf": 1}, "pyerrors.input": {"tf": 1}, "pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.read_dSdm": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.linalg.matmul": {"tf": 1}, "pyerrors.linalg.jack_matmul": {"tf": 1}, "pyerrors.linalg.einsum": {"tf": 1}, "pyerrors.misc.gen_correlated_data": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1.4142135623730951}, "pyerrors.obs.derived_observable": {"tf": 1.7320508075688772}}, "df": 29, "n": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 1}}}}, "s": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.correlators.Corr.prune": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.fit_lin": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}}, "df": 10}}, "u": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.correlators.Corr.fit": {"tf": 1}}, "df": 1}, "u": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.obs.derived_observable": {"tf": 1}}, "df": 1}}}}}}}, "s": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}}, "df": 2}}}}}, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}}, "df": 3}}}}}}, "o": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}}, "df": 2, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1.4142135623730951}}, "df": 1}}}}}}, "e": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}}, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}}, "o": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "e": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.obs.cov_Obs": {"tf": 1}}, "df": 3, "d": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}}}}}, "l": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 3.1622776601683795}, "pyerrors.correlators.Corr": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1}, "pyerrors.obs.CObs": {"tf": 1}}, "df": 5}}}, "o": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 1}}}}}, "u": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.utils.check_idl": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1}}, "df": 3, "l": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.obs.correlate": {"tf": 1}}, "df": 2}}}}}}}, "s": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "m": {"docs": {"pyerrors.correlators.Corr.dump": {"tf": 1}, "pyerrors.misc.dump_object": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}, "pyerrors.obs.Obs.dump": {"tf": 1}}, "df": 5}}}}}, "d": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "a": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}}}}, "c": {"docs": {"pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}}, "df": 4}, "p": {"docs": {"pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}}, "df": 4}, "t": {"docs": {}, "df": 0, "y": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1.7320508075688772}}, "df": 1}}}}}, "n": {"docs": {}, "df": 0, "f": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.4142135623730951}}, "df": 4}}}, "y": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 3}}}}}, "s": {"docs": {}, "df": 0, "v": {"docs": {"pyerrors.input.pandas.dump_df": {"tf": 2}, "pyerrors.input.pandas.load_df": {"tf": 1}}, "df": 2}}, "f": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 1}}, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.misc.pseudo_Obs": {"tf": 1}}, "df": 1}}}}}}}, "o": {"docs": {"pyerrors": {"tf": 2}, "pyerrors.correlators.Corr.prune": {"tf": 1}, "pyerrors.dirac.epsilon_tensor_rank4": {"tf": 1}, "pyerrors.input.bdio.read_ADerrors": {"tf": 1.7320508075688772}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1.7320508075688772}, "pyerrors.input.bdio.read_mesons": {"tf": 1.7320508075688772}, "pyerrors.input.bdio.read_dSdm": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1}}, "df": 8, "f": {"docs": {"pyerrors": {"tf": 10.344080432788601}, "pyerrors.correlators.Corr": {"tf": 2}, "pyerrors.correlators.Corr.__init__": {"tf": 2.449489742783178}, "pyerrors.correlators.Corr.gamma_method": {"tf": 1}, "pyerrors.correlators.Corr.gm": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.Hankel": {"tf": 1}, "pyerrors.correlators.Corr.roll": {"tf": 1}, "pyerrors.correlators.Corr.reverse": {"tf": 1}, "pyerrors.correlators.Corr.thin": {"tf": 1}, "pyerrors.correlators.Corr.correlate": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.reweight": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.T_symmetry": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.deriv": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.second_deriv": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.m_eff": {"tf": 2}, "pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.correlators.Corr.set_prange": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 2.8284271247461903}, "pyerrors.correlators.Corr.spaghetti_plot": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.dump": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.prune": {"tf": 2.6457513110645907}, "pyerrors.covobs.Covobs.__init__": {"tf": 1.7320508075688772}, "pyerrors.covobs.Covobs.errsq": {"tf": 1.4142135623730951}, "pyerrors.dirac.epsilon_tensor": {"tf": 1.4142135623730951}, "pyerrors.dirac.epsilon_tensor_rank4": {"tf": 1.7320508075688772}, "pyerrors.fits.Fit_result": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 3.3166247903554}, "pyerrors.fits.total_least_squares": {"tf": 3.1622776601683795}, "pyerrors.fits.fit_lin": {"tf": 2.449489742783178}, "pyerrors.fits.qqplot": {"tf": 1.4142135623730951}, "pyerrors.fits.error_band": {"tf": 1.4142135623730951}, "pyerrors.fits.ks_test": {"tf": 1.4142135623730951}, "pyerrors.input": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 2.449489742783178}, "pyerrors.input.dobs.write_pobs": {"tf": 2.449489742783178}, "pyerrors.input.dobs.read_pobs": {"tf": 2}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.read_dobs": {"tf": 2}, "pyerrors.input.dobs.create_dobs_string": {"tf": 2.8284271247461903}, "pyerrors.input.dobs.write_dobs": {"tf": 2.8284271247461903}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 2.6457513110645907}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1.7320508075688772}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 5.0990195135927845}, "pyerrors.input.hadrons.Npr_matrix.g5H": {"tf": 1}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1.7320508075688772}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1.7320508075688772}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1.7320508075688772}, "pyerrors.input.json.create_json_string": {"tf": 2.6457513110645907}, "pyerrors.input.json.dump_to_json": {"tf": 2.6457513110645907}, "pyerrors.input.json.import_json_string": {"tf": 1.4142135623730951}, "pyerrors.input.json.load_json": {"tf": 1.7320508075688772}, "pyerrors.input.json.dump_dict_to_json": {"tf": 2.8284271247461903}, "pyerrors.input.json.load_json_dict": {"tf": 1.7320508075688772}, "pyerrors.input.misc.read_pbp": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 2.449489742783178}, "pyerrors.input.openQCD.extract_t0": {"tf": 3.1622776601683795}, "pyerrors.input.openQCD.read_qtop": {"tf": 2.6457513110645907}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 2.6457513110645907}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 2.8284271247461903}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1.7320508075688772}, "pyerrors.input.pandas.to_sql": {"tf": 1}, "pyerrors.input.pandas.read_sql": {"tf": 1}, "pyerrors.input.pandas.dump_df": {"tf": 1.7320508075688772}, "pyerrors.input.pandas.load_df": {"tf": 1.4142135623730951}, "pyerrors.input.sfcf.read_sfcf": {"tf": 4}, "pyerrors.input.utils.check_idl": {"tf": 2}, "pyerrors.linalg.matmul": {"tf": 1}, "pyerrors.linalg.jack_matmul": {"tf": 1.4142135623730951}, "pyerrors.linalg.einsum": {"tf": 1}, "pyerrors.linalg.inv": {"tf": 1}, "pyerrors.linalg.cholesky": {"tf": 1}, "pyerrors.linalg.det": {"tf": 1}, "pyerrors.linalg.eigh": {"tf": 1.4142135623730951}, "pyerrors.linalg.eig": {"tf": 1.4142135623730951}, "pyerrors.linalg.pinv": {"tf": 1.4142135623730951}, "pyerrors.linalg.svd": {"tf": 1.4142135623730951}, "pyerrors.misc.dump_object": {"tf": 1}, "pyerrors.misc.pseudo_Obs": {"tf": 2}, "pyerrors.misc.gen_correlated_data": {"tf": 1.4142135623730951}, "pyerrors.mpm.matrix_pencil_method": {"tf": 2.6457513110645907}, "pyerrors.obs.Obs": {"tf": 2.8284271247461903}, "pyerrors.obs.Obs.__init__": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.gamma_method": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.gm": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.details": {"tf": 1}, "pyerrors.obs.Obs.reweight": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.is_zero_within_error": {"tf": 1}, "pyerrors.obs.Obs.plot_piechart": {"tf": 1}, "pyerrors.obs.Obs.dump": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.export_jackknife": {"tf": 2}, "pyerrors.obs.derived_observable": {"tf": 2.449489742783178}, "pyerrors.obs.reweight": {"tf": 2}, "pyerrors.obs.covariance": {"tf": 3.3166247903554}, "pyerrors.obs.import_jackknife": {"tf": 1}, "pyerrors.obs.merge_obs": {"tf": 1.7320508075688772}, "pyerrors.obs.cov_Obs": {"tf": 1.7320508075688772}, "pyerrors.roots.find_root": {"tf": 1.4142135623730951}}, "df": 101, "f": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}, "s": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.thin": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 2.23606797749979}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 4}}}}, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.correlators.Corr": {"tf": 1}}, "df": 1}}}}, "n": {"docs": {"pyerrors": {"tf": 5.291502622129181}, "pyerrors.correlators.Corr.plottable": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.is_matrix_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.matrix_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.reweight": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1.4142135623730951}, "pyerrors.dirac.epsilon_tensor": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.pandas.read_sql": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1}, "pyerrors.obs.Obs.__init__": {"tf": 1}, "pyerrors.obs.Obs.reweight": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1}, "pyerrors.obs.reweight": {"tf": 1.7320508075688772}, "pyerrors.obs.covariance": {"tf": 1.4142135623730951}, "pyerrors.obs.import_jackknife": {"tf": 1}, "pyerrors.obs.merge_obs": {"tf": 1}, "pyerrors.obs.cov_Obs": {"tf": 1}}, "df": 33, "e": {"docs": {"pyerrors": {"tf": 3}, "pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.correlators.Corr.item": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}, "pyerrors.input.json.import_json_string": {"tf": 1.7320508075688772}, "pyerrors.input.json.load_json": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.obs.Obs.plot_rep_dist": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}, "pyerrors.obs.merge_obs": {"tf": 1}}, "df": 20, "s": {"docs": {"pyerrors.correlators.Corr.Eigenvalue": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 3}}, "l": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.GEVP": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.thin": {"tf": 1}, "pyerrors.correlators.Corr.m_eff": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}, "pyerrors.input.json.import_json_string": {"tf": 2}, "pyerrors.input.json.load_json": {"tf": 2}, "pyerrors.input.json.load_json_dict": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.4142135623730951}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.is_zero_within_error": {"tf": 1}, "pyerrors.obs.Obs.dump": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1}, "pyerrors.obs.correlate": {"tf": 1.4142135623730951}}, "df": 26}}, "t": {"docs": {}, "df": 0, "o": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 1}}, "c": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 1}}}, "p": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.correlators.Corr": {"tf": 1}}, "df": 1, "s": {"docs": {"pyerrors": {"tf": 2.449489742783178}, "pyerrors.obs.Obs": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}}, "df": 3}}}}, "o": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.linalg.matmul": {"tf": 1}}, "df": 1, "s": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 1}, "n": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "{": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "v": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}, "v": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}}}}}}}}}, "n": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.linalg.matmul": {"tf": 1.4142135623730951}, "pyerrors.linalg.jack_matmul": {"tf": 1.4142135623730951}, "pyerrors.linalg.einsum": {"tf": 1}}, "df": 3}}}}}, "n": {"docs": {}, "df": 0, "q": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.openQCD.read_rwms": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 2}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.4142135623730951}}, "df": 5}}}}}, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.4142135623730951}}, "df": 2, "s": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.input.pandas.to_sql": {"tf": 1}}, "df": 2}, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.__init__": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.Hankel": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1.7320508075688772}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.fits.ks_test": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 2.23606797749979}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}, "pyerrors.obs.Obs.__init__": {"tf": 1}}, "df": 20}}}}, "m": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "z": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}}, "df": 1, "d": {"docs": {"pyerrors.fits.error_band": {"tf": 1}}, "df": 1}}}}, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 1}}}}}, "p": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.fit": {"tf": 1}}, "df": 1}}}}}}, "b": {"docs": {}, "df": 0, "s": {"1": {"docs": {"pyerrors": {"tf": 3}, "pyerrors.obs.derived_observable": {"tf": 1.4142135623730951}, "pyerrors.obs.reweight": {"tf": 1}}, "df": 3}, "2": {"docs": {"pyerrors": {"tf": 3}, "pyerrors.obs.derived_observable": {"tf": 1.4142135623730951}, "pyerrors.obs.reweight": {"tf": 1}}, "df": 3}, "3": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.obs.derived_observable": {"tf": 1}, "pyerrors.obs.reweight": {"tf": 1}}, "df": 3}, "docs": {"pyerrors": {"tf": 9.591663046625438}, "pyerrors.correlators.Corr": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.__init__": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.item": {"tf": 1}, "pyerrors.correlators.Corr.correlate": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.reweight": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1.7320508075688772}, "pyerrors.fits.total_least_squares": {"tf": 2.23606797749979}, "pyerrors.fits.fit_lin": {"tf": 2.23606797749979}, "pyerrors.fits.error_band": {"tf": 1}, "pyerrors.input": {"tf": 2.23606797749979}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 2}, "pyerrors.input.dobs.write_pobs": {"tf": 2}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.read_dobs": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.create_dobs_string": {"tf": 2}, "pyerrors.input.dobs.write_dobs": {"tf": 2}, "pyerrors.input.json.create_json_string": {"tf": 2}, "pyerrors.input.json.dump_to_json": {"tf": 2}, "pyerrors.input.json.import_json_string": {"tf": 2}, "pyerrors.input.json.load_json": {"tf": 2}, "pyerrors.input.json.dump_dict_to_json": {"tf": 2}, "pyerrors.input.json.load_json_dict": {"tf": 2}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.qtop_projection": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.4142135623730951}, "pyerrors.input.pandas.to_sql": {"tf": 1}, "pyerrors.input.pandas.read_sql": {"tf": 1.4142135623730951}, "pyerrors.input.pandas.dump_df": {"tf": 1.4142135623730951}, "pyerrors.input.pandas.load_df": {"tf": 1.4142135623730951}, "pyerrors.linalg.matmul": {"tf": 1}, "pyerrors.linalg.jack_matmul": {"tf": 1}, "pyerrors.linalg.einsum": {"tf": 1}, "pyerrors.linalg.inv": {"tf": 1}, "pyerrors.linalg.cholesky": {"tf": 1}, "pyerrors.linalg.det": {"tf": 1}, "pyerrors.linalg.eigh": {"tf": 1}, "pyerrors.linalg.eig": {"tf": 1}, "pyerrors.linalg.pinv": {"tf": 1}, "pyerrors.linalg.svd": {"tf": 1}, "pyerrors.misc.load_object": {"tf": 1}, "pyerrors.misc.pseudo_Obs": {"tf": 2.449489742783178}, "pyerrors.misc.gen_correlated_data": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.__init__": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}, "pyerrors.obs.Obs.details": {"tf": 1}, "pyerrors.obs.Obs.reweight": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.dump": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1.7320508075688772}, "pyerrors.obs.derived_observable": {"tf": 1.7320508075688772}, "pyerrors.obs.reweight": {"tf": 1.7320508075688772}, "pyerrors.obs.correlate": {"tf": 2}, "pyerrors.obs.covariance": {"tf": 1.4142135623730951}, "pyerrors.obs.import_jackknife": {"tf": 1}, "pyerrors.obs.merge_obs": {"tf": 2}, "pyerrors.obs.cov_Obs": {"tf": 1.4142135623730951}, "pyerrors.roots.find_root": {"tf": 2.23606797749979}}, "df": 66, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 3.4641016151377544}, "pyerrors.correlators.Corr.reweight": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.misc.pseudo_Obs": {"tf": 1}, "pyerrors.misc.gen_correlated_data": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs": {"tf": 1}, "pyerrors.obs.Obs.reweight": {"tf": 1}, "pyerrors.obs.Obs.is_zero_within_error": {"tf": 1}, "pyerrors.obs.Obs.is_zero": {"tf": 1}, "pyerrors.obs.CObs": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}, "pyerrors.obs.reweight": {"tf": 1}, "pyerrors.obs.correlate": {"tf": 1.4142135623730951}, "pyerrors.obs.merge_obs": {"tf": 1}}, "df": 21, "s": {"docs": {"pyerrors": {"tf": 3}, "pyerrors.correlators.Corr.reweight": {"tf": 1}, "pyerrors.fits.fit_lin": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.misc.read_pbp": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.misc.gen_correlated_data": {"tf": 1}, "pyerrors.obs.Obs.reweight": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}, "pyerrors.obs.reweight": {"tf": 1.4142135623730951}, "pyerrors.obs.correlate": {"tf": 1.7320508075688772}, "pyerrors.obs.covariance": {"tf": 2.449489742783178}, "pyerrors.obs.merge_obs": {"tf": 1}}, "df": 19}}}}}}}}, "d": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "a": {"docs": {"pyerrors": {"tf": 1.7320508075688772}}, "df": 1}}}}, "[": {"docs": {}, "df": 0, "i": {"docs": {"pyerrors.correlators.Corr.reweight": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.reweight": {"tf": 1.4142135623730951}, "pyerrors.obs.reweight": {"tf": 1.4142135623730951}}, "df": 3}}, "l": {"docs": {"pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}}, "df": 4}}, "j": {"docs": {"pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.misc.dump_object": {"tf": 1}}, "df": 2, "e": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.correlators.Corr.__init__": {"tf": 1}, "pyerrors.correlators.Corr.m_eff": {"tf": 1}, "pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.correlators.Corr.set_prange": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.fits.ks_test": {"tf": 1}, "pyerrors.input": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 2.8284271247461903}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1.7320508075688772}, "pyerrors.misc.dump_object": {"tf": 1.7320508075688772}, "pyerrors.misc.load_object": {"tf": 1.7320508075688772}, "pyerrors.misc.pseudo_Obs": {"tf": 1}, "pyerrors.obs.Obs.__init__": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}, "pyerrors.obs.merge_obs": {"tf": 1}, "pyerrors.roots.find_root": {"tf": 1}}, "df": 21, "s": {"docs": {"pyerrors": {"tf": 3.605551275463989}, "pyerrors.fits.ks_test": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1.4142135623730951}, "pyerrors.input.json.dump_to_json": {"tf": 1.4142135623730951}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1.4142135623730951}, "pyerrors.input.pandas.read_sql": {"tf": 1}, "pyerrors.input.pandas.dump_df": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs": {"tf": 1}}, "df": 9}}}}}, "t": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.pandas.read_sql": {"tf": 1}}, "df": 1, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}}}}}}}, "r": {"docs": {"pyerrors": {"tf": 4.123105625617661}, "pyerrors.correlators.Corr": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.__init__": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.correlators.Corr.correlate": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.T_symmetry": {"tf": 1}, "pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.spaghetti_plot": {"tf": 1}, "pyerrors.covobs.Covobs.__init__": {"tf": 1.7320508075688772}, "pyerrors.fits.least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.fits.fit_lin": {"tf": 1}, "pyerrors.input": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.write_pobs": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.read_pobs": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.import_dobs_string": {"tf": 2}, "pyerrors.input.dobs.read_dobs": {"tf": 2}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.write_dobs": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.7320508075688772}, "pyerrors.input.json.create_json_string": {"tf": 1.4142135623730951}, "pyerrors.input.json.dump_to_json": {"tf": 1.4142135623730951}, "pyerrors.input.json.import_json_string": {"tf": 1.7320508075688772}, "pyerrors.input.json.load_json": {"tf": 1.7320508075688772}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1.4142135623730951}, "pyerrors.input.json.load_json_dict": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}, "pyerrors.input.pandas.to_sql": {"tf": 1}, "pyerrors.input.pandas.read_sql": {"tf": 1}, "pyerrors.input.utils.check_idl": {"tf": 1}, "pyerrors.linalg.matmul": {"tf": 1}, "pyerrors.linalg.jack_matmul": {"tf": 1}, "pyerrors.linalg.einsum": {"tf": 1}, "pyerrors.linalg.inv": {"tf": 1}, "pyerrors.misc.gen_correlated_data": {"tf": 1.4142135623730951}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}, "pyerrors.obs.Obs.__init__": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1.7320508075688772}, "pyerrors.obs.cov_Obs": {"tf": 2}}, "df": 44, "d": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 2.6457513110645907}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 5, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.reverse": {"tf": 1}}, "df": 2}}}, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.Eigenvalue": {"tf": 1}}, "df": 1}}}}}, "t": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}}, "df": 2}}}}}}}}, "i": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.dobs.create_pobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.write_pobs": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.write_dobs": {"tf": 1.4142135623730951}}, "df": 4}}}}}, "t": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors": {"tf": 2}, "pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 1}, "pyerrors.input": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 6, "w": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}}, "df": 4}}}}}}}}, "v": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}}, "df": 4, "l": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}}}}}}, "w": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}}, "df": 2, "s": {"docs": {"pyerrors.obs.Obs": {"tf": 1.7320508075688772}}, "df": 1}}}}}}}}}, "d": {"docs": {"pyerrors.input.json.dump_dict_to_json": {"tf": 1}}, "df": 1, "d": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}}, "u": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.thin": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix.g5H": {"tf": 1}}, "df": 3, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}, "p": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.fit": {"tf": 1.4142135623730951}, "pyerrors.fits.least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.total_least_squares": {"tf": 1.4142135623730951}, "pyerrors.input": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1.7320508075688772}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1.7320508075688772}, "pyerrors.input.json.import_json_string": {"tf": 1.4142135623730951}, "pyerrors.input.json.load_json": {"tf": 1.4142135623730951}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1.7320508075688772}, "pyerrors.input.json.load_json_dict": {"tf": 1}, "pyerrors.input.pandas.dump_df": {"tf": 1.7320508075688772}, "pyerrors.input.sfcf.read_sfcf": {"tf": 2}, "pyerrors.obs.Obs": {"tf": 1}, "pyerrors.obs.Obs.details": {"tf": 1}, "pyerrors.obs.Obs.dump": {"tf": 1}}, "df": 23, "s": {"docs": {"pyerrors.correlators.Corr.plottable": {"tf": 1.4142135623730951}}, "df": 1}}}}}, "r": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}, "m": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}}, "df": 2}}}}}}, "c": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}}, "df": 3}}}}}}, "p": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}}}}}}, "w": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "a": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}}}}}, "l": {"docs": {"pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}}, "df": 2}}, "m": {"docs": {"pyerrors": {"tf": 3.872983346207417}, "pyerrors.correlators.Corr.m_eff": {"tf": 2.449489742783178}, "pyerrors.fits.fit_lin": {"tf": 1.4142135623730951}}, "df": 3, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "k": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "v": {"docs": {"pyerrors": {"tf": 2}}, "df": 1}}}, "q": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}}, "df": 1, "t": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}}}}}}}, "t": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "x": {"docs": {"pyerrors": {"tf": 4.795831523312719}, "pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.correlators.Corr.item": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.Hankel": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.prune": {"tf": 2.8284271247461903}, "pyerrors.covobs.Covobs.__init__": {"tf": 1.4142135623730951}, "pyerrors.dirac.Grid_gamma": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 2.23606797749979}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.linalg.matmul": {"tf": 1}, "pyerrors.linalg.jack_matmul": {"tf": 1}, "pyerrors.linalg.eigh": {"tf": 1}, "pyerrors.linalg.eig": {"tf": 1}, "pyerrors.linalg.pinv": {"tf": 1}, "pyerrors.linalg.svd": {"tf": 1}, "pyerrors.misc.gen_correlated_data": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1.7320508075688772}, "pyerrors.obs.covariance": {"tf": 3}, "pyerrors.obs.cov_Obs": {"tf": 1.7320508075688772}}, "df": 22}, "c": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.correlators.Corr.is_matrix_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.matrix_symmetric": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.linalg.jack_matmul": {"tf": 1}, "pyerrors.linalg.inv": {"tf": 1}, "pyerrors.linalg.cholesky": {"tf": 1}, "pyerrors.linalg.det": {"tf": 1}}, "df": 9}}}, "z": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}}, "df": 1}}}}}, "h": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors": {"tf": 2}, "pyerrors.obs.Obs": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}}, "df": 3}}}}}}}}, "r": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "{": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.obs.covariance": {"tf": 1.7320508075688772}}, "df": 2}}}, "e": {"docs": {}, "df": 0, "x": {"docs": {}, "df": 0, "p": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}, "p": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "j": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1.4142135623730951}}, "df": 1}}}}, "t": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "c": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 1}}}}}}}}}, "c": {"docs": {}, "df": 0, "h": {"docs": {"pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1.4142135623730951}}, "df": 2, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}}, "df": 1}}}}}, "m": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.linalg.jack_matmul": {"tf": 1}}, "df": 1}}}}, "c": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 2}, "pyerrors.obs.CObs.is_zero": {"tf": 1}}, "df": 2}}}}}, "d": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}, "k": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.pandas.dump_df": {"tf": 1}, "pyerrors.linalg.jack_matmul": {"tf": 1}}, "df": 3}}}}, "n": {"docs": {"pyerrors.obs.derived_observable": {"tf": 1}}, "df": 1, "i": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}}}}, "y": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 5}, "u": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}}, "df": 3}}}}}}, "s": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 2}, "pyerrors.correlators.Corr.m_eff": {"tf": 1.7320508075688772}}, "df": 2, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}}}}, "y": {"docs": {"pyerrors": {"tf": 2.6457513110645907}, "pyerrors.correlators.Corr.prune": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 8}, "j": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}}, "df": 2}}}, "x": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.obs.derived_observable": {"tf": 1}}, "df": 1}}}}}}}}}}}}}}}, "o": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 3.4641016151377544}, "pyerrors.obs.Obs.__init__": {"tf": 1}, "pyerrors.obs.Obs.plot_history": {"tf": 1.4142135623730951}}, "df": 3}}, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.correlators.Corr.spaghetti_plot": {"tf": 1}}, "df": 1, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}}, "d": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}}, "df": 1, "s": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 2}}, "u": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 2}}}}, "r": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.plot_rep_dist": {"tf": 1}}, "df": 6}}, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.linalg.pinv": {"tf": 1}}, "df": 2}}}, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 1}}}}}}}, "m": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}}, "df": 4, "a": {"docs": {"pyerrors.input.hadrons.Npr_matrix.g5H": {"tf": 1}}, "df": 1}}}}}, "v": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1.7320508075688772}}, "df": 1}}}, "e": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 4.69041575982343}, "pyerrors.correlators.Corr.gamma_method": {"tf": 1}, "pyerrors.correlators.Corr.gm": {"tf": 1}, "pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.plateau": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1.4142135623730951}, "pyerrors.fits.Fit_result.gamma_method": {"tf": 1}, "pyerrors.fits.Fit_result.gm": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 2}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 2}, "pyerrors.input.pandas.read_sql": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs": {"tf": 1}, "pyerrors.obs.Obs.is_zero_within_error": {"tf": 1}, "pyerrors.obs.CObs.gamma_method": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1.4142135623730951}}, "df": 19, "s": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1}}, "df": 4}}}}, "a": {"docs": {"pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}}, "df": 4}}, "s": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 2.8284271247461903}}, "df": 2, "s": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1.4142135623730951}}, "df": 3}}}}, "a": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.covobs.Covobs.__init__": {"tf": 1.7320508075688772}, "pyerrors.misc.gen_correlated_data": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}, "pyerrors.obs.import_jackknife": {"tf": 1}, "pyerrors.obs.cov_Obs": {"tf": 1.7320508075688772}}, "df": 8, "s": {"docs": {"pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.misc.gen_correlated_data": {"tf": 1}, "pyerrors.obs.cov_Obs": {"tf": 1}}, "df": 4}}, "s": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.4142135623730951}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 5, "s": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop": {"tf": 2}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 2}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.4142135623730951}}, "df": 8}}}}}}}}}, "d": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}}, "df": 1}}, "m": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 2.8284271247461903}}, "df": 1}}}}}, "y": {"docs": {"pyerrors": {"tf": 7.681145747868608}}, "df": 1}, "u": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {"pyerrors": {"tf": 1.7320508075688772}}, "df": 1, "p": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 4}, "y": {"docs": {"pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.linalg.matmul": {"tf": 1}, "pyerrors.linalg.jack_matmul": {"tf": 1}}, "df": 3}, "i": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.linalg.matmul": {"tf": 1}}, "df": 1}}}}}}}}}, "d": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}}}}}}}}}}}}}, "c": {"docs": {}, "df": 0, "h": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}, "i": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.projected": {"tf": 1}}, "df": 2}}, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.fits.least_squares": {"tf": 1.4142135623730951}}, "df": 1}}}}, "n": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "z": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.roots.find_root": {"tf": 1}}, "df": 2}}}}}, "e": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}}, "df": 1, "d": {"docs": {"pyerrors.roots.find_root": {"tf": 1}}, "df": 1}}}}}}, "d": {"docs": {"pyerrors.obs.correlate": {"tf": 1}}, "df": 1}, "u": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}, "s": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.utils.check_idl": {"tf": 1}}, "df": 1, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.utils.check_idl": {"tf": 1}}, "df": 1}}}}}}, "\\": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1.7320508075688772}}, "df": 1}}}}}}, "p": {"docs": {}, "df": 0, "i": {"docs": {"pyerrors": {"tf": 2}}, "df": 1}}, "c": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1, "m": {"docs": {}, "df": 0, "c": {"docs": {"pyerrors.input.bdio.read_ADerrors": {"tf": 1}}, "df": 1}}}, "t": {"docs": {}, "df": 0, "h": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "s": {"1": {"docs": {"pyerrors.input.openQCD.read_rwms": {"tf": 1}}, "df": 1}, "docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 2}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 4}}, "d": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.roots.find_root": {"tf": 2.23606797749979}}, "df": 2, "a": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 4, "a": {"docs": {"pyerrors": {"tf": 4.898979485566356}, "pyerrors.correlators.Corr.__init__": {"tf": 1}, "pyerrors.correlators.Corr.fit": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.fits.residual_plot": {"tf": 1}, "pyerrors.input.bdio.read_ADerrors": {"tf": 1.7320508075688772}, "pyerrors.input.bdio.read_mesons": {"tf": 1.7320508075688772}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 2.23606797749979}, "pyerrors.input.dobs.import_dobs_string": {"tf": 2.449489742783178}, "pyerrors.input.dobs.read_dobs": {"tf": 2.449489742783178}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.write_dobs": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 3.1622776601683795}, "pyerrors.input.json.import_json_string": {"tf": 1.7320508075688772}, "pyerrors.input.json.load_json": {"tf": 1.7320508075688772}, "pyerrors.input.json.load_json_dict": {"tf": 2.8284271247461903}, "pyerrors.input.openQCD.read_rwms": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.extract_t0": {"tf": 2}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 2.449489742783178}, "pyerrors.input.pandas.read_sql": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1.4142135623730951}, "pyerrors.misc.gen_correlated_data": {"tf": 1.4142135623730951}, "pyerrors.mpm.matrix_pencil_method": {"tf": 2.449489742783178}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1.7320508075688772}}, "df": 33, "t": {"docs": {}, "df": 0, "y": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.dump": {"tf": 1}, "pyerrors.obs.Obs.dump": {"tf": 1}}, "df": 3}}}}, "s": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.misc.gen_correlated_data": {"tf": 1}}, "df": 2}}}, "f": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.pandas.to_sql": {"tf": 1.7320508075688772}, "pyerrors.input.pandas.read_sql": {"tf": 1.7320508075688772}, "pyerrors.input.pandas.dump_df": {"tf": 1.7320508075688772}, "pyerrors.input.pandas.load_df": {"tf": 1.7320508075688772}}, "df": 4}}}}}, "b": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.pandas.to_sql": {"tf": 2}, "pyerrors.input.pandas.read_sql": {"tf": 1.7320508075688772}, "pyerrors.input.pandas.load_df": {"tf": 1}}, "df": 3}}}}}, "e": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}}}, "i": {"docs": {}, "df": 0, "f": {"docs": {}, "df": 0, "f": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 2.8284271247461903}, "pyerrors.obs.Obs": {"tf": 1}}, "df": 2, "i": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 2.6457513110645907}, "pyerrors.fits.least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.total_least_squares": {"tf": 1.4142135623730951}, "pyerrors.obs.derived_observable": {"tf": 1.7320508075688772}}, "df": 4}}}}}}, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}}, "df": 2}}}}}}, "c": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 2, "s": {"docs": {"pyerrors.correlators.Corr.deriv": {"tf": 1}, "pyerrors.correlators.Corr.second_deriv": {"tf": 1}}, "df": 2}}}}}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}}}}}, "c": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.read_dobs": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1.4142135623730951}, "pyerrors.input.json.import_json_string": {"tf": 1.4142135623730951}, "pyerrors.input.json.load_json": {"tf": 1.4142135623730951}, "pyerrors.input.json.dump_dict_to_json": {"tf": 2}, "pyerrors.input.json.load_json_dict": {"tf": 2}, "pyerrors.obs.Obs": {"tf": 2.449489742783178}}, "df": 16, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1.7320508075688772}}, "df": 1}}}, "y": {"docs": {"pyerrors.input.bdio.read_mesons": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.read_dSdm": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.plot_piechart": {"tf": 1}}, "df": 5}}}}}}, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "[": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}}, "df": 1}}}}}}, "s": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "z": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}}}}}}}}}}}, "o": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}}, "u": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 1}}}}}}}}, "t": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 2}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 2}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 2}}, "df": 6}}}}, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.fits.qqplot": {"tf": 1}}, "df": 1}}, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.obs.Obs.plot_rep_dist": {"tf": 1}}, "df": 1}}}}}}}}, "i": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}}, "df": 1}}}}}, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}}, "df": 1}}}}}}}}}}}}}}}}}}}, "p": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "y": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1}}, "df": 1}}, "s": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.residual_plot": {"tf": 1}}, "df": 2}}}}}, "k": {"docs": {"pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}}, "df": 3}}, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}}, "df": 2, "l": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}}, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1.7320508075688772}}, "df": 2}, "i": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}}, "df": 1}}}}}}}}}, "d": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "m": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.Hankel": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 3, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors": {"tf": 2}, "pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 3}}, "s": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 2}}}}}}}}, "a": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1.4142135623730951}, "pyerrors.obs.cov_Obs": {"tf": 1}}, "df": 4}}}}, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1.4142135623730951}}, "df": 1}}}}}}, "v": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 1}}}}}}, "e": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 2}, "pyerrors.linalg.cholesky": {"tf": 1}, "pyerrors.linalg.svd": {"tf": 1}}, "df": 3}}}}}, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}}, "df": 1}}}}}}}, "i": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.correlators.Corr.deriv": {"tf": 1}, "pyerrors.correlators.Corr.second_deriv": {"tf": 1}, "pyerrors.correlators.Corr.fit": {"tf": 1}}, "df": 3}}}}, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 1}}}}}}, "a": {"docs": {}, "df": 0, "y": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1.4142135623730951}}, "df": 1}}}}}, "t": {"docs": {"pyerrors": {"tf": 1}}, "df": 1, "a": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 2}, "pyerrors.obs.Obs.details": {"tf": 1}}, "df": 2}}, "s": {"docs": {"pyerrors": {"tf": 3.3166247903554}, "pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.input": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.linalg.einsum": {"tf": 1}, "pyerrors.obs.Obs.details": {"tf": 1}, "pyerrors.obs.Obs.is_zero": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 9}}}}, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 1, "d": {"docs": {"pyerrors": {"tf": 1.7320508075688772}}, "df": 1}, "s": {"docs": {"pyerrors.correlators.Corr.Eigenvalue": {"tf": 1}, "pyerrors.correlators.Corr.Hankel": {"tf": 1}, "pyerrors.correlators.Corr.spaghetti_plot": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}}, "df": 6}}, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.linalg.det": {"tf": 1}}, "df": 2}}}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1}}, "df": 1}}}}}}}, "c": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 2}}}}}}}, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "v": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 2.8284271247461903}, "pyerrors.obs.Obs.plot_history": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1.4142135623730951}, "pyerrors.obs.derived_observable": {"tf": 1.4142135623730951}}, "df": 4}}, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.correlators.Corr.deriv": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.second_deriv": {"tf": 1.4142135623730951}, "pyerrors.obs.derived_observable": {"tf": 1}}, "df": 4, "s": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.obs.derived_observable": {"tf": 1}}, "df": 2}}}}}}}}}, "l": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "a": {"docs": {"pyerrors": {"tf": 2}}, "df": 1, "s": {"docs": {"pyerrors": {"tf": 1.7320508075688772}}, "df": 1}}}}, "s": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}, "c": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}}, "df": 4, "d": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}}, "df": 2}, "s": {"docs": {"pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}}, "df": 8}}}, "p": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.obs.Obs.dump": {"tf": 1.4142135623730951}}, "df": 5}}, "v": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}}, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.correlators.Corr.GEVP": {"tf": 1}}, "df": 1}}}}}}}}, "f": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.fits.least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.total_least_squares": {"tf": 1.4142135623730951}, "pyerrors.roots.find_root": {"tf": 1}}, "df": 4, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 2.449489742783178}, "pyerrors.correlators.Corr.deriv": {"tf": 1}, "pyerrors.correlators.Corr.second_deriv": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}}, "df": 5}}}, "e": {"docs": {"pyerrors.obs.covariance": {"tf": 1.4142135623730951}}, "df": 1}}}, "e": {"docs": {"pyerrors": {"tf": 2}}, "df": 1, "d": {"docs": {"pyerrors": {"tf": 3.3166247903554}, "pyerrors.correlators.Corr.reweight": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.openQCD.qtop_projection": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.obs.Obs.__init__": {"tf": 1}, "pyerrors.obs.Obs.reweight": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1}, "pyerrors.obs.reweight": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1.4142135623730951}, "pyerrors.obs.import_jackknife": {"tf": 1}}, "df": 17}, "s": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 3}}}}, "a": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.correlators.Corr.deriv": {"tf": 1}, "pyerrors.correlators.Corr.second_deriv": {"tf": 1}, "pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.correlators.Corr.dump": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1.4142135623730951}, "pyerrors.fits.least_squares": {"tf": 2.23606797749979}, "pyerrors.fits.total_least_squares": {"tf": 1.7320508075688772}, "pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 2}, "pyerrors.input.bdio.read_dSdm": {"tf": 2}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.qtop_projection": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.4142135623730951}, "pyerrors.input.pandas.read_sql": {"tf": 1.4142135623730951}, "pyerrors.input.pandas.load_df": {"tf": 1.4142135623730951}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.7320508075688772}, "pyerrors.misc.dump_object": {"tf": 1}, "pyerrors.misc.pseudo_Obs": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.gamma_method": {"tf": 2}, "pyerrors.obs.Obs.gm": {"tf": 2}, "pyerrors.obs.Obs.reweight": {"tf": 1}, "pyerrors.obs.Obs.plot_history": {"tf": 1}, "pyerrors.obs.Obs.dump": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}, "pyerrors.obs.reweight": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1.4142135623730951}}, "df": 35, "s": {"docs": {"pyerrors.correlators.Corr.plateau": {"tf": 1}}, "df": 1}}}}}}, "p": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1}}, "df": 1, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr": {"tf": 1}}, "df": 2}, "c": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.correlators.Corr": {"tf": 1}}, "df": 1}}}}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.openQCD.read_qtop": {"tf": 1}}, "df": 1}}}}}}}, "n": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}, "v": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1}}, "df": 1, "s": {"docs": {"pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}}, "df": 2}}}}}}}}, "d": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr": {"tf": 1}}, "df": 1}}}}}}}, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.correlators.Corr": {"tf": 1}}, "df": 1}}}}}, "b": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.openQCD.read_rwms": {"tf": 1}}, "df": 1}}}}}}}}, "o": {"docs": {"pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 4, "w": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}}, "df": 3}}, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 7}}, "m": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "d": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}}, "df": 1}}}}}}}}, "b": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}}, "df": 5}}, "t": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 1}}, "f": {"docs": {"pyerrors.fits.Fit_result": {"tf": 1}}, "df": 1}, "u": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 3}}}}, "n": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.4142135623730951}}, "df": 2}}, "c": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.linalg.einsum": {"tf": 1}, "pyerrors.obs.Obs.is_zero": {"tf": 1}}, "df": 2}}}}}}}}}}}}, "u": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "p": {"docs": {"pyerrors": {"tf": 2}, "pyerrors.misc.dump_object": {"tf": 1}, "pyerrors.obs.Obs.dump": {"tf": 1}}, "df": 3, "s": {"docs": {"pyerrors.correlators.Corr.dump": {"tf": 1}}, "df": 1}, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.pandas.dump_df": {"tf": 1}}, "df": 1}}}}, "e": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 1}}, "t": {"docs": {"pyerrors.correlators.Corr.roll": {"tf": 1.4142135623730951}}, "df": 1, "y": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 3.1622776601683795}}, "df": 1}}}, "r": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.4142135623730951}}, "df": 4}}, "v": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.misc.pseudo_Obs": {"tf": 1.4142135623730951}}, "df": 1, "s": {"docs": {"pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.fits.fit_lin": {"tf": 1.4142135623730951}}, "df": 2}}}}}}, "s": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "m": {"docs": {"pyerrors.input.bdio.read_dSdm": {"tf": 1.4142135623730951}}, "df": 1}}}, "f": {"docs": {"pyerrors.input.pandas.to_sql": {"tf": 1}, "pyerrors.input.pandas.dump_df": {"tf": 1}}, "df": 2}, "b": {"docs": {"pyerrors.input.pandas.to_sql": {"tf": 1}, "pyerrors.input.pandas.read_sql": {"tf": 1}}, "df": 2}}, "b": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.obs.correlate": {"tf": 1}}, "df": 2, "a": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1.7320508075688772}}, "df": 1, "d": {"docs": {"pyerrors": {"tf": 3}, "pyerrors.correlators.Corr.prune": {"tf": 1}, "pyerrors.dirac.epsilon_tensor": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}, "pyerrors.obs.merge_obs": {"tf": 1}, "pyerrors.obs.cov_Obs": {"tf": 1}}, "df": 9}, "m": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "x": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1.7320508075688772}}, "df": 1}}}}}}}, "i": {"docs": {}, "df": 0, "c": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1}}, "df": 2}}}, "r": {"docs": {"pyerrors": {"tf": 1}}, "df": 1, "{": {"docs": {}, "df": 0, "f": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}, "a": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}, "c": {"docs": {}, "df": 0, "k": {"docs": {"pyerrors.correlators.Corr.__init__": {"tf": 1}}, "df": 1, "w": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.deriv": {"tf": 1}}, "df": 1}}}}}}, "n": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.fits.error_band": {"tf": 1.4142135623730951}}, "df": 1}}}, "e": {"docs": {"pyerrors": {"tf": 6.082762530298219}, "pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.correlators.Corr.item": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.correlate": {"tf": 1}, "pyerrors.correlators.Corr.reweight": {"tf": 1}, "pyerrors.correlators.Corr.T_symmetry": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.dump": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1.7320508075688772}, "pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 2.6457513110645907}, "pyerrors.fits.total_least_squares": {"tf": 1.7320508075688772}, "pyerrors.fits.fit_lin": {"tf": 1}, "pyerrors.fits.qqplot": {"tf": 1}, "pyerrors.input.bdio.read_ADerrors": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.read_mesons": {"tf": 2.23606797749979}, "pyerrors.input.bdio.read_dSdm": {"tf": 2.23606797749979}, "pyerrors.input.dobs.create_pobs_string": {"tf": 2}, "pyerrors.input.dobs.write_pobs": {"tf": 2}, "pyerrors.input.dobs.create_dobs_string": {"tf": 2}, "pyerrors.input.dobs.write_dobs": {"tf": 2}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1.7320508075688772}, "pyerrors.input.json.dump_to_json": {"tf": 1.7320508075688772}, "pyerrors.input.json.dump_dict_to_json": {"tf": 2}, "pyerrors.input.json.load_json_dict": {"tf": 1}, "pyerrors.input.misc.read_pbp": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_rwms": {"tf": 2}, "pyerrors.input.openQCD.extract_t0": {"tf": 2.449489742783178}, "pyerrors.input.openQCD.read_qtop": {"tf": 2.449489742783178}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 2.449489742783178}, "pyerrors.input.openQCD.qtop_projection": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 2.6457513110645907}, "pyerrors.input.pandas.to_sql": {"tf": 1}, "pyerrors.input.pandas.read_sql": {"tf": 1}, "pyerrors.input.pandas.dump_df": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 2.23606797749979}, "pyerrors.input.utils.check_idl": {"tf": 1}, "pyerrors.linalg.matmul": {"tf": 1}, "pyerrors.linalg.jack_matmul": {"tf": 1}, "pyerrors.linalg.einsum": {"tf": 1}, "pyerrors.misc.dump_object": {"tf": 1}, "pyerrors.misc.pseudo_Obs": {"tf": 1.7320508075688772}, "pyerrors.misc.gen_correlated_data": {"tf": 2}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.reweight": {"tf": 1}, "pyerrors.obs.Obs.dump": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1.7320508075688772}, "pyerrors.obs.reweight": {"tf": 1}, "pyerrors.obs.correlate": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1.7320508075688772}, "pyerrors.obs.merge_obs": {"tf": 1}, "pyerrors.roots.find_root": {"tf": 1}}, "df": 59, "t": {"docs": {}, "df": 0, "w": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop": {"tf": 2.23606797749979}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 2.23606797749979}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 2.23606797749979}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 10}}}}, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}}, "df": 1}}}, "a": {"docs": {"pyerrors.fits.error_band": {"tf": 1}}, "df": 1}}, "c": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 2}}}}}, "l": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "w": {"docs": {"pyerrors": {"tf": 2}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 3}, "n": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 3}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.covobs.Covobs.__init__": {"tf": 1.4142135623730951}, "pyerrors.obs.cov_Obs": {"tf": 1}}, "df": 3}}}}}}}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}, "e": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.obs.correlate": {"tf": 1}}, "df": 2}}, "f": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.input.pandas.dump_df": {"tf": 1}}, "df": 2}}}}, "h": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.pandas.to_sql": {"tf": 1}}, "df": 1}}}}}, "y": {"docs": {"pyerrors": {"tf": 3}, "pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.roll": {"tf": 1}, "pyerrors.correlators.Corr.reweight": {"tf": 1}, "pyerrors.correlators.Corr.m_eff": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.prune": {"tf": 1.7320508075688772}, "pyerrors.fits.Fit_result": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.total_least_squares": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.read_dobs": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.openQCD.qtop_projection": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.obs.Obs.reweight": {"tf": 1}, "pyerrors.obs.reweight": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1.4142135623730951}}, "df": 24, "t": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1, "s": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 2.23606797749979}}, "df": 1}}}}, "u": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 8}, "i": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "d": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.input.bdio.read_ADerrors": {"tf": 1.7320508075688772}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1.7320508075688772}, "pyerrors.input.bdio.read_mesons": {"tf": 1.7320508075688772}, "pyerrors.input.bdio.read_dSdm": {"tf": 1.7320508075688772}}, "df": 4}}}}}}, "f": {"docs": {}, "df": 0, "f": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 3.1622776601683795}}, "df": 1}}}}, "l": {"docs": {}, "df": 0, "k": {"docs": {"pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}}, "df": 1}}}, "o": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 1, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 2.6457513110645907}}, "df": 4}}}}}, "d": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1.4142135623730951}}, "df": 1}}}}}, "t": {"docs": {}, "df": 0, "h": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.linalg.jack_matmul": {"tf": 1}, "pyerrors.obs.CObs.is_zero": {"tf": 1}}, "df": 3}}, "o": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.Hankel": {"tf": 1}, "pyerrors.correlators.Corr.reweight": {"tf": 1}, "pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.spaghetti_plot": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 2.449489742783178}, "pyerrors.fits.total_least_squares": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.read_dobs": {"tf": 2}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.import_json_string": {"tf": 1.4142135623730951}, "pyerrors.input.json.load_json": {"tf": 1.7320508075688772}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.pandas.to_sql": {"tf": 1}, "pyerrors.input.pandas.read_sql": {"tf": 1}, "pyerrors.input.pandas.dump_df": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1.4142135623730951}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}, "pyerrors.obs.Obs.details": {"tf": 1}, "pyerrors.obs.Obs.reweight": {"tf": 1}, "pyerrors.obs.Obs.plot_history": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}, "pyerrors.obs.reweight": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1.4142135623730951}}, "df": 37}, "k": {"docs": {}, "df": 0, "k": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}}, "df": 5}}}}}}}}}, "x": {"docs": {"pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}}, "df": 1}}, "d": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {"pyerrors.input.bdio.read_ADerrors": {"tf": 2.6457513110645907}, "pyerrors.input.bdio.write_ADerrors": {"tf": 2.6457513110645907}, "pyerrors.input.bdio.read_mesons": {"tf": 2.6457513110645907}, "pyerrors.input.bdio.read_dSdm": {"tf": 2.6457513110645907}}, "df": 4}}}, "i": {"docs": {"pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 1, "l": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}}, "df": 1, "s": {"docs": {"pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}}, "df": 1}}}}}}}, "n": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}}, "df": 1}}}}, "b": {"docs": {"pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 1}}, "b": {"docs": {"pyerrors.input.sfcf.read_sfcf": {"tf": 1.4142135623730951}}, "df": 1}}, "t": {"0": {"docs": {"pyerrors.correlators.Corr.GEVP": {"tf": 2}, "pyerrors.input.openQCD.extract_t0": {"tf": 2.449489742783178}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 5, "p": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "j": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1.4142135623730951}}, "df": 1}}}}}, "2": {"docs": {"pyerrors.fits.Fit_result": {"tf": 1}}, "df": 1}, "docs": {"pyerrors": {"tf": 2.8284271247461903}, "pyerrors.correlators.Corr.GEVP": {"tf": 2.8284271247461903}, "pyerrors.correlators.Corr.Hankel": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.m_eff": {"tf": 2.8284271247461903}, "pyerrors.correlators.Corr.prune": {"tf": 4.47213595499958}, "pyerrors.fits.Fit_result": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 10, "h": {"docs": {"pyerrors.correlators.Corr.thin": {"tf": 1}}, "df": 1, "e": {"docs": {"pyerrors": {"tf": 16.15549442140351}, "pyerrors.correlators.Corr": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.__init__": {"tf": 3}, "pyerrors.correlators.Corr.gamma_method": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.gm": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.projected": {"tf": 2}, "pyerrors.correlators.Corr.item": {"tf": 1}, "pyerrors.correlators.Corr.plottable": {"tf": 2}, "pyerrors.correlators.Corr.symmetric": {"tf": 1}, "pyerrors.correlators.Corr.anti_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.matrix_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 4.58257569495584}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 2.449489742783178}, "pyerrors.correlators.Corr.Hankel": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.roll": {"tf": 1}, "pyerrors.correlators.Corr.reverse": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.thin": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.correlate": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.reweight": {"tf": 2.449489742783178}, "pyerrors.correlators.Corr.T_symmetry": {"tf": 2}, "pyerrors.correlators.Corr.deriv": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.second_deriv": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.m_eff": {"tf": 3.1622776601683795}, "pyerrors.correlators.Corr.fit": {"tf": 2.449489742783178}, "pyerrors.correlators.Corr.plateau": {"tf": 2.6457513110645907}, "pyerrors.correlators.Corr.set_prange": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.show": {"tf": 3.4641016151377544}, "pyerrors.correlators.Corr.spaghetti_plot": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.dump": {"tf": 2}, "pyerrors.correlators.Corr.prune": {"tf": 4.795831523312719}, "pyerrors.covobs.Covobs.__init__": {"tf": 2.23606797749979}, "pyerrors.covobs.Covobs.errsq": {"tf": 1.7320508075688772}, "pyerrors.dirac.epsilon_tensor": {"tf": 1}, "pyerrors.dirac.epsilon_tensor_rank4": {"tf": 1}, "pyerrors.fits.Fit_result": {"tf": 1.4142135623730951}, "pyerrors.fits.Fit_result.gamma_method": {"tf": 1}, "pyerrors.fits.Fit_result.gm": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 4.898979485566356}, "pyerrors.fits.total_least_squares": {"tf": 3.7416573867739413}, "pyerrors.fits.fit_lin": {"tf": 2.449489742783178}, "pyerrors.fits.qqplot": {"tf": 1.7320508075688772}, "pyerrors.fits.residual_plot": {"tf": 1.7320508075688772}, "pyerrors.fits.error_band": {"tf": 1}, "pyerrors.fits.ks_test": {"tf": 1.4142135623730951}, "pyerrors.input": {"tf": 1}, "pyerrors.input.bdio.read_ADerrors": {"tf": 2}, "pyerrors.input.bdio.write_ADerrors": {"tf": 2}, "pyerrors.input.bdio.read_mesons": {"tf": 2.6457513110645907}, "pyerrors.input.bdio.read_dSdm": {"tf": 2.6457513110645907}, "pyerrors.input.dobs.create_pobs_string": {"tf": 3.605551275463989}, "pyerrors.input.dobs.write_pobs": {"tf": 3.872983346207417}, "pyerrors.input.dobs.read_pobs": {"tf": 3}, "pyerrors.input.dobs.import_dobs_string": {"tf": 3.605551275463989}, "pyerrors.input.dobs.read_dobs": {"tf": 3.605551275463989}, "pyerrors.input.dobs.create_dobs_string": {"tf": 4.58257569495584}, "pyerrors.input.dobs.write_dobs": {"tf": 4.58257569495584}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 3.3166247903554}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 2}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 5.830951894845301}, "pyerrors.input.hadrons.Npr_matrix.g5H": {"tf": 2}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 2}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 2}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 2}, "pyerrors.input.json.create_json_string": {"tf": 2.8284271247461903}, "pyerrors.input.json.dump_to_json": {"tf": 3}, "pyerrors.input.json.import_json_string": {"tf": 3}, "pyerrors.input.json.load_json": {"tf": 3}, "pyerrors.input.json.dump_dict_to_json": {"tf": 3.3166247903554}, "pyerrors.input.json.load_json_dict": {"tf": 2.6457513110645907}, "pyerrors.input.misc.read_pbp": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_rwms": {"tf": 3}, "pyerrors.input.openQCD.extract_t0": {"tf": 5.385164807134504}, "pyerrors.input.openQCD.read_qtop": {"tf": 4.58257569495584}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 4.47213595499958}, "pyerrors.input.openQCD.qtop_projection": {"tf": 2}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 4.358898943540674}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 4.47213595499958}, "pyerrors.input.pandas.to_sql": {"tf": 2.23606797749979}, "pyerrors.input.pandas.read_sql": {"tf": 2.449489742783178}, "pyerrors.input.pandas.dump_df": {"tf": 2}, "pyerrors.input.pandas.load_df": {"tf": 2.449489742783178}, "pyerrors.input.sfcf.read_sfcf": {"tf": 4.58257569495584}, "pyerrors.input.utils": {"tf": 1}, "pyerrors.input.utils.check_idl": {"tf": 1}, "pyerrors.linalg.matmul": {"tf": 1}, "pyerrors.linalg.jack_matmul": {"tf": 1}, "pyerrors.linalg.eigh": {"tf": 1}, "pyerrors.linalg.eig": {"tf": 1}, "pyerrors.linalg.pinv": {"tf": 1}, "pyerrors.linalg.svd": {"tf": 1}, "pyerrors.misc.dump_object": {"tf": 1.7320508075688772}, "pyerrors.misc.load_object": {"tf": 1}, "pyerrors.misc.pseudo_Obs": {"tf": 2.23606797749979}, "pyerrors.misc.gen_correlated_data": {"tf": 1.7320508075688772}, "pyerrors.mpm.matrix_pencil_method": {"tf": 2.23606797749979}, "pyerrors.obs.Obs": {"tf": 3.1622776601683795}, "pyerrors.obs.Obs.__init__": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.gamma_method": {"tf": 3.4641016151377544}, "pyerrors.obs.Obs.gm": {"tf": 3.4641016151377544}, "pyerrors.obs.Obs.details": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.reweight": {"tf": 2.449489742783178}, "pyerrors.obs.Obs.is_zero_within_error": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.is_zero": {"tf": 1}, "pyerrors.obs.Obs.plot_tauint": {"tf": 1}, "pyerrors.obs.Obs.plot_rho": {"tf": 1}, "pyerrors.obs.Obs.plot_piechart": {"tf": 2}, "pyerrors.obs.Obs.dump": {"tf": 2}, "pyerrors.obs.Obs.export_jackknife": {"tf": 3.3166247903554}, "pyerrors.obs.CObs.gamma_method": {"tf": 1.7320508075688772}, "pyerrors.obs.derived_observable": {"tf": 2.8284271247461903}, "pyerrors.obs.reweight": {"tf": 2.23606797749979}, "pyerrors.obs.correlate": {"tf": 1.4142135623730951}, "pyerrors.obs.covariance": {"tf": 5.291502622129181}, "pyerrors.obs.import_jackknife": {"tf": 2}, "pyerrors.obs.merge_obs": {"tf": 1.4142135623730951}, "pyerrors.obs.cov_Obs": {"tf": 2}, "pyerrors.roots.find_root": {"tf": 2.449489742783178}}, "df": 117, "i": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}}, "r": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1.4142135623730951}}, "df": 3, "f": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}, "m": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "z": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1.4142135623730951}}, "df": 1}}}}}}}}}}}, "s": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}}, "df": 6}}, "n": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 4}, "y": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1}}, "df": 2}, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 1}}}, "e": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 1}}}}}}}}}}}, "a": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 6.244997998398398}, "pyerrors.correlators.Corr.reweight": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1.7320508075688772}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 2}, "pyerrors.input.dobs.write_pobs": {"tf": 2}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.create_dobs_string": {"tf": 2}, "pyerrors.input.dobs.write_dobs": {"tf": 2}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 2}, "pyerrors.input.hadrons.Npr_matrix.g5H": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1.4142135623730951}, "pyerrors.input.json.dump_to_json": {"tf": 1.4142135623730951}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1.4142135623730951}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1.4142135623730951}, "pyerrors.input.json.load_json_dict": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_rwms": {"tf": 2.23606797749979}, "pyerrors.input.openQCD.extract_t0": {"tf": 2.23606797749979}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.7320508075688772}, "pyerrors.input.pandas.load_df": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}, "pyerrors.obs.Obs.reweight": {"tf": 1}, "pyerrors.obs.reweight": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1.7320508075688772}}, "df": 33}, "n": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.obs.Obs.plot_rep_dist": {"tf": 1}}, "df": 2}}, "i": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 4}, "pyerrors.correlators.Corr": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.__init__": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 2}, "pyerrors.fits.least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.7320508075688772}, "pyerrors.linalg.matmul": {"tf": 1}, "pyerrors.linalg.jack_matmul": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}, "pyerrors.obs.correlate": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1.4142135623730951}}, "df": 27}, "n": {"docs": {"pyerrors.correlators.Corr.thin": {"tf": 1}}, "df": 1}}, "r": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "h": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}}}}, "e": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.plottable": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.fit": {"tf": 1}}, "df": 3}}}}, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}, "u": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.correlators.Corr.reweight": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 2.23606797749979}, "pyerrors.fits.total_least_squares": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.read_dobs": {"tf": 2}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.import_json_string": {"tf": 1.4142135623730951}, "pyerrors.input.json.load_json": {"tf": 1.7320508075688772}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.pandas.to_sql": {"tf": 1}, "pyerrors.input.pandas.read_sql": {"tf": 1}, "pyerrors.input.pandas.dump_df": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1.4142135623730951}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}, "pyerrors.obs.Obs.details": {"tf": 1}, "pyerrors.obs.Obs.reweight": {"tf": 1}, "pyerrors.obs.Obs.plot_history": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}, "pyerrors.obs.reweight": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1.4142135623730951}}, "df": 32}}, "a": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "k": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "n": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 1, "p": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}}}}}, "j": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 1, "e": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.4142135623730951}}, "df": 4}}}}}}}}}}, "i": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}}, "s": {"docs": {"pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}}, "df": 2}}}}}}}, "o": {"docs": {"pyerrors": {"tf": 8.660254037844387}, "pyerrors.correlators.Corr": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.gamma_method": {"tf": 1}, "pyerrors.correlators.Corr.gm": {"tf": 1}, "pyerrors.correlators.Corr.projected": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.item": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.GEVP": {"tf": 2.6457513110645907}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 1}, "pyerrors.correlators.Corr.thin": {"tf": 1}, "pyerrors.correlators.Corr.correlate": {"tf": 1}, "pyerrors.correlators.Corr.reweight": {"tf": 1}, "pyerrors.correlators.Corr.deriv": {"tf": 1}, "pyerrors.correlators.Corr.second_deriv": {"tf": 1}, "pyerrors.correlators.Corr.fit": {"tf": 2.6457513110645907}, "pyerrors.correlators.Corr.plateau": {"tf": 2}, "pyerrors.correlators.Corr.show": {"tf": 2.23606797749979}, "pyerrors.correlators.Corr.spaghetti_plot": {"tf": 1}, "pyerrors.correlators.Corr.dump": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 3.1622776601683795}, "pyerrors.covobs.Covobs.__init__": {"tf": 1.7320508075688772}, "pyerrors.fits.Fit_result.gamma_method": {"tf": 1}, "pyerrors.fits.Fit_result.gm": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 2.8284271247461903}, "pyerrors.fits.total_least_squares": {"tf": 2.6457513110645907}, "pyerrors.fits.fit_lin": {"tf": 1}, "pyerrors.fits.qqplot": {"tf": 1}, "pyerrors.fits.residual_plot": {"tf": 1}, "pyerrors.fits.ks_test": {"tf": 1}, "pyerrors.input.bdio.read_ADerrors": {"tf": 2.23606797749979}, "pyerrors.input.bdio.write_ADerrors": {"tf": 2.6457513110645907}, "pyerrors.input.bdio.read_mesons": {"tf": 2.6457513110645907}, "pyerrors.input.bdio.read_dSdm": {"tf": 2.6457513110645907}, "pyerrors.input.dobs.create_pobs_string": {"tf": 2.23606797749979}, "pyerrors.input.dobs.write_pobs": {"tf": 2.23606797749979}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.read_dobs": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.create_dobs_string": {"tf": 2}, "pyerrors.input.dobs.write_dobs": {"tf": 2}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 2.449489742783178}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1.7320508075688772}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 3.7416573867739413}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1.7320508075688772}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1.7320508075688772}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 2}, "pyerrors.input.json.create_json_string": {"tf": 1.7320508075688772}, "pyerrors.input.json.dump_to_json": {"tf": 1.4142135623730951}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1.7320508075688772}, "pyerrors.input.json.load_json_dict": {"tf": 1.4142135623730951}, "pyerrors.input.misc.read_pbp": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_rwms": {"tf": 2.6457513110645907}, "pyerrors.input.openQCD.extract_t0": {"tf": 3.4641016151377544}, "pyerrors.input.openQCD.read_qtop": {"tf": 2.8284271247461903}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 3}, "pyerrors.input.openQCD.qtop_projection": {"tf": 2}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 3.605551275463989}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 2.8284271247461903}, "pyerrors.input.pandas.to_sql": {"tf": 2.23606797749979}, "pyerrors.input.pandas.read_sql": {"tf": 1.7320508075688772}, "pyerrors.input.pandas.dump_df": {"tf": 2}, "pyerrors.input.pandas.load_df": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 3.3166247903554}, "pyerrors.input.utils.check_idl": {"tf": 1}, "pyerrors.linalg.matmul": {"tf": 1}, "pyerrors.linalg.jack_matmul": {"tf": 1}, "pyerrors.linalg.eigh": {"tf": 1}, "pyerrors.linalg.eig": {"tf": 1}, "pyerrors.misc.dump_object": {"tf": 1}, "pyerrors.misc.load_object": {"tf": 1}, "pyerrors.misc.pseudo_Obs": {"tf": 1.7320508075688772}, "pyerrors.misc.gen_correlated_data": {"tf": 1.7320508075688772}, "pyerrors.mpm.matrix_pencil_method": {"tf": 2}, "pyerrors.obs.Obs.gamma_method": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.gm": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.reweight": {"tf": 1}, "pyerrors.obs.Obs.plot_tauint": {"tf": 1}, "pyerrors.obs.Obs.plot_rho": {"tf": 1}, "pyerrors.obs.Obs.plot_piechart": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.dump": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1.4142135623730951}, "pyerrors.obs.derived_observable": {"tf": 2.23606797749979}, "pyerrors.obs.reweight": {"tf": 1}, "pyerrors.obs.correlate": {"tf": 1.4142135623730951}, "pyerrors.obs.covariance": {"tf": 2.8284271247461903}, "pyerrors.obs.import_jackknife": {"tf": 1}, "pyerrors.obs.merge_obs": {"tf": 1.4142135623730951}, "pyerrors.obs.cov_Obs": {"tf": 1}, "pyerrors.roots.find_root": {"tf": 1.7320508075688772}}, "df": 89, "g": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 3}}}}}}, "t": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 2}}}, "d": {"docs": {}, "df": 0, "o": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}, "p": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.input.openQCD.read_qtop": {"tf": 1}}, "df": 1}}, "c": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.qtop_projection": {"tf": 2}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.7320508075688772}}, "df": 4}}}}}}}}}, "w": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.openQCD.read_qtop": {"tf": 1}}, "df": 1}}}}}, "l": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.obs.Obs.is_zero": {"tf": 1.4142135623730951}}, "df": 1}}}}}}}}, "w": {"docs": {}, "df": 0, "o": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.correlators.Corr.__init__": {"tf": 1}, "pyerrors.correlators.Corr.projected": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.fits.fit_lin": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1.7320508075688772}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop": {"tf": 2.23606797749979}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 2.23606797749979}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 2.23606797749979}, "pyerrors.obs.derived_observable": {"tf": 1}, "pyerrors.obs.correlate": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 20}}, "a": {"docs": {}, "df": 0, "k": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.fits.total_least_squares": {"tf": 1}}, "df": 1, "n": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.prune": {"tf": 1}, "pyerrors.covobs.Covobs.__init__": {"tf": 1}}, "df": 3}, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}, "y": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}, "u": {"docs": {"pyerrors": {"tf": 2.6457513110645907}, "pyerrors.misc.gen_correlated_data": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 2}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1.4142135623730951}}, "df": 6, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}}}}, "i": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}}, "df": 3, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}, "g": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.show": {"tf": 1}}, "df": 2, "s": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}}, "df": 8}}, "r": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1.7320508075688772}, "pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}, "pyerrors.input.openQCD.qtop_projection": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.4142135623730951}}, "df": 7}}}}, "b": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.pandas.to_sql": {"tf": 1.7320508075688772}}, "df": 1}}}}, "i": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 2.6457513110645907}, "pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.reverse": {"tf": 1}, "pyerrors.correlators.Corr.T_symmetry": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.prune": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.obs.Obs.plot_tauint": {"tf": 1}, "pyerrors.obs.Obs.plot_rho": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 10, "s": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1.4142135623730951}, "pyerrors.misc.gen_correlated_data": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 4, "l": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.__init__": {"tf": 1}, "pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.correlators.Corr.item": {"tf": 1}, "pyerrors.correlators.Corr.plottable": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.is_matrix_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.matrix_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 11, "s": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.correlators.Corr.roll": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.fit": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.plateau": {"tf": 1}}, "df": 5}}}}}}}}, "t": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1.4142135623730951}}, "df": 1}}}}, "e": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}, "r": {"docs": {}, "df": 0, "m": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}}}}, "s": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 2.449489742783178}, "pyerrors.fits.ks_test": {"tf": 1}, "pyerrors.misc.pseudo_Obs": {"tf": 1}}, "df": 3}}, "n": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.dirac.epsilon_tensor": {"tf": 1.4142135623730951}, "pyerrors.dirac.epsilon_tensor_rank4": {"tf": 1.4142135623730951}}, "df": 2}}}}}, "y": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 2}, "pyerrors.correlators.Corr.dump": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 2.8284271247461903}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.7320508075688772}}, "df": 6, "s": {"docs": {"pyerrors.correlators.Corr": {"tf": 1}}, "df": 1}}, "i": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 1, "l": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}}, "df": 1}}}}}, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}}}}, "s": {"docs": {"pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 2}, "+": {"1": {"docs": {"pyerrors.correlators.Corr.Hankel": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.m_eff": {"tf": 2}}, "df": 2}, "2": {"docs": {"pyerrors.correlators.Corr.Hankel": {"tf": 1.4142135623730951}}, "df": 1}, "docs": {}, "df": 0, "n": {"docs": {"pyerrors.correlators.Corr.Hankel": {"tf": 1.7320508075688772}}, "df": 1}}, "/": {"2": {"docs": {"pyerrors.correlators.Corr.m_eff": {"tf": 2}, "pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 2}, "docs": {}, "df": 0}, "p": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "j": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 1}}}}, "u": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 2}}, "df": 5}}}}, "^": {"2": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}}, "g": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.GEVP": {"tf": 2}, "pyerrors.correlators.Corr.m_eff": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 2}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.4142135623730951}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.4142135623730951}, "pyerrors.obs.derived_observable": {"tf": 1.4142135623730951}, "pyerrors.obs.reweight": {"tf": 1}}, "df": 15, "a": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "a": {"5": {"docs": {"pyerrors.input.hadrons.read_meson_hd5": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.Npr_matrix.g5H": {"tf": 1}}, "df": 2}, "docs": {"pyerrors": {"tf": 4.242640687119285}, "pyerrors.correlators.Corr.gamma_method": {"tf": 1}, "pyerrors.correlators.Corr.gm": {"tf": 1}, "pyerrors.correlators.Corr.plateau": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.show": {"tf": 1.4142135623730951}, "pyerrors.dirac.Grid_gamma": {"tf": 1}, "pyerrors.fits.Fit_result.gamma_method": {"tf": 1}, "pyerrors.fits.Fit_result.gm": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix.g5H": {"tf": 1}, "pyerrors.input.pandas.read_sql": {"tf": 1.4142135623730951}, "pyerrors.input.pandas.load_df": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs": {"tf": 1}, "pyerrors.obs.Obs.is_zero_within_error": {"tf": 1}, "pyerrors.obs.CObs.gamma_method": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1.7320508075688772}}, "df": 16, "s": {"docs": {"pyerrors.input.hadrons.read_meson_hd5": {"tf": 1.4142135623730951}}, "df": 1}}}}, "p": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "u": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.fits.qqplot": {"tf": 1}}, "df": 1}}}}}}, "r": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}}}}}, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "b": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}, "v": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.error_band": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}, "pyerrors.input.misc.read_pbp": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.4142135623730951}, "pyerrors.linalg.eigh": {"tf": 1}, "pyerrors.linalg.eig": {"tf": 1}, "pyerrors.misc.pseudo_Obs": {"tf": 1}, "pyerrors.misc.gen_correlated_data": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.reweight": {"tf": 1}, "pyerrors.obs.Obs.is_zero": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1}}, "df": 28}, "s": {"docs": {"pyerrors.correlators.Corr.GEVP": {"tf": 1}}, "df": 1}}}}, "t": {"docs": {"pyerrors": {"tf": 8.306623862918075}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 2.8284271247461903}}, "df": 2}, "e": {"docs": {}, "df": 0, "q": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "n": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1}}, "df": 2, "i": {"docs": {}, "df": 0, "z": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}}, "df": 2}}}}}, "t": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.misc.pseudo_Obs": {"tf": 1}, "pyerrors.misc.gen_correlated_data": {"tf": 1}}, "df": 5, "d": {"docs": {"pyerrors.fits.least_squares": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.misc.pseudo_Obs": {"tf": 2}, "pyerrors.misc.gen_correlated_data": {"tf": 1.7320508075688772}}, "df": 4}, "s": {"docs": {"pyerrors.fits.qqplot": {"tf": 1}, "pyerrors.fits.residual_plot": {"tf": 1}}, "df": 2}}, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.obs.derived_observable": {"tf": 1}}, "df": 1}}}}}, "i": {"docs": {}, "df": 0, "c": {"docs": {"pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}}, "df": 2}}}, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.misc.gen_correlated_data": {"tf": 1}}, "df": 1}}}}}}, "v": {"docs": {}, "df": 0, "p": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.prune": {"tf": 2.23606797749979}}, "df": 5}}, "t": {"docs": {"pyerrors.correlators.Corr.projected": {"tf": 1}}, "df": 1}, "o": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "c": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}}}}}}, "l": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1.7320508075688772}}, "df": 2}}}}}, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1.4142135623730951}, "pyerrors.obs.cov_Obs": {"tf": 1}}, "df": 6, "i": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1.4142135623730951}, "pyerrors.obs.cov_Obs": {"tf": 1}}, "df": 5, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}, "n": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 1}}}}}, "o": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "n": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.GEVP": {"tf": 1}}, "df": 1}}}}, "i": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.dirac.Grid_gamma": {"tf": 1}}, "df": 1}}}, "o": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}}, "z": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.dump": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1.4142135623730951}, "pyerrors.input.json.create_json_string": {"tf": 1.4142135623730951}, "pyerrors.input.json.dump_to_json": {"tf": 1.4142135623730951}, "pyerrors.input.json.load_json": {"tf": 1.4142135623730951}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1.4142135623730951}, "pyerrors.input.json.load_json_dict": {"tf": 1.4142135623730951}, "pyerrors.input.pandas.to_sql": {"tf": 1}, "pyerrors.input.pandas.dump_df": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.dump": {"tf": 1.4142135623730951}}, "df": 16, "i": {"docs": {}, "df": 0, "p": {"docs": {"pyerrors": {"tf": 1}}, "df": 1, "p": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}, "pyerrors.input.pandas.to_sql": {"tf": 1}, "pyerrors.input.pandas.dump_df": {"tf": 1.4142135623730951}, "pyerrors.input.pandas.load_df": {"tf": 1}}, "df": 10}}}}}}, "u": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.correlators.Corr.m_eff": {"tf": 1.4142135623730951}, "pyerrors.fits.least_squares": {"tf": 2}, "pyerrors.fits.total_least_squares": {"tf": 1.4142135623730951}, "pyerrors.roots.find_root": {"tf": 1.4142135623730951}}, "df": 4}}}, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}}}}}}}, "^": {"docs": {}, "df": 0, "\\": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 1}}}}}}}, "c": {"docs": {}, "df": 0, "c": {"docs": {"pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}}, "df": 4}}, "f": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}}, "df": 2}}}}, "l": {"docs": {"pyerrors.input.openQCD.read_qtop": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.4142135623730951}}, "df": 3, "a": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1}}, "df": 1, "/": {"0": {"3": {"0": {"6": {"0": {"1": {"7": {"docs": {"pyerrors": {"tf": 2}}, "df": 1}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "9": {"4": {"1": {"2": {"0": {"8": {"7": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "docs": {}, "df": 0}, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.4142135623730951}}, "df": 5}}}}, "e": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 1, "x": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "r": {"docs": {"pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 1}}}, "y": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1.7320508075688772}}, "df": 1}}}}, "b": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1.4142135623730951}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 3, "s": {"docs": {"pyerrors.correlators.Corr.__init__": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1}}, "df": 2}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.dirac.Grid_gamma": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.obs.Obs.__init__": {"tf": 1}}, "df": 6}}}, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}}, "df": 1}}}}}, "s": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.__init__": {"tf": 1}, "pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}, "pyerrors.input.misc.read_pbp": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 10}}, "r": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1.4142135623730951}, "pyerrors.linalg.jack_matmul": {"tf": 1}}, "df": 2, "s": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 2}}, "r": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 2}}}}, "m": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "a": {"docs": {"pyerrors.correlators.Corr.GEVP": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.prune": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}}, "df": 3}}}}}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors": {"tf": 2.449489742783178}, "pyerrors.fits.least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.total_least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.fit_lin": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.4142135623730951}}, "df": 5}}, "s": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1}}, "df": 1}}, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.linalg.eigh": {"tf": 1}, "pyerrors.linalg.eig": {"tf": 1}}, "df": 3}}}}, "s": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 3.1622776601683795}, "pyerrors.correlators.Corr.__init__": {"tf": 2.6457513110645907}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.fit": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.plateau": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.show": {"tf": 2.449489742783178}, "pyerrors.covobs.Covobs.__init__": {"tf": 1.4142135623730951}, "pyerrors.fits.Fit_result": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 2.6457513110645907}, "pyerrors.fits.total_least_squares": {"tf": 2.449489742783178}, "pyerrors.fits.fit_lin": {"tf": 2.23606797749979}, "pyerrors.fits.ks_test": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.create_pobs_string": {"tf": 2.23606797749979}, "pyerrors.input.dobs.write_pobs": {"tf": 2.23606797749979}, "pyerrors.input.dobs.read_pobs": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.read_dobs": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.create_dobs_string": {"tf": 2.23606797749979}, "pyerrors.input.dobs.write_dobs": {"tf": 2.23606797749979}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 2}, "pyerrors.input.json.dump_to_json": {"tf": 2}, "pyerrors.input.json.import_json_string": {"tf": 2.449489742783178}, "pyerrors.input.json.load_json": {"tf": 2.449489742783178}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1.4142135623730951}, "pyerrors.input.misc.read_pbp": {"tf": 2.23606797749979}, "pyerrors.input.openQCD.read_rwms": {"tf": 3.1622776601683795}, "pyerrors.input.openQCD.extract_t0": {"tf": 3}, "pyerrors.input.openQCD.read_qtop": {"tf": 2.449489742783178}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 2.449489742783178}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 2}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 2.8284271247461903}, "pyerrors.input.utils.check_idl": {"tf": 2}, "pyerrors.misc.gen_correlated_data": {"tf": 2.23606797749979}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.__init__": {"tf": 2.449489742783178}, "pyerrors.obs.derived_observable": {"tf": 2}, "pyerrors.obs.reweight": {"tf": 1.7320508075688772}, "pyerrors.obs.covariance": {"tf": 1.4142135623730951}, "pyerrors.obs.merge_obs": {"tf": 2}, "pyerrors.obs.cov_Obs": {"tf": 1.7320508075688772}}, "df": 45, "s": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.correlators.Corr.plottable": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}, "pyerrors.obs.Obs.__init__": {"tf": 1}}, "df": 6}, "[": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.fits.fit_lin": {"tf": 1}, "pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.misc.read_pbp": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.misc.gen_correlated_data": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 11}}}, "s": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}}, "df": 1}}}, "l": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "[": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 1}}}}}}}}}}}, "k": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}}, "df": 2}}, "i": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.fits.fit_lin": {"tf": 1}}, "df": 1}}}, "b": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.input.bdio.read_ADerrors": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.read_mesons": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.read_dSdm": {"tf": 1.4142135623730951}}, "df": 4}}}}, "d": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.input.bdio.read_ADerrors": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.read_mesons": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.read_dSdm": {"tf": 1.4142135623730951}}, "df": 4}}}, "b": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {"pyerrors.input.bdio.read_ADerrors": {"tf": 2.23606797749979}, "pyerrors.input.bdio.write_ADerrors": {"tf": 2.23606797749979}, "pyerrors.input.bdio.read_mesons": {"tf": 2.23606797749979}, "pyerrors.input.bdio.read_dSdm": {"tf": 2.23606797749979}}, "df": 4}}}}}}, "e": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "s": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 3.1622776601683795}, "pyerrors.correlators.Corr.fit": {"tf": 1}}, "df": 2}}, "v": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1, "s": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}}, "s": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "v": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.fits.least_squares": {"tf": 1}}, "df": 2}}}}}, "l": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}}, "df": 4, "s": {"docs": {"pyerrors.mpm.matrix_pencil_method": {"tf": 1.7320508075688772}}, "df": 1}}}}, "n": {"docs": {"pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1.7320508075688772}}, "df": 3, "g": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "h": {"docs": {"pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.correlators.Corr.correlate": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.4142135623730951}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1}}, "df": 9}}}}, "q": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 1}, "f": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 1}}}, "o": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.deriv": {"tf": 1}, "pyerrors.correlators.Corr.second_deriv": {"tf": 1}, "pyerrors.correlators.Corr.m_eff": {"tf": 1.7320508075688772}}, "df": 4, "s": {"docs": {}, "df": 0, "y": {"docs": {}, "df": 0, "m": {"docs": {"pyerrors.correlators.Corr.m_eff": {"tf": 1}}, "df": 1}}, "c": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.spaghetti_plot": {"tf": 1}}, "df": 2}}}}}, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "c": {"docs": {"pyerrors.correlators.Corr.spaghetti_plot": {"tf": 1}}, "df": 1}}}}}}}}}, "n": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}, "a": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.misc.load_object": {"tf": 1}}, "df": 2, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.misc.load_object": {"tf": 1}}, "df": 1}}}}, "w": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}}, "df": 1, "e": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1.4142135623730951}}, "df": 4}}}}}, "t": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}}, "df": 1}, "/": {"docs": {}, "df": 0, "a": {"docs": {"pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 2}}}, "s": {"docs": {"pyerrors": {"tf": 3.1622776601683795}, "pyerrors.correlators.Corr.GEVP": {"tf": 2.6457513110645907}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs": {"tf": 2}, "pyerrors.obs.Obs.gamma_method": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.gm": {"tf": 1.4142135623730951}, "pyerrors.obs.covariance": {"tf": 1.7320508075688772}, "pyerrors.obs.cov_Obs": {"tf": 1.4142135623730951}}, "df": 11, "o": {"docs": {"pyerrors.input.bdio.read_ADerrors": {"tf": 2.23606797749979}, "pyerrors.input.bdio.write_ADerrors": {"tf": 2.23606797749979}, "pyerrors.input.bdio.read_mesons": {"tf": 2.23606797749979}, "pyerrors.input.bdio.read_dSdm": {"tf": 2.23606797749979}, "pyerrors.input.hadrons.Npr_matrix.g5H": {"tf": 1}}, "df": 5, "m": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 2, "t": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}}}}}, "m": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}, "k": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}, "l": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.GEVP": {"tf": 1}}, "df": 1, "r": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.correlators.Corr.Eigenvalue": {"tf": 1}, "pyerrors.correlators.Corr.m_eff": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.prune": {"tf": 1.4142135623730951}}, "df": 3}}}}}, "u": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1.4142135623730951}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 4}}}}, "r": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.GEVP": {"tf": 1.7320508075688772}}, "df": 1, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.GEVP": {"tf": 1}}, "df": 1}}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.correlators.Corr.GEVP": {"tf": 1}}, "df": 1}}}}}, "n": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}}, "df": 4}}}}}, "u": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 2}}, "df": 1}}}}}}}, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "d": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.spaghetti_plot": {"tf": 1}}, "df": 2}}}}, "m": {"docs": {"pyerrors": {"tf": 3.605551275463989}, "pyerrors.obs.covariance": {"tf": 1.4142135623730951}}, "df": 2, "m": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.linalg.einsum": {"tf": 1}}, "df": 1}}}}}}}, "c": {"docs": {}, "df": 0, "h": {"docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.correlators.Corr.prune": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1.4142135623730951}}, "df": 3}, "c": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.bdio.write_ADerrors": {"tf": 1}}, "df": 1, "f": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.input.bdio.write_ADerrors": {"tf": 1}}, "df": 1}}}}}}}}, "r": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "p": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1, "s": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.correlators.Corr.dump": {"tf": 1}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}, "pyerrors.obs.Obs.dump": {"tf": 1}}, "df": 7}}}}, "s": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 2}}}}, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.correlators.Corr.thin": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 2}}}}, "l": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.obs.derived_observable": {"tf": 1}}, "df": 1, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.obs.derived_observable": {"tf": 1}}, "df": 1}}}}}}, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.reweight": {"tf": 1}, "pyerrors.obs.Obs.reweight": {"tf": 1}, "pyerrors.obs.reweight": {"tf": 1}}, "df": 3}}}}}}, "b": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 1}}}}, "c": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.linalg.einsum": {"tf": 1.4142135623730951}}, "df": 1}}}}}}}, "m": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input": {"tf": 1}}, "df": 1}}}}}}}, "f": {"docs": {}, "df": 0, "f": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1.4142135623730951}}, "df": 1}}}}}}}}}}}, "l": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "w": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}}, "df": 3}}}}}, "i": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}}, "df": 2}}}, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}, "i": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1.7320508075688772}}, "df": 1, "s": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}}}}}}}}, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "i": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1, "l": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}, "p": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "f": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}}}, "y": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}, "y": {"docs": {"pyerrors.correlators.Corr": {"tf": 1}}, "df": 1}, "e": {"docs": {"pyerrors.obs.derived_observable": {"tf": 1}}, "df": 1}}}}, "n": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.linalg.svd": {"tf": 1}}, "df": 2}}}}, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 2.449489742783178}, "pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 4}}}, "c": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.prune": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 4}}, "h": {"docs": {"pyerrors.correlators.Corr.m_eff": {"tf": 1.7320508075688772}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}}, "df": 3}, "k": {"docs": {"pyerrors.input.hadrons.read_meson_hd5": {"tf": 1.4142135623730951}}, "df": 1}}, "z": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 2}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 7}}, "g": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "a": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.show": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs": {"tf": 2}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}, "pyerrors.obs.Obs.is_zero_within_error": {"tf": 1.4142135623730951}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 7}}}, "d": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.GEVP": {"tf": 1}}, "df": 1}}, "l": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}}, "df": 3}}}}}, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "f": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}, "p": {"docs": {"pyerrors.input.bdio.read_mesons": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.read_dSdm": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.extract_t0": {"tf": 2}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}}, "df": 7, "s": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 4}}}, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}}, "e": {"docs": {"pyerrors.correlators.Corr.GEVP": {"tf": 2.6457513110645907}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 1.7320508075688772}}, "df": 2, "s": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.prune": {"tf": 1.4142135623730951}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 4}}}, "n": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.correlators.Corr.m_eff": {"tf": 1}, "pyerrors.correlators.Corr.fit": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.show": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.spaghetti_plot": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.linalg.matmul": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 2.449489742783178}, "pyerrors.obs.Obs.gamma_method": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.gm": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.is_zero_within_error": {"tf": 1.4142135623730951}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 12, "i": {"docs": {}, "df": 0, "z": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.pandas.dump_df": {"tf": 1}}, "df": 1}}}}}}}}}, "r": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}, "pyerrors.input.misc.read_pbp": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 10}}, "c": {"docs": {}, "df": 0, "k": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "x": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.dirac.epsilon_tensor": {"tf": 1}, "pyerrors.dirac.epsilon_tensor_rank4": {"tf": 1}}, "df": 2}}}}}}}}}}, "b": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}}, "df": 1}}}}, "d": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}, "r": {"docs": {"pyerrors.correlators.Corr.deriv": {"tf": 1}, "pyerrors.correlators.Corr.second_deriv": {"tf": 1}, "pyerrors.correlators.Corr.m_eff": {"tf": 1}, "pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.dump": {"tf": 1.7320508075688772}, "pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1.7320508075688772}, "pyerrors.input.bdio.read_dSdm": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.create_pobs_string": {"tf": 2.449489742783178}, "pyerrors.input.dobs.write_pobs": {"tf": 2.23606797749979}, "pyerrors.input.dobs.read_pobs": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.read_dobs": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.create_dobs_string": {"tf": 2.449489742783178}, "pyerrors.input.dobs.write_dobs": {"tf": 2.23606797749979}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 2}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1.7320508075688772}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1.7320508075688772}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1.7320508075688772}, "pyerrors.input.json.create_json_string": {"tf": 1.4142135623730951}, "pyerrors.input.json.dump_to_json": {"tf": 1.4142135623730951}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1.7320508075688772}, "pyerrors.input.json.load_json_dict": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_rwms": {"tf": 2}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop": {"tf": 2}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 2.23606797749979}, "pyerrors.input.pandas.to_sql": {"tf": 1.7320508075688772}, "pyerrors.input.pandas.read_sql": {"tf": 1.4142135623730951}, "pyerrors.input.pandas.dump_df": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 2.8284271247461903}, "pyerrors.input.utils.check_idl": {"tf": 1.4142135623730951}, "pyerrors.linalg.einsum": {"tf": 1}, "pyerrors.misc.dump_object": {"tf": 1.4142135623730951}, "pyerrors.misc.load_object": {"tf": 1}, "pyerrors.misc.pseudo_Obs": {"tf": 1}, "pyerrors.misc.gen_correlated_data": {"tf": 1}, "pyerrors.obs.Obs.plot_tauint": {"tf": 1}, "pyerrors.obs.Obs.plot_rho": {"tf": 1}, "pyerrors.obs.Obs.plot_piechart": {"tf": 1}, "pyerrors.obs.Obs.dump": {"tf": 2}, "pyerrors.obs.import_jackknife": {"tf": 1}, "pyerrors.obs.cov_Obs": {"tf": 1}}, "df": 51, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors": {"tf": 3.4641016151377544}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 2}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.import_json_string": {"tf": 2}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.input.utils.check_idl": {"tf": 1}}, "df": 16, "s": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.pandas.to_sql": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1}, "pyerrors.obs.Obs.__init__": {"tf": 1}}, "df": 7}}}, "d": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 2.23606797749979}}, "df": 1}}}, "c": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.obs.correlate": {"tf": 1}}, "df": 1}}}}}, "u": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}}, "df": 1, "u": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 3}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1.4142135623730951}, "pyerrors.input.json.load_json_dict": {"tf": 1}, "pyerrors.input.misc.read_pbp": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 13, "s": {"docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.import_json_string": {"tf": 1.4142135623730951}, "pyerrors.input.json.load_json": {"tf": 1.4142135623730951}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1.4142135623730951}, "pyerrors.input.json.load_json_dict": {"tf": 1.4142135623730951}}, "df": 11}}}}}}}, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}, "o": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 1}}}}}}, "i": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 2}}}, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1, "d": {"docs": {"pyerrors": {"tf": 2.449489742783178}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 2}}, "a": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}}, "df": 3}}}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}, "p": {"docs": {"pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}, "pyerrors.input.misc.read_pbp": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 9}}, "y": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}}, "df": 1}}}}, "c": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "f": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}, "e": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "a": {"docs": {"pyerrors": {"tf": 1.7320508075688772}}, "df": 1}}}}, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.spaghetti_plot": {"tf": 1}}, "df": 2}}}, "i": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}}, "df": 3}}}}, "a": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.fits.error_band": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}}, "df": 4, "s": {"1": {"docs": {"pyerrors": {"tf": 2.23606797749979}}, "df": 1}, "2": {"docs": {"pyerrors": {"tf": 2.23606797749979}}, "df": 1}, "3": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {"pyerrors": {"tf": 3.872983346207417}, "pyerrors.input": {"tf": 1.7320508075688772}, "pyerrors.misc.pseudo_Obs": {"tf": 1.4142135623730951}, "pyerrors.misc.gen_correlated_data": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs": {"tf": 1}, "pyerrors.obs.Obs.__init__": {"tf": 2}, "pyerrors.obs.Obs.export_jackknife": {"tf": 2.23606797749979}, "pyerrors.obs.import_jackknife": {"tf": 1.7320508075688772}}, "df": 8}}}}, "e": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.correlators.Corr.correlate": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1}, "pyerrors.obs.merge_obs": {"tf": 1}}, "df": 12}}, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "f": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}}, "v": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.obs.Obs.plot_tauint": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.plot_rho": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.plot_piechart": {"tf": 1.4142135623730951}}, "df": 4, "d": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.correlators.Corr.dump": {"tf": 1}, "pyerrors.misc.dump_object": {"tf": 1}, "pyerrors.obs.Obs.dump": {"tf": 1}}, "df": 4}, "s": {"docs": {"pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.obs.Obs.plot_tauint": {"tf": 1}, "pyerrors.obs.Obs.plot_rho": {"tf": 1}, "pyerrors.obs.Obs.plot_piechart": {"tf": 1}}, "df": 6}}}, "r": {"docs": {}, "df": 0, "k": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 1}}}}}, "e": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.correlators.Corr.__init__": {"tf": 1}, "pyerrors.correlators.Corr.item": {"tf": 1}, "pyerrors.correlators.Corr.second_deriv": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.obs.correlate": {"tf": 1}}, "df": 7}}}, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}}, "df": 1}}}, "o": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.input.openQCD.qtop_projection": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.7320508075688772}}, "df": 2}}}}, "t": {"docs": {"pyerrors": {"tf": 2.449489742783178}, "pyerrors.correlators.Corr.GEVP": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 13, "s": {"docs": {"pyerrors.correlators.Corr.set_prange": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1}}, "df": 2}}, "e": {"docs": {"pyerrors": {"tf": 3}, "pyerrors.correlators.Corr.m_eff": {"tf": 1}, "pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.input": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 2.23606797749979}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.linalg.einsum": {"tf": 1}, "pyerrors.obs.Obs.is_zero": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 10}, "l": {"docs": {}, "df": 0, "f": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.fit": {"tf": 1}}, "df": 3}}, "q": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr": {"tf": 1}}, "df": 1}}}}}}, "r": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}}, "df": 1}}}, "i": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "z": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.pandas.dump_df": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1}}, "df": 2}}}}}}}}, "p": {"docs": {"pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}}, "df": 1, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 2}, "pyerrors.input.dobs.import_dobs_string": {"tf": 2.449489742783178}, "pyerrors.input.dobs.read_dobs": {"tf": 2.449489742783178}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 9}}, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}}, "df": 3}}}, "e": {"docs": {"pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}}, "df": 1, "s": {"docs": {"pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 1}}}}}}}, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "h": {"docs": {"pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}}, "df": 1}}}}, "v": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 1}}}}}, "m": {"docs": {}, "df": 0, "i": {"docs": {"pyerrors.obs.covariance": {"tf": 1.4142135623730951}}, "df": 1}}}, "p": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "c": {"docs": {"pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}}, "df": 4, "i": {"docs": {}, "df": 0, "f": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 2}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.4142135623730951}}, "df": 14}, "s": {"docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.correlators.Corr.dump": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.qtop_projection": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.4142135623730951}, "pyerrors.misc.dump_object": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}, "pyerrors.obs.Obs.dump": {"tf": 1}}, "df": 10}}, "c": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}}, "y": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1.4142135623730951}, "pyerrors.input.json.load_json_dict": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 12}}, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.input.openQCD.read_rwms": {"tf": 1}}, "df": 1}}}}}, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 4}}}}, "c": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.correlators.Corr.thin": {"tf": 1.7320508075688772}}, "df": 1}}}, "e": {"docs": {"pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}}, "df": 3}}, "g": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {"pyerrors.correlators.Corr.spaghetti_plot": {"tf": 1}}, "df": 1}}}}}}}}, "y": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "c": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.is_matrix_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.deriv": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.second_deriv": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.m_eff": {"tf": 1}}, "df": 5}, "z": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.symmetric": {"tf": 1}, "pyerrors.correlators.Corr.anti_symmetric": {"tf": 1}}, "df": 2, "s": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.matrix_symmetric": {"tf": 1}}, "df": 2}, "d": {"docs": {"pyerrors.correlators.Corr.m_eff": {"tf": 1}}, "df": 1}}}}, "y": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.T_symmetry": {"tf": 1.4142135623730951}}, "df": 2}}}}}, "b": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}}, "df": 4, "s": {"docs": {"pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}}, "df": 4}}}}}, "n": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "x": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}, "h": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "f": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.roll": {"tf": 1}}, "df": 1, "s": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}}}}, "o": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 6}}}, "w": {"docs": {"pyerrors.obs.Obs.plot_history": {"tf": 1}}, "df": 1, "n": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 1}, "s": {"docs": {"pyerrors.obs.Obs.plot_piechart": {"tf": 1}}, "df": 1}}}, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.bdio.read_ADerrors": {"tf": 1.7320508075688772}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1.7320508075688772}, "pyerrors.input.bdio.read_mesons": {"tf": 1.7320508075688772}, "pyerrors.input.bdio.read_dSdm": {"tf": 1.7320508075688772}}, "df": 4}}}, "p": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 2.6457513110645907}}, "df": 1}}}}, "q": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.covobs.Covobs.errsq": {"tf": 1}}, "df": 2, "s": {"docs": {"pyerrors": {"tf": 3}, "pyerrors.correlators.Corr.fit": {"tf": 1}}, "df": 2}, "d": {"docs": {"pyerrors.fits.Fit_result": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 2}}}}}, "r": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 3, "{": {"docs": {}, "df": 0, "j": {"docs": {}, "df": 0, "^": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}, "\\": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "u": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}}}}}, "l": {"docs": {"pyerrors.input.pandas.read_sql": {"tf": 1.7320508075688772}}, "df": 1, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.pandas.to_sql": {"tf": 1.4142135623730951}, "pyerrors.input.pandas.read_sql": {"tf": 1.7320508075688772}, "pyerrors.input.pandas.load_df": {"tf": 1}}, "df": 3}}}}}, "m": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 3, "e": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}}}, "e": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 3}}}}}}, "o": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "h": {"docs": {"pyerrors.obs.covariance": {"tf": 1.4142135623730951}}, "df": 1, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}}}}}, "v": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "k": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "p": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1, "p": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 1}}}}}}, "f": {"docs": {}, "df": 0, "q": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.openQCD.read_qtop": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 2}}, "df": 3}}}, "c": {"docs": {}, "df": 0, "f": {"docs": {"pyerrors.input.sfcf.read_sfcf": {"tf": 2.23606797749979}}, "df": 1}}}}, "n": {"docs": {"pyerrors": {"tf": 2}, "pyerrors.correlators.Corr.Hankel": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.prune": {"tf": 2.8284271247461903}, "pyerrors.fits.fit_lin": {"tf": 1.4142135623730951}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 2}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 2}, "pyerrors.obs.covariance": {"tf": 1.4142135623730951}, "pyerrors.obs.import_jackknife": {"tf": 1}, "pyerrors.obs.cov_Obs": {"tf": 1}}, "df": 12, "o": {"docs": {"pyerrors.fits.least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.fit_lin": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.read_dobs": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 11, "n": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.fits.least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.total_least_squares": {"tf": 1.4142135623730951}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 4, "e": {"docs": {"pyerrors.correlators.Corr.GEVP": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.qqplot": {"tf": 1}, "pyerrors.fits.residual_plot": {"tf": 1}, "pyerrors.fits.ks_test": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 2.23606797749979}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1.4142135623730951}, "pyerrors.input.pandas.to_sql": {"tf": 1}, "pyerrors.input.pandas.dump_df": {"tf": 1}, "pyerrors.misc.dump_object": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 22}}, "t": {"docs": {"pyerrors": {"tf": 2.6457513110645907}, "pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.correlators.Corr.reweight": {"tf": 1}, "pyerrors.correlators.Corr.m_eff": {"tf": 1}, "pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.read_dobs": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.write_dobs": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_rwms": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.7320508075688772}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.obs.Obs.reweight": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}, "pyerrors.obs.reweight": {"tf": 1}, "pyerrors.obs.correlate": {"tf": 1.4142135623730951}, "pyerrors.obs.covariance": {"tf": 1}, "pyerrors.obs.merge_obs": {"tf": 1}}, "df": 28, "e": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}}, "df": 2, "s": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}, "pyerrors.obs.derived_observable": {"tf": 1}, "pyerrors.obs.correlate": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1.4142135623730951}, "pyerrors.obs.merge_obs": {"tf": 1}}, "df": 7}}}, "w": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "r": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "z": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.reweight": {"tf": 1}, "pyerrors.obs.Obs.reweight": {"tf": 1}, "pyerrors.obs.Obs.plot_rho": {"tf": 1}, "pyerrors.obs.reweight": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 5}}, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 1}}}}}}}}}}}, "i": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1.4142135623730951}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1.4142135623730951}}, "df": 2}}}, "e": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}}, "df": 2}}}}}, "f": {"docs": {}, "df": 0, "f": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 1}}}}}}, "e": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "f": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}}, "w": {"docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.7320508075688772}, "pyerrors.obs.derived_observable": {"tf": 1}, "pyerrors.obs.merge_obs": {"tf": 1}, "pyerrors.obs.cov_Obs": {"tf": 1}}, "df": 6}, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.4142135623730951}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 5, "s": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 2}, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 4}}}}, "c": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.obs.correlate": {"tf": 1}}, "df": 2}}}}}}}, "l": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}}, "df": 1}}}}, "x": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.input.openQCD.read_qtop": {"tf": 1}}, "df": 1}}}}}}, "u": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "m": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}}, "df": 3, "p": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 3.4641016151377544}, "pyerrors.fits.least_squares": {"tf": 1.7320508075688772}, "pyerrors.fits.total_least_squares": {"tf": 1.7320508075688772}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 2.23606797749979}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}, "pyerrors.linalg.matmul": {"tf": 1.4142135623730951}, "pyerrors.linalg.jack_matmul": {"tf": 1.4142135623730951}, "pyerrors.linalg.einsum": {"tf": 2}, "pyerrors.misc.gen_correlated_data": {"tf": 1}, "pyerrors.obs.Obs.__init__": {"tf": 1}, "pyerrors.obs.Obs.is_zero": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1.4142135623730951}, "pyerrors.obs.derived_observable": {"tf": 1.4142135623730951}, "pyerrors.obs.covariance": {"tf": 1}, "pyerrors.obs.import_jackknife": {"tf": 1.4142135623730951}, "pyerrors.roots.find_root": {"tf": 1.7320508075688772}}, "df": 21}}, "b": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.roll": {"tf": 1}, "pyerrors.correlators.Corr.T_symmetry": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 2}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.linalg.matmul": {"tf": 1}, "pyerrors.linalg.jack_matmul": {"tf": 1}, "pyerrors.linalg.einsum": {"tf": 1}, "pyerrors.misc.pseudo_Obs": {"tf": 1}, "pyerrors.misc.gen_correlated_data": {"tf": 1.4142135623730951}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}, "pyerrors.obs.Obs.is_zero_within_error": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1}}, "df": 20, "s": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 2}}}}, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.input": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1.4142135623730951}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 6, "l": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}}, "df": 1}}}}}}}}, "d": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "f": {"docs": {}, "df": 0, "f": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.obs.derived_observable": {"tf": 1}}, "df": 1}}}}}}}}}}, "l": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.input.json.dump_to_json": {"tf": 1}}, "df": 1}}}, "p": {"docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.fits.error_band": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 2}, "pyerrors.linalg.eigh": {"tf": 1}, "pyerrors.linalg.eig": {"tf": 1}}, "df": 5, "r": {"docs": {"pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}}, "df": 1}}, "a": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 3.872983346207417}, "pyerrors.correlators.Corr.dump": {"tf": 1}, "pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.read_dSdm": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.write_pobs": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.read_dobs": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.create_dobs_string": {"tf": 2}, "pyerrors.input.dobs.write_dobs": {"tf": 2}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}, "pyerrors.input.pandas.to_sql": {"tf": 1.4142135623730951}, "pyerrors.input.sfcf.read_sfcf": {"tf": 2}, "pyerrors.misc.dump_object": {"tf": 1.4142135623730951}, "pyerrors.misc.pseudo_Obs": {"tf": 1.7320508075688772}, "pyerrors.misc.gen_correlated_data": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.dump": {"tf": 1.4142135623730951}, "pyerrors.obs.import_jackknife": {"tf": 1.4142135623730951}, "pyerrors.obs.cov_Obs": {"tf": 1}}, "df": 24, "s": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.read_dobs": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1.7320508075688772}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1}, "pyerrors.obs.Obs.__init__": {"tf": 1}}, "df": 17, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "m": {"docs": {"pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}}, "df": 4}}}}, "d": {"docs": {"pyerrors.obs.Obs.plot_tauint": {"tf": 1}, "pyerrors.obs.Obs.plot_rho": {"tf": 1}, "pyerrors.obs.Obs.plot_piechart": {"tf": 1}}, "df": 3}}}, "n": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}, "d": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 3.605551275463989}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}, "pyerrors.linalg.matmul": {"tf": 1}, "pyerrors.linalg.jack_matmul": {"tf": 1}, "pyerrors.linalg.einsum": {"tf": 1}, "pyerrors.misc.gen_correlated_data": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}, "pyerrors.obs.import_jackknife": {"tf": 1}}, "df": 15}}}}}, "i": {"docs": {}, "df": 0, "m": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}}, "t": {"docs": {}, "df": 0, "h": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.obs.import_jackknife": {"tf": 1}}, "df": 2}, "r": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "c": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1.7320508075688772}}, "df": 1}}}}}, "x": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.correlators.Corr.Hankel": {"tf": 1}, "pyerrors.obs.cov_Obs": {"tf": 1}}, "df": 2}}, "b": {"docs": {}, "df": 0, "y": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}}}}, "r": {"docs": {}, "df": 0, "w": {"docs": {"pyerrors.input.openQCD.read_rwms": {"tf": 1}}, "df": 1}}}, "x": {"0": {"docs": {"pyerrors": {"tf": 2}, "pyerrors.correlators.Corr.symmetric": {"tf": 1}, "pyerrors.correlators.Corr.anti_symmetric": {"tf": 1}, "pyerrors.correlators.Corr.deriv": {"tf": 1}, "pyerrors.correlators.Corr.second_deriv": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 6}, "1": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.fits.least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.total_least_squares": {"tf": 1.4142135623730951}}, "df": 3}, "2": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.fits.least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.total_least_squares": {"tf": 1.4142135623730951}}, "df": 3}, "docs": {"pyerrors": {"tf": 2.8284271247461903}, "pyerrors.correlators.Corr.show": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.prune": {"tf": 1.4142135623730951}, "pyerrors.fits.least_squares": {"tf": 2.8284271247461903}, "pyerrors.fits.total_least_squares": {"tf": 3}, "pyerrors.fits.fit_lin": {"tf": 1.4142135623730951}, "pyerrors.fits.error_band": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1}, "pyerrors.roots.find_root": {"tf": 1.7320508075688772}}, "df": 10, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.fits.fit_lin": {"tf": 1.4142135623730951}}, "df": 1}}}}}, "m": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.input.dobs.create_pobs_string": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.write_pobs": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.write_dobs": {"tf": 1.7320508075688772}}, "df": 7}, "i": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1.4142135623730951}}, "df": 1}}}, "[": {"0": {"docs": {"pyerrors.obs.derived_observable": {"tf": 1}}, "df": 1}, "1": {"docs": {"pyerrors.obs.derived_observable": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}}, "y": {"docs": {"pyerrors": {"tf": 2}, "pyerrors.correlators.Corr.show": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.spaghetti_plot": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.total_least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.fit_lin": {"tf": 1.4142135623730951}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 7, "o": {"docs": {}, "df": 0, "u": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 3}}, "e": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}, "r": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.fits.fit_lin": {"tf": 1}}, "df": 2}}}}, "t": {"docs": {"pyerrors.obs.correlate": {"tf": 1}}, "df": 1}}, "l": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1}}, "df": 1}}}}}}, "r": {"0": {"1": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "2": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "docs": {"pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 3}, "docs": {"pyerrors": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.misc.read_pbp": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_rwms": {"tf": 2}, "pyerrors.input.openQCD.extract_t0": {"tf": 2.6457513110645907}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.4142135623730951}}, "df": 8, "e": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.linalg.matmul": {"tf": 1}, "pyerrors.linalg.jack_matmul": {"tf": 1}, "pyerrors.linalg.einsum": {"tf": 1}, "pyerrors.misc.gen_correlated_data": {"tf": 1}, "pyerrors.obs.CObs.gamma_method": {"tf": 1}, "pyerrors.obs.CObs.is_zero": {"tf": 1}}, "df": 9}, "d": {"docs": {"pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1.7320508075688772}, "pyerrors.input.bdio.read_dSdm": {"tf": 1.7320508075688772}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 2}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1.7320508075688772}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 2.23606797749979}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 2}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 2}, "pyerrors.input.json.load_json_dict": {"tf": 1.4142135623730951}, "pyerrors.input.misc.read_pbp": {"tf": 2}, "pyerrors.input.openQCD.read_rwms": {"tf": 2.449489742783178}, "pyerrors.input.openQCD.extract_t0": {"tf": 2}, "pyerrors.input.openQCD.read_qtop": {"tf": 2.449489742783178}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 2.23606797749979}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1.7320508075688772}, "pyerrors.input.sfcf.read_sfcf": {"tf": 2.6457513110645907}}, "df": 18, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}}, "df": 3}}}}}, "p": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}}, "l": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "a": {"docs": {"pyerrors": {"tf": 2.6457513110645907}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.read_dobs": {"tf": 1.7320508075688772}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.7320508075688772}, "pyerrors.obs.Obs.details": {"tf": 1}, "pyerrors.obs.Obs.plot_rep_dist": {"tf": 1}}, "df": 11, "s": {"docs": {"pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}}, "df": 1, "/": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 4}}}}}}}}}}}}, "u": {"docs": {}, "df": 0, "m": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.input.misc.read_pbp": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_rwms": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.4142135623730951}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.input.utils.check_idl": {"tf": 1}, "pyerrors.obs.Obs.plot_rep_dist": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1.4142135623730951}, "pyerrors.obs.merge_obs": {"tf": 1}}, "df": 12}}}, "k": {"docs": {}, "df": 0, "a": {"docs": {"pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}}, "df": 1}}}, "a": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.pandas.to_sql": {"tf": 1}}, "df": 4, "s": {"docs": {"pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 1}}}}}, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}}, "s": {"docs": {"pyerrors.fits.Fit_result": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 2}}}}}}}, "s": {"docs": {"pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}}, "df": 2, "[": {"0": {"docs": {"pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}}, "df": 2}, "docs": {}, "df": 0}}}, "s": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.read_dobs": {"tf": 1.4142135623730951}, "pyerrors.misc.pseudo_Obs": {"tf": 1}, "pyerrors.roots.find_root": {"tf": 1}}, "df": 6, "e": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "h": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}, "u": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 4}, "pyerrors.correlators.Corr.show": {"tf": 1.4142135623730951}, "pyerrors.fits.least_squares": {"tf": 1.7320508075688772}, "pyerrors.fits.total_least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.qqplot": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}, "pyerrors.input.json.import_json_string": {"tf": 1.7320508075688772}, "pyerrors.input.json.load_json": {"tf": 1.7320508075688772}, "pyerrors.input.misc.read_pbp": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1}}, "df": 15, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.4142135623730951}}, "df": 3}}}, "s": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.fits.Fit_result": {"tf": 1.4142135623730951}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.ks_test": {"tf": 1}}, "df": 4}}}}, "p": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 2.449489742783178}, "pyerrors.correlators.Corr.deriv": {"tf": 1}, "pyerrors.correlators.Corr.second_deriv": {"tf": 1}}, "df": 3, "i": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}}}}}}, "l": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}}, "df": 2}}}}, "i": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.qqplot": {"tf": 1}, "pyerrors.fits.residual_plot": {"tf": 1}}, "df": 3}}}}}}, "c": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}}}}}, "q": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 2}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 8}, "s": {"docs": {"pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}}, "df": 6}, "m": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}}, "df": 2}}}}}}}}}, "g": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1, "l": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}}, "df": 2}}}}}}}, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.correlators.Corr.__init__": {"tf": 1}, "pyerrors.correlators.Corr.plateau": {"tf": 1.4142135623730951}}, "df": 2}}}}, "t": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.correlators.Corr.T_symmetry": {"tf": 1}, "pyerrors.correlators.Corr.deriv": {"tf": 1}, "pyerrors.correlators.Corr.second_deriv": {"tf": 1}, "pyerrors.covobs.Covobs.errsq": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.total_least_squares": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.bdio.read_dSdm": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}, "pyerrors.roots.find_root": {"tf": 1}}, "df": 12, "s": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.correlators.Corr.item": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.m_eff": {"tf": 1}, "pyerrors.dirac.epsilon_tensor": {"tf": 1}, "pyerrors.dirac.epsilon_tensor_rank4": {"tf": 1}, "pyerrors.dirac.Grid_gamma": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.fit_lin": {"tf": 1.4142135623730951}, "pyerrors.fits.qqplot": {"tf": 1}, "pyerrors.fits.residual_plot": {"tf": 1}, "pyerrors.fits.error_band": {"tf": 1}, "pyerrors.fits.ks_test": {"tf": 1}, "pyerrors.input.bdio.read_ADerrors": {"tf": 1}, "pyerrors.input.bdio.write_ADerrors": {"tf": 1.4142135623730951}, "pyerrors.input.bdio.read_mesons": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}, "pyerrors.input.misc.read_pbp": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.qtop_projection": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}, "pyerrors.input.pandas.to_sql": {"tf": 1}, "pyerrors.input.pandas.read_sql": {"tf": 1}, "pyerrors.input.pandas.dump_df": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.input.utils.check_idl": {"tf": 1}, "pyerrors.misc.dump_object": {"tf": 1}, "pyerrors.misc.load_object": {"tf": 1}, "pyerrors.misc.pseudo_Obs": {"tf": 1}, "pyerrors.misc.gen_correlated_data": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}, "pyerrors.obs.Obs.plot_piechart": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1.4142135623730951}, "pyerrors.obs.import_jackknife": {"tf": 1}, "pyerrors.roots.find_root": {"tf": 1}}, "df": 58}, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.GEVP": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.read_pobs": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.read_dobs": {"tf": 1.4142135623730951}, "pyerrors.input.json.import_json_string": {"tf": 1.4142135623730951}, "pyerrors.input.json.load_json": {"tf": 1.4142135623730951}, "pyerrors.input.json.load_json_dict": {"tf": 1.4142135623730951}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 8}}}}}, "o": {"docs": {"pyerrors.input.openQCD.qtop_projection": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 2}}, "v": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.reverse": {"tf": 1}}, "df": 2, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}, "w": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.reweight": {"tf": 1}, "pyerrors.obs.Obs.reweight": {"tf": 1}, "pyerrors.obs.reweight": {"tf": 1}}, "df": 4, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.reweight": {"tf": 1}, "pyerrors.input.openQCD.qtop_projection": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.obs.Obs.reweight": {"tf": 1}, "pyerrors.obs.reweight": {"tf": 1}, "pyerrors.obs.correlate": {"tf": 1}}, "df": 7}}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.correlators.Corr.reweight": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.obs.Obs.reweight": {"tf": 1.4142135623730951}, "pyerrors.obs.reweight": {"tf": 1.4142135623730951}, "pyerrors.obs.correlate": {"tf": 1}}, "df": 6}}}}}}}}, "i": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.obs.Obs.reweight": {"tf": 1}}, "df": 1}}}}}}}}, "n": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "z": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}}}}}}}}, "f": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 2}}, "df": 4, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.GEVP": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.show": {"tf": 1}}, "df": 2, "s": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1}}, "df": 2}}}}}}}}, "c": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.m_eff": {"tf": 1}}, "df": 1}}}}}}}, "v": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}}, "df": 7}}}}}, "n": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.input.json.import_json_string": {"tf": 1}}, "df": 1, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}}, "df": 2}}}}}}}}}, "r": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 1}}, "g": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "z": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}}, "df": 1}}}}}}}}, "l": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.m_eff": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.dump": {"tf": 1}}, "df": 5}}}}}, "i": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}}, "df": 1}}}}}, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}}, "df": 3}}}}}, "d": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1.4142135623730951}}, "df": 1, "d": {"docs": {"pyerrors.fits.Fit_result": {"tf": 1}}, "df": 1}}}}}, "m": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 1, "d": {"docs": {"pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}}, "df": 4}}}}, "a": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.dobs.read_pobs": {"tf": 1}}, "df": 1}}}}}, "a": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}, "s": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}}, "df": 1}}}}, "n": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 2}, "pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.input.utils.check_idl": {"tf": 1}}, "df": 12, "s": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.obs.Obs.__init__": {"tf": 1}}, "df": 2}}}, "k": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}, "pyerrors.dirac.epsilon_tensor": {"tf": 1.4142135623730951}, "pyerrors.dirac.epsilon_tensor_rank4": {"tf": 1.4142135623730951}}, "df": 3}, "d": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "m": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}}}, "p": {"docs": {"pyerrors": {"tf": 1.7320508075688772}}, "df": 1}, "d": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 3}}}}, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {"pyerrors.obs.derived_observable": {"tf": 1}}, "df": 1}}}}, "h": {"docs": {}, "df": 0, "o": {"docs": {"pyerrors": {"tf": 2}}, "df": 1}}, "o": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "o": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.m_eff": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.roots.find_root": {"tf": 1.7320508075688772}}, "df": 4, "s": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}}}, "u": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.input": {"tf": 1}}, "df": 2}}}}}, "n": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.openQCD.read_qtop": {"tf": 1}}, "df": 1}}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}}}}, "w": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.7320508075688772}}, "df": 2}}, "i": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 2}}}}, "w": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.openQCD.read_rwms": {"tf": 1.4142135623730951}}, "df": 1}}}, "u": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.obs.Obs.is_zero_within_error": {"tf": 1}}, "df": 1}}}, "v": {"1": {"docs": {}, "df": 0, "@": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "@": {"docs": {}, "df": 0, "v": {"2": {"docs": {"pyerrors.correlators.Corr.projected": {"tf": 1}}, "df": 1}, "docs": {}, "df": 0}}}}}, "docs": {"pyerrors.correlators.Corr.GEVP": {"tf": 2.23606797749979}, "pyerrors.correlators.Corr.prune": {"tf": 2.23606797749979}, "pyerrors.obs.covariance": {"tf": 1.4142135623730951}}, "df": 3, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 3.1622776601683795}, "pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.correlators.Corr.plottable": {"tf": 1}, "pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.prune": {"tf": 1.4142135623730951}, "pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.fits.Fit_result": {"tf": 2}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.linalg.svd": {"tf": 1}, "pyerrors.misc.pseudo_Obs": {"tf": 1.7320508075688772}, "pyerrors.misc.gen_correlated_data": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 2.449489742783178}, "pyerrors.obs.Obs.gamma_method": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.gm": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1}, "pyerrors.obs.import_jackknife": {"tf": 1}, "pyerrors.obs.cov_Obs": {"tf": 1}}, "df": 20, "s": {"docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.correlators.Corr.show": {"tf": 1.7320508075688772}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.fits.error_band": {"tf": 1.4142135623730951}, "pyerrors.fits.ks_test": {"tf": 1}, "pyerrors.obs.Obs": {"tf": 1.7320508075688772}}, "df": 7}, "d": {"docs": {"pyerrors": {"tf": 2}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1.4142135623730951}, "pyerrors.input.pandas.to_sql": {"tf": 1}, "pyerrors.input.pandas.read_sql": {"tf": 1}, "pyerrors.input.pandas.dump_df": {"tf": 1}, "pyerrors.linalg.matmul": {"tf": 1}, "pyerrors.linalg.jack_matmul": {"tf": 1}, "pyerrors.linalg.einsum": {"tf": 1}, "pyerrors.linalg.inv": {"tf": 1}, "pyerrors.linalg.cholesky": {"tf": 1}, "pyerrors.linalg.det": {"tf": 1}, "pyerrors.obs.CObs": {"tf": 1}, "pyerrors.roots.find_root": {"tf": 1}}, "df": 13}}}, "i": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}}, "df": 2}}}, "c": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.input": {"tf": 1}}, "df": 2}}}, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors.correlators.Corr.deriv": {"tf": 1}, "pyerrors.correlators.Corr.second_deriv": {"tf": 1}, "pyerrors.correlators.Corr.m_eff": {"tf": 1.4142135623730951}}, "df": 3, "s": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}}, "c": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.covobs.Covobs.__init__": {"tf": 1}, "pyerrors.covobs.Covobs.errsq": {"tf": 1}}, "df": 2}}}, "b": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1.7320508075688772}}, "df": 1}}}}}}}, "n": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.dobs.import_dobs_string": {"tf": 1}}, "df": 1}}}}}}}}, "i": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "a": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}, "a": {"docs": {"pyerrors": {"tf": 3.7416573867739413}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1}, "pyerrors.fits.Fit_result": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.pandas.dump_df": {"tf": 1}, "pyerrors.linalg.matmul": {"tf": 1}}, "df": 8}, "s": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "z": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.obs.Obs": {"tf": 1}}, "df": 1, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}, "e": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1, "d": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1.4142135623730951}}, "df": 1}}}}}}}}, "e": {"docs": {}, "df": 0, "w": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}}, "df": 1}}}, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}}, "df": 1}}}}, "e": {"docs": {}, "df": 0, "x": {"docs": {"pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}}, "df": 1}}}, "y": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 3}, "s": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.openQCD.read_rwms": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.7320508075688772}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 2.23606797749979}, "pyerrors.input.sfcf.read_sfcf": {"tf": 2.449489742783178}}, "df": 6}}}}, "i": {"docs": {}, "df": 0, "f": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}, "b": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.json.load_json_dict": {"tf": 1}}, "df": 3}}}}}, "c": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1.7320508075688772}}, "df": 2, "s": {"docs": {"pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1.4142135623730951}}, "df": 2}}}}}}, "\\": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "\\": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "{": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "}": {"docs": {}, "df": 0, "^": {"docs": {}, "df": 0, "m": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}}}}}}}}}}}}}}, "u": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.m_eff": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.prune": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1.4142135623730951}, "pyerrors.input.pandas.dump_df": {"tf": 1}, "pyerrors.linalg.jack_matmul": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 2}, "pyerrors.roots.find_root": {"tf": 1}}, "df": 15, "d": {"docs": {"pyerrors": {"tf": 2.8284271247461903}, "pyerrors.correlators.Corr.deriv": {"tf": 1}, "pyerrors.correlators.Corr.second_deriv": {"tf": 1}, "pyerrors.correlators.Corr.fit": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1.4142135623730951}, "pyerrors.fits.least_squares": {"tf": 2}, "pyerrors.fits.total_least_squares": {"tf": 1}, "pyerrors.fits.fit_lin": {"tf": 1.4142135623730951}, "pyerrors.fits.qqplot": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_rwms": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.7320508075688772}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}, "pyerrors.obs.Obs.is_zero_within_error": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1.4142135623730951}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 25}, "r": {"docs": {"pyerrors": {"tf": 2}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 2}, "s": {"docs": {"pyerrors.correlators.Corr.m_eff": {"tf": 1.7320508075688772}, "pyerrors.input.hadrons.Npr_matrix.g5H": {"tf": 1}}, "df": 2}, "f": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.input.openQCD.read_rwms": {"tf": 1}}, "df": 1}}}}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors": {"tf": 2}, "pyerrors.correlators.Corr": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}, "pyerrors.obs.derived_observable": {"tf": 1}}, "df": 5}}}, "u": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.correlators.Corr.projected": {"tf": 1}}, "df": 1}}}}}}, "l": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "i": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}, "n": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "q": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 2}}, "df": 1}}}}}, "t": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.4142135623730951}}, "df": 5}}}, "s": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.projected": {"tf": 1.4142135623730951}}, "df": 1}}}}}}}, "c": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}}, "df": 1}}}}}}}}}, "h": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 3}}}}}}}, "l": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}}}, "p": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "k": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}}, "df": 2}}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}}, "df": 1}}}}}}}, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}}, "df": 2}}}}, "p": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 3}, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.utils": {"tf": 1}}, "df": 1}}}}}}}}}, "j": {"docs": {"pyerrors": {"tf": 2}, "pyerrors.correlators.Corr.item": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.prune": {"tf": 1.4142135623730951}, "pyerrors.dirac.epsilon_tensor": {"tf": 1}, "pyerrors.dirac.epsilon_tensor_rank4": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1.7320508075688772}}, "df": 6, "o": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "w": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}, "u": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.plateau": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix.g5H": {"tf": 1}}, "df": 3, "u": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}, "l": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "a": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}}}, "m": {"docs": {}, "df": 0, "p": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1, "s": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}}}, "a": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "c": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "i": {"docs": {"pyerrors.fits.least_squares": {"tf": 1}}, "df": 1, "a": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.obs.derived_observable": {"tf": 1}}, "df": 2}}}}}, "k": {"docs": {}, "df": 0, "k": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "f": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.input": {"tf": 2.23606797749979}, "pyerrors.linalg.jack_matmul": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 2}, "pyerrors.obs.import_jackknife": {"tf": 1.4142135623730951}}, "df": 4}}}}}, "s": {"docs": {"pyerrors.obs.import_jackknife": {"tf": 1}}, "df": 1}}}}, "s": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 3.7416573867739413}, "pyerrors.correlators.Corr.dump": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 2.23606797749979}, "pyerrors.input.json.dump_to_json": {"tf": 2.23606797749979}, "pyerrors.input.json.import_json_string": {"tf": 2}, "pyerrors.input.json.load_json": {"tf": 1.7320508075688772}, "pyerrors.input.json.dump_dict_to_json": {"tf": 2.449489742783178}, "pyerrors.input.json.load_json_dict": {"tf": 1.4142135623730951}, "pyerrors.input.pandas.to_sql": {"tf": 1}, "pyerrors.input.pandas.dump_df": {"tf": 1}, "pyerrors.input.pandas.load_df": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.dump": {"tf": 1.4142135623730951}}, "df": 12}}}, "l": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "}": {"docs": {}, "df": 0, "v": {"docs": {"pyerrors.obs.covariance": {"tf": 1.4142135623730951}}, "df": 1}}, "^": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1}}}, "k": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.dirac.epsilon_tensor": {"tf": 1}, "pyerrors.dirac.epsilon_tensor_rank4": {"tf": 1}, "pyerrors.mpm.matrix_pencil_method": {"tf": 1.7320508075688772}}, "df": 4, "u": {"docs": {}, "df": 0, "b": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "k": {"docs": {}, "df": 0, "i": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}, "h": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}}, "e": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "p": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.correlators.Corr.thin": {"tf": 1}, "pyerrors.obs.correlate": {"tf": 1}}, "df": 3}}, "y": {"docs": {}, "df": 0, "w": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.input.openQCD.read_rwms": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1.4142135623730951}}, "df": 4, "s": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}}}}}}, "o": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, "\u2013": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "v": {"docs": {"pyerrors.fits.ks_test": {"tf": 1}}, "df": 1}}}}}}}}}}}}}}}}}, "a": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "a": {"1": {"docs": {"pyerrors.input.bdio.read_mesons": {"tf": 1}}, "df": 1}, "2": {"docs": {"pyerrors.input.bdio.read_mesons": {"tf": 1}}, "df": 1}, "docs": {"pyerrors.input.bdio.read_dSdm": {"tf": 1}}, "df": 1}}}}, "w": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}, "pyerrors.obs.derived_observable": {"tf": 1.7320508075688772}}, "df": 2}}}}}}, "h": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "p": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 2}, "l": {"docs": {}, "df": 0, "p": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "r": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix.g5H": {"tf": 1.4142135623730951}}, "df": 2}}}, "i": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.linalg.eigh": {"tf": 1}}, "df": 1}}}}}}, "e": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 3}}, "a": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {"pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 2}}}}}, "a": {"docs": {}, "df": 0, "v": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}, "e": {"docs": {"pyerrors": {"tf": 2.23606797749979}, "pyerrors.correlators.Corr": {"tf": 1.4142135623730951}, "pyerrors.correlators.Corr.projected": {"tf": 1}, "pyerrors.correlators.Corr.prune": {"tf": 1}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.json.create_json_string": {"tf": 1}, "pyerrors.input.json.dump_to_json": {"tf": 1}, "pyerrors.input.json.dump_dict_to_json": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.4142135623730951}, "pyerrors.obs.derived_observable": {"tf": 1.4142135623730951}, "pyerrors.obs.correlate": {"tf": 1}, "pyerrors.roots.find_root": {"tf": 1}}, "df": 19}}, "n": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}}, "df": 2, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1, "d": {"docs": {"pyerrors": {"tf": 1.4142135623730951}}, "df": 1}}, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.input.openQCD.extract_t0": {"tf": 1}}, "df": 1}}}}}, "k": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.correlators.Corr.Hankel": {"tf": 1.4142135623730951}}, "df": 1}}}}, "s": {"docs": {"pyerrors": {"tf": 2.6457513110645907}, "pyerrors.correlators.Corr.reweight": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.total_least_squares": {"tf": 1.4142135623730951}, "pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.json.import_json_string": {"tf": 1}, "pyerrors.input.json.load_json": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.4142135623730951}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1}, "pyerrors.obs.Obs.reweight": {"tf": 1}, "pyerrors.obs.reweight": {"tf": 1}, "pyerrors.obs.correlate": {"tf": 1}, "pyerrors.obs.covariance": {"tf": 1}}, "df": 20}, "d": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}}, "df": 5}}}}}}, "i": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 1.4142135623730951}, "pyerrors.obs.Obs.plot_history": {"tf": 1.4142135623730951}}, "df": 2}, "i": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}}, "d": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1.4142135623730951}}, "df": 1, "s": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1}}, "df": 1}}}}, "o": {"docs": {}, "df": 0, "w": {"docs": {"pyerrors": {"tf": 2}, "pyerrors.fits.least_squares": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1}, "pyerrors.input.pandas.to_sql": {"tf": 1}}, "df": 5}, "s": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1}}, "df": 1, "n": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "m": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}}}}, "r": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "z": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "l": {"docs": {"pyerrors.correlators.Corr.show": {"tf": 1}}, "df": 1}}}}}}}}, "l": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors.obs.covariance": {"tf": 1}}, "df": 1, "s": {"docs": {"pyerrors.correlators.Corr.prune": {"tf": 1}}, "df": 1}}}, "t": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "g": {"docs": {"pyerrors.fits.Fit_result": {"tf": 1}}, "df": 1}}}}}}}, "m": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "u": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1}}, "df": 1}}}}}}}}}}, "t": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, ":": {"docs": {}, "df": 0, "/": {"docs": {}, "df": 0, "/": {"docs": {}, "df": 0, "c": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "d": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "g": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "f": {"docs": {"pyerrors.dirac.epsilon_tensor": {"tf": 1}, "pyerrors.dirac.epsilon_tensor_rank4": {"tf": 1}}, "df": 2}}}}}}}}}}}}}}}, "d": {"docs": {}, "df": 0, "f": {"5": {"docs": {"pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_DistillationContraction_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_ExternalLeg_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Bilinear_hd5": {"tf": 1}, "pyerrors.input.hadrons.read_Fourquark_hd5": {"tf": 1}}, "df": 5}, "docs": {}, "df": 0}}, "u": {"docs": {}, "df": 0, "a": {"docs": {"pyerrors.mpm.matrix_pencil_method": {"tf": 1}}, "df": 1}}}, "q": {"docs": {}, "df": 0, "c": {"docs": {"pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1}}, "df": 1, "d": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}, "u": {"docs": {}, "df": 0, "a": {"docs": {}, "df": 0, "d": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}, "n": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors": {"tf": 2.23606797749979}}, "df": 1}, "i": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "s": {"docs": {"pyerrors": {"tf": 2.449489742783178}}, "df": 1}}}}, "l": {"docs": {}, "df": 0, "e": {"docs": {"pyerrors.fits.least_squares": {"tf": 1.4142135623730951}, "pyerrors.fits.qqplot": {"tf": 1.4142135623730951}}, "df": 2}}}, "u": {"docs": {}, "df": 0, "m": {"docs": {"pyerrors.correlators.Corr.T_symmetry": {"tf": 1}}, "df": 1}}}}, "r": {"docs": {}, "df": 0, "k": {"docs": {"pyerrors.input.openQCD.read_ms5_xsf": {"tf": 1.7320508075688772}, "pyerrors.input.sfcf.read_sfcf": {"tf": 1.4142135623730951}}, "df": 2, "s": {"docs": {"pyerrors.input.sfcf.read_sfcf": {"tf": 1.4142135623730951}}, "df": 1}}}}, "o": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 4}}, "df": 1, "e": {"docs": {"pyerrors": {"tf": 1}}, "df": 1}}}, "e": {"docs": {}, "df": 0, "s": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "i": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.hadrons.read_meson_hd5": {"tf": 1}}, "df": 1}}}}}, "r": {"docs": {}, "df": 0, "y": {"docs": {"pyerrors.input.pandas.read_sql": {"tf": 1.4142135623730951}}, "df": 1}}}}, "q": {"docs": {}, "df": 0, "p": {"docs": {}, "df": 0, "l": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "t": {"docs": {"pyerrors": {"tf": 1}, "pyerrors.fits.least_squares": {"tf": 1}}, "df": 2}}}}}, "t": {"docs": {}, "df": 0, "o": {"docs": {}, "df": 0, "p": {"docs": {"pyerrors.input.openQCD.read_qtop": {"tf": 1}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1}}, "df": 2}}}}, "z": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "r": {"docs": {}, "df": 0, "o": {"docs": {"pyerrors": {"tf": 1.7320508075688772}, "pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.correlators.Corr.Eigenvalue": {"tf": 1}, "pyerrors.correlators.Corr.show": {"tf": 1}, "pyerrors.input.hadrons.Npr_matrix": {"tf": 1}, "pyerrors.input.openQCD.extract_t0": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1}, "pyerrors.obs.Obs.gamma_method": {"tf": 1}, "pyerrors.obs.Obs.gm": {"tf": 1}, "pyerrors.obs.Obs.is_zero_within_error": {"tf": 1}, "pyerrors.obs.Obs.is_zero": {"tf": 1}, "pyerrors.obs.CObs.is_zero": {"tf": 1}}, "df": 12, "t": {"docs": {}, "df": 0, "h": {"docs": {"pyerrors.correlators.Corr.GEVP": {"tf": 1}, "pyerrors.obs.Obs.export_jackknife": {"tf": 1}, "pyerrors.obs.import_jackknife": {"tf": 1}}, "df": 3}}, "s": {"docs": {"pyerrors.input.hadrons.Npr_matrix": {"tf": 1.4142135623730951}}, "df": 1}}}, "u": {"docs": {}, "df": 0, "t": {"docs": {}, "df": 0, "h": {"docs": {}, "df": 0, "e": {"docs": {}, "df": 0, "n": {"docs": {"pyerrors.input.dobs.create_pobs_string": {"tf": 1}, "pyerrors.input.dobs.write_pobs": {"tf": 1}, "pyerrors.input.dobs.read_pobs": {"tf": 1}, "pyerrors.input.dobs.import_dobs_string": {"tf": 1}, "pyerrors.input.dobs.read_dobs": {"tf": 1}, "pyerrors.input.dobs.create_dobs_string": {"tf": 1}, "pyerrors.input.dobs.write_dobs": {"tf": 1}, "pyerrors.input.openQCD.read_qtop": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_gf_coupling": {"tf": 1.4142135623730951}, "pyerrors.input.openQCD.read_qtop_sector": {"tf": 1.4142135623730951}}, "df": 10}}}}}}}}}}, "pipeline": ["trimmer"], "_isPrebuiltIndex": true};
// mirrored in build-search-index.js (part 1)
// Also split on html tags. this is a cheap heuristic, but good enough.