# Mage.py -- Pamhyr # Copyright (C) 2023 INRAE # # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, either version 3 of the License, or # (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <https://www.gnu.org/licenses/>. # -*- coding: utf-8 -*- import os import logging import numpy as np from tools import timer, trace from Solver.CommandLine import CommandLineSolver from Checker.Mage import MageNetworkGraphChecker from Model.Results.Results import Results from Model.Results.River.River import River, Reach, Profile logger = logging.getLogger() def mage_file_open(filepath, mode): f = open(filepath, mode) if "w" in mode: # Write header f.write("* This file is generated by PAMHYR, please don't modify\n") return f class Mage(CommandLineSolver): _type = "mage" def __init__(self, name): super(Mage, self).__init__(name) self._type = "mage" self._cmd_input = "" self._cmd_solver = "@path @input -o @output" self._cmd_output = "" @classmethod def default_parameters(cls): lst = super(Mage, cls).default_parameters() lst += [ ("mage_min_timestep", "1.0"), ("mage_timestep_tra", "3600"), ("mage_timestep_bin", "0"), # ("mage_timestep_melissa", "0"), ("mage_implicitation", "0.70"), ("mage_continuity_discretization", "S"), ("mage_qsj_discretization", "B"), ("mage_stop_criterion_iterations", "R"), ("mage_iteration_type", "0"), ("mage_smooth_coef", "0"), ("mage_cfl_max", "-1."), ("mage_min_height", "0.1"), ("mage_max_niter", "10"), ("mage_timestep_reduction_factor", "2"), ("mage_precision_reduction_factor_Z", "1"), ("mage_precision_reduction_factor_Q", "1"), ("mage_niter_max_precision", "99"), ("mage_niter_before_switch", "99"), ("mage_max_froude", "1.5"), ("mage_diffluence_node_height_balance", "-1"), ("mage_compute_reach_volume_balance", "y"), ("mage_max_reach_volume_balance", "0.001"), ("mage_min_reach_volume_to_check", "1000.0"), ("mage_init_internal", " "), ] return lst @classmethod def checkers(cls): lst = [ MageNetworkGraphChecker(connectivity=True), MageNetworkGraphChecker(connectivity=False) ] return lst ########## # Export # ########## def cmd_args(self, study): lst = super(Mage, self).cmd_args(study) lst.append("-r") return lst def input_param(self): name = self._study.name return f"{name}.REP" def output_param(self): name = self._study.name return f"{name}.BIN" def log_file(self): name = self._study.name return f"{name}.TRA" @timer def _export_ST(self, study, repertory, qlog, name="0"): files = [] if qlog is not None: qlog.put("Export ST file") os.makedirs(os.path.join(repertory, "net"), exist_ok=True) # Write header edges = study.river.edges() edges = list( filter( lambda e: e.is_enable(), edges ) ) for edge in edges: name = edge.name.replace(" ", "_") if edge._name == "": name = f"Reach_{edge.id}" with mage_file_open( os.path.join(repertory, "net", f"{name}.ST"), "w+" ) as f: files.append(str(os.path.join("net", f"{name}.ST"))) cnt_num = 1 for profile in edge.reach.profiles: self._export_ST_profile_header( f, files, profile, cnt_num ) cnt_num += 1 # Points for point in profile.points: self._export_ST_point_line( f, files, point ) # Profile last line f.write(f" 999.9990 999.9990 999.9990\n") return files def _export_ST_profile_header(self, wfile, files, profile, cnt): num = f"{cnt:>6}" c1 = f"{profile.code1:>6}" c2 = f"{profile.code2:>6}" t = f"{len(profile.points):>6}" kp = f"{profile.kp:>12f}"[0:12] pname = profile.name if profile.name == "": # Generate name from profile id prefixed with # 'p' (and replace space char with '0' char) pname = f"p{profile.id:>3}".replace(" ", "0") name = f"{pname:<19}" # Generate sediment additional data if available sediment = "" if profile.sl is not None: if not any(filter(lambda f: ".GRA" in f, files)): files.append(self._gra_file) # Number of layers nl = len(profile.sl) sediment = f" {nl:>3}" # Layers data for layer in profile.sl.layers: sediment += ( f" {layer.height:>10} {layer.d50:>10} " + f"{layer.sigma:>10} " + f"{layer.critical_constraint:>10}" ) # Profile header line wfile.write(f"{num}{c1}{c2}{t} {kp} {pname} {sediment}\n") def _export_ST_point_line(self, wfile, files, point): x = f"{point.x:<12.4f}"[0:12] y = f"{point.y:<12.4f}"[0:12] z = f"{point.z:<12.4f}"[0:12] n = f"{point.name:<3}" # Generate sediment additional data if available sediment = "" prev = point.z if point.sl is not None: # Number of layers nl = len(point.sl) sediment = f"{nl:>3}" # Layers data for layer in point.sl.layers: prev = round(prev - layer.height, 5) sediment += ( f" {prev:>10} {layer.d50:>10} " + f"{layer.sigma:>10} " + f"{layer.critical_constraint:>10}" ) # Point line wfile.write(f"{x} {y} {z} {n} {sediment}\n") @timer def _export_BC(self, t, bounds, repertory, qlog, name="0"): files = [] if len(bounds) == 0: return files if qlog is not None: qlog.put(f"Export {t} file") with mage_file_open(os.path.join(repertory, f"{name}.{t}"), "w+") as f: files.append(f"{name}.{t}") for bound in bounds: name = f"{bound.node.id:3}".replace(" ", "x") f.write(f"* {bound.node.name} ({name}) {bound.bctype}\n") f.write(f"${name}\n") header = bound.header f.write(f"*{header[0]:>9}|{header[1]:>10}\n") for d in bound.data: v0 = d[0] v1 = d[1] if t in ["HYD", "QSO", "LIM"]: v0 /= 60 # Convert first column to minute f.write(f"{v0:10}{v1:10}\n") return files @timer def _export_bound_cond(self, study, repertory, qlog, name="0"): files = [] lst = study.river.boundary_condition AVA = [] HYD = [] LIM = [] QSO = [] for tab in ["liquid", "solid", "suspenssion"]: for bound in lst.get_tab(tab): if bound.node is None: continue if bound.bctype == "ZD": AVA.append(bound) elif bound.bctype == "TD" or bound.bctype == "PC": HYD.append(bound) elif bound.bctype == "TZ": LIM.append(bound) elif bound.bctype == "SL": QSO.append(bound) files = files + self._export_BC("AVA", AVA, repertory, qlog, name=name) files = files + self._export_BC("HYD", HYD, repertory, qlog, name=name) files = files + self._export_BC("LIM", LIM, repertory, qlog, name=name) files = files + self._export_QSO(QSO, repertory, qlog, name=name) return files # @timer # def _export_LC(self, lateral, repertory, qlog, name="0"): # files = [] # if qlog is not None: # qlog.put(f"Export LAT file") # with mage_file_open( # os.path.join(repertory, f"{name}.LAT"), # "w+" # ) as f: # files.append(f"{name}.LAT") # name = f"{lateral.node.id:3}".replace(" ", "x") # f.write(f"* {lateral.node.name} ({name}) {lateral.bctype}\n") # f.write(f"${name}\n") # header = lateral.header # f.write(f"*{header[0]:>9}|{header[1]:>10}\n") # for d in lateral.data: # f.write(f"{d[0]:1{name}.3f}{d[1]:10.3f}\n") # return files # @timer # def _export_lateral_contrib(self, study, repertory, qlog, name="0"): # files = [] # lst = study.river.lateral_contribution # for tab in ["liquid", "solid", "suspenssion"]: # for lateral in lst.get_tab(tab): # files = files + self._export_LC(lateral, repertory, qlog) # return files @timer def _export_RUG(self, study, repertory, qlog, name="0"): files = [] if qlog is not None: qlog.put("Export RUG file") # Write header with mage_file_open(os.path.join(repertory, f"{name}.RUG"), "w+") as f: files.append(f"{name}.RUG") edges = study.river.edges() edges = list( filter( lambda e: e.is_enable(), edges ) ) id = 1 for edge in edges: frictions = edge.frictions for friction in frictions.frictions: num = f"{id:>3}" bkp = f"{friction.begin_kp:>10.3f}" ekp = f"{friction.end_kp:>10.3f}" # if friction.begin_kp != friction.end_kp: # print("TODO") strickler = friction.begin_strickler coef_1 = f"{strickler.minor:>10.3f}" coef_2 = f"{strickler.medium:>10.3f}" f.write(f"K{num} {bkp}{ekp}{coef_1}{coef_2}\n") id += 1 return files @timer def _export_INI(self, study, repertory, qlog, name="0"): files = [] if qlog is not None: qlog.put("Export INI file") # Write header with mage_file_open(os.path.join(repertory, f"{name}.INI"), "w+") as f: has_ini = False id = 1 reachs = study.river.enable_edges() # TODO put real date... f.write(f"$ date en minutes : 0.00\n") f.write(f"* IB IS discharge elevation kp\n") id = 1 for reach in reachs: cond = study.river.initial_conditions.get(reach) data = cond.data if len(data) == 0: continue has_ini = True id_sec = 1 for d in data: IR = f"{id}" IS = f"{id_sec}" discharge = f"{d['discharge']:>10.5f}" elevation = f"{d['elevation']:>11.6f}" kp = f"{d['kp']:>9.2f}" f.write(f"{IR} {IS} {discharge} {elevation} {kp}\n") id_sec += 1 id += 1 if has_ini: files.append(f"{name}.INI") return files @timer def _export_CAS(self, study, repertory, qlog, name="0"): files = [] reservoirs = study.river.reservoir.lst if len(reservoirs) == 0: return files if qlog is not None: qlog.put("Export CAS file") with mage_file_open(os.path.join(repertory, f"{name}.CAS"), "w+") as f: files.append(f"{name}.CAS") for reservoir in reservoirs: reservoir.sort() node = reservoir.node name = f"{node.id:3}".replace(" ", "x") f.write(f"* {node.name} ({name}) Reservoir\n") f.write(f"${name}\n") f.write(f"*{'Elev(m)':>9}|{'Area(ha)':>10}\n") for d in reservoir.data: v0 = d[0] v1 = d[1] f.write(f"{v0:>10.3f}{v1:>10.3f}\n") return files @timer def _export_SIN(self, study, repertory, qlog, name="0"): files = [] sin_dict = { "ND": "*", "S1": "D", "S2": "T", "S3": "T", "OR": "O", "OC": "B", "OV": "F", "V1": "V", "V2": "W", "BO": "A", "UD": "X", "PO": "P", } hydraulic_structures = study.river.hydraulic_structures.lst if len(hydraulic_structures) == 0: return files if qlog is not None: qlog.put("Export SIN file") with mage_file_open(os.path.join(repertory, f"{name}.SIN"), "w+") as f: files.append(f"{name}.SIN") for hs in hydraulic_structures: if not hs.input_reach.is_enable(): continue f.write( '* ouvrage au pk ' + f"{hs.input_kp:>12.1f}" + ' ' + hs.name + '\n' ) for bhs in hs.basic_structures: reach_id = study.river.get_edge_id(hs.input_reach) + 1 param_str = ' '.join( [ f'{p:>10.3f}' for p in self._export_SIN_parameters(bhs) ] ) f.write( f"{sin_dict[bhs._type]} " + f"{reach_id} {hs.input_kp:>12.3f} {param_str} " + f"{bhs.name}\n" ) return files def _export_SIN_parameters(self, bhs): res = [9999.999] * 5 if len(bhs) == 5: res = self._export_SIN_parameters_5(bhs) elif len(bhs) == 4: res = self._export_SIN_parameters_4(bhs) elif len(bhs) == 3: res = self._export_SIN_parameters_3(bhs) return res def _export_SIN_parameters_5(self, bhs): # S2, OR, V1, V2, UD return [ bhs._data[0].value, bhs._data[1].value, bhs._data[2].value, bhs._data[3].value, bhs._data[4].value, ] def _export_SIN_parameters_4(self, bhs): # S3, OC res = [ bhs._data[0].value, bhs._data[1].value, bhs._data[2].value, bhs._data[3].value, 0.0, ] if bhs._type == "T": # S3 res = [0.0] + res[:-1] return res def _export_SIN_parameters_3(self, bhs): # S1, BO if bhs._type == "S1": res = [ bhs._data[0].value, bhs._data[1].value, 0.0, bhs._data[2].value, 9999.99, ] else: res = [ bhs._data[0].value, bhs._data[1].value, bhs._data[2].value, 0.0, 0.0, ] return res @timer def _export_DEV(self, study, repertory, qlog, name="0"): files = [] if qlog is not None: qlog.put("Export DEV file") with mage_file_open( os.path.join( repertory, f"{name}.DEV" ), "w+" ) as f: reachs = study.river.enable_edges() id = 1 for reach in reachs: f.write(f"YD{id:3}\n") f.write(f"YG{id:3}\n") id += 1 files.append(f"{name}.DEV") return files @timer def _export_REP(self, study, repertory, files, qlog, name="0"): if qlog is not None: qlog.put("Export REP file") # Write header with mage_file_open( os.path.join( repertory, f"{name}.REP" ), "w+" ) as f: f.write("confirmation=non\n") for file in files: EXT = file.split('.')[1] if EXT not in ["ST", "GRA"]: f.write(f"{EXT} {file}\n") f.write("* OUTPUT\n") f.write(f"TRA {name}.TRA\n") f.write(f"BIN {name}.BIN\n") for file in files: EXT = file.split('.')[1] if EXT in ["GRA"]: f.write(f"{EXT} {file}\n") @timer def export(self, study, repertory, qlog=None): self._study = study name = study.name.replace(" ", "_") # Define GRA file name self._gra_file = f"{name}.GRA" self._bin_file = f"{name}.BIN" self._export_ST(study, repertory, qlog, name=name) return True ########### # RESULTS # ########### def read_bin(self, study, repertory, results, qlog=None, name="0"): return @timer def results(self, study, repertory, qlog=None, name="0"): results = Results( study=study, solver=self, repertory=repertory, name=name, ) self.read_bin(study, repertory, results, qlog, name=name) return results ########## # MAGE 7 # ########## class Mage7(Mage): _type = "mage7" def __init__(self, name): super(Mage7, self).__init__(name) self._type = "mage7" @classmethod def default_parameters(cls): lst = super(Mage7, cls).default_parameters() return lst ########## # MAGE 8 # ########## class Mage8(Mage): _type = "mage8" def __init__(self, name): super(Mage8, self).__init__(name) self._type = "mage8" @classmethod def default_parameters(cls): lst = super(Mage8, cls).default_parameters() # Insert new parameters at specific position names = list(map(lambda t: t[0], lst)) i = names.index("mage_precision_reduction_factor_Q") lst.insert(i+1, ("mage_precision_reduction_factor_r", "1")) # Mage parameter for sediment module (added in DB 0.0.4) lst.append(("mage_sediment_masse_volumique", "2650.0")) lst.append(("mage_sediment_angle_repos", "40.0")) lst.append(("mage_sediment_porosity", "0.40")) lst.append(("mage_distance_Han", "0.0")) lst.append(("mage_distance_chargement_d50", "100.0")) lst.append(("mage_distance_chargement_sigma", "100.0")) lst.append(("mage_methode_modification_geometrie", "1")) lst.append(("mage_shields_critique", "1")) lst.append(("mage_shields_correction", "1")) lst.append(("mage_capacite_solide", "1")) lst.append(("mage_pas_de_temps_charriage", "1")) lst.append(("mage_facteur_multiplicateur", "1.0")) return lst ########## # Export # ########## def cmd_args(self, study): lst = super(Mage8, self).cmd_args(study) if study.river.has_sediment(): lst.append("-c=3") return lst @timer def _export_PAR(self, study, repertory, qlog=None, name="0"): files = [] if qlog is not None: qlog.put("Export PAR file") with mage_file_open(os.path.join(repertory, f"{name}.PAR"), "w+") as f: files.append(f"{name}.PAR") params = study.river.get_params(self.type).parameters for p in params: name = p.name\ .replace("all_", "")\ .replace("mage_", "") value = p.value if name in ["command_line_arguments"]: continue if name == "compute_reach_volume_balance": value = "O" if value.lower() == "y" else "N" if name == "init_internal": value = ("p" if value.lower() in ["y", "yes", "true"] else "") f.write(f"{name} {value}\n") return files @timer def _export_NET(self, study, repertory, qlog=None, name="0"): files = [] if qlog is not None: qlog.put("Export NET file") with mage_file_open(os.path.join(repertory, f"{name}.NET"), "w+") as f: files.append(f"{name}.NET") edges = study.river.edges() edges = list( filter( lambda e: e.is_enable(), edges ) ) for e in edges: name = e.name.replace(" ", "_") if e._name == "": name = f"Reach_{e.id}" id = f"Bief_{e.id+1}" n1 = f"{e.node1.id:3}".replace(" ", "x") n2 = f"{e.node2.id:3}".replace(" ", "x") file = os.path.join("net", name + ".ST") f.write(f"{id} {n1} {n2} {file}\n") return files @timer def _export_QSO(self, bounds, repertory, qlog, name="0"): files = [] if len(bounds) == 0: return files if qlog is not None: qlog.put(f"Export QSO file") with mage_file_open(os.path.join(repertory, f"{name}.QSO"), "w+") as f: files.append(f"{name}.QSO") for bound in bounds: # File header name = f"{bound.node.id:3}".replace(" ", "x") f.write(f"* {bound.node.name} ({name}) {bound.bctype}\n") d50 = bound.d50 sigma = bound.sigma if len(bound.data) == 0: f.write(f"${name} {d50} {sigma} default\n") else: f.write(f"${name} {d50} {sigma}\n") # Table header header = bound.header f.write(f"*{header[0]:>9}|{header[1]:>10}\n") # Data for d in bound.data: f.write(f"{d[0]:10.3f}{d[1]:10.3f}\n") return files @timer def export(self, study, repertory, qlog=None, name="0"): self._study = study name = study.name.replace(" ", "_") # Define GRA file name self._gra_file = f"{name}.GRA" self._bin_file = f"{name}.BIN" # Generate files files = [] files = self._export_ST(study, repertory, qlog, name=name) files = files + self._export_PAR(study, repertory, qlog, name=name) files = files + self._export_NET(study, repertory, qlog, name=name) files = files + \ self._export_bound_cond(study, repertory, qlog, name=name) files = files + self._export_RUG(study, repertory, qlog, name=name) files = files + self._export_INI(study, repertory, qlog, name=name) files = files + self._export_SIN(study, repertory, qlog, name=name) files = files + self._export_CAS(study, repertory, qlog, name=name) files = files + self._export_DEV(study, repertory, qlog, name=name) self._export_REP(study, repertory, files, qlog, name=name) return True ########### # RESULTS # ########### @timer def read_bin(self, study, repertory, results, qlog=None, name="0"): fname = os.path.join(repertory, f"{name}.BIN") logger.info(f"read_bin: Start reading '{fname}' ...") with mage_file_open(fname, "r") as f: def newline(): return np.fromfile(f, dtype=np.int32, count=1) def endline(): return np.fromfile(f, dtype=np.int32, count=1) def read_int(size): return np.fromfile( f, dtype=np.int32, count=size) def read_float(size): return np.fromfile( f, dtype=np.float32, count=size) def read_float64(size): return np.fromfile( f, dtype=np.float64, count=size) # Meta data (1st line) newline() data = read_int(3) nb_reach = data[0] nb_profile = data[1] mage_version = data[2] logger.debug(f"read_bin: nb_reach = {nb_reach}") logger.debug(f"read_bin: nb_profile = {nb_profile}") logger.debug(f"read_bin: mage_version = {mage_version}") if mage_version <= 80: msg = ( "Read BIN files: " + f"Possible incompatible mage version '{mage_version}', " + "please check your solver configuration..." ) logger.warning(msg) if qlog is not None: qlog.put("[WARNING] " + msg) results.set("solver_version", f"Mage8 ({mage_version})") results.set("nb_reach", f"{nb_reach}") results.set("nb_profile", f"{nb_profile}") endline() # Reach information (2nd line) newline() reachs = [] iprofiles = {} reach_offset = {} data = read_int(2*nb_reach) for i in range(nb_reach): # Add results reach to reach list r = results.river.add(i) reachs.append(r) # ID of first and last reach profiles i1 = data[2*i] - 1 i2 = data[2*i+1] - 1 # Add profile id correspondance to reach key = (i1, i2) iprofiles[key] = r # Profile ID offset reach_offset[r] = i1 logger.debug(f"read_bin: iprofiles = {iprofiles}") endline() # X (3rd line) newline() _ = read_float(nb_profile) endline() # Z and Y (4th line) newline() _ = read_float(3*nb_profile) endline() # Data newline() def ip_to_r(i): return iprofiles[ next( filter( lambda k: k[0] <= i <= k[1], iprofiles ) ) ] def ip_to_ri(r, i): return i - reach_offset[r] ts = set() end = False while not end: n = read_int(1)[0] timestamp = read_float64(1)[0] key = bytearray(np.fromfile( f, dtype=np.byte, count=1)).decode() data = read_float(n) logger.debug(f"read_bin: timestamp = {timestamp} sec") ts.add(timestamp) if key in ["Z", "Q"]: for i, d in enumerate(data): # Get reach corresponding to profile ID reach = ip_to_r(i) # Get profile id in reach ri = ip_to_ri(reach, i) # Set data for profile RI reach.set(ri, timestamp, key, d) if key == "Z": profile = study.river\ .current_reach()\ .reach.profile(ri) ptX, ptY = profile.get_water_limits(d) reach.set(ri, timestamp, "ptX", ptX) reach.set(ri, timestamp, "ptY", ptY) endline() end = newline().size <= 0 logger.debug(reachs[0].profiles[0]._data) results.set("timestamps", ts) logger.info(f"read_bin: ... end with {len(ts)} timestamp read") @timer def read_gra(self, study, repertory, results, qlog=None, name="0"): if not study.river.has_sediment(): return fname = os.path.join(repertory, f"{name}.GRA") logger.info(f"read_gra: Start reading '{fname}' ...") with mage_file_open(fname, "r") as f: def newline(): return np.fromfile(f, dtype=np.int32, count=1) def endline(): return np.fromfile(f, dtype=np.int32, count=1) def read_int(size): return np.fromfile( f, dtype=np.int32, count=size) def read_float(size): return np.fromfile( f, dtype=np.float32, count=size) def read_float64(size): return np.fromfile( f, dtype=np.float64, count=size) # Meta data (1st line) newline() data = read_int(3) nb_reach = data[0] nb_profile = data[1] mage_version = data[2] logger.debug(f"read_gra: nb_reach = {nb_reach}") logger.debug(f"read_gra: nb_profile = {nb_profile}") logger.debug(f"read_gra: mage_version = {mage_version}") if mage_version < 80: msg = ( "Read GRA files: " + f"Possible incompatible mage version '{mage_version}', " + "please check your solver configuration..." ) logger.warning(msg) if qlog is not None: qlog.put("[WARNING] " + msg) results.set("gra_solver_version", f"Mage8 ({mage_version})") results.set("gra_nb_reach", f"{nb_reach}") results.set("gra_nb_profile", f"{nb_profile}") endline() # Reach information (2nd line) newline() reachs = [] iprofiles = {} reach_offset = {} data = read_int(2*nb_reach) for i in range(nb_reach): # Get results reach to reach list r = results.river.reach(i) reachs.append(r) # ID of first and last reach profiles i1 = data[2*i] - 1 i2 = data[2*i+1] - 1 # Add profile id correspondance to reach key = (i1, i2) iprofiles[key] = r # Profile ID offset reach_offset[r] = i1 logger.debug(f"read_gra: iprofiles = {iprofiles}") endline() # X (3rd line) newline() _ = read_float(nb_profile) endline() # npts (4th line) newline() _ = read_int(nb_profile) endline() # Data def ip_to_r(i): return iprofiles[ next( filter( lambda k: k[0] <= i <= k[1], iprofiles ) ) ] def ip_to_ri(r, i): return i - reach_offset[r] ts = set() end = False newline() while not end: n = read_int(1)[0] timestamp = read_float64(1)[0] with_bedload = read_int(1)[0] logger.debug(f"read_gra: Number of cross section: {n}") logger.debug(f"read_gra: Timestamp: {timestamp}") logger.debug(f"read_gra: Type of bedload: {with_bedload}") endline() npts = [1] * n if with_bedload == 1: newline() npts = read_int(n) endline() sum_npts = sum(npts) logger.debug(f"read_gra: Number of points: {sum_npts}") newline() nsl = read_int(sum_npts) logger.debug(f"read_gra: Number of sedimentary layers: {nsl}") endline() newline() data = read_float64(sum(nsl) * 3) endline() ts.add(timestamp) i_pts = 0 i_data = 0 # Loop on cross section for i in range(n): sec_sl = [] reach = ip_to_r(i) p_i = ip_to_ri(reach, i) # Loop on cross section points for j in range(npts[i]): sl = [] # Loop on sediment layers for k in range(nsl[i_pts]): h = data[i_data] d50 = data[i_data + 1] sigma = data[i_data + 2] sl.append((h, d50, sigma)) i_data += 3 i_pts += 1 sec_sl.append(sl) reach.set(p_i, timestamp, "sl", sec_sl) logger.debug( f"read_gra: data size = {len(data)} ({i_data} readed)" ) end = newline().size <= 0 results.set("sediment_timestamps", ts) logger.info(f"read_gra: ... end with {len(ts)} timestamp read") @timer def results(self, study, repertory, qlog=None): self._study = study name = study.name.replace(" ", "_") results = super(Mage8, self).results(study, repertory, qlog, name=name) self.read_gra(study, repertory, results, qlog, name=name) return results