Loading ivas_processing_scripts/processing/processing.py +3 −2 Original line number Diff line number Diff line Loading @@ -156,12 +156,13 @@ def concat_setup(cfg: TestConfig, chain, logger: logging.Logger): cfg.items_list = [cfg.concat_file] # write out splits with open(cfg.concat_file.with_suffix(".splits.log"), "w") as f: splits_info_file = cfg.concat_file.parent.joinpath(Path("undo_concat.txt")) with open(splits_info_file, "w") as f: print(", ".join([str(s) for s in cfg.splits]), file=f) print(", ".join([str(sn) for sn in cfg.split_names]), file=f) print(f"{fs}", file=f) logger.info(f"Splits written to file {cfg.concat_file.with_suffix('.splits.log')}") logger.info(f"Splits written to file {splits_info_file}") def concat_teardown(x, splits, out_fmt, fs, in_fs, meta, logger: logging.Logger): Loading ivas_processing_scripts/processing/processing_splitting_scaling.py +1 −8 Original line number Diff line number Diff line Loading @@ -208,14 +208,7 @@ class Processing_splitting_scaling(Processing): # reverse concatenation if self.concatenate_input: # read out splits file -> start/end, names, sampling rate if noerror: splits_info_file = Path( f"{in_file.parent.parent.joinpath('tmp_preprocessing_2').joinpath(Path(Path(in_file.stem).stem).stem)}.splits.log" ) else: splits_info_file = Path( f"{in_file.parent.parent.joinpath('tmp_preprocessing_2').joinpath(Path(in_file.stem).stem)}.splits.log" ) splits_info_file = Path(in_file.parent.parent.joinpath('tmp_preprocessing_2').joinpath("undo_concat.txt")) splits, split_names, split_fs = read_splits_file(splits_info_file) # split file Loading Loading
ivas_processing_scripts/processing/processing.py +3 −2 Original line number Diff line number Diff line Loading @@ -156,12 +156,13 @@ def concat_setup(cfg: TestConfig, chain, logger: logging.Logger): cfg.items_list = [cfg.concat_file] # write out splits with open(cfg.concat_file.with_suffix(".splits.log"), "w") as f: splits_info_file = cfg.concat_file.parent.joinpath(Path("undo_concat.txt")) with open(splits_info_file, "w") as f: print(", ".join([str(s) for s in cfg.splits]), file=f) print(", ".join([str(sn) for sn in cfg.split_names]), file=f) print(f"{fs}", file=f) logger.info(f"Splits written to file {cfg.concat_file.with_suffix('.splits.log')}") logger.info(f"Splits written to file {splits_info_file}") def concat_teardown(x, splits, out_fmt, fs, in_fs, meta, logger: logging.Logger): Loading
ivas_processing_scripts/processing/processing_splitting_scaling.py +1 −8 Original line number Diff line number Diff line Loading @@ -208,14 +208,7 @@ class Processing_splitting_scaling(Processing): # reverse concatenation if self.concatenate_input: # read out splits file -> start/end, names, sampling rate if noerror: splits_info_file = Path( f"{in_file.parent.parent.joinpath('tmp_preprocessing_2').joinpath(Path(Path(in_file.stem).stem).stem)}.splits.log" ) else: splits_info_file = Path( f"{in_file.parent.parent.joinpath('tmp_preprocessing_2').joinpath(Path(in_file.stem).stem)}.splits.log" ) splits_info_file = Path(in_file.parent.parent.joinpath('tmp_preprocessing_2').joinpath("undo_concat.txt")) splits, split_names, split_fs = read_splits_file(splits_info_file) # split file Loading