Loading ivas_processing_scripts/__init__.py +4 −4 Original line number Diff line number Diff line Loading @@ -95,16 +95,16 @@ def main(args): if hasattr(args, "multiprocessing"): cfg.multiprocessing = args.multiprocessing # set up logging logger = logging_init(args, cfg) # set up processing chains chains.init_processing_chains(cfg, logger) chains.init_processing_chains(cfg) # context manager to create output directories and clean up temporary directories with DirManager( cfg.out_dirs + cfg.tmp_dirs, cfg.tmp_dirs if cfg.delete_tmp else [] ): # set up logging logger = logging_init(args, cfg) # Re-ordering items based on concatenation order if hasattr(cfg, "preprocessing_2"): if ( Loading ivas_processing_scripts/processing/chains.py +6 −7 Original line number Diff line number Diff line Loading @@ -30,7 +30,6 @@ # the United Nations Convention on Contracts on the International Sales of Goods. # import logging from pathlib import Path from shutil import copyfile, rmtree from typing import Optional Loading @@ -51,7 +50,7 @@ from ivas_processing_scripts.processing.processing_splitting_scaling import ( from ivas_processing_scripts.utils import get_abs_path, list_audio, parse_gain def init_processing_chains(cfg: TestConfig, logger: logging.Logger) -> None: def init_processing_chains(cfg: TestConfig) -> None: """initialise processing chains for each condition and list items to process""" # pre-processing - special case; should only be run once if hasattr(cfg, "preprocessing"): Loading Loading @@ -112,7 +111,7 @@ def init_processing_chains(cfg: TestConfig, logger: logging.Logger) -> None: cfg.tmp_dirs.append(cfg.output_path.joinpath(f"tmp_{chain['name']}")) # delete output files if requested clean_outputs(cfg, logger) clean_outputs(cfg) # validate input files for correct format and sampling rate validate_input_files(cfg) Loading Loading @@ -710,7 +709,7 @@ def validate_input_files(cfg: TestConfig): cfg.input_path = output_dir def clean_outputs(cfg: TestConfig, logger: logging.Logger) -> None: def clean_outputs(cfg: TestConfig) -> None: if cfg.delete_output: deletion_list = [ d Loading @@ -722,10 +721,10 @@ def clean_outputs(cfg: TestConfig, logger: logging.Logger) -> None: if Path(d).exists() ] if deletion_list: logger.warning( warn( "\nWARNING! The configuration key to delete output directories was specified!" ) logger.warning( warn( f"The following directories will be REMOVED from {cfg.output_path}:\n {', '.join([d.name for d in deletion_list])}\n" ) confirm = input( Loading @@ -735,6 +734,6 @@ def clean_outputs(cfg: TestConfig, logger: logging.Logger) -> None: for dir in deletion_list: rmtree(dir) else: logger.warning( print( "Deletion was canceled. Please remove the output directories manually." ) Loading
ivas_processing_scripts/__init__.py +4 −4 Original line number Diff line number Diff line Loading @@ -95,16 +95,16 @@ def main(args): if hasattr(args, "multiprocessing"): cfg.multiprocessing = args.multiprocessing # set up logging logger = logging_init(args, cfg) # set up processing chains chains.init_processing_chains(cfg, logger) chains.init_processing_chains(cfg) # context manager to create output directories and clean up temporary directories with DirManager( cfg.out_dirs + cfg.tmp_dirs, cfg.tmp_dirs if cfg.delete_tmp else [] ): # set up logging logger = logging_init(args, cfg) # Re-ordering items based on concatenation order if hasattr(cfg, "preprocessing_2"): if ( Loading
ivas_processing_scripts/processing/chains.py +6 −7 Original line number Diff line number Diff line Loading @@ -30,7 +30,6 @@ # the United Nations Convention on Contracts on the International Sales of Goods. # import logging from pathlib import Path from shutil import copyfile, rmtree from typing import Optional Loading @@ -51,7 +50,7 @@ from ivas_processing_scripts.processing.processing_splitting_scaling import ( from ivas_processing_scripts.utils import get_abs_path, list_audio, parse_gain def init_processing_chains(cfg: TestConfig, logger: logging.Logger) -> None: def init_processing_chains(cfg: TestConfig) -> None: """initialise processing chains for each condition and list items to process""" # pre-processing - special case; should only be run once if hasattr(cfg, "preprocessing"): Loading Loading @@ -112,7 +111,7 @@ def init_processing_chains(cfg: TestConfig, logger: logging.Logger) -> None: cfg.tmp_dirs.append(cfg.output_path.joinpath(f"tmp_{chain['name']}")) # delete output files if requested clean_outputs(cfg, logger) clean_outputs(cfg) # validate input files for correct format and sampling rate validate_input_files(cfg) Loading Loading @@ -710,7 +709,7 @@ def validate_input_files(cfg: TestConfig): cfg.input_path = output_dir def clean_outputs(cfg: TestConfig, logger: logging.Logger) -> None: def clean_outputs(cfg: TestConfig) -> None: if cfg.delete_output: deletion_list = [ d Loading @@ -722,10 +721,10 @@ def clean_outputs(cfg: TestConfig, logger: logging.Logger) -> None: if Path(d).exists() ] if deletion_list: logger.warning( warn( "\nWARNING! The configuration key to delete output directories was specified!" ) logger.warning( warn( f"The following directories will be REMOVED from {cfg.output_path}:\n {', '.join([d.name for d in deletion_list])}\n" ) confirm = input( Loading @@ -735,6 +734,6 @@ def clean_outputs(cfg: TestConfig, logger: logging.Logger) -> None: for dir in deletion_list: rmtree(dir) else: logger.warning( print( "Deletion was canceled. Please remove the output directories manually." )