From aacac4e0a09f2a476ad558eb4789d1c4e3ed6bc2 Mon Sep 17 00:00:00 2001 From: Patrick Jentsch Date: Thu, 27 Jan 2022 13:48:16 +0100 Subject: [PATCH] Codestyle enhancements --- file-setup | 50 +++++++++++++++++++++++++++++++++++--------------- 1 file changed, 35 insertions(+), 15 deletions(-) diff --git a/file-setup b/file-setup index 7d14b9b..f61aa7b 100755 --- a/file-setup +++ b/file-setup @@ -7,7 +7,6 @@ __version__ = '0.1.0' from argparse import ArgumentParser from pyflow import WorkflowRunner import json -import multiprocessing import os import sys @@ -52,7 +51,9 @@ class CreatePDFWorkflow(WorkflowRunner): os.path.join(self.job.output_dir, '{}.pdf'.format(self.job.name)) ) cmd += ' && ' - cmd += 'rm "{}"'.format(os.path.join(self.job.output_dir, 'inputs.txt')) # noqa + cmd += 'rm "{}"'.format( + os.path.join(self.job.output_dir, 'inputs.txt') + ) self.addTask( 'convert', command=cmd, @@ -68,6 +69,7 @@ class MainWorkflow(WorkflowRunner): self.jobs = [] def collect_jobs(self): + self.jobs = [] for dir in os.listdir(self.input_dir): if not os.path.isdir(os.path.join(self.input_dir, dir)): continue @@ -102,38 +104,53 @@ class MainWorkflow(WorkflowRunner): create_pdf_tasks.append(task) self.waitForTasks() - output_files = [] + outputs = [] for job in self.jobs: # Track output files - relative_output_dir = os.path.relpath(job.output_dir, start=self.output_dir) # noqa - output_files.append( + relative_output_dir = os.path.relpath( + job.output_dir, + start=self.output_dir + ) + outputs.append( { 'description': 'PDF file without text layer.', - 'file': os.path.join(relative_output_dir, '{}.pdf'.format(job.name)), # noqa + 'file': os.path.join( + relative_output_dir, + '{}.pdf'.format(job.name) + ), 'mimetype': 'application/pdf' } ) - with open(os.path.join(self.output_dir, 'output_records.json'), 'w') as f: # noqa - json.dump(output_files, f, indent=4) + with open(os.path.join(self.output_dir, 'outputs.json'), 'w') as f: + json.dump(outputs, f, indent=4) def parse_args(): parser = ArgumentParser(description='Pipeline for merging images') parser.add_argument( - '-i', '--input-dir', help='Input directory', required=True) + '-i', '--input-dir', + help='Input directory', + required=True + ) parser.add_argument( - '-o', '--output-dir', help='Output directory', required=True) + '-o', '--output-dir', + help='Output directory', + required=True + ) parser.add_argument( - '--log-dir', help='Logging directory (Default: --output-dir)') + '--log-dir', + help='Logging directory (Default: --output-dir)' + ) parser.add_argument( '--mem-mb', - help='Amount of system memory to be used (Default: min(--n-cores * 256, available system memory))', # noqa + help='Amount of system memory to be used ' + '(Default: min(--n-cores * 256, available system memory))', type=int ) parser.add_argument( '--n-cores', - default=min(2, multiprocessing.cpu_count()), - help='Number of CPU threads to be used (Default: min(2, CPU count))', + default=1, + help='Number of CPU threads to be used', type=int ) parser.add_argument( @@ -162,7 +179,10 @@ def main(): main_workflow = MainWorkflow(args.input_dir, args.output_dir) main_workflow.collect_jobs() retval = main_workflow.run( - dataDirRoot=args.log_dir, memMb=args.mem_mb, nCores=args.n_cores) # noqa + dataDirRoot=args.log_dir, + memMb=args.mem_mb, + nCores=args.n_cores + ) sys.exit(retval)