forked from EisenRa/holoflow
-
Notifications
You must be signed in to change notification settings - Fork 0
/
metagenomics_IB_TMP-Compress.py
221 lines (170 loc) · 8.15 KB
/
metagenomics_IB_TMP-Compress.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
import argparse
import subprocess
import os
import sys
###########################
#Argument parsing
###########################
# Gather input files and variables from command line
parser = argparse.ArgumentParser(description='Runs holoflow pipeline.')
parser.add_argument('-f', help="input.txt file", dest="input_txt", required=True)
parser.add_argument('-d', help="temp files directory path", dest="work_dir", required=True)
parser.add_argument('-c', help="config file", dest="config_file", required=False)
parser.add_argument('-k', help="keep tmp directories", dest="keep", action='store_true')
parser.add_argument('-l', help="pipeline log file", dest="log", required=False)
parser.add_argument('-t', help="threads", dest="threads", required=True)
parser.add_argument('-N', help="JOB ID", dest="job", required=True)
parser.add_argument('-W', help="rewrite everything", dest="REWRITE", action='store_true')
args = parser.parse_args()
in_f=args.input_txt
path=args.work_dir
cores=args.threads
job=args.job
# retrieve current directory
file = os.path.dirname(sys.argv[0])
curr_dir = os.path.abspath(file)
# If the user does not specify a config file, provide default file in GitHub
if not (args.config_file):
cpconfigCmd= 'cp '+curr_dir+'/workflows/individual_binning/config.yaml '+path+'/'+job+'_config.yaml'
subprocess.Popen(cpconfigCmd,shell=True).wait()
config = path+'/'+job+'_config.yaml'
else:
config=args.config_file
# If the user does not specify a log file, provide default path
if not (args.log):
log = os.path.join(path,"Holoflow_individualA_metagenomics.log")
else:
log=args.log
# Load dependencies
loaddepCmd='module unload gcc && module load tools anaconda3/4.4.0'
subprocess.Popen(loaddepCmd,shell=True).wait()
#Append current directory to .yaml config for standalone calling
# see preprocessing.py for verbose description
import ruamel.yaml
yaml = ruamel.yaml.YAML()
yaml.explicit_start = True
with open(str(config), 'r') as config_file:
data = yaml.load(config_file)
if data == None:
data = {}
with open(str(config), 'w') as config_file:
data['threads'] = str(cores)
data['holopath'] = str(curr_dir)
data['logpath'] = str(log)
dump = yaml.dump(data, config_file)
###########################
## Functions
###########################
###########################
###### METAGENOMICS FUNCTIONS
def in_out_metagenomics(path,in_f):
"""Generate output names files from input.txt. Rename and move
input files where snakemake expects to find them if necessary."""
in_dir_0 = os.path.join(path,"PPR_03-MappedToReference")
if not os.path.exists(in_dir_0):
os.makedirs(in_dir_0)
with open(in_f,'r') as in_file:
# Define variables
output_files=''
final_temp_dir="MIB_04-BinMerging"
all_lines = in_file.readlines() # Read input.txt lines
# remove empty lines
all_lines = map(lambda s: s.strip(), all_lines)
lines = list(filter(None, list(all_lines)))
if os.path.exists(in_dir_0): # Already run for: same job (wants to continue/Rewrite), for another job
# Define specific job dir
in_dir=in_dir_0+'/'+job
# Define specific job final output dir - for snakemake (needs output files)
final_temp_dir=final_temp_dir+'/'+job
# If user wants to remove previous runs' data and run from scratch
if args.REWRITE:
if os.path.exists(in_dir):
rmCmd='rm -rf '+in_dir+''
subprocess.Popen(rmCmd,shell=True).wait()
if not os.path.exists(in_dir): # if specific job input directory does not exist
os.makedirs(in_dir)
else: # already exists and don't want to rewrite, then pass
pass
# If directory is empty, do all - otherwise, just save output names
if len(os.listdir(in_dir) ) == 0:
for line in lines:# for line in lines in input file, do:
### Skip line if starts with # (comment line)
if not (line.startswith('#')):
line = line.strip('\n').split(' ') # Create a list of each line
sample_name=line[0]
in_for=line[1]# input for (read1) file
in_rev=line[2] # input reverse (read2) file
# Define input file
in1=in_dir+'/'+sample_name+'_1.fastq.gz'
# Check if input files already in desired dir
if os.path.isfile(in1):
pass
else:
#If the file is not in the working directory, create soft link in it
if os.path.isfile(in_for):
if in_for.endswith('.gz'):# if compressed, decompress in standard dir with std ID
read1Cmd = 'ln -s '+in_for+' '+in1+''
subprocess.Popen(read1Cmd, shell=True).wait()
else:
read1Cmd = 'gzip -c '+in_for+' > '+in1+''
subprocess.Popen(read1Cmd, shell=True).wait()
# Define input file
in2=in_dir+'/'+sample_name+'_2.fastq.gz'
# Check if input files already in desired dir
if os.path.isfile(in2):
pass
else:
#If the file is not in the working directory, transfer it
if os.path.isfile(in_rev):
if in_for.endswith('.gz'):
read2Cmd = 'ln -s '+in_rev+' '+in2+''
subprocess.Popen(read2Cmd, shell=True).wait()
else:
read2Cmd = 'gzip -c '+in_rev+' > '+in2+''
subprocess.Popen(read2Cmd, shell=True).wait()
output_files+=(path+"/"+final_temp_dir+"/"+sample_name+"_DASTool_files ")
else: # the input directory already exists and is full, don't want to create it again, just re-run from last step
for line in lines:
### Skip line if starts with # (comment line)
if not (line.startswith('#')):
line = line.strip('\n').split(' ') # Create a list of each line
sample_name=line[0]
in_for=line[1]
in_rev=line[2]
output_files+=(path+"/"+final_temp_dir+"/"+sample_name+"_DASTool_files ")
return output_files
def run_metagenomics(in_f, path, config, cores):
"""Run snakemake on shell"""
# Define output names
out_files = in_out_metagenomics(path,in_f)
curr_dir = os.path.dirname(sys.argv[0])
holopath = os.path.abspath(curr_dir)
path_snkf = os.path.join(holopath,'workflows/metagenomics/individual_binning/Snakefile')
# Run snakemake
log_file = open(str(log),'w+')
log_file.write("Have a nice run!\n\t\tHOLOFOW Metagenomics-IndividualBinning starting")
log_file.close()
mtg_snk_Cmd = 'snakemake -s '+path_snkf+' -k '+out_files+' --configfile '+config+' --cores '+cores+''
subprocess.check_call(mtg_snk_Cmd, shell=True)
log_file = open(str(log),'a+')
log_file.write("\n\t\tHOLOFOW Metagenomics-IndividualBinning has finished :)")
log_file.close()
# Keep temp dirs / remove all
if args.keep: # If -k, True: keep
pass
else: # If not -k, keep only last dir
exist=list()
for file in out_files.split(" "):
exist.append(os.path.isfile(file))
if all(exist): # all output files exist
rmCmd='cd '+path+' | grep -v '+final_temp_dir+' | xargs rm -rf && mv '+final_temp_dir+' MIB_Holoflow'
subprocess.Popen(rmCmd,shell=True).wait()
else: # all expected output files don't exist: keep tmp dirs
log_file = open(str(log),'a+')
log_file.write("Looks like something went wrong...\n\t\t The temporal directories have been kept, you should have a look...")
log_file.close()
###########################
#### Workflows running
###########################
# 2 # Metagenomics workflow
run_metagenomics(in_f, path, config, cores)