0
+ − 1 import argparse
+ − 2 import os
+ − 3 import pandas
+ − 4 import pypandoc
+ − 5 import re
+ − 6 import subprocess
+ − 7 import sys
+ − 8
+ − 9 from Bio import SeqIO
+ − 10 from datetime import date
+ − 11 from mdutils.mdutils import MdUtils
+ − 12
+ − 13 CDC_ADVISORY = 'The analysis and report presented here should be treated as preliminary. Please contact the CDC/BDRD with any results regarding _Bacillus anthracis_.'
+ − 14
+ − 15
+ − 16 class PimaReport:
+ − 17
1
+ − 18 def __init__(self, analysis_name=None, amr_deletions_file=None, amr_matrix_files=None, assembly_fasta_file=None,
2
+ − 19 assembly_name=None, blastn_version=None, compute_sequence_length_file=None, contig_coverage_file=None,
+ − 20 dbkey=None, dnadiff_snps_file=None, dnadiff_version=None, feature_bed_files=None, feature_png_files=None,
+ − 21 flye_assembly_info_file=None, flye_version=None, genome_insertions_file=None, gzipped=None,
+ − 22 illumina_fastq_file=None, kraken2_report_file=None, kraken2_version=None, mutation_regions_bed_file=None,
+ − 23 mutation_regions_tsv_files=None, pima_css=None, plasmids_file=None, reference_insertions_file=None):
0
+ − 24 self.ofh = open("process_log.txt", "w")
+ − 25
1
+ − 26 self.ofh.write("amr_deletions_file: %s\n" % str(amr_deletions_file))
+ − 27 self.ofh.write("amr_matrix_files: %s\n" % str(amr_matrix_files))
0
+ − 28 self.ofh.write("analysis_name: %s\n" % str(analysis_name))
+ − 29 self.ofh.write("assembly_fasta_file: %s\n" % str(assembly_fasta_file))
+ − 30 self.ofh.write("assembly_name: %s\n" % str(assembly_name))
2
+ − 31 self.ofh.write("blastn_version: %s\n" % str(blastn_version))
1
+ − 32 self.ofh.write("compute_sequence_length_file: %s\n" % str(compute_sequence_length_file))
+ − 33 self.ofh.write("contig_coverage_file: %s\n" % str(contig_coverage_file))
+ − 34 self.ofh.write("dbkey: %s\n" % str(dbkey))
+ − 35 self.ofh.write("dnadiff_snps_file: %s\n" % str(dnadiff_snps_file))
2
+ − 36 self.ofh.write("dnadiff_version: %s\n" % str(dnadiff_version))
0
+ − 37 self.ofh.write("feature_bed_files: %s\n" % str(feature_bed_files))
+ − 38 self.ofh.write("feature_png_files: %s\n" % str(feature_png_files))
1
+ − 39 self.ofh.write("flye_assembly_info_file: %s\n" % str(flye_assembly_info_file))
+ − 40 self.ofh.write("flye_version: %s\n" % str(flye_version))
0
+ − 41 self.ofh.write("gzipped: %s\n" % str(gzipped))
1
+ − 42 self.ofh.write("genome_insertions_file: %s\n" % str(genome_insertions_file))
0
+ − 43 self.ofh.write("illumina_fastq_file: %s\n" % str(illumina_fastq_file))
2
+ − 44 self.ofh.write("kraken2_report_file: %s\n" % str(kraken2_report_file))
+ − 45 self.ofh.write("kraken2_version: %s\n" % str(kraken2_version))
0
+ − 46 self.ofh.write("mutation_regions_bed_file: %s\n" % str(mutation_regions_bed_file))
+ − 47 self.ofh.write("mutation_regions_tsv_files: %s\n" % str(mutation_regions_tsv_files))
+ − 48 self.ofh.write("pima_css: %s\n" % str(pima_css))
1
+ − 49 self.ofh.write("plasmids_file: %s\n" % str(plasmids_file))
+ − 50 # self.ofh.write("reference_genome: %s\n" % str(reference_genome))
+ − 51 self.ofh.write("reference_insertions_file: %s\n" % str(reference_insertions_file))
0
+ − 52
+ − 53 # General
+ − 54 self.doc = None
+ − 55 self.report_md = 'pima_report.md'
+ − 56
+ − 57 # Inputs
1
+ − 58 self.amr_deletions_file = amr_deletions_file
+ − 59 self.amr_matrix_files = amr_matrix_files
0
+ − 60 self.analysis_name = analysis_name
+ − 61 self.assembly_fasta_file = assembly_fasta_file
+ − 62 self.assembly_name = assembly_name
2
+ − 63 self.blastn_version = blastn_version
1
+ − 64 self.compute_sequence_length_file = compute_sequence_length_file
+ − 65 self.contig_coverage_file = contig_coverage_file
+ − 66 self.dbkey = dbkey
+ − 67 self.dnadiff_snps_file = dnadiff_snps_file
2
+ − 68 self.dnadiff_version = dnadiff_version
0
+ − 69 self.feature_bed_files = feature_bed_files
+ − 70 self.feature_png_files = feature_png_files
1
+ − 71 self.flye_assembly_info_file = flye_assembly_info_file
+ − 72 self.flye_version = flye_version
0
+ − 73 self.gzipped = gzipped
1
+ − 74 self.genome_insertions_file = genome_insertions_file
0
+ − 75 self.illumina_fastq_file = illumina_fastq_file
2
+ − 76 self.kraken2_report_file = kraken2_report_file
+ − 77 self.kraken2_version = kraken2_version
0
+ − 78 self.mutation_regions_bed_file = mutation_regions_bed_file
+ − 79 self.mutation_regions_tsv_files = mutation_regions_tsv_files
+ − 80 self.read_type = 'Illumina'
+ − 81 self.ont_bases = None
+ − 82 self.ont_n50 = None
+ − 83 self.ont_read_count = None
+ − 84 self.pima_css = pima_css
1
+ − 85 self.plasmids_file = plasmids_file
+ − 86 # self.reference_genome = reference_genome
+ − 87 self.reference_insertions_file = reference_insertions_file
0
+ − 88
+ − 89 # Titles
+ − 90 self.alignment_title = 'Comparison with reference'
+ − 91 self.alignment_notes_title = 'Alignment notes'
+ − 92 self.amr_matrix_title = 'AMR matrix'
+ − 93 self.assembly_methods_title = 'Assembly'
+ − 94 self.assembly_notes_title = 'Assembly notes'
+ − 95 self.basecalling_title = 'Basecalling'
+ − 96 self.basecalling_methods_title = 'Basecalling'
+ − 97 self.contamination_methods_title = 'Contamination check'
+ − 98 self.contig_alignment_title = 'Alignment vs. reference contigs'
+ − 99 self.feature_title = 'Features found in the assembly'
+ − 100 self.feature_methods_title = 'Feature annotation'
+ − 101 self.feature_plot_title = 'Feature annotation plots'
+ − 102 self.large_indel_title = 'Large insertions & deletions'
+ − 103 self.methods_title = 'Methods'
+ − 104 self.mutation_title = 'Mutations found in the sample'
+ − 105 self.mutation_methods_title = 'Mutation screening'
+ − 106 self.plasmid_methods_title = 'Plasmid annotation'
1
+ − 107 self.plasmid_title = 'Plasmid annotation'
0
+ − 108 self.reference_methods_title = 'Reference comparison'
+ − 109 self.snp_indel_title = 'SNPs and small indels'
+ − 110 self.summary_title = 'Analysis of %s' % analysis_name
+ − 111
2
+ − 112 # Contamination
+ − 113 self.kraken_fracs = pandas.Series(dtype=object)
+ − 114
0
+ − 115 # Methods
+ − 116 self.methods = pandas.Series(dtype='float64')
+ − 117 self.methods[self.contamination_methods_title] = pandas.Series(dtype='float64')
+ − 118 self.methods[self.assembly_methods_title] = pandas.Series(dtype='float64')
+ − 119 self.methods[self.reference_methods_title] = pandas.Series(dtype='float64')
+ − 120 self.methods[self.mutation_methods_title] = pandas.Series(dtype='float64')
+ − 121 self.methods[self.feature_methods_title] = pandas.Series(dtype='float64')
+ − 122 self.methods[self.plasmid_methods_title] = pandas.Series(dtype='float64')
+ − 123
+ − 124 # Notes
+ − 125 self.assembly_notes = pandas.Series(dtype=object)
+ − 126 self.alignment_notes = pandas.Series(dtype=object)
+ − 127 self.contig_alignment = pandas.Series(dtype=object)
+ − 128
+ − 129 # Values
+ − 130 self.assembly_size = 0
+ − 131 self.contig_info = None
+ − 132 self.did_medaka_ont_assembly = False
+ − 133 self.feature_hits = pandas.Series(dtype='float64')
+ − 134 self.illumina_length_mean = 0
+ − 135 self.illumina_read_count = 0
+ − 136 self.illumina_bases = 0
+ − 137 self.mean_coverage = 0
+ − 138 self.num_assembly_contigs = 0
1
+ − 139 # TODO: should the following 2 values be passed as parameters?
+ − 140 self.ont_n50_min = 2500
+ − 141 self.ont_coverage_min = 30
+ − 142 self.quast_indels = 0
+ − 143 self.quast_mismatches = 0
0
+ − 144
+ − 145 # Actions
+ − 146 self.did_guppy_ont_fast5 = False
+ − 147 self.did_qcat_ont_fastq = False
+ − 148 self.info_illumina_fastq()
+ − 149 self.load_contig_info()
+ − 150
+ − 151 def run_command(self, command):
+ − 152 self.ofh.write("\nXXXXXX In run_command, command:\n%s\n\n" % str(command))
+ − 153 try:
+ − 154 return re.split('\\n', subprocess.check_output(command, shell=True).decode('utf-8'))
+ − 155 except Exception:
+ − 156 message = 'Command %s failed: exiting...' % command
+ − 157 sys.exit(message)
+ − 158
+ − 159 def format_kmg(self, number, decimals=0):
+ − 160 self.ofh.write("\nXXXXXX In format_kmg, number:\n%s\n" % str(number))
+ − 161 self.ofh.write("XXXXXX In format_kmg, decimals:\n%s\n\n" % str(decimals))
+ − 162 if number == 0:
+ − 163 return '0'
+ − 164 magnitude_powers = [10**9, 10**6, 10**3, 1]
+ − 165 magnitude_units = ['G', 'M', 'K', '']
+ − 166 for i in range(len(magnitude_units)):
+ − 167 if number >= magnitude_powers[i]:
+ − 168 magnitude_power = magnitude_powers[i]
+ − 169 magnitude_unit = magnitude_units[i]
+ − 170 return ('{:0.' + str(decimals) + 'f}').format(number / magnitude_power) + magnitude_unit
+ − 171
+ − 172 def load_contig_info(self):
+ − 173 self.contig_info = pandas.Series(dtype=object)
+ − 174 self.contig_info[self.read_type] = pandas.read_csv(self.contig_coverage_file, header=None, index_col=None, sep='\t').sort_values(1, axis=0, ascending=False)
+ − 175 self.contig_info[self.read_type].columns = ['contig', 'size', 'coverage']
+ − 176 self.mean_coverage = (self.contig_info[self.read_type].iloc[:, 1] * self.contig_info[self.read_type].iloc[:, 2]).sum() / self.contig_info[self.read_type].iloc[:, 1].sum()
1
+ − 177 if self.mean_coverage <= self.ont_coverage_min:
+ − 178 warning = '%s mean coverage ({:.0f}X) is less than the recommended minimum ({:.0f}X).'.format(self.mean_coverage, self.ont_coverage_min) % self.read_type
+ − 179 self.assembly_notes = self.assembly_notes.append(pandas.Series(warning))
+ − 180 # Report if some contigs have low coverage.
+ − 181 low_coverage = self.contig_info[self.read_type].loc[self.contig_info[self.read_type]['coverage'] < self.ont_coverage_min, :]
+ − 182 if low_coverage.shape[0] >= 0:
+ − 183 for contig_i in range(low_coverage.shape[0]):
+ − 184 warning = '%s coverage of {:s} ({:.0f}X) is less than the recommended minimum ({:.0f}X).'.format(low_coverage.iloc[contig_i, 0], low_coverage.iloc[contig_i, 2], self.ont_coverage_min) % self.read_type
+ − 185 self.assembly_notes = self.assembly_notes.append(pandas.Series(warning))
+ − 186 # See if some contigs have anolously low coverage.
+ − 187 fold_coverage = self.contig_info[self.read_type]['coverage'] / self.mean_coverage
+ − 188 low_coverage = self.contig_info[self.read_type].loc[fold_coverage < 1 / 5, :]
+ − 189 if low_coverage.shape[0] >= 0 :
+ − 190 for contig_i in range(low_coverage.shape[0]):
+ − 191 warning = '%s coverage of {:s} ({:.0f}X) is less than 1/5 the mean coverage ({:.0f}X).'.format(low_coverage.iloc[contig_i, 0], low_coverage.iloc[contig_i, 2], self.mean_coverage) % self.read_type
+ − 192 self.assembly_notes = self.assembly_notes.append(pandas.Series(warning))
0
+ − 193
+ − 194 def load_fasta(self, fasta):
+ − 195 sequence = pandas.Series(dtype=object)
+ − 196 for contig in SeqIO.parse(fasta, 'fasta'):
+ − 197 sequence[contig.id] = contig
+ − 198 return sequence
+ − 199
+ − 200 def load_assembly(self):
+ − 201 self.assembly = self.load_fasta(self.assembly_fasta_file)
+ − 202 self.num_assembly_contigs = len(self.assembly)
+ − 203 for i in self.assembly:
+ − 204 self.assembly_size += len(i.seq)
+ − 205 self.assembly_size = self.format_kmg(self.assembly_size, decimals=1)
+ − 206
+ − 207 def info_illumina_fastq(self):
+ − 208 self.ofh.write("\nXXXXXX In info_illumina_fastq\n\n")
+ − 209 if self.gzipped:
+ − 210 opener = 'gunzip -c'
+ − 211 else:
+ − 212 opener = 'cat'
+ − 213 command = ' '.join([opener,
+ − 214 self.illumina_fastq_file,
+ − 215 '| awk \'{getline;s += length($1);getline;getline;}END{print s/(NR/4)"\t"(NR/4)"\t"s}\''])
+ − 216 output = self.run_command(command)
+ − 217 self.ofh.write("output:\n%s\n" % str(output))
+ − 218 self.ofh.write("re.split('\\t', self.run_command(command)[0]:\n%s\n" % str(re.split('\\t', self.run_command(command)[0])))
+ − 219 values = []
+ − 220 for i in re.split('\\t', self.run_command(command)[0]):
+ − 221 if i == '':
+ − 222 values.append(float('nan'))
+ − 223 else:
+ − 224 values.append(float(i))
+ − 225 self.ofh.write("values:\n%s\n" % str(values))
+ − 226 self.ofh.write("values[0]:\n%s\n" % str(values[0]))
+ − 227 self.illumina_length_mean += values[0]
+ − 228 self.ofh.write("values[1]:\n%s\n" % str(values[1]))
+ − 229 self.illumina_read_count += int(values[1])
+ − 230 self.ofh.write("values[2]:\n%s\n" % str(values[2]))
+ − 231 self.illumina_bases += int(values[2])
+ − 232 # The original PIMA code inserts self.illumina_fastq into
+ − 233 # a list for no apparent reason. We don't do that here.
+ − 234 # self.illumina_length_mean /= len(self.illumina_fastq)
+ − 235 self.illumina_length_mean /= 1
+ − 236 self.illumina_bases = self.format_kmg(self.illumina_bases, decimals=1)
+ − 237
+ − 238 def start_doc(self):
+ − 239 self.doc = MdUtils(file_name=self.report_md, title='')
+ − 240
+ − 241 def add_run_information(self):
+ − 242 self.ofh.write("\nXXXXXX In add_run_information\n\n")
+ − 243 self.doc.new_line()
+ − 244 self.doc.new_header(1, 'Run information')
+ − 245 # Tables in md.utils are implemented as a wrapping function.
+ − 246 Table_list = [
+ − 247 "Category",
+ − 248 "Information",
+ − 249 "Date",
+ − 250 date.today(),
+ − 251 "ONT FAST5",
+ − 252 "N/A",
+ − 253 "ONT FASTQ",
+ − 254 "N/A",
+ − 255 "Illumina FASTQ",
+ − 256 self.wordwrap_markdown(self.analysis_name),
+ − 257 "Assembly",
+ − 258 self.wordwrap_markdown(self.assembly_name),
+ − 259 "Reference",
+ − 260 self.wordwrap_markdown(self.dbkey),
+ − 261 ]
+ − 262 self.doc.new_table(columns=2, rows=7, text=Table_list, text_align='left')
+ − 263 self.doc.new_line()
+ − 264 self.doc.new_line()
+ − 265
+ − 266 def add_ont_library_information(self):
+ − 267 self.ofh.write("\nXXXXXX In add_ont_library_information\n\n")
+ − 268 if self.ont_n50 is None:
+ − 269 return
+ − 270 self.doc.new_line()
+ − 271 self.doc.new_header(2, 'ONT library statistics')
+ − 272 Table_List = [
+ − 273 "Category",
+ − 274 "Quantity",
+ − 275 "ONT N50",
+ − 276 '{:,}'.format(self.ont_n50),
+ − 277 "ONT reads",
+ − 278 '{:,}'.format(self.ont_read_count),
+ − 279 "ONT bases",
+ − 280 '{:s}'.format(self.ont_bases),
+ − 281 "Illumina FASTQ",
+ − 282 self.wordwrap_markdown(self.illumina_fastq_file),
+ − 283 "Assembly",
+ − 284 self.wordwrap_markdown(self.assembly_name),
+ − 285 "Reference",
+ − 286 self.wordwrap_markdown(self.dbkey),
+ − 287 ]
+ − 288 self.doc.new_table(columns=2, rows=7, text=Table_List, text_align='left')
+ − 289 self.doc.new_line()
+ − 290
+ − 291 def add_illumina_library_information(self):
+ − 292 self.ofh.write("\nXXXXXX In add_illumina_library_information\n\n")
+ − 293 if self.illumina_length_mean is None:
+ − 294 return
+ − 295 self.doc.new_line()
+ − 296 self.doc.new_header(2, 'Illumina library statistics')
+ − 297 Table_List = [
+ − 298 "Illumina Info.",
+ − 299 "Quantity",
+ − 300 'Illumina mean length',
+ − 301 '{:.1f}'.format(self.illumina_length_mean),
+ − 302 'Illumina reads',
+ − 303 '{:,}'.format(self.illumina_read_count),
+ − 304 'Illumina bases',
+ − 305 '{:s}'.format(self.illumina_bases)
+ − 306 ]
+ − 307 self.doc.new_table(columns=2, rows=4, text=Table_List, text_align='left')
+ − 308
1
+ − 309 def evaluate_assembly(self) :
+ − 310 assembly_info = pandas.read_csv(self.compute_sequence_length_file, sep='\t', header=None)
+ − 311 assembly_info.columns = ['contig', 'length']
+ − 312 self.contig_sizes = assembly_info
+ − 313 # Take a look at the number of contigs, their sizes,
+ − 314 # and circularity. Warn if things don't look good.
+ − 315 if assembly_info.shape[0] > 4:
+ − 316 warning = 'Assembly produced {:d} contigs, more than ususally expected; assembly may be fragmented'.format(assembly_info.shape[0])
+ − 317 self.assembly_notes = self.assembly_notes.append(pandas.Series(warning))
+ − 318 small_contigs = assembly_info.loc[assembly_info['length'] <= 3000, :]
+ − 319 if small_contigs.shape[0] > 0:
+ − 320 warning = 'Assembly produced {:d} small contigs ({:s}); assembly may include spurious sequences.'.format(small_contigs.shape[0], ', '.join(small_contigs['contig']))
+ − 321 self.assembly_notes = self.assembly_notes.append(pandas.Series(warning))
+ − 322
0
+ − 323 def add_assembly_information(self):
+ − 324 self.ofh.write("\nXXXXXX In add_assembly_information\n\n")
+ − 325 if self.assembly_fasta_file is None:
+ − 326 return
+ − 327 self.load_assembly()
+ − 328 self.doc.new_line()
+ − 329 self.doc.new_header(2, 'Assembly statistics')
+ − 330 Table_List = [
+ − 331 "Category",
+ − 332 "Information",
+ − 333 "Contigs",
+ − 334 str(self.num_assembly_contigs),
+ − 335 "Assembly size",
+ − 336 str(self.assembly_size),
+ − 337 ]
+ − 338 self.doc.new_table(columns=2, rows=3, text=Table_List, text_align='left')
+ − 339
+ − 340 def info_ont_fastq(self, fastq_file):
+ − 341 self.ofh.write("\nXXXXXX In info_ont_fastq, fastq_file:\n%s\n\n" % str(fastq_file))
+ − 342 opener = 'cat'
+ − 343 if self.gzipped:
+ − 344 opener = 'gunzip -c'
+ − 345 else:
+ − 346 opener = 'cat'
+ − 347 command = ' '.join([opener,
+ − 348 fastq_file,
+ − 349 '| awk \'{getline;print length($0);s += length($1);getline;getline;}END{print "+"s}\'',
+ − 350 '| sort -gr',
+ − 351 '| awk \'BEGIN{bp = 0;f = 0}',
+ − 352 '{if(NR == 1){sub(/+/, "", $1);s=$1}else{bp += $1;if(bp > s / 2 && f == 0){n50 = $1;f = 1}}}',
+ − 353 'END{printf "%d\\t%d\\t%d\\n", n50, (NR - 1), s;exit}\''])
+ − 354 result = list(re.split('\\t', self.run_command(command)[0]))
+ − 355 if result[1] == '0':
+ − 356 self.error_out('No ONT reads found')
+ − 357 ont_n50, ont_read_count, ont_raw_bases = [int(i) for i in result]
+ − 358 command = ' '.join([opener,
+ − 359 fastq_file,
+ − 360 '| awk \'{getline;print length($0);getline;getline;}\''])
+ − 361 result = self.run_command(command)
+ − 362 result = list(filter(lambda x: x != '', result))
1
+ − 363 # TODO: the following are not yet used...
+ − 364 # ont_read_lengths = [int(i) for i in result]
+ − 365 # ont_bases = self.format_kmg(ont_raw_bases, decimals=1)
+ − 366 if ont_n50 <= self.ont_n50_min:
+ − 367 warning = 'ONT N50 (%s) is less than the recommended minimum (%s)' % (str(ont_n50), str(self.ont_n50_min))
+ − 368 self.assembly_notes = self.assembly_notes.append(pandas.Series(warning))
0
+ − 369
+ − 370 def wordwrap_markdown(self, string):
+ − 371 if string:
+ − 372 if len(string) < 35:
+ − 373 return string
+ − 374 else:
+ − 375 if '/' in string:
+ − 376 adjust = string.split('/')
+ − 377 out = ''
+ − 378 max = 35
+ − 379 for i in adjust:
+ − 380 out = out + '/' + i
+ − 381 if len(out) > max:
+ − 382 out += '<br>'
+ − 383 max += 35
+ − 384 return out
+ − 385 else:
+ − 386 out = [string[i:i + 35] for i in range(0, len(string), 50)]
+ − 387 return '<br>'.join(out)
+ − 388 else:
+ − 389 return string
+ − 390
+ − 391 def add_contig_info(self):
+ − 392 self.ofh.write("\nXXXXXX In add_contig_info\n\n")
+ − 393 if self.contig_info is None:
+ − 394 return
+ − 395 for method in ['ONT', 'Illumina']:
+ − 396 if method not in self.contig_info.index:
+ − 397 continue
+ − 398 self.doc.new_line()
+ − 399 self.doc.new_header(2, 'Assembly coverage by ' + method)
+ − 400 Table_List = ["Contig", "Length (bp)", "Coverage (X)"]
+ − 401 formatted = self.contig_info[method].copy()
+ − 402 formatted.iloc[:, 1] = formatted.iloc[:, 1].apply(lambda x: '{:,}'.format(x))
+ − 403 for i in range(self.contig_info[method].shape[0]):
+ − 404 Table_List = Table_List + formatted.iloc[i, :].values.tolist()
+ − 405 row_count = int(len(Table_List) / 3)
+ − 406 self.doc.new_table(columns=3, rows=row_count, text=Table_List, text_align='left')
+ − 407
+ − 408 def add_assembly_notes(self):
+ − 409 self.ofh.write("\nXXXXXX In add_assembly_notes\n\n")
+ − 410 if len(self.assembly_notes) == 0:
+ − 411 return
+ − 412 self.doc.new_line()
+ − 413 self.doc.new_line('<div style="page-break-after: always;"></div>')
+ − 414 self.doc.new_line()
+ − 415 self.doc.new_header(2, self.assembly_notes_title)
1
+ − 416 for note in self.assembly_notes:
+ − 417 self.doc.new_line(note)
0
+ − 418
+ − 419 def add_contamination(self):
+ − 420 self.ofh.write("\nXXXXXX In add_contamination\n\n")
2
+ − 421 if self.kraken2_report_file is None:
0
+ − 422 return
2
+ − 423 # Read in the Kraken fractions and pull out the useful parts
+ − 424 self.kraken_fracs = pandas.read_csv(self.kraken_report_file, delimiter='\t', header=None)
+ − 425 self.kraken_fracs.index = self.kraken_fracs.iloc[:, 4].values
+ − 426 self.kraken_fracs = self.kraken_fracs.loc[self.kraken_fracs.iloc[:, 3].str.match('[UG]1?'), :]
+ − 427 self.kraken_fracs = self.kraken_fracs.loc[(self.kraken_fracs.iloc[:, 0] >= 1) | (self.kraken_fracs.iloc[:, 3] == 'U'), :]
+ − 428 self.kraken_fracs = self.kraken_fracs.iloc[:, [0, 1, 3, 5]]
+ − 429 self.kraken_fracs.columns = ['Fraction', 'Reads', 'Level', 'Taxa']
+ − 430 self.kraken_fracs['Fraction'] = (self.kraken_fracs['Fraction'] / 100).round(4)
+ − 431 self.kraken_fracs.sort_values(by='Fraction', inplace=True, ascending=False)
+ − 432 self.kraken_fracs['Taxa'] = self.kraken_fracs['Taxa'].str.lstrip()
0
+ − 433 self.doc.new_line()
+ − 434 self.doc.new_header(2, 'Contamination check')
2
+ − 435 for kraken_fracs in self.kraken_fracs.iteritems():
+ − 436 self.doc.new_line(self.read_type + ' classifications')
0
+ − 437 self.doc.new_line()
+ − 438 Table_List = ["Percent of Reads", "Reads", "Level", "Label"]
+ − 439 for index, row in kraken_fracs.iterrows():
+ − 440 Table_List = Table_List + row.tolist()
+ − 441 row_count = int(len(Table_List) / 4)
+ − 442 self.doc.new_table(columns=4, rows=row_count, text=Table_List, text_align='left')
+ − 443 if self.contamination_methods_title not in self.methods:
+ − 444 self.methods[self.contamination_methods_title] = ''
2
+ − 445 method = 'Kraken2 version %s was used to assign the raw reads into taxa.' % self.kraken2_version
0
+ − 446 self.methods[self.contamination_methods_title] = self.methods[self.contamination_methods_title].append(pandas.Series(method))
+ − 447
+ − 448 def add_alignment(self):
+ − 449 self.ofh.write("\nXXXXXX In add_alignment\n\n")
+ − 450 # TODO: implement the draw_circos function for this.
+ − 451 if len(self.contig_alignment) > 0:
+ − 452 alignments = self.contig_alignment
+ − 453 else:
+ − 454 return
+ − 455 self.doc.new_line()
+ − 456 self.doc.new_header(level=2, title=self.alignment_title)
+ − 457 self.doc.new_line()
+ − 458 self.doc.new_header(level=3, title=self.snp_indel_title)
+ − 459 Table_1 = [
+ − 460 "Category",
+ − 461 "Quantity",
+ − 462 'SNPs',
1
+ − 463 '{:,}'.format(self.quast_mismatches),
0
+ − 464 'Small indels',
1
+ − 465 '{:,}'.format(self.quast_indels)
0
+ − 466 ]
+ − 467 self.doc.new_table(columns=2, rows=3, text=Table_1, text_align='left')
+ − 468 self.doc.new_line('<div style="page-break-after: always;"></div>')
+ − 469 self.doc.new_line()
+ − 470 if len(self.alignment_notes) > 0:
+ − 471 self.doc.new_header(level=3, title=self.alignment_notes_title)
+ − 472 for note in self.alignment_notes:
+ − 473 self.doc.new_line(note)
+ − 474 for contig in alignments.index.tolist():
+ − 475 contig_title = 'Alignment to %s' % contig
+ − 476 image_png = alignments[contig]
+ − 477 self.doc.new_line()
+ − 478 self.doc.new_header(level=3, title=contig_title)
+ − 479 self.doc.new_line(self.doc.new_inline_image(text='contig_title', path=os.path.abspath(image_png)))
+ − 480 self.doc.new_line('<div style="page-break-after: always;"></div>')
+ − 481 self.doc.new_line()
2
+ − 482 method = 'The genome assembly was aligned against the reference sequencing using dnadiff version %s.' % self.dnadiff_version
0
+ − 483 self.methods[self.reference_methods_title] = self.methods[self.reference_methods_title].append(pandas.Series(method))
+ − 484
+ − 485 def add_features(self):
+ − 486 self.ofh.write("\nXXXXXX In add_features\n\n")
+ − 487 if len(self.feature_bed_files) == 0:
+ − 488 return
+ − 489 for bbf in self.feature_bed_files:
+ − 490 if os.path.getsize(bbf) > 0:
+ − 491 best = pandas.read_csv(filepath_or_buffer=bbf, sep='\t', header=None)
+ − 492 self.feature_hits[os.path.basename(bbf)] = best
+ − 493 if len(self.feature_hits) == 0:
+ − 494 return
+ − 495 self.ofh.write("self.feature_hits: %s\n" % str(self.feature_hits))
+ − 496 self.doc.new_line()
+ − 497 self.doc.new_header(level=2, title=self.feature_title)
+ − 498 for feature_name in self.feature_hits.index.tolist():
+ − 499 self.ofh.write("feature_name: %s\n" % str(feature_name))
+ − 500 features = self.feature_hits[feature_name].copy()
+ − 501 self.ofh.write("features: %s\n" % str(features))
+ − 502 if features.shape[0] == 0:
+ − 503 continue
+ − 504 features.iloc[:, 1] = features.iloc[:, 1].apply(lambda x: '{:,}'.format(x))
+ − 505 features.iloc[:, 2] = features.iloc[:, 2].apply(lambda x: '{:,}'.format(x))
+ − 506 self.doc.new_line()
+ − 507 self.doc.new_header(level=3, title=feature_name)
+ − 508 if (features.shape[0] == 0):
+ − 509 continue
+ − 510 for contig in pandas.unique(features.iloc[:, 0]):
+ − 511 self.ofh.write("contig: %s\n" % str(contig))
+ − 512 self.doc.new_line(contig)
+ − 513 contig_features = features.loc[(features.iloc[:, 0] == contig), :]
+ − 514 self.ofh.write("contig_features: %s\n" % str(contig_features))
+ − 515 Table_List = ['Start', 'Stop', 'Feature', 'Identity (%)', 'Strand']
+ − 516 for i in range(contig_features.shape[0]):
+ − 517 self.ofh.write("i: %s\n" % str(i))
+ − 518 feature = contig_features.iloc[i, :].copy(deep=True)
+ − 519 self.ofh.write("feature: %s\n" % str(feature))
+ − 520 feature[4] = '{:.3f}'.format(feature[4])
2
+ − 521 self.ofh.write("feature[1:].values.tolist(): %s\n" % str(feature[1:].values.tolist()))
+ − 522 Table_List = Table_List + feature[1:].values.tolist()
0
+ − 523 self.ofh.write("Table_List: %s\n" % str(Table_List))
+ − 524 row_count = int(len(Table_List) / 5)
+ − 525 self.ofh.write("row_count: %s\n" % str(row_count))
+ − 526 self.doc.new_line()
1
+ − 527 self.ofh.write("Before new_table, len(Table_List):: %s\n" % str(len(Table_List)))
+ − 528 self.doc.new_table(columns=5, rows=row_count, text=Table_List, text_align='left')
2
+ − 529 if self.blastn_version is not None:
+ − 530 blastn_version = 'The genome assembly was queried for features using blastn version %s.' % self.blastn_version
0
+ − 531 bedtools_version = 'Feature hits were clustered using bedtools and the highest scoring hit for each cluster was reported.'
+ − 532 method = '%s %s' % (blastn_version, bedtools_version)
+ − 533 self.methods[self.feature_methods_title] = self.methods[self.feature_methods_title].append(pandas.Series(method))
+ − 534
+ − 535 def add_feature_plots(self):
+ − 536 self.ofh.write("\nXXXXXX In add_feature_plots\n\n")
+ − 537 if len(self.feature_png_files) == 0:
+ − 538 return
+ − 539 self.doc.new_line()
+ − 540 self.doc.new_header(level=2, title='Feature Plots')
+ − 541 self.doc.new_paragraph('Only contigs with features are shown')
+ − 542 for feature_png_file in self.feature_png_files:
+ − 543 self.doc.new_line(self.doc.new_inline_image(text='Analysis', path=os.path.abspath(feature_png_file)))
+ − 544
+ − 545 def add_mutations(self):
+ − 546 self.ofh.write("\nXXXXXX In add_mutations\n\n")
+ − 547 if len(self.mutation_regions_tsv_files) == 0:
+ − 548 return
1
+ − 549 try :
0
+ − 550 mutation_regions = pandas.read_csv(self.mutation_regions_bed_file, sep='\t', header=0, index_col=False)
+ − 551 except Exception:
+ − 552 # Likely an empty file.
+ − 553 return
1
+ − 554 # TODO: this is the only place where reference_genome is used,
+ − 555 # so I'm commenting it out for now. We need to confirm if these
+ − 556 # errors that require the reference genmoe being passed are necessary.
+ − 557 # If so, we'll need to implement data tables in this tool.
+ − 558 # Make sure that the positions in the BED file fall within
+ − 559 # the chromosomes provided in the reference sequence.
+ − 560 """
+ − 561 for mutation_region in range(mutation_regions.shape[0]):
+ − 562 mutation_region = mutation_regions.iloc[mutation_region, :]
+ − 563 if not (mutation_region[0] in self.reference_genome):
+ − 564 self.ofh.write("\nMutation region: %s not found in reference genome.\n" % ' '.join(mutation_region.astype(str)))
+ − 565 continue
+ − 566 if not isinstance(mutation_region[1], int):
+ − 567 self.ofh.write("\nNon-integer found in mutation region start (column 2): %s.\n" % str(mutation_region[1]))
+ − 568 break
+ − 569 elif not isinstance(mutation_region[2], int):
+ − 570 self.ofh.write("\nNon-integer found in mutation region start (column 3): %s.\n" % str(mutation_region[2]))
+ − 571 break
+ − 572 if mutation_region[1] <= 0 or mutation_region[2] <= 0:
+ − 573 self.ofh.write("\nMutation region %s starts before the reference sequence.\n" % ' '.join(mutation_region.astype(str)))
+ − 574 if mutation_region[1] > len(self.reference_genome[mutation_region[0]].seq) or mutation_region[2] > len(self.reference_genome[mutation_region[0]].seq):
+ − 575 self.ofh.write("\nMutation region %s ends after the reference sequence.\n" % ' '.join(mutation_region.astype(str)))
+ − 576 """
0
+ − 577 amr_mutations = pandas.Series(dtype=object)
+ − 578 for region_i in range(mutation_regions.shape[0]):
+ − 579 region = mutation_regions.iloc[region_i, :]
+ − 580 region_name = str(region['name'])
+ − 581 self.ofh.write("Processing mutations for region %s\n" % region_name)
+ − 582 region_mutations_tsv_name = '%s_mutations.tsv' % region_name
+ − 583 if region_mutations_tsv_name not in self.mutation_regions_tsv_files:
+ − 584 continue
+ − 585 region_mutations_tsv = self.mutation_regions_tsv_files[region_mutations_tsv_name]
1
+ − 586 try :
0
+ − 587 region_mutations = pandas.read_csv(region_mutations_tsv, sep='\t', header=0, index_col=False)
+ − 588 except Exception:
+ − 589 region_mutations = pandas.DataFrame()
+ − 590 if region_mutations.shape[0] == 0:
+ − 591 continue
1
+ − 592 # Figure out what kind of mutations are in this region.
0
+ − 593 region_mutation_types = pandas.Series(['snp'] * region_mutations.shape[0], name='TYPE', index=region_mutations.index)
+ − 594 region_mutation_types[region_mutations['REF'].str.len() != region_mutations['ALT'].str.len()] = 'small-indel'
+ − 595 region_mutation_drugs = pandas.Series(region['drug'] * region_mutations.shape[0], name='DRUG', index=region_mutations.index)
+ − 596 region_notes = pandas.Series(region['note'] * region_mutations.shape[0], name='NOTE', index=region_mutations.index)
+ − 597 region_mutations = pandas.concat([region_mutations, region_mutation_types, region_mutation_drugs, region_notes], axis=1)
+ − 598 region_mutations = region_mutations[['#CHROM', 'POS', 'TYPE', 'REF', 'ALT', 'DRUG', 'NOTE']]
+ − 599 amr_mutations[region['name']] = region_mutations
2
+ − 600 if (amr_mutations.shape[0] > 0):
+ − 601 # Report the mutations.
0
+ − 602 self.doc.new_line()
2
+ − 603 self.doc.new_header(level=2, title=self.mutation_title)
+ − 604 for region_name in amr_mutations.index.tolist():
+ − 605 region_mutations = amr_mutations[region_name].copy()
+ − 606 self.doc.new_line()
+ − 607 self.doc.new_header(level=3, title=region_name)
+ − 608 if (region_mutations.shape[0] == 0):
+ − 609 self.doc.append('None')
+ − 610 continue
+ − 611 region_mutations.iloc[:, 1] = region_mutations.iloc[:, 1].apply(lambda x: '{:,}'.format(x))
+ − 612 Table_List = ['Reference contig', 'Position', 'Reference', 'Alternate', 'Drug', 'Note']
+ − 613 for i in range(region_mutations.shape[0]):
+ − 614 Table_List = Table_List + region_mutations.iloc[i, [0, 1, 3, 4, 5, 6]].values.tolist()
+ − 615 row_count = int(len(Table_List) / 6)
+ − 616 self.doc.new_table(columns=6, rows=row_count, text=Table_List, text_align='left')
0
+ − 617 method = '%s reads were mapped to the reference sequence using minimap2.' % self.read_type
+ − 618 self.methods[self.mutation_methods_title] = self.methods[self.mutation_methods_title].append(pandas.Series(method))
+ − 619 method = 'Mutations were identified using samtools mpileup and varscan.'
+ − 620 self.methods[self.mutation_methods_title] = self.methods[self.mutation_methods_title].append(pandas.Series(method))
+ − 621
+ − 622 def add_amr_matrix(self):
+ − 623 self.ofh.write("\nXXXXXX In add_amr_matrix\n\n")
+ − 624 # Make sure that we have an AMR matrix to plot
1
+ − 625 if len(self.amr_matrix_files) == 0:
+ − 626 return
+ − 627 self.doc.new_line()
+ − 628 self.doc.new_header(level=2, title=self.amr_matrix_title)
+ − 629 self.doc.new_line('AMR genes and mutations with their corresponding drugs')
+ − 630 for amr_matrix_file in self.amr_matrix_files:
+ − 631 self.doc.new_line(self.doc.new_inline_image(text='AMR genes and mutations with their corresponding drugs',
+ − 632 path=os.path.abspath(amr_matrix_file)))
0
+ − 633
+ − 634 def add_large_indels(self):
+ − 635 self.ofh.write("\nXXXXXX In add_large_indels\n\n")
1
+ − 636 large_indels = pandas.Series(dtype='float64')
+ − 637 # Pull in insertions.
+ − 638 try:
+ − 639 reference_insertions = pandas.read_csv(filepath_or_buffer=self.reference_insertions_file, sep='\t', header=None)
+ − 640 except Exception:
+ − 641 reference_insertions = pandas.DataFrame()
+ − 642 try:
+ − 643 genome_insertions = pandas.read_csv(filepath_or_buffer=self.genome_insertions_file, sep='\t', header=None)
+ − 644 except Exception:
+ − 645 genome_insertions = pandas.DataFrame()
+ − 646 large_indels['Reference insertions'] = reference_insertions
+ − 647 large_indels['Query insertions'] = genome_insertions
+ − 648 # TODO: we don't seem to be reporting snps and deletions for some reason...
+ − 649 # Pull in the number of SNPs and small indels.
+ − 650 try:
+ − 651 snps = pandas.read_csv(filepath_or_buffer=self.dnadiff_snps_file, sep='\t', header=None)
+ − 652 # TODO: the following is not used...
+ − 653 # small_indels = snps.loc[(snps.iloc[:, 1] == '.') | (snps.iloc[:, 2] == '.'), :]
+ − 654 snps = snps.loc[(snps.iloc[:, 1] != '.') & (snps.iloc[:, 2] != '.'), :]
+ − 655 except Exception:
+ − 656 snps = pandas.DataFrame()
+ − 657 # Pull in deletions.
+ − 658 try:
+ − 659 amr_deletions = pandas.read_csv(filepath_or_buffer=self.amr_deletion_file, sep='\t', header=None)
+ − 660 except Exception:
+ − 661 amr_deletions = pandas.DataFrame()
+ − 662 if amr_deletions.shape[0] > 0:
+ − 663 amr_deletions.columns = ['contig', 'start', 'stop', 'name', 'type', 'drug', 'note']
+ − 664 amr_deletions = amr_deletions.loc[amr_deletions['type'].isin(['large-deletion', 'any']), :]
+ − 665 self.doc.new_line()
+ − 666 self.doc.new_header(level=2, title=self.large_indel_title)
+ − 667 for genome in ['Reference insertions', 'Query insertions']:
+ − 668 genome_indels = large_indels[genome].copy()
+ − 669 self.doc.new_line()
+ − 670 self.doc.new_header(level=3, title=genome)
+ − 671 if (genome_indels.shape[0] == 0):
+ − 672 continue
+ − 673 genome_indels.iloc[:, 1] = genome_indels.iloc[:, 1].apply(lambda x: '{:,}'.format(x))
+ − 674 genome_indels.iloc[:, 2] = genome_indels.iloc[:, 2].apply(lambda x: '{:,}'.format(x))
+ − 675 genome_indels.iloc[:, 3] = genome_indels.iloc[:, 3].apply(lambda x: '{:,}'.format(x))
+ − 676 Table_List = [
+ − 677 'Reference contig', 'Start', 'Stop', 'Size (bp)'
+ − 678 ]
+ − 679 for i in range(genome_indels.shape[0]):
+ − 680 Table_List = Table_List + genome_indels.iloc[i, :].values.tolist()
+ − 681 row_count = int(len(Table_List) / 4)
+ − 682 self.doc.new_table(columns=4, rows=row_count, text=Table_List, text_align='left')
+ − 683 method = 'Large insertions or deletions were found as the complement of aligned regions using bedtools.'
+ − 684 self.methods[self.reference_methods_title] = self.methods[self.reference_methods_title].append(pandas.Series(method))
+ − 685 self.doc.new_line()
+ − 686 self.doc.new_line('<div style="page-break-after: always;"></div>')
+ − 687 self.doc.new_line()
0
+ − 688
+ − 689 def add_plasmids(self):
1
+ − 690 try :
+ − 691 plasmids = pandas.read_csv(filepath_or_buffer=self.plasmids_file, sep='\t', header=0)
+ − 692 except Exception:
0
+ − 693 return
+ − 694 plasmids = plasmids.copy()
+ − 695 self.doc.new_line()
1
+ − 696 self.doc.new_header(level=2, title=self.plasmid_title)
0
+ − 697 if (plasmids.shape[0] == 0):
+ − 698 self.doc.new_line('None')
+ − 699 return
+ − 700 plasmids.iloc[:, 3] = plasmids.iloc[:, 3].apply(lambda x: '{:,}'.format(x))
+ − 701 plasmids.iloc[:, 4] = plasmids.iloc[:, 4].apply(lambda x: '{:,}'.format(x))
+ − 702 plasmids.iloc[:, 5] = plasmids.iloc[:, 5].apply(lambda x: '{:,}'.format(x))
1
+ − 703 Table_List = ['Genome contig', 'Plasmid hit', 'Plasmid acc.', 'Contig size', 'Aliged', 'Plasmid size']
0
+ − 704 for i in range(plasmids.shape[0]):
+ − 705 Table_List = Table_List + plasmids.iloc[i, 0:6].values.tolist()
+ − 706 row_count = int(len(Table_List) / 6)
+ − 707 self.doc.new_table(columns=6, rows=row_count, text=Table_List, text_align='left')
+ − 708 method = 'The plasmid reference database was queried against the genome assembly using minimap2.'
+ − 709 self.methods[self.plasmid_methods_title] = self.methods[self.plasmid_methods_title].append(pandas.Series(method))
2
+ − 710 method = 'The resulting BAM was converted to a PSL using a custom version of sam2psl.'
0
+ − 711 self.methods[self.plasmid_methods_title] = self.methods[self.plasmid_methods_title].append(pandas.Series(method))
+ − 712 method = 'Plasmid-to-genome hits were resolved using the pChunks algorithm.'
+ − 713 self.methods[self.plasmid_methods_title] = self.methods[self.plasmid_methods_title].append(pandas.Series(method))
+ − 714
+ − 715 def add_methods(self):
+ − 716 self.ofh.write("\nXXXXXX In add_methods\n\n")
+ − 717 self.doc.new_line('<div style="page-break-after: always;"></div>')
+ − 718 self.doc.new_line()
+ − 719 if len(self.methods) == 0:
+ − 720 return
+ − 721 self.doc.new_line()
+ − 722 self.doc.new_header(level=2, title=self.methods_title)
+ − 723 for methods_section in self.methods.index.tolist():
+ − 724 if self.methods[methods_section] is None or len(self.methods[methods_section]) == 0:
+ − 725 continue
+ − 726 self.doc.new_line()
+ − 727 self.doc.new_header(level=3, title=methods_section)
+ − 728 self.doc.new_paragraph(' '.join(self.methods[methods_section]))
+ − 729
+ − 730 def add_summary(self):
+ − 731 self.ofh.write("\nXXXXXX In add_summary\n\n")
+ − 732 # Add summary title
+ − 733 self.doc.new_header(level=1, title=self.summary_title)
+ − 734 # First section of Summary
+ − 735 self.doc.new_header(level=1, title='CDC Advisory')
+ − 736 self.doc.new_paragraph(CDC_ADVISORY)
+ − 737 self.doc.new_line()
+ − 738 self.add_run_information()
+ − 739 self.add_ont_library_information()
+ − 740 methods = []
+ − 741 if self.did_guppy_ont_fast5:
+ − 742 methods += ['ONT reads were basecalled using guppy']
+ − 743 if self.did_qcat_ont_fastq:
+ − 744 methods += ['ONT reads were demultiplexed and trimmed using qcat']
+ − 745 self.methods[self.basecalling_methods_title] = pandas.Series(methods)
+ − 746 self.add_illumina_library_information()
1
+ − 747 self.add_contig_info()
+ − 748 self.evaluate_assembly()
0
+ − 749 self.add_assembly_information()
1
+ − 750 if self.flye_assembly_info_file is not None:
+ − 751 method = 'ONT reads were assembled using %s' % self.flye_version
0
+ − 752 self.methods[self.assembly_methods_title] = self.methods[self.assembly_methods_title].append(pandas.Series(method))
1
+ − 753 # Pull in the assembly summary and look at the coverage.
+ − 754 assembly_info = pandas.read_csv(self.flye_assembly_info_file, header=0, index_col=0, sep='\t')
+ − 755 # Look for non-circular contigs.
+ − 756 open_contigs = assembly_info.loc[assembly_info['circ.'] == 'N', :]
+ − 757 if open_contigs.shape[0] > 0:
+ − 758 open_contig_ids = open_contigs.index.values
+ − 759 warning = 'Flye reported {:d} open contigs ({:s}); assembly may be incomplete.'.format(open_contigs.shape[0], ', '.join(open_contig_ids))
+ − 760 self.assembly_notes = self.assembly_notes.append(pandas.Series(warning))
0
+ − 761 if self.did_medaka_ont_assembly:
+ − 762 method = 'the genome assembly was polished using ont reads and medaka.'
+ − 763 self.methods[self.assembly_methods_title] = self.methods[self.assembly_methods_title].append(pandas.series(method))
1
+ − 764 self.info_ont_fastq(self.illumina_fastq_file)
+ − 765 self.add_assembly_notes()
0
+ − 766
+ − 767 def make_tex(self):
+ − 768 self.doc.new_table_of_contents(table_title='detailed run information', depth=2, marker="tableofcontents")
+ − 769 text = self.doc.file_data_text
+ − 770 text = text.replace("##--[", "")
+ − 771 text = text.replace("]--##", "")
+ − 772 self.doc.file_data_text = text
+ − 773 self.doc.create_md_file()
+ − 774
+ − 775 def make_report(self):
+ − 776 self.ofh.write("\nXXXXXX In make_report\n\n")
+ − 777 self.start_doc()
+ − 778 self.add_summary()
+ − 779 self.add_contamination()
+ − 780 self.add_alignment()
+ − 781 self.add_features()
+ − 782 self.add_feature_plots()
+ − 783 self.add_mutations()
+ − 784 self.add_large_indels()
+ − 785 self.add_plasmids()
+ − 786 self.add_amr_matrix()
+ − 787 # self.add_snps()
+ − 788 self.add_methods()
+ − 789 self.make_tex()
+ − 790 # It took me quite a long time to find out that the value of the -t
+ − 791 # (implied) argument in the following command must be 'html' instead of
+ − 792 # the more logical 'pdf'. see the answer from snsn in this thread:
+ − 793 # https://github.com/jessicategner/pypandoc/issues/186
+ − 794 self.ofh.write("\nXXXXX In make_report, calling pypandoc.convert_file...\n\n")
+ − 795 pypandoc.convert_file(self.report_md,
+ − 796 'html',
+ − 797 extra_args=['--pdf-engine=weasyprint', '-V', '-css=%s' % self.pima_css],
+ − 798 outputfile='pima_report.pdf')
+ − 799 self.ofh.close()
+ − 800
+ − 801
+ − 802 parser = argparse.ArgumentParser()
+ − 803
1
+ − 804 parser.add_argument('--amr_deletions_file', action='store', dest='amr_deletions_file', help='AMR deletions BED file')
+ − 805 parser.add_argument('--amr_matrix_png_dir', action='store', dest='amr_matrix_png_dir', help='Directory of AMR matrix PNG files')
0
+ − 806 parser.add_argument('--analysis_name', action='store', dest='analysis_name', help='Sample identifier')
+ − 807 parser.add_argument('--assembly_fasta_file', action='store', dest='assembly_fasta_file', help='Assembly fasta file')
+ − 808 parser.add_argument('--assembly_name', action='store', dest='assembly_name', help='Assembly identifier')
2
+ − 809 parser.add_argument('--blastn_version', action='store', dest='blastn_version', default=None, help='Blastn version string')
1
+ − 810 parser.add_argument('--compute_sequence_length_file', action='store', dest='compute_sequence_length_file', help='Comnpute sequence length tabular file')
+ − 811 parser.add_argument('--contig_coverage_file', action='store', dest='contig_coverage_file', help='Contig coverage TSV file')
+ − 812 parser.add_argument('--dbkey', action='store', dest='dbkey', help='Reference genome identifier')
+ − 813 parser.add_argument('--dnadiff_snps_file', action='store', dest='dnadiff_snps_file', help='DNAdiff snps tabular file')
2
+ − 814 parser.add_argument('--dnadiff_version', action='store', dest='dnadiff_version', help='DNAdiff version string')
0
+ − 815 parser.add_argument('--feature_bed_dir', action='store', dest='feature_bed_dir', help='Directory of best feature hits bed files')
+ − 816 parser.add_argument('--feature_png_dir', action='store', dest='feature_png_dir', help='Directory of best feature hits png files')
1
+ − 817 parser.add_argument('--flye_assembly_info_file', action='store', dest='flye_assembly_info_file', default=None, help='Flye assembly info tabular file')
+ − 818 parser.add_argument('--flye_version', action='store', dest='flye_version', default=None, help='Flye version string')
+ − 819 parser.add_argument('--genome_insertions_file', action='store', dest='genome_insertions_file', help='Genome insertions BED file')
0
+ − 820 parser.add_argument('--gzipped', action='store_true', dest='gzipped', default=False, help='Input sample is gzipped')
+ − 821 parser.add_argument('--illumina_fastq_file', action='store', dest='illumina_fastq_file', help='Input sample')
2
+ − 822 parser.add_argument('--kraken2_report_file', action='store', dest='kraken2_report_file', default=None, help='kraken2 report file')
+ − 823 parser.add_argument('--kraken2_version', action='store', dest='kraken2_version', default=None, help='kraken2 version string')
0
+ − 824 parser.add_argument('--mutation_regions_bed_file', action='store', dest='mutation_regions_bed_file', help='AMR mutation regions BRD file')
+ − 825 parser.add_argument('--mutation_regions_dir', action='store', dest='mutation_regions_dir', help='Directory of mutation regions TSV files')
+ − 826 parser.add_argument('--pima_css', action='store', dest='pima_css', help='PIMA css stypesheet')
1
+ − 827 parser.add_argument('--plasmids_file', action='store', dest='plasmids_file', help='pChunks plasmids TSV file')
+ − 828 parser.add_argument('--reference_insertions_file', action='store', dest='reference_insertions_file', help='Reference insertions BED file')
+ − 829 # parser.add_argument('--reference_genome', action='store', dest='reference_genome', help='Reference genome fasta file')
0
+ − 830
+ − 831 args = parser.parse_args()
+ − 832
1
+ − 833 # Prepare the AMR matrix PNG files.
+ − 834 amr_matrix_files = []
+ − 835 for file_name in sorted(os.listdir(args.amr_matrix_png_dir)):
+ − 836 file_path = os.path.abspath(os.path.join(args.amr_matrix_png_dir, file_name))
+ − 837 amr_matrix_files.append(file_path)
0
+ − 838 # Prepare the features BED files.
+ − 839 feature_bed_files = []
+ − 840 for file_name in sorted(os.listdir(args.feature_bed_dir)):
+ − 841 file_path = os.path.abspath(os.path.join(args.feature_bed_dir, file_name))
+ − 842 feature_bed_files.append(file_path)
+ − 843 # Prepare the features PNG files.
+ − 844 feature_png_files = []
+ − 845 for file_name in sorted(os.listdir(args.feature_png_dir)):
+ − 846 file_path = os.path.abspath(os.path.join(args.feature_png_dir, file_name))
+ − 847 feature_png_files.append(file_path)
+ − 848 # Prepare the mutation regions TSV files.
+ − 849 mutation_regions_files = []
+ − 850 for file_name in sorted(os.listdir(args.mutation_regions_dir)):
+ − 851 file_path = os.path.abspath(os.path.join(args.feature_png_dir, file_name))
+ − 852 mutation_regions_files.append(file_path)
+ − 853
+ − 854 markdown_report = PimaReport(args.analysis_name,
1
+ − 855 args.amr_deletions_file,
+ − 856 amr_matrix_files,
0
+ − 857 args.assembly_fasta_file,
+ − 858 args.assembly_name,
2
+ − 859 args.blastn_version,
1
+ − 860 args.compute_sequence_length_file,
+ − 861 args.contig_coverage_file,
+ − 862 args.dbkey,
+ − 863 args.dnadiff_snps_file,
2
+ − 864 args.dnadiff_version,
0
+ − 865 feature_bed_files,
+ − 866 feature_png_files,
1
+ − 867 args.flye_assembly_info_file,
+ − 868 args.flye_version,
+ − 869 args.genome_insertions_file,
0
+ − 870 args.gzipped,
+ − 871 args.illumina_fastq_file,
2
+ − 872 args.kraken2_report_file,
+ − 873 args.kraken2_version,
0
+ − 874 args.mutation_regions_bed_file,
+ − 875 mutation_regions_files,
1
+ − 876 args.pima_css,
+ − 877 args.plasmids_file,
+ − 878 args.reference_insertions_file)
0
+ − 879 markdown_report.make_report()