Mercurial > repos > yufei-luo > s_mart
comparison SMART/Java/Python/mapperAnalyzer.py @ 6:769e306b7933
Change the repository level.
author | yufei-luo |
---|---|
date | Fri, 18 Jan 2013 04:54:14 -0500 |
parents | |
children |
comparison
equal
deleted
inserted
replaced
5:ea3082881bf8 | 6:769e306b7933 |
---|---|
1 #! /usr/bin/env python | |
2 # | |
3 # Copyright INRA-URGI 2009-2010 | |
4 # | |
5 # This software is governed by the CeCILL license under French law and | |
6 # abiding by the rules of distribution of free software. You can use, | |
7 # modify and/ or redistribute the software under the terms of the CeCILL | |
8 # license as circulated by CEA, CNRS and INRIA at the following URL | |
9 # "http://www.cecill.info". | |
10 # | |
11 # As a counterpart to the access to the source code and rights to copy, | |
12 # modify and redistribute granted by the license, users are provided only | |
13 # with a limited warranty and the software's author, the holder of the | |
14 # economic rights, and the successive licensors have only limited | |
15 # liability. | |
16 # | |
17 # In this respect, the user's attention is drawn to the risks associated | |
18 # with loading, using, modifying and/or developing or reproducing the | |
19 # software by the user in light of its specific status of free software, | |
20 # that may mean that it is complicated to manipulate, and that also | |
21 # therefore means that it is reserved for developers and experienced | |
22 # professionals having in-depth computer knowledge. Users are therefore | |
23 # encouraged to load and test the software's suitability as regards their | |
24 # requirements in conditions enabling the security of their systems and/or | |
25 # data to be ensured and, more generally, to use and operate it in the | |
26 # same conditions as regards security. | |
27 # | |
28 # The fact that you are presently reading this means that you have had | |
29 # knowledge of the CeCILL license and that you accept its terms. | |
30 # | |
31 """ | |
32 Read a mapping file (many formats supported) and select some of them | |
33 Mappings should be sorted by read names | |
34 """ | |
35 import os, random, shelve | |
36 from optparse import OptionParser, OptionGroup | |
37 from commons.core.parsing.ParserChooser import ParserChooser | |
38 from commons.core.parsing.FastaParser import FastaParser | |
39 from commons.core.parsing.FastqParser import FastqParser | |
40 from commons.core.parsing.GffParser import GffParser | |
41 from commons.core.writer.BedWriter import BedWriter | |
42 from commons.core.writer.UcscWriter import UcscWriter | |
43 from commons.core.writer.GbWriter import GbWriter | |
44 from commons.core.writer.Gff2Writer import Gff2Writer | |
45 from commons.core.writer.Gff3Writer import Gff3Writer | |
46 from commons.core.writer.FastaWriter import FastaWriter | |
47 from commons.core.writer.FastqWriter import FastqWriter | |
48 from commons.core.writer.MySqlTranscriptWriter import MySqlTranscriptWriter | |
49 from SMART.Java.Python.mySql.MySqlConnection import MySqlConnection | |
50 from SMART.Java.Python.mySql.MySqlTable import MySqlTable | |
51 from SMART.Java.Python.misc.RPlotter import RPlotter | |
52 from SMART.Java.Python.misc.Progress import Progress | |
53 from SMART.Java.Python.misc.UnlimitedProgress import UnlimitedProgress | |
54 | |
55 | |
56 distanceExons = 20 | |
57 exonSize = 20 | |
58 | |
59 | |
60 class MapperAnalyzer(object): | |
61 """ | |
62 Analyse the output of a parser | |
63 """ | |
64 | |
65 def __init__(self, verbosity = 0): | |
66 self.verbosity = verbosity | |
67 self.mySqlConnection = MySqlConnection(verbosity) | |
68 self.tooShort = 0 | |
69 self.tooManyMismatches = 0 | |
70 self.tooManyGaps = 0 | |
71 self.tooShortExons = 0 | |
72 self.tooManyMappings = 0 | |
73 self.nbMappings = 0 | |
74 self.nbSequences = 0 | |
75 self.nbAlreadyMapped = 0 | |
76 self.nbAlreadyMappedSequences = 0 | |
77 self.nbWrittenMappings = 0 | |
78 self.nbWrittenSequences = 0 | |
79 self.parser = None | |
80 self.logHandle = None | |
81 self.randomNumber = random.randint(0, 100000) | |
82 self.gff3Writer = None | |
83 self.alreadyMappedReader = None | |
84 self.unmatchedWriter = None | |
85 self.sequenceListParser = None | |
86 self.sequences = None | |
87 self.alreadyMapped = None | |
88 self.mappedNamesTable = None | |
89 self.minSize = None | |
90 self.minId = None | |
91 self.maxMismatches = None | |
92 self.maxGaps = None | |
93 self.maxMappings = None | |
94 self.merge = False | |
95 self.checkExons = False | |
96 self.suffix = None | |
97 self.tmpDirectory = "%s%s" % (os.environ["SMARTMPPATH"], os.sep) if "SMARTMPPATH" in os.environ else "" | |
98 | |
99 | |
100 def __del__(self): | |
101 if self.sequences != None: | |
102 self.sequences.close() | |
103 if self.alreadyMapped != None: | |
104 self.alreadyMapped.close() | |
105 if self.mappedNamesTable != None: | |
106 self.mappedNamesTable.remove() | |
107 if self.gff3Writer != None: | |
108 self.gff3Writer.close() | |
109 | |
110 if self.logHandle != None: | |
111 self.logHandle.close() | |
112 | |
113 | |
114 def setMappingFile(self, fileName, format): | |
115 parserChooser = ParserChooser(self.verbosity) | |
116 parserChooser.findFormat(format, "mapping") | |
117 self.parser = parserChooser.getParser(fileName) | |
118 | |
119 | |
120 def setSequenceFile(self, fileName, format): | |
121 if format == "fasta": | |
122 self.sequenceListParser = FastaParser(fileName, self.verbosity) | |
123 elif format == "fastq": | |
124 self.sequenceListParser = FastqParser(fileName, self.verbosity) | |
125 else: | |
126 raise Exception("Do not understand sequence format %s" % (format)) | |
127 | |
128 | |
129 def setOutputFile(self, fileName, title): | |
130 self.gff3Writer = Gff3Writer(fileName, self.verbosity) | |
131 self.gff3Writer.setTitle(title) | |
132 | |
133 | |
134 def setAlreadyMatched(self, fileName): | |
135 self.alreadyMappedReader = GffParser(fileName, self.verbosity) | |
136 | |
137 | |
138 def setRemainingFile(self, fileName, format): | |
139 if format == "fasta": | |
140 self.unmatchedWriter = FastaWriter("%s_unmatched.fasta" % (fileName), self.verbosity) | |
141 elif format == "fastq": | |
142 self.unmatchedWriter = FastqWriter("%s_unmatched.fastq" % (fileName), self.verbosity) | |
143 else: | |
144 raise Exception("Do not understand %s format." % (format)) | |
145 self.mappedNamesTable = MySqlTable(self.mySqlConnection, "mappedNames_%d" % (self.randomNumber), self.verbosity) | |
146 self.mappedNamesTable.create(["name"], {"name": "char"}, {"name": 50}) | |
147 self.mappedNamesTable.createIndex("iNameMapped", ["name", ], True) | |
148 | |
149 | |
150 def setLog(self, fileName): | |
151 self.logHandle = open(fileName, "w") | |
152 | |
153 | |
154 def setMinSize(self, size): | |
155 self.minSize = size | |
156 | |
157 | |
158 def setMinId(self, id): | |
159 self.minId = id | |
160 | |
161 | |
162 def setMaxMismatches(self, mismatches): | |
163 self.maxMismatches = mismatches | |
164 | |
165 | |
166 def setMaxGaps(self, gaps): | |
167 self.maxGaps = gaps | |
168 | |
169 | |
170 def setMaxMappings(self, mappings): | |
171 self.maxMappings = mappings | |
172 | |
173 | |
174 def mergeExons(self, b): | |
175 self.merge = b | |
176 | |
177 | |
178 def acceptShortExons(self, b): | |
179 self.checkExons = not b | |
180 | |
181 | |
182 def countMappings(self): | |
183 self.nbMappings = self.parser.getNbMappings() | |
184 if self.verbosity > 0: | |
185 print "%i matches found" % (self.nbMappings) | |
186 | |
187 | |
188 def storeAlreadyMapped(self): | |
189 self.alreadyMapped = shelve.open("%stmpAlreadyMapped_%d" % (self.tmpDirectory, self.randomNumber)) | |
190 progress = Progress(self.alreadyMappedReader.getNbTranscripts(), "Reading already mapped reads", self.verbosity) | |
191 self.nbAlreadyMappedSequences = 0 | |
192 for transcript in self.alreadyMappedReader.getIterator(): | |
193 if not self.alreadyMapped.has_key(transcript.getName()): | |
194 self.alreadyMapped[transcript.getName()] = 1 | |
195 self.nbAlreadyMappedSequences += 1 | |
196 progress.inc() | |
197 progress.done() | |
198 self.nbAlreadyMapped = self.alreadyMappedReader.getNbTranscripts() | |
199 | |
200 | |
201 def storeSequences(self): | |
202 self.sequences = shelve.open("%stmpSequences_%d" % (self.tmpDirectory, self.randomNumber)) | |
203 progress = Progress(self.sequenceListParser.getNbSequences(), "Reading sequences", self.verbosity) | |
204 for sequence in self.sequenceListParser.getIterator(): | |
205 self.sequences[sequence.getName().split(" ")[0]] = len(sequence.getSequence()) | |
206 self.nbSequences += 1 | |
207 progress.inc() | |
208 progress.done() | |
209 if self.verbosity > 0: | |
210 print "%i sequences read" % (self.nbSequences) | |
211 | |
212 | |
213 def checkOrder(self): | |
214 names = shelve.open("%stmpNames_%d" % (self.tmpDirectory, self.randomNumber)) | |
215 previousName = None | |
216 progress = Progress(self.nbMappings, "Checking mapping file", self.verbosity) | |
217 for mapping in self.parser.getIterator(): | |
218 name = mapping.queryInterval.getName() | |
219 if name != previousName and previousName != None: | |
220 if names.has_key(previousName): | |
221 raise Exception("Error! Input mapping file is not ordered! (Name '%s' occurs at least twice)" % (previousName)) | |
222 names[previousName] = 1 | |
223 previousName = name | |
224 progress.inc() | |
225 progress.done() | |
226 names.close() | |
227 | |
228 | |
229 def checkPreviouslyMapped(self, name): | |
230 if self.alreadyMappedReader == None: | |
231 return False | |
232 return self.alreadyMapped.has_key(name) | |
233 | |
234 | |
235 def findOriginalSize(self, name): | |
236 alternate = "%s/1" % (name) | |
237 if (self.suffix == None) or (not self.suffix): | |
238 if self.sequences.has_key(name): | |
239 self.suffix = False | |
240 return self.sequences[name] | |
241 if self.suffix == None: | |
242 self.suffix = True | |
243 else: | |
244 raise Exception("Cannot find name %n" % (name)) | |
245 if (self.suffix): | |
246 if self.sequences.has_key(alternate): | |
247 return self.sequences[alternate] | |
248 raise Exception("Cannot find name %s" % (name)) | |
249 | |
250 | |
251 def checkErrors(self, mapping): | |
252 accepted = True | |
253 # short size | |
254 if self.minSize != None and mapping.size * 100 < self.minSize * mapping.queryInterval.size: | |
255 self.tooShort += 1 | |
256 accepted = False | |
257 if self.logHandle != None: | |
258 self.logHandle.write("size of mapping %s is too short (%i instead of %i)\n" % (str(mapping), mapping.queryInterval.size, mapping.size)) | |
259 # low identity | |
260 if self.minId != None and mapping.getTagValue("identity") < self.minId: | |
261 self.tooManyMismatches += 1 | |
262 accepted = False | |
263 if self.logHandle != None: | |
264 self.logHandle.write("mapping %s has a low identity rate\n" % (str(mapping))) | |
265 # too many mismatches | |
266 if self.maxMismatches != None and mapping.getTagValue("nbMismatches") > self.maxMismatches: | |
267 self.tooManyMismatches += 1 | |
268 accepted = False | |
269 if self.logHandle != None: | |
270 self.logHandle.write("mapping %s has more mismatches than %i\n" % (str(mapping), self.maxMismatches)) | |
271 # too many gaps | |
272 if self.maxGaps != None and mapping.getTagValue("nbGaps") > self.maxGaps: | |
273 self.tooManyGaps += 1 | |
274 accepted = False | |
275 if self.logHandle != None: | |
276 self.logHandle.write("mapping %s has more gaps than %i\n" % (str(mapping), self.maxGaps)) | |
277 # short exons | |
278 if self.checkExons and len(mapping.subMappings) > 1 and min([subMapping.targetInterval.getSize() for subMapping in mapping.subMappings]) < exonSize: | |
279 self.tooShortExons += 1 | |
280 accepted = False | |
281 if self.logHandle != None: | |
282 self.logHandle.write("sequence %s maps as too short exons\n" % (mapping)) | |
283 return accepted | |
284 | |
285 | |
286 def checkNbMappings(self, mappings): | |
287 nbOccurrences = 0 | |
288 for mapping in mappings: | |
289 nbOccurrences += 1 if "nbOccurrences" not in mapping.getTagNames() else mapping.getTagValue("nbOccurrences") | |
290 if (self.maxMappings != None and nbOccurrences > self.maxMappings): | |
291 self.tooManyMappings += 1 | |
292 if self.logHandle != None: | |
293 self.logHandle.write("sequence %s maps %i times\n" % (mappings[0].queryInterval.getName(), nbOccurrences)) | |
294 return False | |
295 return (nbOccurrences > 0) | |
296 | |
297 | |
298 def sortMappings(self, mappings): | |
299 nbOccurrences = 0 | |
300 for mapping in mappings: | |
301 nbOccurrences += 1 if "nbOccurrences" not in mapping.getTagNames() else mapping.getTagValue("nbOccurrences") | |
302 | |
303 orderedMappings = sorted(mappings, key = lambda mapping: mapping.getErrorScore()) | |
304 cpt = 1 | |
305 rank = 1 | |
306 previousMapping = None | |
307 previousScore = None | |
308 wasLastTie = False | |
309 rankedMappings = [] | |
310 bestRegion = "%s:%d-%d" % (orderedMappings[0].targetInterval.getChromosome(), orderedMappings[0].targetInterval.getStart(), orderedMappings[0].targetInterval.getEnd()) | |
311 for mapping in orderedMappings: | |
312 mapping.setNbOccurrences(nbOccurrences) | |
313 mapping.setOccurrence(cpt) | |
314 | |
315 score = mapping.getErrorScore() | |
316 if previousScore != None and previousScore == score: | |
317 if "Rank" in previousMapping.getTagNames(): | |
318 if not wasLastTie: | |
319 previousMapping.setRank("%sTie" % (rank)) | |
320 mapping.setRank("%sTie" % (rank)) | |
321 wasLastTie = True | |
322 else: | |
323 rank = cpt | |
324 mapping.setRank(rank) | |
325 wasLastTie = False | |
326 if cpt != 1: | |
327 mapping.setBestRegion(bestRegion) | |
328 | |
329 rankedMappings.append(mapping) | |
330 previousMapping = mapping | |
331 previousScore = score | |
332 cpt += 1 | |
333 return rankedMappings | |
334 | |
335 | |
336 def processMappings(self, mappings): | |
337 if not mappings: | |
338 return | |
339 selectedMappings = [] | |
340 name = mappings[0].queryInterval.getName() | |
341 size = self.findOriginalSize(name) | |
342 for mapping in mappings: | |
343 if self.merge: | |
344 mapping.mergeExons(distanceExons) | |
345 mapping.queryInterval.size = size | |
346 if self.checkErrors(mapping): | |
347 selectedMappings.append(mapping) | |
348 | |
349 if self.checkNbMappings(selectedMappings): | |
350 if self.unmatchedWriter != None: | |
351 query = self.mySqlConnection.executeQuery("INSERT INTO %s (name) VALUES ('%s')" % (self.mappedNamesTable.name, name if not self.suffix else "%s/1" % (name))) | |
352 self.nbWrittenSequences += 1 | |
353 mappings = self.sortMappings(selectedMappings) | |
354 for mapping in mappings: | |
355 self.nbWrittenMappings += 1 | |
356 self.gff3Writer.addTranscript(mapping.getTranscript()) | |
357 | |
358 | |
359 def readMappings(self): | |
360 previousQueryName = None | |
361 mappings = [] | |
362 self.parser.reset() | |
363 progress = Progress(self.nbMappings, "Reading mappings", self.verbosity) | |
364 for mapping in self.parser.getIterator(): | |
365 queryName = mapping.queryInterval.getName().split(" ")[0] | |
366 if self.checkPreviouslyMapped(queryName): | |
367 if self.logHandle != None: | |
368 self.logHandle.write("Mapping %s has already been mapped.\n" % (queryName)) | |
369 else: | |
370 if previousQueryName == queryName: | |
371 mappings.append(mapping) | |
372 else: | |
373 if previousQueryName != None: | |
374 self.processMappings(mappings) | |
375 previousQueryName = queryName | |
376 mappings = [mapping, ] | |
377 progress.inc() | |
378 self.processMappings(mappings) | |
379 self.gff3Writer.write() | |
380 self.gff3Writer.close() | |
381 progress.done() | |
382 | |
383 | |
384 def writeUnmatched(self): | |
385 progress = Progress(self.nbSequences, "Reading unmatched sequences", self.verbosity) | |
386 for sequence in self.sequenceListParser.getIterator(): | |
387 name = sequence.getName().split(" ")[0] | |
388 query = self.mySqlConnection.executeQuery("SELECT * FROM %s WHERE name = '%s' LIMIT 1" % (self.mappedNamesTable.name, name)) | |
389 if query.isEmpty(): | |
390 self.unmatchedWriter.addSequence(sequence) | |
391 progress.inc() | |
392 progress.done() | |
393 | |
394 | |
395 def analyze(self): | |
396 self.countMappings() | |
397 self.checkOrder() | |
398 self.storeSequences() | |
399 if self.alreadyMappedReader != None: | |
400 self.storeAlreadyMapped() | |
401 self.readMappings() | |
402 if self.unmatchedWriter != None: | |
403 self.writeUnmatched() | |
404 | |
405 | |
406 | |
407 | |
408 if __name__ == "__main__": | |
409 | |
410 # parse command line | |
411 description = "Mapper Analyzer v1.0.1: Read the output of an aligner, print statistics and possibly translate into BED or GBrowse formats. [Category: Conversion]" | |
412 | |
413 parser = OptionParser(description = description) | |
414 compGroup = OptionGroup(parser, "Compulsory options") | |
415 filtGroup = OptionGroup(parser, "Filtering options") | |
416 tranGroup = OptionGroup(parser, "Transformation options") | |
417 outpGroup = OptionGroup(parser, "Output options") | |
418 otheGroup = OptionGroup(parser, "Other options") | |
419 compGroup.add_option("-i", "--input", dest="inputFileName", action="store", type="string", help="input file (output of the tool) [compulsory] [format: file in mapping format given by -f]") | |
420 compGroup.add_option("-f", "--format", dest="format", action="store", default="seqmap", type="string", help="format of the file [compulsory] [format: mapping file format]") | |
421 compGroup.add_option("-q", "--sequences", dest="sequencesFileName", action="store", type="string", help="file of the sequences [compulsory] [format: file in sequence format given by -k]") | |
422 compGroup.add_option("-k", "--seqFormat", dest="sequenceFormat", action="store", default="fasta", type="string", help="format of the sequences: fasta or fastq [default: fasta] [format: sequence file format]") | |
423 compGroup.add_option("-o", "--output", dest="outputFileName", action="store", type="string", help="output file [compulsory] [format: output file in GFF3 format]") | |
424 filtGroup.add_option("-n", "--number", dest="number", action="store", default=None, type="int", help="max. number of occurrences of a sequence [format: int]") | |
425 filtGroup.add_option("-s", "--size", dest="size", action="store", default=None, type="int", help="minimum pourcentage of size [format: int]") | |
426 filtGroup.add_option("-d", "--identity", dest="identity", action="store", default=None, type="int", help="minimum pourcentage of identity [format: int]") | |
427 filtGroup.add_option("-m", "--mismatch", dest="mismatch", action="store", default=None, type="int", help="maximum number of mismatches [format: int]") | |
428 filtGroup.add_option("-p", "--gap", dest="gap", action="store", default=None, type="int", help="maximum number of gaps [format: int]") | |
429 tranGroup.add_option("-e", "--mergeExons", dest="mergeExons", action="store_true", default=False, help="merge exons when introns are short [format: bool] [default: false]") | |
430 tranGroup.add_option("-x", "--removeExons", dest="removeExons", action="store_true", default=False, help="remove transcripts when exons are short [format: bool] [default: false]") | |
431 outpGroup.add_option("-t", "--title", dest="title", action="store", default="SMART", type="string", help="title of the UCSC track [format: string] [default: SMART]") | |
432 outpGroup.add_option("-r", "--remaining", dest="remaining", action="store_true", default=False, help="print the unmatched sequences [format: bool] [default: false]") | |
433 otheGroup.add_option("-a", "--append", dest="appendFileName", action="store", default=None, type="string", help="append to GFF3 file [format: file in GFF3 format]") | |
434 otheGroup.add_option("-v", "--verbosity", dest="verbosity", action="store", default=1, type="int", help="trace level [default: 1] [format: int]") | |
435 otheGroup.add_option("-l", "--log", dest="log", action="store_true", default=False, help="write a log file [format: bool] [default: false]") | |
436 parser.add_option_group(compGroup) | |
437 parser.add_option_group(filtGroup) | |
438 parser.add_option_group(tranGroup) | |
439 parser.add_option_group(outpGroup) | |
440 parser.add_option_group(otheGroup) | |
441 (options, args) = parser.parse_args() | |
442 | |
443 | |
444 analyzer = MapperAnalyzer(options.verbosity) | |
445 analyzer.setMappingFile(options.inputFileName, options.format) | |
446 analyzer.setSequenceFile(options.sequencesFileName, options.sequenceFormat) | |
447 analyzer.setOutputFile(options.outputFileName, options.title) | |
448 if options.appendFileName != None: | |
449 analyzer.setAlreadyMatched(options.appendFileName) | |
450 if options.remaining: | |
451 analyzer.setRemainingFile(options.outputFileName, options.sequenceFormat) | |
452 if options.number != None: | |
453 analyzer.setMaxMappings(options.number) | |
454 if options.size != None: | |
455 analyzer.setMinSize(options.size) | |
456 if options.identity != None: | |
457 analyzer.setMinId(options.identity) | |
458 if options.mismatch != None: | |
459 analyzer.setMaxMismatches(options.mismatch) | |
460 if options.gap != None: | |
461 analyzer.setMaxGaps(options.gap) | |
462 if options.mergeExons: | |
463 analyzer.mergeExons(True) | |
464 if options.removeExons: | |
465 analyzer.acceptShortExons(False) | |
466 if options.log: | |
467 analyzer.setLog("%s.log" % (options.outputFileName)) | |
468 analyzer.analyze() | |
469 | |
470 if options.verbosity > 0: | |
471 print "kept %i sequences over %s (%f%%)" % (analyzer.nbWrittenSequences, analyzer.nbSequences, float(analyzer.nbWrittenSequences) / analyzer.nbSequences * 100) | |
472 if options.appendFileName != None: | |
473 print "kept %i sequences over %s (%f%%) including already mapped sequences" % (analyzer.nbWrittenSequences + analyzer.nbAlreadyMappedSequences, analyzer.nbSequences, float(analyzer.nbWrittenSequences + analyzer.nbAlreadyMappedSequences) / analyzer.nbSequences * 100) | |
474 print "kept %i mappings over %i (%f%%)" % (analyzer.nbWrittenMappings, analyzer.nbMappings, float(analyzer.nbWrittenMappings) / analyzer.nbMappings * 100) | |
475 if options.appendFileName != None: | |
476 print "kept %i mappings over %i (%f%%) including already mapped" % (analyzer.nbWrittenMappings + analyzer.nbAlreadyMapped, analyzer.nbMappings, float(analyzer.nbWrittenMappings + analyzer.nbAlreadyMapped) / analyzer.nbMappings * 100) | |
477 print "removed %i too short mappings (%f%%)" % (analyzer.tooShort, float(analyzer.tooShort) / analyzer.nbMappings * 100) | |
478 print "removed %i mappings with too many mismatches (%f%%)" % (analyzer.tooManyMismatches, float(analyzer.tooManyMismatches) / analyzer.nbMappings * 100) | |
479 print "removed %i mappings with too many gaps (%f%%)" % (analyzer.tooManyGaps, float(analyzer.tooManyGaps) / analyzer.nbMappings * 100) | |
480 print "removed %i mappings with too short exons (%f%%)" % (analyzer.tooShortExons, float(analyzer.tooShortExons) / analyzer.nbMappings * 100) | |
481 print "removed %i sequences with too many hits (%f%%)" % (analyzer.tooManyMappings, float(analyzer.tooManyMappings) / analyzer.nbSequences * 100) | |
482 print "%i sequences have no mapping (%f%%)" % (analyzer.nbSequences - analyzer.nbWrittenSequences, float(analyzer.nbSequences - analyzer.nbWrittenSequences) / analyzer.nbSequences * 100) | |
483 if options.appendFileName != None: | |
484 print "%i sequences have no mapping (%f%%) excluding already mapped sequences" % (analyzer.nbSequences - analyzer.nbWrittenSequences - analyzer.nbAlreadyMappedSequences, float(analyzer.nbSequences - analyzer.nbWrittenSequences - analyzer.nbAlreadyMappedSequences) / analyzer.nbSequences * 100) | |
485 | |
486 |