comparison operation_filter.py @ 0:5bc2dacbe729

Imported from capsule None
author devteam
date Tue, 01 Apr 2014 10:52:51 -0400
parents
children ecb36112b056
comparison
equal deleted inserted replaced
-1:000000000000 0:5bc2dacbe729
1 # runs after the job (and after the default post-filter)
2 import os
3 from galaxy import eggs
4 from galaxy import jobs
5 from galaxy.tools.parameters import DataToolParameter
6
7 from galaxy.jobs.handler import JOB_ERROR
8
9 # Older py compatibility
10 try:
11 set()
12 except:
13 from sets import Set as set
14
15 #def exec_before_process(app, inp_data, out_data, param_dict, tool=None):
16 # """Sets the name of the data"""
17 # dbkeys = sets.Set( [data.dbkey for data in inp_data.values() ] )
18 # if len(dbkeys) != 1:
19 # raise Exception, '<p><font color="yellow">Both Queries must be from the same genome build</font></p>'
20
21 def validate_input( trans, error_map, param_values, page_param_map ):
22 dbkeys = set()
23 data_param_names = set()
24 data_params = 0
25 for name, param in page_param_map.iteritems():
26 if isinstance( param, DataToolParameter ):
27 # for each dataset parameter
28 if param_values.get(name, None) != None:
29 dbkeys.add( param_values[name].dbkey )
30 data_params += 1
31 # check meta data
32 try:
33 param = param_values[name]
34 if isinstance( param.datatype, trans.app.datatypes_registry.get_datatype_by_extension( 'gff' ).__class__ ):
35 # TODO: currently cannot validate GFF inputs b/c they are not derived from interval.
36 pass
37 else: # Validate interval datatype.
38 startCol = int( param.metadata.startCol )
39 endCol = int( param.metadata.endCol )
40 chromCol = int( param.metadata.chromCol )
41 if param.metadata.strandCol is not None:
42 strandCol = int ( param.metadata.strandCol )
43 else:
44 strandCol = 0
45 except:
46 error_msg = "The attributes of this dataset are not properly set. " + \
47 "Click the pencil icon in the history item to set the chrom, start, end and strand columns."
48 error_map[name] = error_msg
49 data_param_names.add( name )
50 if len( dbkeys ) > 1:
51 for name in data_param_names:
52 error_map[name] = "All datasets must belong to same genomic build, " \
53 "this dataset is linked to build '%s'" % param_values[name].dbkey
54 if data_params != len(data_param_names):
55 for name in data_param_names:
56 error_map[name] = "A dataset of the appropriate type is required"
57
58 # Commented out by INS, 5/30/2007. What is the PURPOSE of this?
59 def exec_after_process(app, inp_data, out_data, param_dict, tool=None, stdout=None, stderr=None):
60 """Verify the output data after each run"""
61 items = out_data.items()
62
63 for name, data in items:
64 try:
65 if stderr and len( stderr ) > 0:
66 raise Exception( stderr )
67
68 except Exception, exc:
69 data.blurb = JOB_ERROR
70 data.state = JOB_ERROR
71
72 ## def exec_after_process(app, inp_data, out_data, param_dict, tool=None, stdout=None, stderr=None):
73 ## pass
74
75
76 def exec_after_merge(app, inp_data, out_data, param_dict, tool=None, stdout=None, stderr=None):
77 exec_after_process(
78 app, inp_data, out_data, param_dict, tool=tool, stdout=stdout, stderr=stderr)
79
80 # strip strand column if clusters were merged
81 items = out_data.items()
82 for name, data in items:
83 if param_dict['returntype'] == True:
84 data.metadata.chromCol = 1
85 data.metadata.startCol = 2
86 data.metadata.endCol = 3
87 # merge always clobbers strand
88 data.metadata.strandCol = None
89
90
91 def exec_after_cluster(app, inp_data, out_data, param_dict, tool=None, stdout=None, stderr=None):
92 exec_after_process(
93 app, inp_data, out_data, param_dict, tool=tool, stdout=stdout, stderr=stderr)
94
95 # strip strand column if clusters were merged
96 if param_dict["returntype"] == '1':
97 items = out_data.items()
98 for name, data in items:
99 data.metadata.strandCol = None