Mercurial > repos > devteam > join
comparison operation_filter.py @ 0:e56b47dce68a
Imported from capsule None
author | devteam |
---|---|
date | Tue, 01 Apr 2014 10:54:03 -0400 |
parents | |
children | ffbd1de29c28 |
comparison
equal
deleted
inserted
replaced
-1:000000000000 | 0:e56b47dce68a |
---|---|
1 # runs after the job (and after the default post-filter) | |
2 import os | |
3 from galaxy import eggs | |
4 from galaxy import jobs | |
5 from galaxy.tools.parameters import DataToolParameter | |
6 | |
7 from galaxy.jobs.handler import JOB_ERROR | |
8 | |
9 # Older py compatibility | |
10 try: | |
11 set() | |
12 except: | |
13 from sets import Set as set | |
14 | |
15 #def exec_before_process(app, inp_data, out_data, param_dict, tool=None): | |
16 # """Sets the name of the data""" | |
17 # dbkeys = sets.Set( [data.dbkey for data in inp_data.values() ] ) | |
18 # if len(dbkeys) != 1: | |
19 # raise Exception, '<p><font color="yellow">Both Queries must be from the same genome build</font></p>' | |
20 | |
21 def validate_input( trans, error_map, param_values, page_param_map ): | |
22 dbkeys = set() | |
23 data_param_names = set() | |
24 data_params = 0 | |
25 for name, param in page_param_map.iteritems(): | |
26 if isinstance( param, DataToolParameter ): | |
27 # for each dataset parameter | |
28 if param_values.get(name, None) != None: | |
29 dbkeys.add( param_values[name].dbkey ) | |
30 data_params += 1 | |
31 # check meta data | |
32 try: | |
33 param = param_values[name] | |
34 if isinstance( param.datatype, trans.app.datatypes_registry.get_datatype_by_extension( 'gff' ).__class__ ): | |
35 # TODO: currently cannot validate GFF inputs b/c they are not derived from interval. | |
36 pass | |
37 else: # Validate interval datatype. | |
38 startCol = int( param.metadata.startCol ) | |
39 endCol = int( param.metadata.endCol ) | |
40 chromCol = int( param.metadata.chromCol ) | |
41 if param.metadata.strandCol is not None: | |
42 strandCol = int ( param.metadata.strandCol ) | |
43 else: | |
44 strandCol = 0 | |
45 except: | |
46 error_msg = "The attributes of this dataset are not properly set. " + \ | |
47 "Click the pencil icon in the history item to set the chrom, start, end and strand columns." | |
48 error_map[name] = error_msg | |
49 data_param_names.add( name ) | |
50 if len( dbkeys ) > 1: | |
51 for name in data_param_names: | |
52 error_map[name] = "All datasets must belong to same genomic build, " \ | |
53 "this dataset is linked to build '%s'" % param_values[name].dbkey | |
54 if data_params != len(data_param_names): | |
55 for name in data_param_names: | |
56 error_map[name] = "A dataset of the appropriate type is required" | |
57 | |
58 # Commented out by INS, 5/30/2007. What is the PURPOSE of this? | |
59 def exec_after_process(app, inp_data, out_data, param_dict, tool=None, stdout=None, stderr=None): | |
60 """Verify the output data after each run""" | |
61 items = out_data.items() | |
62 | |
63 for name, data in items: | |
64 try: | |
65 if stderr and len( stderr ) > 0: | |
66 raise Exception( stderr ) | |
67 | |
68 except Exception, exc: | |
69 data.blurb = JOB_ERROR | |
70 data.state = JOB_ERROR | |
71 | |
72 ## def exec_after_process(app, inp_data, out_data, param_dict, tool=None, stdout=None, stderr=None): | |
73 ## pass | |
74 | |
75 | |
76 def exec_after_merge(app, inp_data, out_data, param_dict, tool=None, stdout=None, stderr=None): | |
77 exec_after_process( | |
78 app, inp_data, out_data, param_dict, tool=tool, stdout=stdout, stderr=stderr) | |
79 | |
80 # strip strand column if clusters were merged | |
81 items = out_data.items() | |
82 for name, data in items: | |
83 if param_dict['returntype'] == True: | |
84 data.metadata.chromCol = 1 | |
85 data.metadata.startCol = 2 | |
86 data.metadata.endCol = 3 | |
87 # merge always clobbers strand | |
88 data.metadata.strandCol = None | |
89 | |
90 | |
91 def exec_after_cluster(app, inp_data, out_data, param_dict, tool=None, stdout=None, stderr=None): | |
92 exec_after_process( | |
93 app, inp_data, out_data, param_dict, tool=tool, stdout=stdout, stderr=stderr) | |
94 | |
95 # strip strand column if clusters were merged | |
96 if param_dict["returntype"] == '1': | |
97 items = out_data.items() | |
98 for name, data in items: | |
99 data.metadata.strandCol = None |