Mercurial > repos > devteam > subtract
comparison operation_filter.py @ 3:ecb36112b056 draft
planemo upload for repository https://github.com/galaxyproject/tools-devteam/tree/master/tool_collections/gops/subtract commit a1517c9d22029095120643bbe2c8fa53754dd2b7
author | devteam |
---|---|
date | Wed, 11 Nov 2015 12:49:24 -0500 |
parents | 5bc2dacbe729 |
children | 0145969324c4 |
comparison
equal
deleted
inserted
replaced
2:c19a2a29c561 | 3:ecb36112b056 |
---|---|
1 # runs after the job (and after the default post-filter) | 1 # runs after the job (and after the default post-filter) |
2 import os | |
3 from galaxy import eggs | |
4 from galaxy import jobs | |
5 from galaxy.tools.parameters import DataToolParameter | 2 from galaxy.tools.parameters import DataToolParameter |
6 | 3 |
7 from galaxy.jobs.handler import JOB_ERROR | 4 from galaxy.jobs.handler import JOB_ERROR |
8 | 5 |
9 # Older py compatibility | 6 # Older py compatibility |
10 try: | 7 try: |
11 set() | 8 set() |
12 except: | 9 except: |
13 from sets import Set as set | 10 from sets import Set as set |
14 | 11 |
15 #def exec_before_process(app, inp_data, out_data, param_dict, tool=None): | |
16 # """Sets the name of the data""" | |
17 # dbkeys = sets.Set( [data.dbkey for data in inp_data.values() ] ) | |
18 # if len(dbkeys) != 1: | |
19 # raise Exception, '<p><font color="yellow">Both Queries must be from the same genome build</font></p>' | |
20 | 12 |
21 def validate_input( trans, error_map, param_values, page_param_map ): | 13 def validate_input( trans, error_map, param_values, page_param_map ): |
22 dbkeys = set() | 14 dbkeys = set() |
23 data_param_names = set() | 15 data_param_names = set() |
24 data_params = 0 | 16 data_params = 0 |
25 for name, param in page_param_map.iteritems(): | 17 for name, param in page_param_map.iteritems(): |
26 if isinstance( param, DataToolParameter ): | 18 if isinstance( param, DataToolParameter ): |
27 # for each dataset parameter | 19 # for each dataset parameter |
28 if param_values.get(name, None) != None: | 20 if param_values.get(name, None) is not None: |
29 dbkeys.add( param_values[name].dbkey ) | 21 dbkeys.add( param_values[name].dbkey ) |
30 data_params += 1 | 22 data_params += 1 |
31 # check meta data | 23 # check meta data |
32 try: | 24 try: |
33 param = param_values[name] | 25 param = param_values[name] |
34 if isinstance( param.datatype, trans.app.datatypes_registry.get_datatype_by_extension( 'gff' ).__class__ ): | 26 if isinstance( param.datatype, trans.app.datatypes_registry.get_datatype_by_extension( 'gff' ).__class__ ): |
35 # TODO: currently cannot validate GFF inputs b/c they are not derived from interval. | 27 # TODO: currently cannot validate GFF inputs b/c they are not derived from interval. |
36 pass | 28 pass |
37 else: # Validate interval datatype. | 29 else: # Validate interval datatype. |
38 startCol = int( param.metadata.startCol ) | 30 int( param.metadata.startCol ) |
39 endCol = int( param.metadata.endCol ) | 31 int( param.metadata.endCol ) |
40 chromCol = int( param.metadata.chromCol ) | 32 int( param.metadata.chromCol ) |
41 if param.metadata.strandCol is not None: | 33 if param.metadata.strandCol is not None: |
42 strandCol = int ( param.metadata.strandCol ) | 34 int( param.metadata.strandCol ) |
43 else: | |
44 strandCol = 0 | |
45 except: | 35 except: |
46 error_msg = "The attributes of this dataset are not properly set. " + \ | 36 error_msg = "The attributes of this dataset are not properly set. " + \ |
47 "Click the pencil icon in the history item to set the chrom, start, end and strand columns." | 37 "Click the pencil icon in the history item to set the chrom, start, end and strand columns." |
48 error_map[name] = error_msg | 38 error_map[name] = error_msg |
49 data_param_names.add( name ) | 39 data_param_names.add( name ) |
50 if len( dbkeys ) > 1: | 40 if len( dbkeys ) > 1: |
51 for name in data_param_names: | 41 for name in data_param_names: |
52 error_map[name] = "All datasets must belong to same genomic build, " \ | 42 error_map[name] = "All datasets must belong to same genomic build, " \ |
53 "this dataset is linked to build '%s'" % param_values[name].dbkey | 43 "this dataset is linked to build '%s'" % param_values[name].dbkey |
54 if data_params != len(data_param_names): | 44 if data_params != len(data_param_names): |
55 for name in data_param_names: | 45 for name in data_param_names: |
56 error_map[name] = "A dataset of the appropriate type is required" | 46 error_map[name] = "A dataset of the appropriate type is required" |
57 | 47 |
48 | |
58 # Commented out by INS, 5/30/2007. What is the PURPOSE of this? | 49 # Commented out by INS, 5/30/2007. What is the PURPOSE of this? |
59 def exec_after_process(app, inp_data, out_data, param_dict, tool=None, stdout=None, stderr=None): | 50 def exec_after_process(app, inp_data, out_data, param_dict, tool=None, stdout=None, stderr=None): |
60 """Verify the output data after each run""" | 51 """Verify the output data after each run""" |
61 items = out_data.items() | 52 for data in out_data.values(): |
62 | |
63 for name, data in items: | |
64 try: | 53 try: |
65 if stderr and len( stderr ) > 0: | 54 if stderr and len( stderr ) > 0: |
66 raise Exception( stderr ) | 55 raise Exception( stderr ) |
67 | 56 |
68 except Exception, exc: | 57 except Exception: |
69 data.blurb = JOB_ERROR | 58 data.blurb = JOB_ERROR |
70 data.state = JOB_ERROR | 59 data.state = JOB_ERROR |
71 | |
72 ## def exec_after_process(app, inp_data, out_data, param_dict, tool=None, stdout=None, stderr=None): | |
73 ## pass | |
74 | 60 |
75 | 61 |
76 def exec_after_merge(app, inp_data, out_data, param_dict, tool=None, stdout=None, stderr=None): | 62 def exec_after_merge(app, inp_data, out_data, param_dict, tool=None, stdout=None, stderr=None): |
77 exec_after_process( | 63 exec_after_process( |
78 app, inp_data, out_data, param_dict, tool=tool, stdout=stdout, stderr=stderr) | 64 app, inp_data, out_data, param_dict, tool=tool, stdout=stdout, stderr=stderr) |
79 | 65 |
80 # strip strand column if clusters were merged | 66 # strip strand column if clusters were merged |
81 items = out_data.items() | 67 for data in out_data.values(): |
82 for name, data in items: | 68 if param_dict['returntype'] is True: |
83 if param_dict['returntype'] == True: | |
84 data.metadata.chromCol = 1 | 69 data.metadata.chromCol = 1 |
85 data.metadata.startCol = 2 | 70 data.metadata.startCol = 2 |
86 data.metadata.endCol = 3 | 71 data.metadata.endCol = 3 |
87 # merge always clobbers strand | 72 # merge always clobbers strand |
88 data.metadata.strandCol = None | 73 data.metadata.strandCol = None |
89 | 74 |
90 | 75 |
91 def exec_after_cluster(app, inp_data, out_data, param_dict, tool=None, stdout=None, stderr=None): | 76 def exec_after_cluster(app, inp_data, out_data, param_dict, tool=None, stdout=None, stderr=None): |
92 exec_after_process( | 77 exec_after_process( |
93 app, inp_data, out_data, param_dict, tool=tool, stdout=stdout, stderr=stderr) | 78 app, inp_data, out_data, param_dict, tool=tool, stdout=stdout, stderr=stderr) |
94 | 79 |
95 # strip strand column if clusters were merged | 80 # strip strand column if clusters were merged |
96 if param_dict["returntype"] == '1': | 81 if param_dict["returntype"] == '1': |
97 items = out_data.items() | 82 for data in out_data.values(): |
98 for name, data in items: | |
99 data.metadata.strandCol = None | 83 data.metadata.strandCol = None |