Re: [galaxy-dev] Setup galaxy to run on SGE cluster

2012-01-09 Thread Ryan Golhar
Whoops.  That's what I mean...a submission host, not an exec host.

On Mon, Jan 9, 2012 at 11:13 PM, Derrick LIN  wrote:

> Thanks Ryan,
>
> I have been following that guide. I can configure the galaxy host to be an
> exec host, although it makes more sense to be a submission host.
>
> I hope this can be clarified in the future revise of the documentation.
>
> Cheers,
> Derrick
>
>
> On Tue, Jan 10, 2012 at 12:57 PM, Ryan Golhar  > wrote:
>
>> Derrick,
>>
>> Take a look at the section entitled "Job runner URL" on
>> http://wiki.g2.bx.psu.edu/Admin/Config/Performance/Cluster.  But as I
>> understand it, you don't need to do anything since the drmaa library is
>> part of your SGE installation.  Just make the node Galaxy runs on an exec
>> host.
>>
>> Ryan
>>
>> On Mon, Jan 9, 2012 at 8:45 PM, Derrick LIN  wrote:
>>
>>> Hi guys,
>>>
>>> I am trying to configure our test galaxy to run on our SGE cluster
>>> http://wiki.g2.bx.psu.edu/Admin/Config/Performance/Cluster.
>>>
>>> This is my first time to come across to drmaa, so a bit confused.
>>>
>>> I installed teh galaxy on a NFS location where can be accessed
>>> clsuter-wide. I have configured in universe_wsgi.ini:
>>>
>>> start_job_runners = drmaa
>>> default_cluster_job_runner = drmaa:///
>>>
>>> and export the libdrmaa path:
>>> export DRMAA_LIBRARY_PATH=/opt/sge/lib/lx-amd64/libdrmaa.so.1.0
>>>
>>> Sound like that's all. But I am not sure how galaxy can actually bind to
>>> my cluster? Anywhere I need to specify the qmaster URL etc?
>>>
>>> Or simply configure galaxy host to be a submission host of the cluster,
>>> the galaxy will fetch all information by itself?
>>>
>>> Kind regards,
>>> Derrick
>>>
>>> ___
>>> Please keep all replies on the list by using "reply all"
>>> in your mail client.  To manage your subscriptions to this
>>> and other Galaxy lists, please use the interface at:
>>>
>>>  http://lists.bx.psu.edu/
>>>
>>
>>
>
___
Please keep all replies on the list by using "reply all"
in your mail client.  To manage your subscriptions to this
and other Galaxy lists, please use the interface at:

  http://lists.bx.psu.edu/

Re: [galaxy-dev] Setup galaxy to run on SGE cluster

2012-01-09 Thread Derrick LIN
Thanks Ryan,

I have been following that guide. I can configure the galaxy host to be an
exec host, although it makes more sense to be a submission host.

I hope this can be clarified in the future revise of the documentation.

Cheers,
Derrick


On Tue, Jan 10, 2012 at 12:57 PM, Ryan Golhar
wrote:

> Derrick,
>
> Take a look at the section entitled "Job runner URL" on
> http://wiki.g2.bx.psu.edu/Admin/Config/Performance/Cluster.  But as I
> understand it, you don't need to do anything since the drmaa library is
> part of your SGE installation.  Just make the node Galaxy runs on an exec
> host.
>
> Ryan
>
> On Mon, Jan 9, 2012 at 8:45 PM, Derrick LIN  wrote:
>
>> Hi guys,
>>
>> I am trying to configure our test galaxy to run on our SGE cluster
>> http://wiki.g2.bx.psu.edu/Admin/Config/Performance/Cluster.
>>
>> This is my first time to come across to drmaa, so a bit confused.
>>
>> I installed teh galaxy on a NFS location where can be accessed
>> clsuter-wide. I have configured in universe_wsgi.ini:
>>
>> start_job_runners = drmaa
>> default_cluster_job_runner = drmaa:///
>>
>> and export the libdrmaa path:
>> export DRMAA_LIBRARY_PATH=/opt/sge/lib/lx-amd64/libdrmaa.so.1.0
>>
>> Sound like that's all. But I am not sure how galaxy can actually bind to
>> my cluster? Anywhere I need to specify the qmaster URL etc?
>>
>> Or simply configure galaxy host to be a submission host of the cluster,
>> the galaxy will fetch all information by itself?
>>
>> Kind regards,
>> Derrick
>>
>> ___
>> Please keep all replies on the list by using "reply all"
>> in your mail client.  To manage your subscriptions to this
>> and other Galaxy lists, please use the interface at:
>>
>>  http://lists.bx.psu.edu/
>>
>
>
___
Please keep all replies on the list by using "reply all"
in your mail client.  To manage your subscriptions to this
and other Galaxy lists, please use the interface at:

  http://lists.bx.psu.edu/

Re: [galaxy-dev] Setup galaxy to run on SGE cluster

2012-01-09 Thread Ryan Golhar
Derrick,

Take a look at the section entitled "Job runner URL" on
http://wiki.g2.bx.psu.edu/Admin/Config/Performance/Cluster.  But as I
understand it, you don't need to do anything since the drmaa library is
part of your SGE installation.  Just make the node Galaxy runs on an exec
host.

Ryan

On Mon, Jan 9, 2012 at 8:45 PM, Derrick LIN  wrote:

> Hi guys,
>
> I am trying to configure our test galaxy to run on our SGE cluster
> http://wiki.g2.bx.psu.edu/Admin/Config/Performance/Cluster.
>
> This is my first time to come across to drmaa, so a bit confused.
>
> I installed teh galaxy on a NFS location where can be accessed
> clsuter-wide. I have configured in universe_wsgi.ini:
>
> start_job_runners = drmaa
> default_cluster_job_runner = drmaa:///
>
> and export the libdrmaa path:
> export DRMAA_LIBRARY_PATH=/opt/sge/lib/lx-amd64/libdrmaa.so.1.0
>
> Sound like that's all. But I am not sure how galaxy can actually bind to
> my cluster? Anywhere I need to specify the qmaster URL etc?
>
> Or simply configure galaxy host to be a submission host of the cluster,
> the galaxy will fetch all information by itself?
>
> Kind regards,
> Derrick
>
> ___
> Please keep all replies on the list by using "reply all"
> in your mail client.  To manage your subscriptions to this
> and other Galaxy lists, please use the interface at:
>
>  http://lists.bx.psu.edu/
>
___
Please keep all replies on the list by using "reply all"
in your mail client.  To manage your subscriptions to this
and other Galaxy lists, please use the interface at:

  http://lists.bx.psu.edu/

[galaxy-dev] Setup galaxy to run on SGE cluster

2012-01-09 Thread Derrick LIN
Hi guys,

I am trying to configure our test galaxy to run on our SGE cluster
http://wiki.g2.bx.psu.edu/Admin/Config/Performance/Cluster.

This is my first time to come across to drmaa, so a bit confused.

I installed teh galaxy on a NFS location where can be accessed
clsuter-wide. I have configured in universe_wsgi.ini:

start_job_runners = drmaa
default_cluster_job_runner = drmaa:///

and export the libdrmaa path:
export DRMAA_LIBRARY_PATH=/opt/sge/lib/lx-amd64/libdrmaa.so.1.0

Sound like that's all. But I am not sure how galaxy can actually bind to my
cluster? Anywhere I need to specify the qmaster URL etc?

Or simply configure galaxy host to be a submission host of the cluster, the
galaxy will fetch all information by itself?

Kind regards,
Derrick
___
Please keep all replies on the list by using "reply all"
in your mail client.  To manage your subscriptions to this
and other Galaxy lists, please use the interface at:

  http://lists.bx.psu.edu/

Re: [galaxy-dev] SAM/BAM Hybrid Selection Metrics

2012-01-09 Thread Ross
Hi Ryan,

Yes, the Picard tool mandates a bizarre bait/target format file for
reasons which might best be addressed to the Picard devs - they may
have some very good reasons although I can't imagine what they are.
:)

Yes, automated conversion of any valid Galaxy bed dataset into the
strange format required by the Picard tool is a very good idea. We're
already half way there because the tool wrapper adds the (IMHO really
silly) required SAM header automagically.

A new datatype (eg "picardBaitTarget") and an automated converter
would make the tool much easier to use - it's far from ideal to force
Galaxy users to comply with the strange Picard format requirements if
we can automate a converter.

I thought about implementing one but stopped when I realized that am
not sure what an automated converter should do if the user supplies a
valid Galaxy bed lacking strand information - generally, making up
strand is not a good idea. I don't have enough insight into the way
the stats are calculated to know whether bad things might happen if
(eg) we assume all the bait and target regions are on the + strand if
they're not - but if someone can describe how to automate the
conversion, it would definitely be an improvement to the usability of
the Picard tool.

Suggestions welcomed!


On Tue, Jan 10, 2012 at 8:03 AM, Ryan Golhar
 wrote:
> Hi all - I think there is a problem with the Picard HSMetrics wrapper in
> Galaxy.  The wrapper accepts a BAM files and a BED file.  However the BED
> file isn't really in a BED format...it requires a SAM header before the BED
> lines.  This really isn't a BED file format.  I'm not quite sure how Galaxy
> should deal with this...maybe a file format specific for Picard formatted
> BED file.
>
>
> ___
> Please keep all replies on the list by using "reply all"
> in your mail client.  To manage your subscriptions to this
> and other Galaxy lists, please use the interface at:
>
>  http://lists.bx.psu.edu/



-- 
Ross Lazarus MBBS MPH;
Associate Professor, Harvard Medical School;
Head, Medical Bioinformatics, BakerIDI; Tel: +61 385321444;

___
Please keep all replies on the list by using "reply all"
in your mail client.  To manage your subscriptions to this
and other Galaxy lists, please use the interface at:

  http://lists.bx.psu.edu/


Re: [galaxy-dev] error on viewing workflow details via API

2012-01-09 Thread Dannon Baker
Glad you got it working, and thanks for the additional information.  Now that I 
know what's different about the non-working workflows I can probably recreate 
and fix this on my end.  Extracted workflows *should* function via the API.

-Dannon

On Jan 9, 2012, at 5:06 PM, Leon Mei wrote:

> Hi Dannon,
> 
> Thanks for the hint. I got it working by removing "input dataset" step and 
> adding it back in the workflow canvas. The previous workflow in question was 
> purely extracted from history.
> 
> So now I can check the details via API: 
> 
> ./display.py  
> http://galaxy.nbic.nl/galaxy/api/workflows/e037fdb493429c2a
> Member Information
> --
> url: /galaxy/api/workflows/e037fdb493429c2a
> inputs: {'8648': {'value': '', 'label': 'Input Dataset'}}
> id: e037fdb493429c2a
> name: test
> 
> I still don't understand why it is not working with an extracted workflow. 
> But I can live with this work around.
> 
> Cheers,
> Leon
> 
> 
> On Mon, Jan 9, 2012 at 3:18 PM, Dannon Baker  wrote:
> Leon,
> 
> I'm not seeing this behavior in any of my instances.  I assume they work 
> correctly through the standard galaxy interface, but is there anything 
> special about the workflows in question?  Is there perhaps a tool with no 
> inputs in use?
> 
> Thanks!
> 
> -Dannon
> 
> 
> 
> 
> On Jan 6, 2012, at 11:59 AM, Leon Mei wrote:
> 
> > Hi,
> >
> > I received an error when I was trying to view the details of existing 
> > workflows through API.
> >
> > I can see 4 workflows owned by myself successfully. But viewing the details 
> > of any workflow got me a 500 error.
> >
> > ./display.py   http://galaxy.nbic.nl/galaxy/api/workflows
> > Collection Members
> > --
> > #1: /galaxy/api/workflows/e037fdb493429c2a
> >   name: test
> >   id: e037fdb493429c2a
> > #2: /galaxy/api/workflows/13120e62d0fbb985
> >   name: imported: Partial Workflow for Student Project (1)
> >   id: 13120e62d0fbb985
> > #3: /galaxy/api/workflows/c9468fdb6dc5c5f1
> >   name: GAPSS SNP calling
> >   id: c9468fdb6dc5c5f1
> > #4: /galaxy/api/workflows/3f5830403180d620
> >   name: GAPPS CAGE
> >   id: 3f5830403180d620
> >
> > 4 element(s) in collection
> >
> > ./display.py  
> > http://galaxy.nbic.nl/galaxy/api/workflows/e037fdb493429c2a
> > HTTP Error 500: Internal Server Error
> > 500 Internal Server Error
> > The server has either erred or is incapable of performing
> > the requested operation.
> >
> > Here is the error message in paster.log:
> > /*
> > galaxy.web.framework ERROR 2012-01-06 17:46:53,349 Uncaught exception in 
> > exposed API method:
> > Traceback (most recent call last):
> >   File 
> > "/opt/galaxy/prog/galaxy-2011-12-5/lib/galaxy/web/framework/__init__.py", 
> > line 147, in decorator
> > return simplejson.dumps( func( self, trans, *args, **kwargs ) )
> >   File "/opt/galaxy/prog/galaxy-2011-12-5/lib/galaxy/web/api/workflows.py", 
> > line 68, in show
> > inputs[step.id] = {'label':step.tool_inputs['name'], 'value':""}
> > TypeError: 'NoneType' object is unsubscriptable
> >  - - [06/Jan/2012:17:46:53 +0200] "GET 
> > /galaxy/api/workflows/e037fdb493429c2a?key= HTTP/1.1" 500 - "-" 
> > "Python-urllib/2.6"
> > */
> >
> > Any suggestion on how to solve this problem?
> >
> > Thanks,
> > Leon
> >
> > --
> > Hailiang (Leon) Mei
> > Netherlands Bioinformatics Center
> > BioAssist NGS Taskforce
> >  - https://wiki.nbic.nl/index.php/Next_Generation_Sequencing
> > Skype: leon_meiMobile: +31 6 41709231
> > ___
> > Please keep all replies on the list by using "reply all"
> > in your mail client.  To manage your subscriptions to this
> > and other Galaxy lists, please use the interface at:
> >
> >  http://lists.bx.psu.edu/
> 
> 
> 
> 
> -- 
> Hailiang (Leon) Mei
> Netherlands Bioinformatics Center   
> BioAssist NGS Taskforce 
>  - https://wiki.nbic.nl/index.php/Next_Generation_Sequencing
> Skype: leon_meiMobile: +31 6 41709231

___
Please keep all replies on the list by using "reply all"
in your mail client.  To manage your subscriptions to this
and other Galaxy lists, please use the interface at:

  http://lists.bx.psu.edu/


Re: [galaxy-dev] error on viewing workflow details via API

2012-01-09 Thread Leon Mei
Hi Dannon,

Thanks for the hint. I got it working by removing "input dataset" step and
adding it back in the workflow canvas. The previous workflow in question
was purely extracted from history.

So now I can check the details via API:

./display.py 
http://galaxy.nbic.nl/galaxy/api/workflows/e037fdb493429c2a
Member Information
--
url: /galaxy/api/workflows/e037fdb493429c2a
inputs: {'8648': {'value': '', 'label': 'Input Dataset'}}
id: e037fdb493429c2a
name: test

I still don't understand why it is not working with an extracted workflow.
But I can live with this work around.

Cheers,
Leon


On Mon, Jan 9, 2012 at 3:18 PM, Dannon Baker  wrote:

> Leon,
>
> I'm not seeing this behavior in any of my instances.  I assume they work
> correctly through the standard galaxy interface, but is there anything
> special about the workflows in question?  Is there perhaps a tool with no
> inputs in use?
>
> Thanks!
>
> -Dannon
>
>
>
>
> On Jan 6, 2012, at 11:59 AM, Leon Mei wrote:
>
> > Hi,
> >
> > I received an error when I was trying to view the details of existing
> workflows through API.
> >
> > I can see 4 workflows owned by myself successfully. But viewing the
> details of any workflow got me a 500 error.
> >
> > ./display.py   http://galaxy.nbic.nl/galaxy/api/workflows
> > Collection Members
> > --
> > #1: /galaxy/api/workflows/e037fdb493429c2a
> >   name: test
> >   id: e037fdb493429c2a
> > #2: /galaxy/api/workflows/13120e62d0fbb985
> >   name: imported: Partial Workflow for Student Project (1)
> >   id: 13120e62d0fbb985
> > #3: /galaxy/api/workflows/c9468fdb6dc5c5f1
> >   name: GAPSS SNP calling
> >   id: c9468fdb6dc5c5f1
> > #4: /galaxy/api/workflows/3f5830403180d620
> >   name: GAPPS CAGE
> >   id: 3f5830403180d620
> >
> > 4 element(s) in collection
> >
> > ./display.py 
> http://galaxy.nbic.nl/galaxy/api/workflows/e037fdb493429c2a
> > HTTP Error 500: Internal Server Error
> > 500 Internal Server Error
> > The server has either erred or is incapable of performing
> > the requested operation.
> >
> > Here is the error message in paster.log:
> > /*
> > galaxy.web.framework ERROR 2012-01-06 17:46:53,349 Uncaught exception
> in exposed API method:
> > Traceback (most recent call last):
> >   File
> "/opt/galaxy/prog/galaxy-2011-12-5/lib/galaxy/web/framework/__init__.py",
> line 147, in decorator
> > return simplejson.dumps( func( self, trans, *args, **kwargs ) )
> >   File
> "/opt/galaxy/prog/galaxy-2011-12-5/lib/galaxy/web/api/workflows.py", line
> 68, in show
> > inputs[step.id] = {'label':step.tool_inputs['name'], 'value':""}
> > TypeError: 'NoneType' object is unsubscriptable
> >  - - [06/Jan/2012:17:46:53 +0200] "GET
> /galaxy/api/workflows/e037fdb493429c2a?key= HTTP/1.1" 500 - "-"
> "Python-urllib/2.6"
> > */
> >
> > Any suggestion on how to solve this problem?
> >
> > Thanks,
> > Leon
> >
> > --
> > Hailiang (Leon) Mei
> > Netherlands Bioinformatics Center
> > BioAssist NGS Taskforce
> >  - https://wiki.nbic.nl/index.php/Next_Generation_Sequencing
> > Skype: leon_meiMobile: +31 6 41709231
> > ___
> > Please keep all replies on the list by using "reply all"
> > in your mail client.  To manage your subscriptions to this
> > and other Galaxy lists, please use the interface at:
> >
> >  http://lists.bx.psu.edu/
>
>


-- 
Hailiang (Leon) Mei
Netherlands Bioinformatics Center
BioAssist NGS Taskforce
 - https://wiki.nbic.nl/index.php/Next_Generation_Sequencing
Skype: leon_meiMobile: +31 6 41709231
___
Please keep all replies on the list by using "reply all"
in your mail client.  To manage your subscriptions to this
and other Galaxy lists, please use the interface at:

  http://lists.bx.psu.edu/

[galaxy-dev] SAM/BAM Hybrid Selection Metrics

2012-01-09 Thread Ryan Golhar
Hi all - I think there is a problem with the Picard HSMetrics wrapper in
Galaxy.  The wrapper accepts a BAM files and a BED file.  However the BED
file isn't really in a BED format...it requires a SAM header before the BED
lines.  This really isn't a BED file format.  I'm not quite sure how Galaxy
should deal with this...maybe a file format specific for Picard formatted
BED file.
___
Please keep all replies on the list by using "reply all"
in your mail client.  To manage your subscriptions to this
and other Galaxy lists, please use the interface at:

  http://lists.bx.psu.edu/

Re: [galaxy-dev] Status on importing BAM file into Library does not update

2012-01-09 Thread Ryan Golhar
On Fri, Jan 6, 2012 at 12:55 PM, Ryan Golhar wrote:

>
>> This indicates that set_meta is running locally, in the runner process.
>>  Can you make sure there's not a typo in your config?  The other
>> possibility is that external metadata setting failed and it's being retried
>> internally (if that was true, you'd see messages indicated such in the
>> server log).
>>
>> I'm pretty sure there isn't a typo.  Here is anything meta related (with
> comment lines removed) in my universe_wsgi.*.ini files:
>
> [galaxy@bic galaxy-dist]$ grep set_meta *.ini
> universe_wsgi.ini:set_metadata_externally = True
> universe_wsgi.runner.ini:set_metadata_externally = True
> universe_wsgi.webapp.ini:set_metadata_externally = True
>
> [galaxy@bic galaxy-dist]$ grep meta *.ini
> universe_wsgi.ini:set_metadata_externally = True
> universe_wsgi.runner.ini:set_metadata_externally = True
> universe_wsgi.webapp.ini:set_metadata_externally = True
>
> I just tried it again on some BAM files, and nothing comes up in
> /var/log/messages or /var/log/httpd/error_log.  Runner0.log also doesn't
> show anything except for the upload job being completed.
>


I'm still trying to track this one down.  Can I add a debug output string
to show what the value of set_metadata_externally is when its read in?  If
so, where would I do this?
___
Please keep all replies on the list by using "reply all"
in your mail client.  To manage your subscriptions to this
and other Galaxy lists, please use the interface at:

  http://lists.bx.psu.edu/

Re: [galaxy-dev] Fwd: tool restrict access

2012-01-09 Thread Jeremy Goecks
Ivan,

> 
> 
>table browser
>
>#if $__user_email__ == ""
>You are not authorized to use this tool
>#else
>data_source.py $output $__app__.config.output_size_limit
>#end if
> 
> 
> 
> 
> ... and I got the error below, which seems connected to
> the cheetah syntax. Any idea of what I'm doing wrong?

As is tradition in python, you need to put semi-colons after conditionals. E.g.

--
#if $__user_email__ == "":
   You are not authorized to use this tool
#else:
   data_source.py $output $__app__.config.output_size_limit
#end if
--


J.


___
Please keep all replies on the list by using "reply all"
in your mail client.  To manage your subscriptions to this
and other Galaxy lists, please use the interface at:

  http://lists.bx.psu.edu/

Re: [galaxy-dev] Using workflows with steps with several inputs through the API

2012-01-09 Thread Carlos Borroto
On Mon, Jan 9, 2012 at 9:12 AM, Dannon Baker  wrote:
> Carlos,
>
> The method you describe below for mapping inputs is precisely the intended 
> approach.  One reason for the long identifier being used instead of a simple 
> step number is that step number (ordering as you would see in the usual run 
> workflow dialog) can change without it being obvious to the user.  The 
> ordering is based on two factors- a topological sort of the workflow graph, 
> and the distance within the editor from the top left corner.  In other words, 
> it's possible to edit a workflow in display position only (drag the boxes 
> around) and change step numbers.  One approach here would be to use the step 
> label ('label': 'Left-hand Reads') in your case and programmatically fetch 
> and use the input id.
>

Thanks for your response, is nice to get a confirmation I'm on the
right track. I'll look into using the label.

> For your last issue-- that library dataset is actually marked as deleted on 
> main, so the error is appropriate.  I'm not yet sure why that is the case, 
> but I'll look into it.  If you retry the operation on a known non-deleted 
> library dataset everything should work just fine.
>

Oh, I always try to use data from the shared libraries when testing my
workflows on main or test server, this one is from "Illumina iDEA
Datasets (sub-sampled)". I didn't realized it could have been deleted
but still show in the data library.

Thanks again,
Carlos

___
Please keep all replies on the list by using "reply all"
in your mail client.  To manage your subscriptions to this
and other Galaxy lists, please use the interface at:

  http://lists.bx.psu.edu/

Re: [galaxy-dev] (no subject)

2012-01-09 Thread Hans-Rudolf Hotz



On 01/09/2012 05:04 AM, Vidya .H.K wrote:

Hi,

I am new user to galaxy. I have installed galaxy on the Linux local server
for ChIP-seq analysis. While running MACS, an error is flagged
  "An error occurred running this job: /bin/sh: macs: not found" .
I have set the path for macs in /bin. But still the error persists. It
would be great if any one help me resolve this problem.



Hi Vidya

Is macs executable for the galaxy user (ie the user the galaxy server is 
running as) ?


Regards, Hans


Thanks in advance.


Regards
vidya



___
Please keep all replies on the list by using "reply all"
in your mail client.  To manage your subscriptions to this
and other Galaxy lists, please use the interface at:

   http://lists.bx.psu.edu/

___
Please keep all replies on the list by using "reply all"
in your mail client.  To manage your subscriptions to this
and other Galaxy lists, please use the interface at:

 http://lists.bx.psu.edu/


Re: [galaxy-dev] Fwd: tool restrict access

2012-01-09 Thread Ivan Merelli

Hi Jeremy,

thank you for your answer. I followed your hint modifyng
(for example) the  ucsc_tablebrowser.xml as follows (not
sure about the  tag to have a simple error
message, but this is another problem):


table browser

#if $__user_email__ == ""
You are not authorized to use this tool
#else
data_source.py $output $__app__.config.output_size_limit
#end if




... and I got the error below, which seems connected to
the cheetah syntax. Any idea of what I'm doing wrong?
Cheers,
I.




Traceback (most recent call last):
  File "/home/galaxy/galaxy-dist/lib/galaxy/jobs/runners/local.py", 
line 58, in run_job

job_wrapper.prepare()
  File "/home/galaxy/galaxy-dist/lib/galaxy/jobs/__init__.py", line 
411, in prepare

self.command_line = self.tool.build_command_line( param_dict )
  File "/home/galaxy/galaxy-dist/lib/galaxy/tools/__init__.py", line 
1691, in build_command_line

command_line = fill_template( self.command, context=param_dict )
  File "/home/galaxy/galaxy-dist/lib/galaxy/util/template.py", line 9, 
in fill_template

return str( Template( source=template_text, searchList=[context] ) )
  File 
"/home/galaxy/galaxy-dist/eggs/Cheetah-2.2.2-py2.5-linux-x86_64-ucs4.egg/Cheetah/Template.py", 
line 1244, in __init__

self._compile(source, file, compilerSettings=compilerSettings)
  File 
"/home/galaxy/galaxy-dist/eggs/Cheetah-2.2.2-py2.5-linux-x86_64-ucs4.egg/Cheetah/Template.py", 
line 1538, in _compile

keepRefToGeneratedCode=True)
  File 
"/home/galaxy/galaxy-dist/eggs/Cheetah-2.2.2-py2.5-linux-x86_64-ucs4.egg/Cheetah/Template.py", 
line 745, in compile

compiler.compile()
  File 
"/home/galaxy/galaxy-dist/eggs/Cheetah-2.2.2-py2.5-linux-x86_64-ucs4.egg/Cheetah/Compiler.py", 
line 1670, in compile

self._parser.parse()
  File 
"/home/galaxy/galaxy-dist/eggs/Cheetah-2.2.2-py2.5-linux-x86_64-ucs4.egg/Cheetah/Parser.py", 
line 1496, in parse

self.assertEmptyOpenDirectivesStack()
  File 
"/home/galaxy/galaxy-dist/eggs/Cheetah-2.2.2-py2.5-linux-x86_64-ucs4.egg/Cheetah/Parser.py", 
line 2658, in assertEmptyOpenDirectivesStack

raise ParseError(self, msg=errorMsg)
ParseError:

Some #directives are missing their corresponding #end ___ tag: if
Line 2, column 8

Line|Cheetah Code
|-
2   |
^



Ivan,


#if $__user_email__ == ""
You are not authorized to use this tool
#else


data_source.py $output $__app__.config.output_size_limit



To make this approach work, the email check should go in the command tag.

More information:

If you're looking to require users to login before using any tools, you
can use this flag in the universe config file:

# Force everyone to log in (disable anonymous access).
#require_login = False

If you're looking to implement tool-based access control, the best
approach is probably to use the same role-based approach that libraries use:

https://bitbucket.org/galaxy/galaxy-central/issue/269/use-galaxy-security-to-restrict-tool

Thanks,
J.


___
Please keep all replies on the list by using "reply all"
in your mail client.  To manage your subscriptions to this
and other Galaxy lists, please use the interface at:

 http://lists.bx.psu.edu/


[galaxy-dev] Python/R Script Not Integrating with Galaxy Correctly

2012-01-09 Thread Christopher Fletez-Brant
Hello Galaxy-Dev List,

I am having some issues in getting a script to work correctly in Galaxy.  My 
script is written in Python, uses RPy2 to pass some data to the R environment 
and then calls a custom R function; the output of this function, and of the 
greater Python/R tool, is a file.  I am having 2 problems: the first is that R 
cannot find the R file holding the function, and the second is that Galaxy 
cannot find the output file.  Before I elaborate on these issues, I should 
point out that when called from the command line this script executes exactly 
as I intend it to.

First issue:

When called by Galaxy, the Python portion of the script runs fine, and the RPy2 
portion devoted to setting up the R environment so that the script can run 
seems to execute just fine.  The script fails when it tells R to call the 
custom function, with the following output:

Python output:

Traceback (most recent call last):
  File 
"/Users/kipperfletez-brant/galaxy-dist/tools/myTools/svm_enh/svm_train_classify_cv.py",
 line 865, in 
if __name__=='__main__': main()
  File 
"/Users/kipperfletez-brant/galaxy-dist/tools/myTools/svm_enh/svm_train_classify_cv.py",
 line 851, in main
robjects.r.source(current)
  File 
"/opt/local/Library/Frameworks/Python.framework/Versions/2.6/lib/python2.6/site-packages/rpy2/robjects/functions.py",
 line 82, in __call__
return super(SignatureTranslatedFunction, self).__call__(*args, **kwargs)
  File 
"/opt/local/Library/Frameworks/Python.framework/Versions/2.6/lib/python2.6/site-packages/rpy2/robjects/functions.py",
 line 34, in __call__
res = super(Function, self).__call__(*new_args, **new_kwargs)
rpy2.rinterface.RRuntimeError: Error in file(file, "r", encoding = encoding) : 
  cannot open the connection

R output:

/Users/kipperfletez-brant/galaxy-dist/database/job_working_directory/106/rocprc.R
[1] "/Users/kipperfletez-brant/galaxy-dist/database/job_working_directory/106"

Error in file(file, "r", encoding = encoding) : 
  cannot open the connection
In addition: Warning message:
In file(file, "r", encoding = encoding) :
  cannot open file 
'/Users/kipperfletez-brant/galaxy-dist/database/job_working_directory/106/rocprc.R':
 No such file or directory

The first line of the R output is the return of a call to the Python function 
os.path.abspath('rocprc.R') (note that os.path.realpath('rocprc.R') returns the 
same path), while the second line, for comparison, is a call to R's getwd() - 
the R file is in the same directory as the R environment's setting, but for 
some reason R cannot find the file.  Because this works from the command line, 
I suspect that I am not taking some Galaxy setting into account.

Second issue:

Trying to identify even the kind of error, I hard-coded the directory in which 
I know the R file to reside in a call to R's setwd().  This time, the script 
runs correctly, but there is no file returned by Galaxy.  The script execution 
ends with the following output message:

empty
format: pdf, database: ?
Info: [1] "/Users/kipperfletez-brant/galaxy-dist/tools/myTools/svm_enh"

Watching the file in which the R script is held, I know that the output PDF is 
saved in the same file as that specified in my tool's XML, which, for other 
tools I've written, has resulted in an output file.  Looking at the log, the 
following DEBUG message is written:

galaxy.jobs DEBUG 2012-01-09 10:16:49,899 finish(): Could not move 
/Users/kipperfletez-brant/galaxy-dist/database/job_working_directory/111/rocprc.pdf
 to /Users/kipperfletez-brant/galaxy-dist/database/files/000/dataset_111.dat as 
directed by from_work_dir

As with the first issue, I suspect there is some sort of Galaxy setting I am 
unaware of.

Generally, because both issues relate to finding files not directly 
created/referenced by the caller Python script, I suspect that there will be 
some commonality between the resolutions of the two issues, but I am need help 
to see that.

Thank you for reading,
Kipper Fletez-Brant


___
Please keep all replies on the list by using "reply all"
in your mail client.  To manage your subscriptions to this
and other Galaxy lists, please use the interface at:

  http://lists.bx.psu.edu/


Re: [galaxy-dev] error on viewing workflow details via API

2012-01-09 Thread Dannon Baker
Leon,

I'm not seeing this behavior in any of my instances.  I assume they work 
correctly through the standard galaxy interface, but is there anything special 
about the workflows in question?  Is there perhaps a tool with no inputs in use?

Thanks!

-Dannon




On Jan 6, 2012, at 11:59 AM, Leon Mei wrote:

> Hi,
> 
> I received an error when I was trying to view the details of existing 
> workflows through API.
> 
> I can see 4 workflows owned by myself successfully. But viewing the details 
> of any workflow got me a 500 error. 
> 
> ./display.py   http://galaxy.nbic.nl/galaxy/api/workflows
> Collection Members
> --
> #1: /galaxy/api/workflows/e037fdb493429c2a
>   name: test
>   id: e037fdb493429c2a
> #2: /galaxy/api/workflows/13120e62d0fbb985
>   name: imported: Partial Workflow for Student Project (1)
>   id: 13120e62d0fbb985
> #3: /galaxy/api/workflows/c9468fdb6dc5c5f1
>   name: GAPSS SNP calling
>   id: c9468fdb6dc5c5f1
> #4: /galaxy/api/workflows/3f5830403180d620
>   name: GAPPS CAGE
>   id: 3f5830403180d620
> 
> 4 element(s) in collection
> 
> ./display.py  
> http://galaxy.nbic.nl/galaxy/api/workflows/e037fdb493429c2a
> HTTP Error 500: Internal Server Error
> 500 Internal Server Error
> The server has either erred or is incapable of performing
> the requested operation.
> 
> Here is the error message in paster.log:
> /*
> galaxy.web.framework ERROR 2012-01-06 17:46:53,349 Uncaught exception in 
> exposed API method:
> Traceback (most recent call last):
>   File 
> "/opt/galaxy/prog/galaxy-2011-12-5/lib/galaxy/web/framework/__init__.py", 
> line 147, in decorator
> return simplejson.dumps( func( self, trans, *args, **kwargs ) )
>   File "/opt/galaxy/prog/galaxy-2011-12-5/lib/galaxy/web/api/workflows.py", 
> line 68, in show
> inputs[step.id] = {'label':step.tool_inputs['name'], 'value':""}
> TypeError: 'NoneType' object is unsubscriptable
>  - - [06/Jan/2012:17:46:53 +0200] "GET 
> /galaxy/api/workflows/e037fdb493429c2a?key= HTTP/1.1" 500 - "-" 
> "Python-urllib/2.6"
> */
> 
> Any suggestion on how to solve this problem?
> 
> Thanks,
> Leon
> 
> -- 
> Hailiang (Leon) Mei
> Netherlands Bioinformatics Center   
> BioAssist NGS Taskforce 
>  - https://wiki.nbic.nl/index.php/Next_Generation_Sequencing
> Skype: leon_meiMobile: +31 6 41709231
> ___
> Please keep all replies on the list by using "reply all"
> in your mail client.  To manage your subscriptions to this
> and other Galaxy lists, please use the interface at:
> 
>  http://lists.bx.psu.edu/

___
Please keep all replies on the list by using "reply all"
in your mail client.  To manage your subscriptions to this
and other Galaxy lists, please use the interface at:

  http://lists.bx.psu.edu/


Re: [galaxy-dev] Using workflows with steps with several inputs through the API

2012-01-09 Thread Dannon Baker
Carlos,

The method you describe below for mapping inputs is precisely the intended 
approach.  One reason for the long identifier being used instead of a simple 
step number is that step number (ordering as you would see in the usual run 
workflow dialog) can change without it being obvious to the user.  The ordering 
is based on two factors- a topological sort of the workflow graph, and the 
distance within the editor from the top left corner.  In other words, it's 
possible to edit a workflow in display position only (drag the boxes around) 
and change step numbers.  One approach here would be to use the step label 
('label': 'Left-hand Reads') in your case and programmatically fetch and use 
the input id.

For your last issue-- that library dataset is actually marked as deleted on 
main, so the error is appropriate.  I'm not yet sure why that is the case, but 
I'll look into it.  If you retry the operation on a known non-deleted library 
dataset everything should work just fine.

Thanks!

Dannon

On Jan 5, 2012, at 5:35 PM, Carlos Borroto wrote:

> Hi,
> 
> Looking at workflow_execute.py, it seems to me there can be only one
> input for step, is this correct? or is there a wait to map more than
> one input and match them to an specific position in a  particularly
> step?
> 
> This is the workaround I found.
> 
> I created 'Input Dataset' boxes for all of my inputs, then I linked
> them to the step where they are needed. This is actually very
> convenient if the same input is used in more than one step, as when I
> run the workflow manually I only need to select each input once. I
> tried doing this, as each 'Input Dataset' would have his own step.
> 
> This is an example workflow:
> http://main.g2.bx.psu.edu/u/cjav/w/fastq-joiner---input-datasets
> 
> Initially I ran into this issue:
> $ ./workflow_execute.py my_key http://main.g2.bx.psu.edu/api/workflows
> fcfa9ae643d37c32 'Fastq joiner - Input Datasets'
> '1=ldda=b4c5be159c8057c2' '2=ldda=ea0e32961d454539'
> HTTP Error 400: Bad Request
> "Workflow cannot be run because an expected input step '517620' has no
> input dataset."
> 
> I then found you can do:
> $ ./display.py my_key http://main.g2.bx.psu.edu/api/workflows/fcfa9ae643d37c32
> Member Information
> --
> url: /api/workflows/fcfa9ae643d37c32
> inputs: {'517620': {'value': '', 'label': 'Left-hand Reads'},
> '517622': {'value': '', 'label': 'Right-hand Reads'}}
> id: fcfa9ae643d37c32
> name: Fastq joiner - Input Datasets
> 
> 
> I see inputs get different keys and I can use them when executing the 
> workflow:
> $ ./workflow_execute.py my_key http://main.g2.bx.psu.edu/api/workflows
> fcfa9ae643d37c32 'hist_id=9fbfd3a1be8b7f8e'
> '517620=hda=8d89405199368bb0' '517622=hda=5c93f30078e5ac25'
> Response
> 
> {'outputs': ['f7c4b0900d8f3b1b'], 'history': '9fbfd3a1be8b7f8e'}
> 
> This worked perfectly. I still think it would be easier to be able to
> use the step number, but this works.
> 
> 
> Lastly, I did run into one issue where I had to import the datasets
> firsts, as trying to use ldda like this:
> $ ./workflow_execute.py my_key http://main.g2.bx.psu.edu/api/workflows
> fcfa9ae643d37c32 'Fastq joiner - Input Datasets'
> '517620=ldda=b4c5be159c8057c2' '517622=ldda=ea0e32961d454539'
> Response
> 
> {'outputs': ['44c033fad737acc5'], 'history': '9fbfd3a1be8b7f8e'}
> 
> I got this error in the history:
> input data 1 (file:
> /galaxy/main_database/files/002/780/dataset_2780871.dat) was deleted
> before the job started
> 
> This didn't happen in our local instance using galaxy-central.
> 
> Thanks,
> Carlos
> ___
> Please keep all replies on the list by using "reply all"
> in your mail client.  To manage your subscriptions to this
> and other Galaxy lists, please use the interface at:
> 
>  http://lists.bx.psu.edu/

___
Please keep all replies on the list by using "reply all"
in your mail client.  To manage your subscriptions to this
and other Galaxy lists, please use the interface at:

  http://lists.bx.psu.edu/


Re: [galaxy-dev] broken pipe when running VCF-merge.

2012-01-09 Thread James Boocock

Hi,

We are attempting to integrate vcftools specifically vcf-merge into 
galaxy. We have two or files that we
are attempting to merge, the command when run through terminal does not 
give a broken pipe, but
when galaxy runs the command a broken pipe is given at the beginning and 
the start of the execution.



The output from the merge appears to be exactly the same. But the broken 
pipe error causes galaxy
to go read on the output wondering if anyone can help that would be much 
appreciated.



The shell script  is pasted below.


_SHELL SCRIPT THAT GALAXY RUNS
_
  1 #!/bin/bash
  2 # @Date 9/01/2012
  3 # @Author Ed Hills, James Boocock
  4 #
  5 # This file will take a white space separated list of file names,
  6 # and run the vcf-merge tool and merge all files into a new file.
  7 #
  8 # INPUTS
  9 # $1 = First input File
 10 # $2 = Second input File
 11 # $N = Extra input files
 12
 13 FILE_LIST=""
 14
 15 cat $1 | bgzip -c > ~tmp1.vcf.gz
 16 cat $2 | bgzip -c > ~tmp2.vcf.gz
 17
 18 tabix -p vcf ~tmp1.vcf.gz
 19 tabix -p vcf ~tmp2.vcf.gz
 20
 21 FILE_LIST="~tmp1.vcf.gz ~tmp2.vcf.gz"
 22
 23 if [ $# > 2 ]
 24 then
 25 for ((i=3; i <= $#; i++))
 26 do
 27 eval EXTRA_FILE=\$${i}
 28 cat $EXTRA_FILE | bgzip -c > ~tmp${i}.vcf.gz
 29 tabix -p vcf ~tmp${i}.vcf.gz
 30 FILE_LIST="${FILE_LIST} ~tmp${i}.vcf.gz"
 31 done
 32 fi
 33
 34 perl ~/galaxy-dist/tool-data/shared/vcfperltools/vcf-merge 
${FILE_LIST} 2> /dev/null

 35
 36 rm -f ~tmp*
 37


_ERROR MESSAGE_

Warning: The AC tag(s) will not be merged correctly for multiallelic sites.
   To be handled correctly, please redefine with Number=. or set the environment
   variable DONT_FIX_VCF40_AG_TAGS=0.

gzip: stdout: Broken pipe
Using column name 'HG00096' for ~tmp1.vcf.gz:HG00096
gzip: stdout: Broken pipe
Using column name 'NA06986' for ~tmp2.vcf.gz:NA06986



Cheers James.
___
Please keep all replies on the list by using "reply all"
in your mail client.  To manage your subscriptions to this
and other Galaxy lists, please use the interface at:

  http://lists.bx.psu.edu/

[galaxy-dev] (no subject)

2012-01-09 Thread Vidya .H.K
Hi,

I am new user to galaxy. I have installed galaxy on the Linux local server
for ChIP-seq analysis. While running MACS, an error is flagged
 "An error occurred running this job: /bin/sh: macs: not found" .
I have set the path for macs in /bin. But still the error persists. It
would be great if any one help me resolve this problem.

Thanks in advance.


Regards
vidya



___
Please keep all replies on the list by using "reply all"
in your mail client.  To manage your subscriptions to this
and other Galaxy lists, please use the interface at:

  http://lists.bx.psu.edu/