[galaxy-dev] ulimit problems after update

2014-11-10 Thread Ido Tamir
Hi,
I updated to the latest galaxy distribution (after one year). And now
every job fails with:
/home/imba/solexa/.profile.sh: line 118: ulimit: virtual memory: cannot modify 
limit: Operation not permitted

The limit is ridiculously high:
 ulimit -v 6000

Its just to prevent some badly programmed in house galaxy tools to crash the 
server.
I think the problem happens after the set_metadata stage.

Any advice?

thank you very much,
ido
___
Please keep all replies on the list by using reply all
in your mail client.  To manage your subscriptions to this
and other Galaxy lists, please use the interface at:
  http://lists.bx.psu.edu/

To search Galaxy mailing lists use the unified search at:
  http://galaxyproject.org/search/mailinglists/


Re: [galaxy-dev] Jobs stuck in new state - Data Library datasets to blame?

2014-11-10 Thread Ido Tamir
did you check if the metadata on the input was set correctly/at all?
This happens sometimes in our galaxy instance. Metadata is not set correctly, 
and jobs run after metadata is set by hand.
No reupload necessary.

best,
ido

On 06 Nov 2014, at 17:13, Lance Parsons lpars...@princeton.edu wrote:

 I'v run into this same issue again (just with some other Data Library 
 datasets).  This time, there are a few users involved with quite a few 
 stuck jobs.  Does anyone have any advice on pushing these jobs through?  
 Maybe even a pointer to the relevant code?  I'm running latest_2014.08.11.  
 Thanks in advance.
 
 Lance
 
 Lance Parsons wrote:
 Thanks, that was the first thing I checked.  However, restarting the handler 
 didn't help.  Downloading the offending data and re-uploading as a new data 
 set and then rerunning using the new dataset as input did work.  Also, all 
 other jobs continued to run fine.
 
 Lance
 
 Kandalaft, Iyad wrote:
 I’ve had jobs get stuck in the new state when one of the handler servers 
 crashes.  If you have dedicated handlers, check to make sure they are still 
 running.
 Restart the handler to see if the jobs get resumed automatically.
  
  
  
 Iyad Kandalaft
 
  
 From: galaxy-dev-boun...@lists.bx.psu.edu 
 [mailto:galaxy-dev-boun...@lists.bx.psu.edu] On Behalf Of Aaron Petkau
 Sent: Wednesday, October 01, 2014 5:32 PM
 To: Lance Parsons
 Cc: galaxy-dev@lists.bx.psu.edu
 Subject: Re: [galaxy-dev] Jobs stuck in new state - Data Library datasets 
 to blame?
  
 Are you attempting to upload datasets to a Data Library, and then copy to a 
 history and run jobs on them right away?  I've run into issues before where 
 if I attempt to run a job on a dataset in a library before it is finished 
 being uploaded and processed, then the job gets stuck in a queued state and 
 never executes.
 
 Aaron
  
 On Wed, Oct 1, 2014 at 2:51 PM, Lance Parsons lpars...@princeton.edu 
 wrote:
 Recently, I updated our Galaxy instance to use two processes (one for web, 
 the other as a job handler).  This has been working well, except in a few 
 cases.  I've noticed that a number of jobs get stuck in the new status.
 
 In a number of cases, I've resolved the issue by downloading and uploading 
 one of the input files and rerunning the job using the newly uploaded file. 
  In at least one of these cases, the offending input file was one that was 
 copied from a Data Library.
 
 Can anyone point me to something to look for in the database, etc. that 
 would cause a job to think a dataset was not ready for use as a job input?  
 I'd very much like to fix these datasets since having to re-upload data 
 libraries would be very tedious.
 
 Thanks in advance.
 
 -- 
 Lance Parsons - Scientific Programmer
 134 Carl C. Icahn Laboratory
 Lewis-Sigler Institute for Integrative Genomics
 Princeton University
 
 ___
 Please keep all replies on the list by using reply all
 in your mail client.  To manage your subscriptions to this
 and other Galaxy lists, please use the interface at:
  http://lists.bx.psu.edu/
 
 To search Galaxy mailing lists use the unified search at:
  http://galaxyproject.org/search/mailinglists/
  
 
 -- 
 Lance Parsons - Scientific Programmer
 134 Carl C. Icahn Laboratory
 Lewis-Sigler Institute for Integrative Genomics
 Princeton University
 
 
 -- 
 Lance Parsons - Scientific Programmer
 134 Carl C. Icahn Laboratory
 Lewis-Sigler Institute for Integrative Genomics
 Princeton University
 
 ___
 Please keep all replies on the list by using reply all
 in your mail client.  To manage your subscriptions to this
 and other Galaxy lists, please use the interface at:
  http://lists.bx.psu.edu/
 
 To search Galaxy mailing lists use the unified search at:
  http://galaxyproject.org/search/mailinglists/


___
Please keep all replies on the list by using reply all
in your mail client.  To manage your subscriptions to this
and other Galaxy lists, please use the interface at:
  http://lists.bx.psu.edu/

To search Galaxy mailing lists use the unified search at:
  http://galaxyproject.org/search/mailinglists/


Re: [galaxy-dev] A proposed modules extension for toolshed wrappers

2014-04-25 Thread Ido Tamir
Hi,
has anything changed in galaxy in this regard?
Any way to modify an environment before a tools is run?

I now have a tool relying on R-devel and bioconductor devel, both of which I 
can load in a module.
The tool comes from the toolshed with xml like:

command interpreter=Rscript 
…


I don't want to hack around in the tool itself, but simply load the necessary 
R-version.

thank you very much,
ido


On Sep 13, 2013, at 3:23 AM, Guest, Simon simon.gu...@agresearch.co.nz 
wrote:

 Just been reading a bit more about the Galaxy packaging system.  Here's a 
 slight modification to what I was suggesting that might fit in a bit better.  
 Apologies for not being more familiar with the existing system before 
 proposing extensions.
 
 Recall that my goal is to support using a system-installed (native) package, 
 at a defined version, which I aim to achieve by loading the appropriate 
 environment module before running a tool.
 
 We still have tool_dependencies.xml defining a package at a particular 
 version, but rather than download and build the source code, there's just a 
 directive that says how to pick up the correct program version at runtime, 
 e.g. which environment module to load.
 
 So instead of the tool_dependencies.xml fragment:
 tool_dependency
package name=bwa version=0.6.2
install version=1.0
actions
action 
 type=download_by_urlhttp://downloads.sourceforge.net/project/bio-bwa/bwa-0.6.2.tar.bz2/action
action type=shell_commandmake/action
action type=move_file
sourcebwa/source
destination$INSTALL_DIR/bin/destination
/action
action type=set_environment
environment_variable name=PATH 
 action=prepend_to$INSTALL_DIR/bin/environment_variable
/action
/actions
/install
/package
 /tool_dependency
 
 We have something like this (NB: element and attribute names are for 
 illustrative purposes only):
 
 tool_dependency
package name=bwa version=0.6.2
use_native
actions
action type=module_loadbwa/0.6.2/action
/actions
/use_native
/package
 /tool_dependency
 
 This causes the right thing (module load bwa/0.6.2) to be stuck into the 
 dependencies env.sh file when this package is installed from the toolshed.  
 We could call this toolshed tool native_package_bwa_0_6_2, to avoid confusion 
 with the existing download-and-make one.
 
 We might want a bit of flexibility on what actions are supported (in case we 
 want to support Software Collections, for example).
 
 What do you think?
 
 cheers,
 Simon
 
 PS: In case it wasn't already clear, solving this problem well is quite 
 important to us here at AgResearch.  ;-)
 
 
 ===
 Attention: The information contained in this message and/or attachments
 from AgResearch Limited is intended only for the persons or entities
 to which it is addressed and may contain confidential and/or privileged
 material. Any review, retransmission, dissemination or other use of, or
 taking of any action in reliance upon, this information by persons or
 entities other than the intended recipients is prohibited by AgResearch
 Limited. If you have received this message in error, please notify the
 sender immediately.
 ===
 
 ___
 Please keep all replies on the list by using reply all
 in your mail client.  To manage your subscriptions to this
 and other Galaxy lists, please use the interface at:
  http://lists.bx.psu.edu/
 
 To search Galaxy mailing lists use the unified search at:
  http://galaxyproject.org/search/mailinglists/


___
Please keep all replies on the list by using reply all
in your mail client.  To manage your subscriptions to this
and other Galaxy lists, please use the interface at:
  http://lists.bx.psu.edu/

To search Galaxy mailing lists use the unified search at:
  http://galaxyproject.org/search/mailinglists/


Re: [galaxy-dev] A proposed modules extension for toolshed wrappers

2014-04-25 Thread Ido Tamir
Dear  Björn,

maybe I could change the tool, which is not mine, and which I don't want to 
maintain to use a specific R-version that is already available on our cluster 
and
which I can put into my path with module load R/3.1.0-devel ( 
http://modules.sourceforge.net/)

there even is a requirements in the wrapper which is not fulfilled (these 
versions are not available at the moment) and still it installed without 
problems.
  requirements
requirement type=R-module version=3.5.27edgeR/requirement
requirement type=R-module version=3.18.13limma/requirement
  /requirements

This requirements tags looks also rather inflexible to me. With an additional 
level of user configurable indirection it would be possible to make the tools
fit to different infrastructures without having to use binaries provided by 
somebody else, taking up space for just one tool etc...

Currently for jobs that are not run locally there is a general 
environment_setup_file. There could be one optional environment_setup_file for 
every job (or destination).

In the end I created yet another wrapper.

best,
ido


On Apr 25, 2014, at 2:34 PM, Björn Grüning bjoern.gruen...@gmail.com wrote:

 Hi Ido,
 
 I do not get your question in all detail, but it is possible to define a 
 tool_dependencies.xml with a specific R version, + R libraries and use only 
 that specific version from your tool with requirement tags.
 
 For an example please see various R tools from:
 
 https://github.com/bgruening/galaxytools
 
 Cheers,
 Bjoern
 
 Am 25.04.2014 14:25, schrieb Ido Tamir:
 Hi,
 has anything changed in galaxy in this regard?
 Any way to modify an environment before a tools is run?
 
 I now have a tool relying on R-devel and bioconductor devel, both of which I 
 can load in a module.
 The tool comes from the toolshed with xml like:
 
 command interpreter=Rscript
 …
 
 
 I don't want to hack around in the tool itself, but simply load the 
 necessary R-version.
 
 thank you very much,
 ido
 
 
 On Sep 13, 2013, at 3:23 AM, Guest, Simon simon.gu...@agresearch.co.nz 
 wrote:
 
 Just been reading a bit more about the Galaxy packaging system.  Here's a 
 slight modification to what I was suggesting that might fit in a bit 
 better.  Apologies for not being more familiar with the existing system 
 before proposing extensions.
 
 Recall that my goal is to support using a system-installed (native) 
 package, at a defined version, which I aim to achieve by loading the 
 appropriate environment module before running a tool.
 
 We still have tool_dependencies.xml defining a package at a particular 
 version, but rather than download and build the source code, there's just a 
 directive that says how to pick up the correct program version at runtime, 
 e.g. which environment module to load.
 
 So instead of the tool_dependencies.xml fragment:
 tool_dependency
package name=bwa version=0.6.2
install version=1.0
actions
action 
 type=download_by_urlhttp://downloads.sourceforge.net/project/bio-bwa/bwa-0.6.2.tar.bz2/action
action type=shell_commandmake/action
action type=move_file
sourcebwa/source
destination$INSTALL_DIR/bin/destination
/action
action type=set_environment
environment_variable name=PATH 
 action=prepend_to$INSTALL_DIR/bin/environment_variable
/action
/actions
/install
/package
 /tool_dependency
 
 We have something like this (NB: element and attribute names are for 
 illustrative purposes only):
 
 tool_dependency
package name=bwa version=0.6.2
use_native
actions
action type=module_loadbwa/0.6.2/action
/actions
/use_native
/package
 /tool_dependency
 
 This causes the right thing (module load bwa/0.6.2) to be stuck into the 
 dependencies env.sh file when this package is installed from the toolshed.  
 We could call this toolshed tool native_package_bwa_0_6_2, to avoid 
 confusion with the existing download-and-make one.
 
 We might want a bit of flexibility on what actions are supported (in case 
 we want to support Software Collections, for example).
 
 What do you think?
 
 cheers,
 Simon
 
 PS: In case it wasn't already clear, solving this problem well is quite 
 important to us here at AgResearch.  ;-)
 
 
 ===
 Attention: The information contained in this message and/or attachments
 from AgResearch Limited is intended only for the persons or entities
 to which it is addressed and may contain confidential and/or privileged
 material. Any review, retransmission, dissemination or other use of, or
 taking of any action in reliance upon, this information by persons or
 entities other than the intended recipients is prohibited by AgResearch
 Limited. If you have received this message in error, please notify

Re: [galaxy-dev] A proposed modules extension for toolshed wrappers

2014-04-25 Thread Ido Tamir

On Apr 25, 2014, at 3:55 PM, John Chilton jmchil...@gmail.com wrote:

 Additionally (perhaps more pertinently), I have an open pull request
 (created just this week) to allow setting environment variables on a
 per destination basis - this is sort of what you are getting at and
 might be a better way to go -
 https://bitbucket.org/galaxy/galaxy-central/pull-request/378/allow-specification-of-environment.
 With this modification you could hack up the path to reflect the
 changes made by the module load. That is a bit of hack but would work
 for this case. That said I think you make a good point about the
 environment file, I will modify the new env tags to allow a file
 attribute and just source that file. This will give you your per
 destination environment files - hopefully this is satisfactory.
 
 

That sounds great!

best,
ido

 
 
 On Fri, Apr 25, 2014 at 8:22 AM, Ido Tamir ta...@imp.ac.at wrote:
 Dear  Björn,
 
 maybe I could change the tool, which is not mine, and which I don't want to 
 maintain to use a specific R-version that is already available on our 
 cluster and
 which I can put into my path with module load R/3.1.0-devel ( 
 http://modules.sourceforge.net/)
 
 there even is a requirements in the wrapper which is not fulfilled (these 
 versions are not available at the moment) and still it installed without 
 problems.
  requirements
requirement type=R-module version=3.5.27edgeR/requirement
requirement type=R-module version=3.18.13limma/requirement
  /requirements
 
 This requirements tags looks also rather inflexible to me. With an 
 additional level of user configurable indirection it would be possible to 
 make the tools
 fit to different infrastructures without having to use binaries provided by 
 somebody else, taking up space for just one tool etc...
 
 Currently for jobs that are not run locally there is a general 
 environment_setup_file. There could be one optional environment_setup_file 
 for every job (or destination).
 
 In the end I created yet another wrapper.
 
 best,
 ido
 
 
 On Apr 25, 2014, at 2:34 PM, Björn Grüning bjoern.gruen...@gmail.com wrote:
 
 Hi Ido,
 
 I do not get your question in all detail, but it is possible to define a 
 tool_dependencies.xml with a specific R version, + R libraries and use only 
 that specific version from your tool with requirement tags.
 
 For an example please see various R tools from:
 
 https://github.com/bgruening/galaxytools
 
 Cheers,
 Bjoern
 
 Am 25.04.2014 14:25, schrieb Ido Tamir:
 Hi,
 has anything changed in galaxy in this regard?
 Any way to modify an environment before a tools is run?
 
 I now have a tool relying on R-devel and bioconductor devel, both of which 
 I can load in a module.
 The tool comes from the toolshed with xml like:
 
 command interpreter=Rscript
 …
 
 
 I don't want to hack around in the tool itself, but simply load the 
 necessary R-version.
 
 thank you very much,
 ido
 
 
 On Sep 13, 2013, at 3:23 AM, Guest, Simon simon.gu...@agresearch.co.nz 
 wrote:
 
 Just been reading a bit more about the Galaxy packaging system.  Here's a 
 slight modification to what I was suggesting that might fit in a bit 
 better.  Apologies for not being more familiar with the existing system 
 before proposing extensions.
 
 Recall that my goal is to support using a system-installed (native) 
 package, at a defined version, which I aim to achieve by loading the 
 appropriate environment module before running a tool.
 
 We still have tool_dependencies.xml defining a package at a particular 
 version, but rather than download and build the source code, there's just 
 a directive that says how to pick up the correct program version at 
 runtime, e.g. which environment module to load.
 
 So instead of the tool_dependencies.xml fragment:
 tool_dependency
   package name=bwa version=0.6.2
   install version=1.0
   actions
   action 
 type=download_by_urlhttp://downloads.sourceforge.net/project/bio-bwa/bwa-0.6.2.tar.bz2/action
   action type=shell_commandmake/action
   action type=move_file
   sourcebwa/source
   destination$INSTALL_DIR/bin/destination
   /action
   action type=set_environment
   environment_variable name=PATH 
 action=prepend_to$INSTALL_DIR/bin/environment_variable
   /action
   /actions
   /install
   /package
 /tool_dependency
 
 We have something like this (NB: element and attribute names are for 
 illustrative purposes only):
 
 tool_dependency
   package name=bwa version=0.6.2
   use_native
   actions
   action type=module_loadbwa/0.6.2/action
   /actions
   /use_native
   /package
 /tool_dependency
 
 This causes the right thing (module load bwa/0.6.2) to be stuck into the 
 dependencies env.sh file when this package is installed from the 
 toolshed.  We could call this toolshed tool native_package_bwa_0_6_2, to 
 avoid

[galaxy-dev] select both chromosome size and mappable genome size

2014-04-17 Thread Ido Tamir
Hi,
my tool needs both a file with chromosome lengths and the total mappable genome 
size.
I would like to hardcode the mappable genome size per build somewhere.
However the user should have to use only one drop down to select both. How can 
I achieve this?

e.g. currently:
param name=FAI type=select label=Select chromsome sizes 
   options from_file=chrsize.loc
column name=name index=0/
column name=value index=1/
   /options
 /param

with chrsize.loc:
mm9 /groups/csf-ngs/wfsys/galaxy/tool-data/shared/ucsc/chrom/mm9.len
mm10/groups/csf-ngs/wfsys/galaxy/tool-data/shared/ucsc/chrom/mm10.len
hg19/groups/csf-ngs/wfsys/galaxy/tool-data/shared/ucsc/chrom/hg19.len
dm3 /groups/csf-ngs/wfsys/galaxy/tool-data/shared/ucsc/chrom/dm3.len

Currently I multiply the genome size from the fai file by 0.8 to get the 
mappable size within my tool,
but I would like to use more exact numbers per build.

thank you very much,
ido


___
Please keep all replies on the list by using reply all
in your mail client.  To manage your subscriptions to this
and other Galaxy lists, please use the interface at:
  http://lists.bx.psu.edu/

To search Galaxy mailing lists use the unified search at:
  http://galaxyproject.org/search/mailinglists/


Re: [galaxy-dev] jobs stuck in new state

2014-03-31 Thread Ido Tamir
In my case it was uncompleted metadata in one of the input files.
(but maybe it was not new state but something else?)

HTH,
ido

On Mar 26, 2014, at 5:25 PM, David Hoover hoove...@helix.nih.gov wrote:

 I have many jobs stuck in the 'new' state on our local Galaxy instance.  The 
 jobs can't be stopped using the Admin-Manage jobs tool.  First, does anyone 
 know why a job would get stuck in the 'new' state for weeks?  I have cleaned 
 things up by manually setting their states to 'error' in the MySQL database.  
 Is there a better way of dealing with 'new' jobs?
 
 BTW, our Galaxy instance was updated about two weeks ago.
 
 Wondering,
 David Hoover
 Helix Systems Staff
 ___
 Please keep all replies on the list by using reply all
 in your mail client.  To manage your subscriptions to this
 and other Galaxy lists, please use the interface at:
  http://lists.bx.psu.edu/
 
 To search Galaxy mailing lists use the unified search at:
  http://galaxyproject.org/search/mailinglists/


___
Please keep all replies on the list by using reply all
in your mail client.  To manage your subscriptions to this
and other Galaxy lists, please use the interface at:
  http://lists.bx.psu.edu/

To search Galaxy mailing lists use the unified search at:
  http://galaxyproject.org/search/mailinglists/


Re: [galaxy-dev] universe_wsgi.ini question for Cistrome

2014-02-28 Thread Ido Tamir
I have CEAS as a tool. I wrapped the standalone CEAS version. I tried adding a 
binary datatype for the annotation db but even with the binary datatype upload 
somehow did not work (IIRC galaxy tried to convert it to ascii).
I then put the database as a selectable parameter in a dropdown list, because 
users anyway would not create and upload their own annotation db.

best,
ido

On Feb 27, 2014, at 7:13 AM, ruiwang.sz ruiwang...@gmail.com wrote:

 Hi Guys,
 
 Are there anyone who is using Cistrome?
 
 I tried to merge Cistrome's setting into our own galaxy instance. A new entry 
 is
 
 # Path to the static library files for assembly, ceaslib, chromLen, 
 conservation, liftOver and MAT-lib
 # Then in other tool configuration xml file, we can use 'from galaxy import 
 config' then '$config.Configuration().cistrome_static_library_path'
 # default would be in tool-data/ folder
 cistrome_static_library_path = 
 /home/bioinfoadmin/app/cistrome-apps/cistrome_library
 
 However, I got error when I tried CEAS: Enrichment on chromosome and 
 annotation:
  
 *
 Dataset 23: CEAS: Enrichment on chromosome and annotation on data 16 and data 
 13
 The Galaxy framework encountered the following error while attempting to run 
 the tool:
  
 Traceback (most recent call last):
   File 
 /home/bioinfoadmin/app/galaxy-dist/lib/galaxy/jobs/runners/__init__.py, 
 line 121, in prepare_job
 job_wrapper.prepare()
   File /home/bioinfoadmin/app/galaxy-dist/lib/galaxy/jobs/__init__.py, line 
 707, in prepare
 config_filenames = self.tool.build_config_files( param_dict, 
 self.working_directory )
   File /home/bioinfoadmin/app/galaxy-dist/lib/galaxy/tools/__init__.py, 
 line 2609, in build_config_files
 f.write( fill_template( template_text, context=param_dict ) )
   File /home/bioinfoadmin/app/galaxy-dist/lib/galaxy/util/template.py, line 
 9, in fill_template
 return str( Template( source=template_text, searchList=[context] ) )
   File 
 /home/bioinfoadmin/app/galaxy-dist/eggs/Cheetah-2.2.2-py2.7-linux-x86_64-ucs4.egg/Cheetah/Template.py,
  line 1004, in __str__
 return getattr(self, mainMethName)()
   File cheetah_DynamicallyCompiledCheetahTemplate_1391037616_15_43258.py, 
 line 274, in respond
 NotFound: cannot find 'cistrome_static_library_path' while searching for 
 '__app__.config.cistrome_static_library_path'
  
 Tool execution generated the following error message:
  
 failure preparing job
  
 
 
 I'm quite confused since '__app__.config.cistrome_static_library_path' should 
 be defined? or right now __app__ is dropped
 as Galaxy evolves? Besides, what does the comment mean by
 
 # Then in other tool configuration xml file, we can use 'from galaxy import 
 config' then '$config.Configuration().cistrome_static_library_path'
 
 Does that mean that we should not be using __app__.config...but how to use 
 this? I tried to insert 
 
 'from galaxy import config' 
 
 into the xml file but only got error...
 
 I'll appreciate any input.
 
 Thanks,
 Rui
 
 
 ___
 Please keep all replies on the list by using reply all
 in your mail client.  To manage your subscriptions to this
 and other Galaxy lists, please use the interface at:
  http://lists.bx.psu.edu/
 
 To search Galaxy mailing lists use the unified search at:
  http://galaxyproject.org/search/mailinglists/


___
Please keep all replies on the list by using reply all
in your mail client.  To manage your subscriptions to this
and other Galaxy lists, please use the interface at:
  http://lists.bx.psu.edu/

To search Galaxy mailing lists use the unified search at:
  http://galaxyproject.org/search/mailinglists/


Re: [galaxy-dev] Security vulnerability in Galaxy filtering tools

2013-11-05 Thread Ido Tamir
This seems to happen often e.g. 
http://wiki.galaxyproject.org/DevNewsBriefs/2012_10_23#Compute_Tool_Security_Fix

a) are there general guidelines in the wiki on how to avoid these problems when 
creating tools?
b) is there a way to check automatically if all input fields are correctly 
escaped in a tool?

A search for security in the wiki brings up:
• Admin/Data Libraries/Library Security
0.0k - rev: 1 (current) last modified: 2013-01-02 23:54:33
• Admin/DataLibraries/LibrarySecurity
19.2k - rev: 4 (current) last modified: 2013-01-03 00:12:36
• HelpOnConfiguration/SecurityPolicy
1.9k - rev: 1 (current) last modified: 0
• Learn/Security Features
7.0k - rev: 3 (current) last modified: 2011-09-13 16:52:08
• News/2013_04_08_Galaxy_Security_Release
1.3k - rev: 3 (current) last modified: 2013-04-08 16:56:41

escape does not bring up anything.

thank you very much,
ido


On Nov 5, 2013, at 12:45 AM, Nate Coraor n...@bx.psu.edu wrote:

 A security vulnerability was recently discovered by John Chilton with 
 Galaxy's Filter data on any column using simple expressions and Filter on 
 ambiguities in polymorphism datasets tools that can allow for arbitrary 
 execution of code on the command line.
 
 The fix for these tools has been committed to the Galaxy source.  The timing 
 of this commit coincides with the next Galaxy stable release (which has also 
 been pushed out today).
 
 To apply the fix and simultaneously update to the new Galaxy stable release, 
 ensure you are on the stable branch and upgrade to the latest changeset:
 
 % hg branch
 stable
 
 % hg pull -u
 
 For Galaxy installations that administrators are not yet ready to upgrade to 
 the latest release, there are three workarounds.
 
 First, for Galaxy installations running on a relatively new version of the 
 stable release (e.g. release_2013.08.12), Galaxy can be updated to the 
 specific changeset that that contains the fix.  This will include all of the 
 stable (non-feature) commits that have been accumulated since the 8/12 
 release plus any new features included with (and prior to) the 8/12 release, 
 but without all of the new features included in the 11/4 release.  Ensure you 
 are on the stable branch and then upgrade to the specific changeset:
 
 % hg pull -u -r e094c73fed4d
 
 Second, the patch can be downloaded and applied manually:
 
 % wget -o security.patch 
 https://bitbucket.org/galaxy/galaxy-central/commits/e094c73fed4dc66b589932edb83412cb8b827cd3raw/
 
 and then:
 
 % hg patch security.patch
 
 or:
 
 % patch -p1  security.patch
 
 Third, the tools can be completely disabled by removing them from the tool 
 configuration file (by default, tool_conf.xml) and restarting all Galaxy 
 server processes.  The relevant lines in tool_conf.xml are:
 
   tool file=stats/dna_filtering.xml /
   tool file=stats/filtering.xml /
 
 The full 11/4 Galaxy Distribution News Brief will be available later today 
 and will contain details of changes since the last release.
 
 --nate
 Galaxy Team
 ___
 Please keep all replies on the list by using reply all
 in your mail client.  To manage your subscriptions to this
 and other Galaxy lists, please use the interface at:
  http://lists.bx.psu.edu/
 
 To search Galaxy mailing lists use the unified search at:
  http://galaxyproject.org/search/mailinglists/


___
Please keep all replies on the list by using reply all
in your mail client.  To manage your subscriptions to this
and other Galaxy lists, please use the interface at:
  http://lists.bx.psu.edu/

To search Galaxy mailing lists use the unified search at:
  http://galaxyproject.org/search/mailinglists/


Re: [galaxy-dev] Security vulnerability in Galaxy filtering tools

2013-11-05 Thread Ido Tamir

On Nov 5, 2013, at 6:28 PM, Nate Coraor n...@bx.psu.edu wrote:

 Hi Ido,
 
 Thanks for the feedback.  Replies below.
 
 On Nov 5, 2013, at 9:54 AM, Ido Tamir wrote:
 
 This seems to happen often e.g. 
 http://wiki.galaxyproject.org/DevNewsBriefs/2012_10_23#Compute_Tool_Security_Fix
 
 I'm not sure I'd agree that it's often - we've had 4 or 5 vulnerabilities 
 over the life of the project.  2 allowed arbitrary code execution, the others 
 were less severe.
 
But these were written by experts, not by people like me, that don't know what 
the galaxy framework really does/does not do with the input, so I guess I make 
many more mistakes.

 a) are there general guidelines in the wiki on how to avoid these problems 
 when creating tools?
 
 The guidelines for writing a Galaxy tool are no different from best practices 
 for writing secure code.  In specific for this vulnerability, execution of 
 user input should be handled with extreme care, and this tool had some gaps 
 in its input validation and sanitization.  For what it's worth, the filter 
 tool (on which the other vulnerable tools were based) is one of the few tools 
 surviving from the very early days of Galaxy, and would not be implemented 
 the same way if written today.

I think it would be nice to have a small outline on the wiki of what galaxy 
does with the input data and how it could affect a tool.
What sanitisation is there by default so I don't have to worry about it, but 
what could happen I if I don't care to check/remove sanitise ' | or  ..., 
maybe with examples.

 b) is there a way to check automatically if all input fields are correctly 
 escaped in a tool?
 
 I am not sure how Galaxy could do this.  Galaxy sanitizes the command line so 
 that input fields passed to a tool as command line arguments cannot be 
 crafted to exploit the shell's parsing rules.  
Thats good

best,
ido


 What the tool itself does with its inputs are out of Galaxy's control.
 
 --nate
 
 
 A search for security in the wiki brings up:
  • Admin/Data Libraries/Library Security
 0.0k - rev: 1 (current) last modified: 2013-01-02 23:54:33
  • Admin/DataLibraries/LibrarySecurity
 19.2k - rev: 4 (current) last modified: 2013-01-03 00:12:36
  • HelpOnConfiguration/SecurityPolicy
 1.9k - rev: 1 (current) last modified: 0
  • Learn/Security Features
 7.0k - rev: 3 (current) last modified: 2011-09-13 16:52:08
  • News/2013_04_08_Galaxy_Security_Release
 1.3k - rev: 3 (current) last modified: 2013-04-08 16:56:41
 
 escape does not bring up anything.
 
 thank you very much,
 ido
 
 
 On Nov 5, 2013, at 12:45 AM, Nate Coraor n...@bx.psu.edu wrote:
 
 A security vulnerability was recently discovered by John Chilton with 
 Galaxy's Filter data on any column using simple expressions and Filter 
 on ambiguities in polymorphism datasets tools that can allow for arbitrary 
 execution of code on the command line.
 
 The fix for these tools has been committed to the Galaxy source.  The 
 timing of this commit coincides with the next Galaxy stable release (which 
 has also been pushed out today).
 
 To apply the fix and simultaneously update to the new Galaxy stable 
 release, ensure you are on the stable branch and upgrade to the latest 
 changeset:
 
 % hg branch
 stable
 
 % hg pull -u
 
 For Galaxy installations that administrators are not yet ready to upgrade 
 to the latest release, there are three workarounds.
 
 First, for Galaxy installations running on a relatively new version of the 
 stable release (e.g. release_2013.08.12), Galaxy can be updated to the 
 specific changeset that that contains the fix.  This will include all of 
 the stable (non-feature) commits that have been accumulated since the 8/12 
 release plus any new features included with (and prior to) the 8/12 
 release, but without all of the new features included in the 11/4 release.  
 Ensure you are on the stable branch and then upgrade to the specific 
 changeset:
 
 % hg pull -u -r e094c73fed4d
 
 Second, the patch can be downloaded and applied manually:
 
 % wget -o security.patch 
 https://bitbucket.org/galaxy/galaxy-central/commits/e094c73fed4dc66b589932edb83412cb8b827cd3raw/
 
 and then:
 
 % hg patch security.patch
 
 or:
 
 % patch -p1  security.patch
 
 Third, the tools can be completely disabled by removing them from the tool 
 configuration file (by default, tool_conf.xml) and restarting all Galaxy 
 server processes.  The relevant lines in tool_conf.xml are:
 
 tool file=stats/dna_filtering.xml /
 tool file=stats/filtering.xml /
 
 The full 11/4 Galaxy Distribution News Brief will be available later today 
 and will contain details of changes since the last release.
 
 --nate
 Galaxy Team
 ___
 Please keep all replies on the list by using reply all
 in your mail client.  To manage your subscriptions to this
 and other Galaxy lists, please use the interface at:
 http://lists.bx.psu.edu/
 
 To search Galaxy mailing lists

Re: [galaxy-dev] Test Toolshed Biopython package dependency Atlas fails to install (Was: Re: UnboundLocalError: local variable 'prior_installation_required' referenced before assignment)

2013-09-20 Thread Ido Tamir
Yes this tar is broken at least on OSX. 
Other people have the same issue:
http://code.google.com/p/libarchive/issues/detail?id=299



On Sep 20, 2013, at 10:41 AM, Bjoern Gruening bjoern.gruen...@gmail.com wrote:

 Hi Carlos,
 
 Hi Peter and Carlos,
 
 On Mon, Sep 16, 2013 at 8:57 PM, Carlos Borroto
 carlos.borr...@gmail.com wrote:
 I did an extra test. Started with a clean 'galaxy-dist'. This time
 both repositories fail with the same error. I guess before something
 was cached for the repository with version 0.1.4.
 
 I used biopython repository as a guide to write my tool dependency 
 definition:
 http://testtoolshed.g2.bx.psu.edu/view/biopython/package_biopython_1_61
 
 I can confirm biopython repository is failing to install for me with
 exactly the same error. I wonder if a recently addition in the test
 toolshed broke the treatment of prior_installation_required.
 
 Thanks,
 Carlos
 
 Could be - note that currently Biopython isn't currently
 installing properly on the Test Tool Shed due to ALTAS
 failing (a requirement of NumPy which is a requirement
 of Biopython). Dave and Bjoern are I think looking at this
 already...
 
 I can't do much I tested it again and for me its working fine on my
 computers I have at hand ... sorry.
 
 
 In case it helps, this is how the INSTALLATION.log file ends on OS X 10.8.4:
 $ tail -n 3 
 ~/src/shed_tools_dependencies.central/atlas/3.10.1/iuc/package_atlas_3_10/3508de0ebae1/INSTALLATION.log
 x ATLAS/tune/lapack/lanbsrch.c
 tar: Error exit delayed from previous errors.
 #
 
 This is the relevant part I can find in Galaxy's log:
 [localhost] local: tar xfvj atlas3.10.1.tar.bz2
 
 Warning: local() encountered an error (return code 1) while executing
 'tar xfvj atlas3.10.1.tar.bz2'
 
 After noticing this I got what I'm guessing is the original file from
 sourceforge:
 http://downloads.sourceforge.net/project/math-atlas/Stable/3.10.1/atlas3.10.1.tar.bz2
 
 I can confirm that trying to untar this file fails with the exact same
 error. However, on Ubuntu 13.04 untaring this file works just fine.
 
 That is new to me. How can that happen? Anyone with an OS-X can confirm
 that?
 
 On Ubuntu 13.04 the error I see is:
 STDERR
 It appears you have cpu throttling enabled, which makes timings
 unreliable and an ATLAS install nonsensical.  Aborting.
 See ATLAS/INSTALL.txt for further information
 #
 
 Björn, you said cpu throttling can be easily disable on Ubuntu. Can
 you comment how? Do I need to disable it completely, or should the
 install script do it when installing?
 
 ATLAS (once untared ;-)) needs cpu-throttling to be disabled to
 optimizes its library. If it is not disabled ATLAS compilation will
 fail. For OS-X I found that:
 
 http://apple.stackexchange.com/questions/41045/how-can-i-disable-cpu-throttling-and-cpu-disabling
 
 Sorry, I never touched a OS-X. Nevertheless, If its not disabled, it is
 supposed to fail silently and downstream packages will not be affected.
 But if its crashing during untarring, I can't do much. What I can do is
 to repack the tarball and host it elsewhere. What brings me to:
 
 http://gmod.827538.n3.nabble.com/RFC-Storing-of-tarballs-and-patches-for-tool-dependencies-to-enable-reproducibility-td4036591.html
 
 Bad news for a Friday morning, sorry :(
 Bjoern
 
 
 Thanks,
 Carlos
 
 ___
 Please keep all replies on the list by using reply all
 in your mail client.  To manage your subscriptions to this
 and other Galaxy lists, please use the interface at:
  http://lists.bx.psu.edu/
 
 To search Galaxy mailing lists use the unified search at:
  http://galaxyproject.org/search/mailinglists/
 
 
 
 ___
 Please keep all replies on the list by using reply all
 in your mail client.  To manage your subscriptions to this
 and other Galaxy lists, please use the interface at:
  http://lists.bx.psu.edu/
 
 To search Galaxy mailing lists use the unified search at:
  http://galaxyproject.org/search/mailinglists/


___
Please keep all replies on the list by using reply all
in your mail client.  To manage your subscriptions to this
and other Galaxy lists, please use the interface at:
  http://lists.bx.psu.edu/

To search Galaxy mailing lists use the unified search at:
  http://galaxyproject.org/search/mailinglists/


Re: [galaxy-dev] Test Toolshed Biopython package dependency Atlas fails to install (Was: Re: UnboundLocalError: local variable 'prior_installation_required' referenced before assignment)

2013-09-20 Thread Ido Tamir
tar xvfj atlas3.11.11.tar.bz2
shows no errors on OSX and creates one ATLAS folder.

best,
ido

On Sep 20, 2013, at 11:17 AM, Bjoern Gruening bjoern.gruen...@gmail.com wrote:

 Hi Ido and Carlos,
 
 can you check if that tarball is working?
 
 http://downloads.sourceforge.net/project/math-atlas/Developer%20%28unstable%29/3.11.11/atlas3.11.11.tar.bz2
 
 The chance is low, but if its working for you I will consider to create a new 
 version for it.
 Thanks,
 Bjoern
 
 Yes this tar is broken at least on OSX. 
 Other people have the same issue:
 
 http://code.google.com/p/libarchive/issues/detail?id=299
 
 
 
 
 On Sep 20, 2013, at 10:41 AM, Bjoern Gruening 
 bjoern.gruen...@gmail.com
  wrote:
 
 
  Hi Carlos,
  
  Hi Peter and Carlos,
  
  On Mon, Sep 16, 2013 at 8:57 PM, Carlos Borroto
  carlos.borr...@gmail.com wrote:
  I did an extra test. Started with a clean 'galaxy-dist'. This time
  both repositories fail with the same error. I guess before something
  was cached for the repository with version 0.1.4.
  
  I used biopython repository as a guide to write my tool dependency 
  definition:
  http://testtoolshed.g2.bx.psu.edu/view/biopython/package_biopython_1_61
  
  I can confirm biopython repository is failing to install for me with
  exactly the same error. I wonder if a recently addition in the test
  toolshed broke the treatment of prior_installation_required.
  
  Thanks,
  Carlos
  
  Could be - note that currently Biopython isn't currently
  installing properly on the Test Tool Shed due to ALTAS
  failing (a requirement of NumPy which is a requirement
  of Biopython). Dave and Bjoern are I think looking at this
  already...
  
  I can't do much I tested it again and for me its working fine on my
  computers I have at hand ... sorry.
  
  
  In case it helps, this is how the INSTALLATION.log file ends on OS X 
  10.8.4:
  $ tail -n 3 
  ~/src/shed_tools_dependencies.central/atlas/3.10.1/iuc/package_atlas_3_10/3508de0ebae1/INSTALLATION.log
  x ATLAS/tune/lapack/lanbsrch.c
  tar: Error exit delayed from previous errors.
  #
  
  This is the relevant part I can find in Galaxy's log:
  [localhost] local: tar xfvj atlas3.10.1.tar.bz2
  
  Warning: local() encountered an error (return code 1) while executing
  'tar xfvj atlas3.10.1.tar.bz2'
  
  After noticing this I got what I'm guessing is the original file from
  sourceforge:
  http://downloads.sourceforge.net/project/math-atlas/Stable/3.10.1/atlas3.10.1.tar.bz2
  
  I can confirm that trying to untar this file fails with the exact same
  error. However, on Ubuntu 13.04 untaring this file works just fine.
  
  That is new to me. How can that happen? Anyone with an OS-X can confirm
  that?
  
  On Ubuntu 13.04 the error I see is:
  STDERR
  It appears you have cpu throttling enabled, which makes timings
  unreliable and an ATLAS install nonsensical.  Aborting.
  See ATLAS/INSTALL.txt for further information
  #
  
  Björn, you said cpu throttling can be easily disable on Ubuntu. Can
  you comment how? Do I need to disable it completely, or should the
  install script do it when installing?
  
  ATLAS (once untared ;-)) needs cpu-throttling to be disabled to
  optimizes its library. If it is not disabled ATLAS compilation will
  fail. For OS-X I found that:
  
  http://apple.stackexchange.com/questions/41045/how-can-i-disable-cpu-throttling-and-cpu-disabling
  
  Sorry, I never touched a OS-X. Nevertheless, If its not disabled, it is
  supposed to fail silently and downstream packages will not be affected.
  But if its crashing during untarring, I can't do much. What I can do is
  to repack the tarball and host it elsewhere. What brings me to:
  
  http://gmod.827538.n3.nabble.com/RFC-Storing-of-tarballs-and-patches-for-tool-dependencies-to-enable-reproducibility-td4036591.html
  
  Bad news for a Friday morning, sorry :(
  Bjoern
  
  
  Thanks,
  Carlos
  
  ___
  Please keep all replies on the list by using reply all
  in your mail client.  To manage your subscriptions to this
  and other Galaxy lists, please use the interface at:
   http://lists.bx.psu.edu/
  
  To search Galaxy mailing lists use the unified search at:
   http://galaxyproject.org/search/mailinglists/
  
  
  
  ___
  Please keep all replies on the list by using reply all
  in your mail client.  To manage your subscriptions to this
  and other Galaxy lists, please use the interface at:
   http://lists.bx.psu.edu/
  
  To search Galaxy mailing lists use the unified search at:
   http://galaxyproject.org/search/mailinglists/
 
 
 
 
 


___
Please keep all replies on the list by using reply all
in your mail client.  To manage your subscriptions to this
and other Galaxy lists, please use the interface at:
  http://lists.bx.psu.edu/

To 

Re: [galaxy-dev] Towards Galaxy Linux (not) ? [was [RFC] Storing of tarballs and patches for tool_dependencies to enable reproducibility]

2013-09-19 Thread Ido Tamir
If I might chime in, I am a bit worried about all the automatic installation 
going on in galaxy, and it seems that the trend is to enhance this.
A small R or python script calling into well known libraries that come from 
well known repositories (bioconductor etc… ) I can check.
(Of course I install too much stuff from github, bioconductor etc… without 
checking).
 
 I'm not sure it is comparable to a entire Linux distribution, its more
 like an Appstore, like pypi, bioconductor or gems, and yes that is

The app stores are checked by Apple or google for malicious code, the apps are 
sandboxed.
There are many eyes for python, bioconductor packages and gems because much 
more people interact with
them directly compared to galaxy-tools.

 Sorry maybe I was misleading. I only want a central storage for
 binaries/tarballs where the source can not be trusted for long term.
 'long term' and 'trusted' needs to be defined in such a discussion here.
 I do not think we should copy python packages that are stored in pypi.
 We should make it easy as possible to install them in our repository. If
 you do not trust pypi, we can offer a mirror. Some goes for gems.

Trusted for me means I trust the source not having dangerous code. I trust pypi
more than some mirror, bioconductor base packages from more than some freshly 
published package 
that few people have used, tools from galaxy core developers more than from 
tool-shed etc…
I know this is not the type of trust you were talking about.

best,
ido
___
Please keep all replies on the list by using reply all
in your mail client.  To manage your subscriptions to this
and other Galaxy lists, please use the interface at:
  http://lists.bx.psu.edu/

To search Galaxy mailing lists use the unified search at:
  http://galaxyproject.org/search/mailinglists/


Re: [galaxy-dev] tools installed from a toolshed repository displayed in an arbitrary order

2013-09-16 Thread Ido Tamir
IIRC then you can change the order of tools manually by copying around the 
individual tools in shed_tool_conf.xml
In my case it builds the integrated_tool_panel.xml in reverse order within the 
section.

I still have labels between the section defined in the tool_conf.xml.
But no subsections.

best,
ido

On Sep 16, 2013, at 6:40 AM, yulia.arzha...@csiro.au wrote:

 Hello guys,
  
 After porting our Galaxy tools into a tool shed repository the tool sections 
 got lost. I know there is a possibility to assign a repository to a section 
 inside the tools panel when the repository is getting installed. However, we 
 would still like to have sections inside the repository – that makes 
 navigating the tools much neater. Also, the way shed_tool_conf.xml is 
 populated with tools during repository installation seems arbitrary to me. Is 
 there any way to control this, to have repository tools sorted/displayed in a 
 certain order in Tools panel, in the very least, alphabetically? I’ve got 
 many tools in my repository (and no, it is not possible to divide them into 
 several repositories), and the way they are sorted or divided into sections 
 makes navigating the tools much easier for a user.
 Thanks,
 Yulia
  
 ___
 Please keep all replies on the list by using reply all
 in your mail client.  To manage your subscriptions to this
 and other Galaxy lists, please use the interface at:
  http://lists.bx.psu.edu/
 
 To search Galaxy mailing lists use the unified search at:
  http://galaxyproject.org/search/mailinglists/


___
Please keep all replies on the list by using reply all
in your mail client.  To manage your subscriptions to this
and other Galaxy lists, please use the interface at:
  http://lists.bx.psu.edu/

To search Galaxy mailing lists use the unified search at:
  http://galaxyproject.org/search/mailinglists/


Re: [galaxy-dev] Galaxy Release Cycle Length

2013-08-21 Thread Ido Tamir
Why the dislike for quick turnover? Could somebody present the arguments for 
people not having been at the BOF?

People don't have to upgrade - unless its breaking changes that e.g. disable 
the possibility to download from the public toolshed which forced me to upgrade.

The alternative to frequent releases are not better tested and documented 
releases. I think its impossible for the galaxy team to test on all the diverse 
configurations
galaxy is deployed on. Then you really have accumulated many bugs, feedback 
from users comes in at once creating coordination need for developers going all 
at once into bugfixing mode for features they have worked on months ago etc… 
Then the fixes for these bugs also come only with the next release which is 
still many months away etc….  This makes it necessary to split development into 
major releases (say once a year) and minor, bugfix, releases (1-2 months after 
the major)  - where is the gain for the users?

I would not even split things between breaking changes and minor changes. I 
think this slows down development of the platform and what concerns
people most, the tools, are developed independently of the platform and one can 
upgrade them any time.

To give an example, the job_config is now much better than before and its good 
that I did not have to wait months from its development to deployment at our 
site.
Now small additional features like setting the number of threads dynamically 
are suggested, and then I would have to wait again many,many months until the 
next release. 

Upgrading galaxy was o.k. and while its unfortunate to have to learn new 
settings and remember the old ones from sometimes not so good (but improving!) 
documentation,
there is no way around it. I think I will now upgrade more often, because it 
went so well. I like release early, release often.

best,
ido

On Aug 20, 2013, at 8:36 PM, Dave Clements cleme...@galaxyproject.org wrote:

 Hello all,
 
 At one of the GCC2013 Birds of a Feather sessions the group was very clear 
 that they would like to see less frequent releases of Galaxy.  We're 
 currently aiming to do a release every 2 months and have been pretty 
 successful at making that target.  In the past, we have tried doing releases 
 more often and less often.
 
 Is there a sweet spot for the time between releases?
 
 Please reply to the group.  We are interested in a discussion.
 
 Thanks,
 
 Dave C
 
 -- 
 http://galaxyproject.org/GCC2013
 http://galaxyproject.org/
 http://getgalaxy.org/
 http://usegalaxy.org/
 http://wiki.galaxyproject.org/
 ___
 Please keep all replies on the list by using reply all
 in your mail client.  To manage your subscriptions to this
 and other Galaxy lists, please use the interface at:
  http://lists.bx.psu.edu/
 
 To search Galaxy mailing lists use the unified search at:
  http://galaxyproject.org/search/mailinglists/


___
Please keep all replies on the list by using reply all
in your mail client.  To manage your subscriptions to this
and other Galaxy lists, please use the interface at:
  http://lists.bx.psu.edu/

To search Galaxy mailing lists use the unified search at:
  http://galaxyproject.org/search/mailinglists/


Re: [galaxy-dev] shall us specify which tool run in cluster?

2013-07-31 Thread Ido Tamir

On Jul 31, 2013, at 8:52 AM, shenwiyn shenw...@gmail.com wrote:

 Hi Thon Deboer ,
 I am a newer in Galaxy.I installed my Galaxy with Torque2.5.0 ,and Galaxy 
 uses the pbs_modoule to interface with TORQUE.But I have some question of the 
  job_conf.xml :
 1.)In your  job_conf.xml ,you use regularjobs,longjobs,shortjobs...to run 
 different jobs ,how our Galaxy know which tool belongs to regularjobs or 
 longjobs.And what is the meaning of nativeSpecification?

by specifying, as Thon did, in tools the id of the tool and its destination 
which are the settings.
the nativeSpecification allows you to set additional parameters that are passed 
with the call.
e.g.  -pe smp 4 tells the grid engine to use the parallel environment smp with 
4 cores.

 2.)Shall us use toolscollection of tool id=bwa_wrapper 
 destination=multicorejobs4/to specify bwa ?Does it mean the bwa belong to 
 multicorejobs4,and run in cluster?

exactly

 3.)Does every tool need us to specify which job it belong to?
  I saw http://wiki.galaxyproject.org/Admin/Config/Jobs about this,but I am 
 not sure above.Could you help me please?
  

fortunately there is a default 


 shenwiyn
  
 From: Thon Deboer
 Date: 2013-07-18 14:31
 To: galaxy-dev
 Subject: [galaxy-dev] Jobs remain in queue until restart
 Hi,
  
 I have noticed that from time to time, the job queue seems to be “stuck” and 
 can only be unstuck by restarting galaxy.
 The jobs seem to be in the queue state and the python job handler processes 
 are hardly ticking over and the cluster is empty.
  
 When I restart, the startup procedure realizes all jobs are in the a “new 
 state” and it then assigns a jobhandler after which the jobs start fine….
  
 Any ideas?
  Torque
  
 Thon
  
 P.S I am using the june version of galaxy and I DO set limits on my users in 
 job_conf.xml as so: (Maybe it is related? Before it went into dormant mode, 
 this user had started lots of jobs and may have hit the limit, but I assumed 
 this limit was the number of running jobs at one time, right?)
  
 ?xml version=1.0?
 job_conf
 plugins workers=4
 !-- workers is the number of threads for the runner's work queue.
  The default from plugins is used if not defined for a plugin.
   --
 plugin id=local type=runner 
 load=galaxy.jobs.runners.local:LocalJobRunner workers=2/
 plugin id=drmaa type=runner 
 load=galaxy.jobs.runners.drmaa:DRMAAJobRunner workers=8/
 plugin id=cli type=runner 
 load=galaxy.jobs.runners.cli:ShellJobRunner workers=2/
 /plugins
 handlers default=handlers
 !-- Additional job handlers - the id should match the name of a
  [server:id] in universe_wsgi.ini.
  --
 handler id=handler0 tags=handlers/
 handler id=handler1 tags=handlers/
 handler id=handler2 tags=handlers/
 handler id=handler3 tags=handlers/
 !-- handler id=handler10 tags=handlers/
 handler id=handler11 tags=handlers/
 handler id=handler12 tags=handlers/
 handler id=handler13 tags=handlers/
 --
 /handlers
 destinations default=regularjobs
 !-- Destinations define details about remote resources and how jobs
  should be executed on those remote resources.
  --
 destination id=local runner=local/
 destination id=regularjobs runner=drmaa tags=cluster
 !-- These are the parameters for qsub, such as queue etc. --
 param id=nativeSpecification-V -q long.q -pe smp 1/param
 /destination
 destination id=longjobs runner=drmaa tags=cluster,long_jobs
 !-- These are the parameters for qsub, such as queue etc. --
 param id=nativeSpecification-V -q long.q -pe smp 1/param
 /destination
 destination id=shortjobs runner=drmaa tags=cluster,short_jobs
 !-- These are the parameters for qsub, such as queue etc. --
 param id=nativeSpecification-V -q short.q -pe smp 1/param
 /destination
 destination id=multicorejobs4 runner=drmaa 
 tags=cluster,multicore_jobs
 !-- These are the parameters for qsub, such as queue etc. --
 param id=nativeSpecification-V -q long.q -pe smp 4/param
 /destination
  
 !-- destination id=real_user_cluster runner=drmaa
 param 
 id=galaxy_external_runjob_scriptscripts/drmaa_external_runner.py/param
 param 
 id=galaxy_external_killjob_scriptscripts/drmaa_external_killer.py/param
 param 
 id=galaxy_external_chown_scriptscripts/external_chown_script.py/param
 /destination --
  
 destination id=dynamic runner=dynamic
 !-- A destination that represents a method in the dynamic 
 runner. --
 param id=typepython/param
 param id=functioninteractiveOrCluster/param
 /destination
 /destinations
 tools
 !-- Tools can be configured to use specific destinations or 

[galaxy-dev] job_conf.xml workers

2013-07-24 Thread Ido Tamir
Hi,

what is the relationship between workers defined in universe and job_conf.xml? 
e.g.:

[server:handler0]
use = egg:Paste#http
port = 9092
host = 127.0.0.1
use_threadpool = true
threadpool_workers = 5


plugins workers=4
!-- workers is the number of threads for the runner's work queue.
 The default from plugins is used if not defined for a plugin.
  --
plugin id=local type=runner 
load=galaxy.jobs.runners.local:LocalJobRunner/
plugin id=drmaa type=runner 
load=galaxy.jobs.runners.drmaa:DRMAAJobRunner workers=10/
/plugins

a) do I have to specify the sum of the workers in job_conf.xml in universe?
e.g. 4 local, 10 drmaa = 14 threadpool_workers per handler? Or is universe 
overridden by job_conf.xml?

b)
How many concurrent jobs can I then expect to have concurrently? 14 per handler?


thank you very much,
ido




___
Please keep all replies on the list by using reply all
in your mail client.  To manage your subscriptions to this
and other Galaxy lists, please use the interface at:
  http://lists.bx.psu.edu/

To search Galaxy mailing lists use the unified search at:
  http://galaxyproject.org/search/mailinglists/


Re: [galaxy-dev] job_conf.xml workers

2013-07-24 Thread Ido Tamir
Ah yes,
this was local_job_queue_workers and cluster_job_queue_workers in universe.ini

thank you very much,
ido

On Jul 24, 2013, at 5:10 PM, James Taylor ja...@jamestaylor.org wrote:

 use_threadpool and threadpool_workers specify the number of threads
 for handling web requests and have nothing to do with job running.
 
 --
 James Taylor, Assistant Professor, Biology/CS, Emory University
 
 
 On Wed, Jul 24, 2013 at 11:02 AM, Ido Tamir ta...@imp.ac.at wrote:
 Hi,
 
 what is the relationship between workers defined in universe and 
 job_conf.xml? e.g.:
 
 [server:handler0]
 use = egg:Paste#http
 port = 9092
 host = 127.0.0.1
 use_threadpool = true
 threadpool_workers = 5
 
 
 plugins workers=4
!-- workers is the number of threads for the runner's work queue.
 The default from plugins is used if not defined for a plugin.
  --
plugin id=local type=runner 
 load=galaxy.jobs.runners.local:LocalJobRunner/
plugin id=drmaa type=runner 
 load=galaxy.jobs.runners.drmaa:DRMAAJobRunner workers=10/
 /plugins
 
 a) do I have to specify the sum of the workers in job_conf.xml in universe?
 e.g. 4 local, 10 drmaa = 14 threadpool_workers per handler? Or is universe 
 overridden by job_conf.xml?
 
 b)
 How many concurrent jobs can I then expect to have concurrently? 14 per 
 handler?
 
 
 thank you very much,
 ido
 
 
 
 
 ___
 Please keep all replies on the list by using reply all
 in your mail client.  To manage your subscriptions to this
 and other Galaxy lists, please use the interface at:
  http://lists.bx.psu.edu/
 
 To search Galaxy mailing lists use the unified search at:
  http://galaxyproject.org/search/mailinglists/


___
Please keep all replies on the list by using reply all
in your mail client.  To manage your subscriptions to this
and other Galaxy lists, please use the interface at:
  http://lists.bx.psu.edu/

To search Galaxy mailing lists use the unified search at:
  http://galaxyproject.org/search/mailinglists/


[galaxy-dev] job_conf.xml questions

2013-07-23 Thread Ido Tamir
Hi,
I work myself through the job_conf.xml and have a question:

a)
In your advanced example you have:

destination id=remote_cluster runner=drmaa tags=longjobs/
 destination id=real_user_cluster runner=drmaa
!-- TODO: The real user options should maybe not be considered 
runner params. --
param 
id=galaxy_external_runjob_scriptscripts/drmaa_external_runner.py/param
param 
id=galaxy_external_killjob_scriptscripts/drmaa_external_killer.py/param
param 
id=galaxy_external_chown_scriptscripts/external_chown_script.py/param
 /destination

Does this mean that remote_cluster jobs can not be killed unless I add the 3 
scripts to this destination?
Or does this have to do with the run jobs as real user that I don't need 
currently. I am using the galaxy user for all jobs for the moment?

To phrase this differently: Do I need these three scripts?

b)
Can I set a wall time limit for only for my local runners?

like:
destination id=local runner=local/
   param id=Resource_Listwalltime=72:00:00/param
/destination

When I set it in limits, will it work for local runners and then I have to 
override it in 
every destination?


thank you very much,
ido





___
Please keep all replies on the list by using reply all
in your mail client.  To manage your subscriptions to this
and other Galaxy lists, please use the interface at:
  http://lists.bx.psu.edu/

To search Galaxy mailing lists use the unified search at:
  http://galaxyproject.org/search/mailinglists/


[galaxy-dev] no metadata for tools in toolshed

2013-07-10 Thread Ido Tamir
Hi,
I grabbed some xml files from Hannons website for his fastx toolkit (improved 
over galaxys default version).
Then I created a repository, uploaded the xml files,
but the toolshed complains that it could not find the tools for 
creating metadata in the xml files.

Could somebody please tell me what the problem is with these files?
They look like valid tool.xml files to me.

http://toolshed.g2.bx.psu.edu/view/idot/fastx_toolkit2
http://i...@toolshed.g2.bx.psu.edu/repos/idot/fastx_toolkit2

thank you very much,
ido




___
Please keep all replies on the list by using reply all
in your mail client.  To manage your subscriptions to this
and other Galaxy lists, please use the interface at:
  http://lists.bx.psu.edu/

To search Galaxy mailing lists use the unified search at:
  http://galaxyproject.org/search/mailinglists/


Re: [galaxy-dev] deleted_new kills handlers

2013-01-22 Thread Ido Tamir
No the problem is still there, but we have not updated yet since beginning of 
december.

maybe this is related:
http://gmod.827538.n3.nabble.com/segfault-in-libdrmaa-gt-galaxy-front-end-failure-td4027349.html

best,
ido

On Jan 21, 2013, at 5:45 AM, Derrick Lin wrote:

 Hi Ido,
 
 I just found you post, and I think we are having the similar issue as your. 
 (I posted it in the mailing yesterday, you can find it for the detail of my 
 problem).
 
 I am wondering if you are able to fix your problem?
 
 Cheers,
 Derrick
 
 
 
 
 On Tue, Dec 11, 2012 at 12:13 AM, Ido Tamir ta...@imp.ac.at wrote:
 Dear galaxy maintainers,
 
 we have the problem that killing jobs running on drmaa before they run 
 leading to a deleted_new state,
 kills our handlers (sometimes?). The job gets sent to the handler, but the 
 handler never acknowledges the jobs (local and drmaa) and
 all submitted jobs afterwards stay in submitted state.
 
 Normally submission works fine, running also.
 
 Is such a problem known to anybody else?
 
 in my universe_wsgi.ini:
 enable_job_recovery = False
 I don't know which other settings could affect this.
 
 Could it be an old drmaa library, maybe because the SGE was updated
 after galaxy installation? Other ideas?
 
 
 thank you very much,
 ido
 
 
 
 ___
 Please keep all replies on the list by using reply all
 in your mail client.  To manage your subscriptions to this
 and other Galaxy lists, please use the interface at:
 
   http://lists.bx.psu.edu/
 


___
Please keep all replies on the list by using reply all
in your mail client.  To manage your subscriptions to this
and other Galaxy lists, please use the interface at:

  http://lists.bx.psu.edu/


Re: [galaxy-dev] Job handler keeps crashing

2013-01-22 Thread Ido Tamir
Is it possible to backport this onto the latest distribution?
Yes, I'm lazy, but there are also others that are still updating
within the next weeks and will have problems without them being
aware of this fix.

best,
ido


On Jan 21, 2013, at 9:50 PM, Nate Coraor wrote:

 Hi all,
 
 The commit[1] that fixes this is not in the January 11 distribution.  It'll 
 be part of the next distribution.
 
 --nate
 
 [1] 
 https://bitbucket.org/galaxy/galaxy-central/commits/c015b82b3944f967e2c859d5552c00e3e38a2da0
 
 On Jan 21, 2013, at 3:10 PM, Anthonius deBoer wrote:
 
 I have seen this same issue exactly. Python just dies without any errors in 
 the log. Using the latest galaxy-dist
 
 Sent from my iPhone
 
 On Jan 20, 2013, at 8:35 PM, Derrick Lin klin...@gmail.com wrote:
 
 Update to the 11 Jan 2013 dist does not help with this issue. :(
 
 I checked the database and have the look at the job entries that handler0 
 tried to stop then shutdown:
 
 | 3088 | 2013-01-03 14:25:38 | 2013-01-03 14:27:05 |531 | 
 toolshed.g2.bx.psu.edu/repos/kevyin/homer/homer_findPeaks/0.1.2
 | 0.1.2| deleted_new | Job output deleted by user before job 
 completed. | NULL | NULL   | NULL| NULL   | NULL   
 | NULL  |   1659 | drmaa://-V -j n -R y -q intel.q/ | NULL  
  |  NULL |  76 |0 | NULL| NULL  
  | handler0 |  NULL |
 | 3091 | 2013-01-04 10:52:19 | 2013-01-07 09:14:34 |531 | 
 toolshed.g2.bx.psu.edu/repos/kevyin/homer/homer_findPeaks/0.1.2
 | 0.1.2| deleted_new | Job output deleted by user before job 
 completed. | NULL | NULL   | NULL| NULL   | NULL   
 | NULL  |   1659 | drmaa://-V -j n -R y -q intel.q/ | NULL  
  |  NULL |  76 |0 | NULL| NULL  
  | handler0 |  NULL |
 | 3093 | 2013-01-07 22:02:21 | 2013-01-07 22:16:27 |531 | 
 toolshed.g2.bx.psu.edu/repos/kevyin/homer/homer_pos2bed/1.0.0  
 | 1.0.0| deleted_new | Job output deleted by user before job 
 completed. | NULL | NULL   | NULL| NULL   | NULL   
 | NULL  |   1749 | drmaa://-V -j n -R y -q intel.q/ | NULL  
  |  NULL |  76 |0 | NULL| NULL  
  | handler0 |  NULL |
 
 So basically the job table has several of these entries what assigned to 
 handler0 and marked as deleted_new. When the handler0 is up, it starts 
 stopping these jobs, after the first job has been stopped, handler0 went 
 crash and died. But that job was then marked as deleted.
 
 I think if I manually change the job state from deleted_new to deleted 
 in the db, the handler0 will become fine. I am just concerned about how 
 these jobs were created (like assigned to a handler but marked as 
 deleted_new). 
 
 Cheers,
 D
 
 
 On Mon, Jan 21, 2013 at 1:49 PM, Derrick Lin klin...@gmail.com wrote:
 I had a close look at the code in 
 
 galaxy-dist / lib / galaxy / jobs / handler.py
 galaxy-dist / lib / galaxy / jobs / runners / drmaa.py
 
 and found that stopping deleted and deleted_new seems to be normal 
 routine for the job handler. Could not find any exception that caused the 
 shutdown.
 
 I do notice in the galaxy-dist on bitbucket, there is one commit with 
 comment Fix shutdown on python = 2.6.2 by calling setDaemon when creating 
 threads (these are still..., it seems to be relevant?
 
 I will do the update to 11 Jan release and see if it fixes the issue.
 
 D
 
 
 On Fri, Jan 18, 2013 at 4:03 PM, Derrick Lin klin...@gmail.com wrote:
 Hi guys,
 
 We have updated our galaxy to 20 Dec 2012 release. Recently we found that 
 some submitted jobs could not start (stay gray forever).
 
 We found that it was caused by the job manager sent jobs to a handler 
 (handler0) whose python process crashed and died.
 
 From the handler log we found the last messages right before the crash:
 
 galaxy.jobs.handler DEBUG 2013-01-18 15:00:34,481 Stopping job 3032:
 galaxy.jobs.handler DEBUG 2013-01-18 15:00:34,481 stopping job 3032 in 
 drmaa runner
 
 We restarted the galaxy, handler0 is up for few seconds then died again 
 with the same error messages except the job number moved to the next one.
 
 We observed that the jobs it was trying to stop are all previous jobs whose 
 status is either deleted or deleted_new.
 
 We have never seen this in the past, so wondering if there is bugs in the 
 new release?
 
 Cheers,
 Derrick
 
 
 ___
 Please keep all replies on the list by using reply all
 in your mail client.  To manage your subscriptions to this
 and other Galaxy lists, please use the interface at:
 
 http://lists.bx.psu.edu/
 ___
 Please keep all replies on the list by using reply all
 in your mail client.  To manage your subscriptions to this
 and other Galaxy lists, 

Re: [galaxy-dev] In which Encryption form does Galaxy store password?

2013-01-07 Thread Ido Tamir
Its hashed (sha1) and you can not reverse the hash. Storing passwords in a 
reversible manner is bad practice. 
http://galaxy-central.readthedocs.org/en/latest/_modules/galaxy/util/hash_util.html

You could maybe authenticate your users from the other program with the same 
method against the hashed passwords.

best,
ido

On Jan 7, 2013, at 12:43 PM, Sachit Adhikari wrote:

 Hi,
 
 In which encryption galaxy store the password? It's really urgent. I want to 
 decode the password and use in the other program. Thanks
 ___
 Please keep all replies on the list by using reply all
 in your mail client.  To manage your subscriptions to this
 and other Galaxy lists, please use the interface at:
 
 http://lists.bx.psu.edu/


___
Please keep all replies on the list by using reply all
in your mail client.  To manage your subscriptions to this
and other Galaxy lists, please use the interface at:

  http://lists.bx.psu.edu/


Re: [galaxy-dev] let galaxy use more hardware resouses

2012-12-10 Thread Ido Tamir
If you did what the links posted previously suggested (inc. parallelization of 
tools, mutliple server processes),
there is nothing more you can do.

Most of the tools galaxy just calls, so its the tool that you have to change.
Often genomic tools are also IO bound, so a faster storage system could lead to 
better CPU utilization - if this is the bottleneck.

HTH,
ido

On Dec 10, 2012, at 9:51 AM, 泽 蔡 wrote:

 Hi all,
  
 I run a local instance of galaxy. Today, I take a look at the system monitor, 
 I found the galaxy just use 4% of a CPU and a little bit of the memory of the 
 PC. But the PC has twelve CPUs and 126G memory. So, how to configure the 
 galaxy to use more hardware resouces to calculate quickly?
 ___
 Please keep all replies on the list by using reply all
 in your mail client.  To manage your subscriptions to this
 and other Galaxy lists, please use the interface at:
 
  http://lists.bx.psu.edu/


___
Please keep all replies on the list by using reply all
in your mail client.  To manage your subscriptions to this
and other Galaxy lists, please use the interface at:

  http://lists.bx.psu.edu/

Re: [galaxy-dev] binary data types get corrupted on upload

2012-12-07 Thread Ido Tamir
The specification does not mention zip, gzip or bz2 [1] and
I am pretty confident, that it does not in addition compresses
its content with zip etc..
It starts with a header SQLite format 3\000 
which is visible in my file with less.

ido

[1]http://www.sqlite.org/fileformat.html

On Dec 7, 2012, at 3:50 PM, John Chilton wrote:

 Hi Ido,
 
 Is .sqlite a common compressed file type (i.e. zip, gzip, or bz2) just
 with a different extension?
 
 If it is, I think upload.py is always going to try to decompress it.
 That that has long been a behavior I would like to be able to disable
 on a per file type basis. If this is the problem and the galaxy team
 agrees to accept a patch for address the issue, I would be happy to
 try to implement the fix.
 
 -John
 
 On Fri, Dec 7, 2012 at 8:39 AM, Ido Tamir ta...@imp.ac.at wrote:
 
 from http://wiki.galaxyproject.org/Admin/Datatypes/Adding%20Datatypes
 I thought that adding subclass=True in datatypes_conf.xml would be enough 
 to register my new
 type, but it seems like I stil have to add an entry in binary.py for it to 
 be recognized. But its still corrupted.
 
 thank you very much,
 ido
 
 --
 Tool: Upload File
 Name:   hg19.sqlite
 Created:Dec 07, 2012
 Filesize:   11.1 MB
 Dbkey:  hg19
 Format: sqlite


___
Please keep all replies on the list by using reply all
in your mail client.  To manage your subscriptions to this
and other Galaxy lists, please use the interface at:

  http://lists.bx.psu.edu/


[galaxy-dev] from shed to shed

2012-12-04 Thread Ido Tamir
Hi,
how do I get tools from one local toolshed (or the main galaxy toolshed)
into another tool-shed? Is it possible to clone the content of a complete 
tool-shed - all repositories
at once into another tool-shed?
Can I then import all the tools at once into a local galaxy instance?

We have a group with their own tool-shed. Unfortunately, I accessed it directly
without a proxy on its port which leads to a problem described in issue 825 in 
galaxy-central.

I want to 
a) clone the tool-shed as it is into my own tool-shed (rsync?)
b) selectively or bulk install these tools into my galaxy server.

thank you very much,
ido



___
Please keep all replies on the list by using reply all
in your mail client.  To manage your subscriptions to this
and other Galaxy lists, please use the interface at:

  http://lists.bx.psu.edu/