i personally only know of adding a .jar file via add archive but my experience there is very limited. i believe if you 'add file' and the file is a directory it'll recursively take everything underneath but i know of nothing that inflates or un tars things on the remote end automatically.
i would 'add file' your python script and then within that untar your tarball to get at your model data. its just the matter of figuring out the path to that tarball that's kinda up in the air when its added as 'add file'. Yeah. "local downlooads directory". What's the literal path is what i'd like to know. :) On Thu, Jun 20, 2013 at 8:37 AM, Stephen Boesch <java...@gmail.com> wrote: > > @Stephen: given the 'relative' path for hive is from a local downloads > directory on each local tasktracker in the cluster, it was my thought that > if the archive were actually being expanded then > somedir/somefileinthearchive should work. I will go ahead and test this > assumption. > > In the meantime, is there any facility available in hive for making > archived files available to hive jobs? archive or hadoop archive ("har") > etc? > > > 2013/6/20 Stephen Sprague <sprag...@gmail.com> > >> what would be interesting would be to run a little experiment and find >> out what the default PATH is on your data nodes. How much of a pain would >> it be to run a little python script to print to stderr the value of the >> environmental variable $PATH and $PWD (or the shell command 'pwd') ? >> >> that's of course going through normal channels of "add file". >> >> the thing is given you're using a relative path "hive/parse_qx.py" you >> need to know what the "current directory" is when the process runs on the >> data nodes. >> >> >> >> >> On Thu, Jun 20, 2013 at 5:32 AM, Stephen Boesch <java...@gmail.com>wrote: >> >>> >>> We have a few dozen files that need to be made available to all >>> mappers/reducers in the cluster while running hive transformation steps . >>> >>> It seems the "add archive" does not make the entries unarchived and >>> thus available directly on the default file path - and that is what we are >>> looking for. >>> >>> To illustrate: >>> >>> add file modelfile.1; >>> add file modelfile.2; >>> .. >>> add file modelfile.N; >>> >>> Then, our model that is invoked during the transformation step *does *have >>> correct access to its model files in the defaul path. >>> >>> But .. those model files take low *minutes* to all load.. >>> >>> instead when we try: >>> add archive modelArchive.tgz. >>> >>> The problem is the archive does not get exploded apparently .. >>> >>> I have an archive for example that contains shell scripts under the >>> "hive" directory stored inside. I am *not *able to access >>> hive/my-shell-script.sh after adding the archive. Specifically the >>> following fails: >>> >>> $ tar -tvf appm*.tar.gz | grep launch-quixey_to_xml >>> -rwxrwxr-x stephenb/stephenb 664 2013-06-18 17:46 >>> appminer/bin/launch-quixey_to_xml.sh >>> >>> from (select transform (aappname,qappname) >>> *using *'*hive/parse_qx.py*' as (aappname2 string, qappname2 string) >>> from eqx ) o insert overwrite table c select o.aappname2, o.qappname2; >>> >>> Cannot run program "hive/parse_qx.py": java.io.IOException: error=2, No >>> such file or directory >>> >>> >>> >>> >> >