RE: libhdfs on windows

2012-10-31 Thread Peter Marron
Hi Colin,

I have a requirement to be able to run all the Hadoop functionality
that I build from a variety of platforms. This was the original
motivation for wanting to use libhdfs. I followed your
suggestion and looked into using webhdfs and it's looking
promising. Thanks for that. However I also need to be able
to launch Map/Reduce jobs from any platform.
In particular from Windows.  I looked into this by hacking
the bin/hadoop  script to extract the required class path
and various arguments so that I could launch a Map/Reduce job
just by invoking java with the correct arguments.
However I ran into HADOOP-7682.
I can see that there is a workaround here
https://github.com/congainc/patch-hadoop_7682-1.0.x-win
but it suggests that this is not really appropriate for
deployment. I suspect that I can get it to work reliably
by using cygwin and making loads of modifications
but that all seems rather a large effort, error-prone and
difficult to maintain.

Given that I plan to have a relatively small repertoire of Map/Reduce
jobs that I need to launch, I'm tempted to have all the jars pre-packed
on the Name-Node and have the ability to run them there. Then
have a daemon running so that I can use any appropriate ad-hoc RPC
mechanism from Windows to launch them.

Am I missing something? Is there a way to launch Map/Reduce
jobs in a platform neutral way, which runs out of the box, on Windows?

Again, any suggestions welcome.

Peter Marron

 -Original Message-
 From: Peter Marron
 [mailto:Peter.Marron@trilliumsoftware.
 com]
 Sent: 26 October 2012 00:53
 To: common-dev@hadoop.apache.org
 Subject: RE: libhdfs on windows
 
 Hi Colin,
 
 OK, I didn't know there was a hdfs-dev.
 I'm happy to ask there.
 (However there's a lot of mail on
 dev@hadoop and user@hadoop as well
 as user@hive and it's a bit of a
 commitment to track them all.) As for
 webhdfs, I did think about that, and in
 some ways it's a beautiful solution as it
 gives me a platform- and language-
 neutral access mechanism. I was just a
 little worried about the HTTP overhead if
 I am reading a single record at a time.
 Also I will need some way to launch my
 Map/Reduce jobs as well. So I'll probably
 end up using the C++/JNI/Java route to
 do that anyway. Unless there's a better
 way?
 Is there a web Map/Reduce interface?
 
 Many thanks,
 
 Z
 
  -Original Message-
  From: rarecac...@gmail.com
  [mailto:rarecac...@gmail.com] On
  Behalf Of Colin McCabe
  Sent: 25 October 2012 18:24
  To: common-dev@hadoop.apache.org
  Subject: Re: libhdfs on windows
 
  Hi Peter,
 
  This might be a good question for hdfs-
 dev?
 
  As Harsh pointed out below, HDFS-573
  was never committed.  I don't even see
 a patch attached, although
  there is some discussion.
 
  In the mean time, might I suggest using
 the webhdfs interface on
  Windows?
  webhdfs was intended as a stable REST
  interface that can be accessed from
 any platform.
 
  cheers,
  Colin
 
 
  On Thu, Oct 25, 2012 at 7:19 AM, Peter
 Marron
  peter.mar...@trilliumsoftware.com
  wrote:
   Hi,
  
   I've been looking at using libhdfs and
 I
  would like to use it on windows.
   I have found HDFS-573 and the
  information on this page:
  
 
 http://issues.apache.org/jira/browse/HD
  FS-573?page=com.atlassian.jira.
   plugin.system.issuetabpanels:all-
  tabpanel
   which suggests that quite a lot of
 work
  was done on this way back in 2009.
   So is there some source from this
  effort retained somewhere? If so,
  where?
   Or do I have to start from scratch?
  Apologies if this has already been asked
 recently.
  
   Any help appreciated.
  
   Peter Marron
 
 




libhdfs on windows

2012-10-25 Thread Peter Marron
Hi,

I've been looking at using libhdfs and I would like to use it on windows.
I have found HDFS-573 and the information on this page:
http://issues.apache.org/jira/browse/HDFS-573?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
which suggests that quite a lot of work was done on this way back in 2009.
So is there some source from this effort retained somewhere? If so, where?
Or do I have to start from scratch? Apologies if this has already been asked 
recently.

Any help appreciated.

Peter Marron


RE: libhdfs on windows

2012-10-25 Thread Peter Marron
Hi Colin,

OK, I didn't know there was a hdfs-dev.
I'm happy to ask there.
(However there's a lot of mail on dev@hadoop
and user@hadoop as well as user@hive
and it's a bit of a commitment to track them all.)
As for webhdfs, I did think about that, and in
some ways it's a beautiful solution as it
gives me a platform- and language-neutral
access mechanism. I was just a little worried
about the HTTP overhead if I am reading a single
record at a time. Also I will need some way to launch
my Map/Reduce jobs as well. So I'll probably
end up using the C++/JNI/Java route to do
that anyway. Unless there's a better way?
Is there a web Map/Reduce interface?

Many thanks,

Z

 -Original Message-
 From: rarecac...@gmail.com
 [mailto:rarecac...@gmail.com] On
 Behalf Of Colin McCabe
 Sent: 25 October 2012 18:24
 To: common-dev@hadoop.apache.org
 Subject: Re: libhdfs on windows
 
 Hi Peter,
 
 This might be a good question for hdfs-
 dev?
 
 As Harsh pointed out below, HDFS-573
 was never committed.  I don't even see a
 patch attached, although there is some
 discussion.
 
 In the mean time, might I suggest using
 the webhdfs interface on Windows?
 webhdfs was intended as a stable REST
 interface that can be accessed from any
 platform.
 
 cheers,
 Colin
 
 
 On Thu, Oct 25, 2012 at 7:19 AM, Peter
 Marron
 peter.mar...@trilliumsoftware.com
 wrote:
  Hi,
 
  I've been looking at using libhdfs and I
 would like to use it on windows.
  I have found HDFS-573 and the
 information on this page:
 
 http://issues.apache.org/jira/browse/HD
 FS-573?page=com.atlassian.jira.
  plugin.system.issuetabpanels:all-
 tabpanel
  which suggests that quite a lot of work
 was done on this way back in 2009.
  So is there some source from this
 effort retained somewhere? If so,
 where?
  Or do I have to start from scratch?
 Apologies if this has already been asked
 recently.
 
  Any help appreciated.
 
  Peter Marron