I tried hdfs:/// and hdfs://cldbnode:7222/ Neither worked (examples below)
I really think the hdfs vs other prefixes should be looked at. Like I said
above, the tachyon project just added a env variable to address this.



hdfs://cldbnode:7222/

WARNING: Logging before InitGoogleLogging() is written to STDERR
I0815 19:14:17.101666 22022 fetcher.cpp:76] Fetching URI
'hdfs://hadoopmapr1:7222/mesos/hadoop-0.20.2-mapr-4.0.0.tgz'
I0815 19:14:17.101780 22022 fetcher.cpp:105] Downloading resource from
'hdfs://hadoopmapr1:7222/mesos/hadoop-0.20.2-mapr-4.0.0.tgz' to
'/tmp/mesos/slaves/20140815-103603-1677764800-5050-24315-2/frameworks/20140815-154511-1677764800-5050-7162-0003/executors/executor_Task_Tracker_5/runs/b3174e72-75ea-48be-bbb8-a9a6cc605018/hadoop-0.20.2-mapr-4.0.0.tgz'
E0815 19:14:17.778833 22022 fetcher.cpp:109] HDFS copyToLocal failed:
hadoop fs -copyToLocal
'hdfs://hadoopmapr1:7222/mesos/hadoop-0.20.2-mapr-4.0.0.tgz'
'/tmp/mesos/slaves/20140815-103603-1677764800-5050-24315-2/frameworks/20140815-154511-1677764800-5050-7162-0003/executors/executor_Task_Tracker_5/runs/b3174e72-75ea-48be-bbb8-a9a6cc605018/hadoop-0.20.2-mapr-4.0.0.tgz'
WARNING: org.apache.hadoop.metrics.jvm.EventCounter is deprecated.
Please use org.apache.hadoop.log.metrics.EventCounter in all the
log4j.properties files.
-copyToLocal: Wrong FS:
maprfs://hadoopmapr1:7222/mesos/hadoop-0.20.2-mapr-4.0.0.tgz,
expected: hdfs://hadoopmapr1:7222/mesos/hadoop-0.20.2-mapr-4.0.0.tgz
Usage: hadoop fs [generic options] -copyToLocal [-p] [-ignoreCrc]
[-crc] <src> ... <localdst>
Failed to fetch: hdfs://hadoopmapr1:7222/mesos/hadoop-0.20.2-mapr-4.0.0.tgz
Failed to synchronize with slave (it's probably exited)




hdfs:///


I0815 19:10:45.006803 21508 fetcher.cpp:76] Fetching URI
'hdfs:///mesos/hadoop-0.20.2-mapr-4.0.0.tgz'
I0815 19:10:45.007099 21508 fetcher.cpp:105] Downloading resource from
'hdfs:///mesos/hadoop-0.20.2-mapr-4.0.0.tgz' to
'/tmp/mesos/slaves/20140815-103603-1677764800-5050-24315-2/frameworks/20140815-154511-1677764800-5050-7162-0002/executors/executor_Task_Tracker_2/runs/22689054-aff6-4f7c-9746-a068a11ff000/hadoop-0.20.2-mapr-4.0.0.tgz'
E0815 19:10:45.681922 21508 fetcher.cpp:109] HDFS copyToLocal failed:
hadoop fs -copyToLocal 'hdfs:///mesos/hadoop-0.20.2-mapr-4.0.0.tgz'
'/tmp/mesos/slaves/20140815-103603-1677764800-5050-24315-2/frameworks/20140815-154511-1677764800-5050-7162-0002/executors/executor_Task_Tracker_2/runs/22689054-aff6-4f7c-9746-a068a11ff000/hadoop-0.20.2-mapr-4.0.0.tgz'
WARNING: org.apache.hadoop.metrics.jvm.EventCounter is deprecated.
Please use org.apache.hadoop.log.metrics.EventCounter in all the
log4j.properties files.
-copyToLocal: Wrong FS: maprfs:/mesos/hadoop-0.20.2-mapr-4.0.0.tgz,
expected: hdfs:///mesos/hadoop-0.20.2-mapr-4.0.0.tgz
Usage: hadoop fs [generic options] -copyToLocal [-p] [-ignoreCrc]
[-crc] <src> ... <localdst>
Failed to fetch: hdfs:///mesos/hadoop-0.20.2-mapr-4.0.0.tgz
Failed to synchronize with slave (it's probably exited)



On Fri, Aug 15, 2014 at 5:38 PM, John Omernik <j...@omernik.com> wrote:

> I am away from my cluster right now, I trued doing a hadoop fs -ls
> maprfs:// and that worked.   When I tries hadoop fs -ls hdfs:/// it failed
> with wrong fs type.  With that error I didn't try it in the mapred-site.  I
> will try it.  Still...why hard code the file prefixes? I guess I am curious
> on how glusterfs would work, or others as they pop up.
>  On Aug 15, 2014 5:04 PM, "Adam Bordelon" <a...@mesosphere.io> wrote:
>
>> Can't you just use the hdfs:// protocol for maprfs? That should work just
>> fine.
>>
>>
>> On Fri, Aug 15, 2014 at 2:50 PM, John Omernik <j...@omernik.com> wrote:
>>
>>> Thanks all.
>>>
>>> I realized MapR has a work around for me that I will try soon in that I
>>> have MapR fs NFS mounted on each node, I.e. I should be able to get the tar
>>> from there.
>>>
>>> That said, perhaps someone with better coding skills than me could
>>> provide an env variable where a user could provide the HDFS prefixes to
>>> try. I know we did that with the tachyon project and it works well for
>>> other HDFS compatible fs implementations, perhaps that would work here?
>>> Hard coding a pluggable system seems like a long term issue that will keep
>>> coming up.
>>>  On Aug 15, 2014 4:02 PM, "Tim St Clair" <tstcl...@redhat.com> wrote:
>>>
>>>> The uri doesn't currently start with any of the known types (at least
>>>> on 1st grok).
>>>> You could redirect via a proxy that does the job for you.
>>>>
>>>> | if you had some fuse mount that would work too.
>>>>
>>>> Cheers,
>>>> Tim
>>>>
>>>> ------------------------------
>>>>
>>>> *From: *"John Omernik" <j...@omernik.com>
>>>> *To: *user@mesos.apache.org
>>>> *Sent: *Friday, August 15, 2014 3:55:02 PM
>>>> *Subject: *Alternate HDFS Filesystems + Hadoop on Mesos
>>>>
>>>> I am on a wonderful journey trying to get hadoop on Mesos working with
>>>> MapR.   I feel like I am close, but when the slaves try to run the packaged
>>>> Hadoop, I get the error below.  The odd thing is,  I KNOW I got Spark
>>>> running on Mesos pulling both data and the packages from MapRFS.  So I am
>>>> confused why there is and issue with the fetcher.cpp here. Granted, when I
>>>> got spark working, it was on 0.19.0, and I am trying a "fresh" version from
>>>> git (0.20.0?) that I just pulled today. I am not sure if that work, but
>>>> when I have more time I will try spark again.
>>>>
>>>> Any thoughts on this error? Thanks.
>>>>
>>>>
>>>>
>>>>
>>>> Error:
>>>>
>>>>
>>>>
>>>>
>>>>
>>>> WARNING: Logging before InitGoogleLogging() is written to STDERR
>>>> I0815 15:48:35.446071 20636 fetcher.cpp:76] Fetching URI 
>>>> 'maprfs:///mesos/hadoop-0.20.2-mapr-4.0.0.tgz'
>>>> E0815 15:48:35.446184 20636 fetcher.cpp:161] A relative path was passed 
>>>> for the resource but the environment variable MESOS_FRAMEWORKS_HOME is not 
>>>> set. Please either specify this config option or avoid using a relative 
>>>> path
>>>> Failed to fetch: maprfs:///mesos/hadoop-0.20.2-mapr-4.0.0.tgz
>>>> Failed to synchronize with slave (it's probably exited)
>>>>
>>>>
>>>>
>>>>
>>>> --
>>>> Cheers,
>>>> Timothy St. Clair
>>>> Red Hat Inc.
>>>>
>>>
>>

Reply via email to