Connor, Adam, Tim, Great! A JIRA (https://tachyon.atlassian.net/browse/TACHYON) and a PR (Java) to Tachyon repo will be great!
Thanks, Haoyuan On Mon, Aug 25, 2014 at 8:17 AM, Tim St Clair <tstcl...@redhat.com> wrote: > Conner - > > Awesome! > > +1 re: java + PR to tachyon repo > > I should have some time later this week to eval, I'm curious how you > handled the mounting details. > > Cheers, > Tim > > ------------------------------ > > *From: *"Connor Doyle" <connor....@gmail.com> > *To: *user@mesos.apache.org > *Cc: *"Haoyuan Li" <haoyuan...@gmail.com>, "Huamin Chen" <hc...@redhat.com>, > "Brad Childs" <bchi...@redhat.com>, "Adam Bordelon" <a...@mesosphere.io> > *Sent: *Sunday, August 24, 2014 6:46:46 PM > *Subject: *Re: Alternate HDFS Filesystems + Hadoop on Mesos > > > Also, fwiw I'm interested in rallying folks on a Tachyon Framework in the > not-too-distant future, for anyone who is interested. Probably follow the > spark model and try to push upstream. > > > Hi Tim, late follow-up: > > The not-too distant future is here! Adam and I took a stab at a Tachyon > framework during the MesosCon hackathon ( > http://github.com/mesosphere/tachyon-mesos). > We started writing in Scala, but not at all opposed to switching to Java, > especially if the work can be upstreamed. > -- > Connor > > > > On Fri, Aug 15, 2014 at 5:16 PM, John Omernik <j...@omernik.com> wrote: > >> I tried hdfs:/// and hdfs://cldbnode:7222/ Neither worked (examples >> below) I really think the hdfs vs other prefixes should be looked at. Like >> I said above, the tachyon project just added a env variable to address >> this. >> >> >> >> hdfs://cldbnode:7222/ >> >> WARNING: Logging before InitGoogleLogging() is written to STDERR >> I0815 19:14:17.101666 22022 fetcher.cpp:76] Fetching URI >> 'hdfs://hadoopmapr1:7222/mesos/hadoop-0.20.2-mapr-4.0.0.tgz' >> I0815 19:14:17.101780 22022 fetcher.cpp:105] Downloading resource from >> 'hdfs://hadoopmapr1:7222/mesos/hadoop-0.20.2-mapr-4.0.0.tgz' to >> '/tmp/mesos/slaves/20140815-103603-1677764800-5050-24315-2/frameworks/20140815-154511-1677764800-5050-7162-0003/executors/executor_Task_Tracker_5/runs/b3174e72-75ea-48be-bbb8-a9a6cc605018/hadoop-0.20.2-mapr-4.0.0.tgz' >> E0815 19:14:17.778833 22022 fetcher.cpp:109] HDFS copyToLocal failed: hadoop >> fs -copyToLocal 'hdfs://hadoopmapr1:7222/mesos/hadoop-0.20.2-mapr-4.0.0.tgz' >> '/tmp/mesos/slaves/20140815-103603-1677764800-5050-24315-2/frameworks/20140815-154511-1677764800-5050-7162-0003/executors/executor_Task_Tracker_5/runs/b3174e72-75ea-48be-bbb8-a9a6cc605018/hadoop-0.20.2-mapr-4.0.0.tgz' >> WARNING: org.apache.hadoop.metrics.jvm.EventCounter is deprecated. Please >> use org.apache.hadoop.log.metrics.EventCounter in all the log4j.properties >> files. >> -copyToLocal: Wrong FS: >> maprfs://hadoopmapr1:7222/mesos/hadoop-0.20.2-mapr-4.0.0.tgz, expected: >> hdfs://hadoopmapr1:7222/mesos/hadoop-0.20.2-mapr-4.0.0.tgz >> Usage: hadoop fs [generic options] -copyToLocal [-p] [-ignoreCrc] [-crc] >> <src> ... <localdst> >> Failed to fetch: hdfs://hadoopmapr1:7222/mesos/hadoop-0.20.2-mapr-4.0.0.tgz >> Failed to synchronize with slave (it's probably exited) >> >> >> >> >> hdfs:/// >> >> >> I0815 19:10:45.006803 21508 fetcher.cpp:76] Fetching URI >> 'hdfs:///mesos/hadoop-0.20.2-mapr-4.0.0.tgz' >> I0815 19:10:45.007099 21508 fetcher.cpp:105] Downloading resource from >> 'hdfs:///mesos/hadoop-0.20.2-mapr-4.0.0.tgz' to >> '/tmp/mesos/slaves/20140815-103603-1677764800-5050-24315-2/frameworks/20140815-154511-1677764800-5050-7162-0002/executors/executor_Task_Tracker_2/runs/22689054-aff6-4f7c-9746-a068a11ff000/hadoop-0.20.2-mapr-4.0.0.tgz' >> E0815 19:10:45.681922 21508 fetcher.cpp:109] HDFS copyToLocal failed: hadoop >> fs -copyToLocal 'hdfs:///mesos/hadoop-0.20.2-mapr-4.0.0.tgz' >> '/tmp/mesos/slaves/20140815-103603-1677764800-5050-24315-2/frameworks/20140815-154511-1677764800-5050-7162-0002/executors/executor_Task_Tracker_2/runs/22689054-aff6-4f7c-9746-a068a11ff000/hadoop-0.20.2-mapr-4.0.0.tgz' >> WARNING: org.apache.hadoop.metrics.jvm.EventCounter is deprecated. Please >> use org.apache.hadoop.log.metrics.EventCounter in all the log4j.properties >> files. >> -copyToLocal: Wrong FS: maprfs:/mesos/hadoop-0.20.2-mapr-4.0.0.tgz, >> expected: hdfs:///mesos/hadoop-0.20.2-mapr-4.0.0.tgz >> Usage: hadoop fs [generic options] -copyToLocal [-p] [-ignoreCrc] [-crc] >> <src> ... <localdst> >> Failed to fetch: hdfs:///mesos/hadoop-0.20.2-mapr-4.0.0.tgz >> Failed to synchronize with slave (it's probably exited) >> >> >> >> On Fri, Aug 15, 2014 at 5:38 PM, John Omernik <j...@omernik.com> wrote: >> >>> I am away from my cluster right now, I trued doing a hadoop fs -ls >>> maprfs:// and that worked. When I tries hadoop fs -ls hdfs:/// it failed >>> with wrong fs type. With that error I didn't try it in the mapred-site. I >>> will try it. Still...why hard code the file prefixes? I guess I am curious >>> on how glusterfs would work, or others as they pop up. >>> On Aug 15, 2014 5:04 PM, "Adam Bordelon" <a...@mesosphere.io> wrote: >>> >>>> Can't you just use the hdfs:// protocol for maprfs? That should work >>>> just fine. >>>> >>>> >>>> On Fri, Aug 15, 2014 at 2:50 PM, John Omernik <j...@omernik.com> wrote: >>>> >>>>> Thanks all. >>>>> >>>>> I realized MapR has a work around for me that I will try soon in that >>>>> I have MapR fs NFS mounted on each node, I.e. I should be able to get the >>>>> tar from there. >>>>> >>>>> That said, perhaps someone with better coding skills than me could >>>>> provide an env variable where a user could provide the HDFS prefixes to >>>>> try. I know we did that with the tachyon project and it works well for >>>>> other HDFS compatible fs implementations, perhaps that would work here? >>>>> Hard coding a pluggable system seems like a long term issue that will keep >>>>> coming up. >>>>> On Aug 15, 2014 4:02 PM, "Tim St Clair" <tstcl...@redhat.com> wrote: >>>>> >>>>>> The uri doesn't currently start with any of the known types (at least >>>>>> on 1st grok). >>>>>> You could redirect via a proxy that does the job for you. >>>>>> >>>>>> | if you had some fuse mount that would work too. >>>>>> >>>>>> Cheers, >>>>>> Tim >>>>>> >>>>>> ------------------------------ >>>>>> >>>>>> *From: *"John Omernik" <j...@omernik.com> >>>>>> *To: *user@mesos.apache.org >>>>>> *Sent: *Friday, August 15, 2014 3:55:02 PM >>>>>> *Subject: *Alternate HDFS Filesystems + Hadoop on Mesos >>>>>> >>>>>> I am on a wonderful journey trying to get hadoop on Mesos working >>>>>> with MapR. I feel like I am close, but when the slaves try to run the >>>>>> packaged Hadoop, I get the error below. The odd thing is, I KNOW I got >>>>>> Spark running on Mesos pulling both data and the packages from MapRFS. >>>>>> So >>>>>> I am confused why there is and issue with the fetcher.cpp here. Granted, >>>>>> when I got spark working, it was on 0.19.0, and I am trying a "fresh" >>>>>> version from git (0.20.0?) that I just pulled today. I am not sure if >>>>>> that >>>>>> work, but when I have more time I will try spark again. >>>>>> >>>>>> Any thoughts on this error? Thanks. >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> Error: >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> WARNING: Logging before InitGoogleLogging() is written to STDERR >>>>>> I0815 15:48:35.446071 20636 fetcher.cpp:76] Fetching URI >>>>>> 'maprfs:///mesos/hadoop-0.20.2-mapr-4.0.0.tgz' >>>>>> E0815 15:48:35.446184 20636 fetcher.cpp:161] A relative path was passed >>>>>> for the resource but the environment variable MESOS_FRAMEWORKS_HOME is >>>>>> not set. Please either specify this config option or avoid using a >>>>>> relative path >>>>>> Failed to fetch: maprfs:///mesos/hadoop-0.20.2-mapr-4.0.0.tgz >>>>>> Failed to synchronize with slave (it's probably exited) >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> -- >>>>>> Cheers, >>>>>> Timothy St. Clair >>>>>> Red Hat Inc. >>>>>> >>>>> >>>> >> > > > > -- > Cheers, > Timothy St. Clair > Red Hat Inc. > > > > > -- > Cheers, > Timothy St. Clair > Red Hat Inc. > -- Haoyuan Li AMPLab, EECS, UC Berkeley http://www.cs.berkeley.edu/~haoyuan/