to Josephs point… hdfs and s3 challenges are dcos issues not a mesos issue.     
We do however need Mesos to support custom protocols for the fetcher.   At our 
current pace of releases it sounds not too far away.

ken
> On May 10, 2016, at 2:20 PM, Joseph Wu <jos...@mesosphere.io> wrote:
> 
> Mesos does not explicitly support HDFS and S3.  Rather, Mesos will assume you 
> have a hadoop binary and use it (blindly) for certain types of URIs.  If the 
> hadoop binary is not present, the mesos-fetcher will fail to fetch your HDFS 
> or S3 URIs.
> 
> Mesos does not ship/package hadoop, so these URIs are not expected to work 
> out of the box (for plain Mesos distributions).  In all cases, the operator 
> must preconfigure hadoop on each node (similar to how Docker in Mesos works).
> 
> Here's the epic tracking the modularization of the mesos-fetcher (I estimate 
> it'll be done by 0.30):
> https://issues.apache.org/jira/browse/MESOS-3918 
> <https://issues.apache.org/jira/browse/MESOS-3918>
> 
> ^ Once done, it should be easier to plug in more fetchers, such as one for 
> your use-case.
> 
> On Tue, May 10, 2016 at 11:21 AM, Briant, James 
> <james.bri...@thermofisher.com <mailto:james.bri...@thermofisher.com>> wrote:
> I’m happy to have default IAM role on the box that can read-only fetch from 
> my s3 bucket. s3a gets the credentials from AWS instance metadata. It works.
> 
> If hadoop is gone, does that mean that hfds: URIs don’t work either?
> 
> Are you saying dcos and mesos are diverging? Mesos explicitly supports hdfs 
> and s3.
> 
> In the absence of S3, how do you propose I make large binaries available to 
> my cluster, and only to my cluster, on AWS?
> 
> Jamie
> 
> From: Cody Maloney <c...@mesosphere.io <mailto:c...@mesosphere.io>>
> Reply-To: "user@mesos.apache.org <mailto:user@mesos.apache.org>" 
> <user@mesos.apache.org <mailto:user@mesos.apache.org>>
> Date: Tuesday, May 10, 2016 at 10:58 AM
> To: "user@mesos.apache.org <mailto:user@mesos.apache.org>" 
> <user@mesos.apache.org <mailto:user@mesos.apache.org>>
> Subject: Re: Enable s3a for fetcher
> 
> The s3 fetcher stuff inside of DC/OS is not supported. The `hadoop` binary 
> has been entirely removed from DC/OS 1.8 already. There have been various 
> proposals to make it so the mesos fetcher is much more pluggable / extensible 
> (https://issues.apache.org/jira/browse/MESOS-2731 
> <https://issues.apache.org/jira/browse/MESOS-2731> for instance). 
> 
> Generally speaking people want a lot of different sorts of fetching, and 
> there are all sorts of questions of how to properly get auth to the various 
> chunks (if you're using s3a:// presumably you need to get credentials there 
> somehow. Otherwise you could just use http://). Need to design / build that 
> into Mesos and DC/OS to be able to use this stuff.
> 
> Cody
> 
> On Tue, May 10, 2016 at 9:55 AM Briant, James <james.bri...@thermofisher.com 
> <mailto:james.bri...@thermofisher.com>> wrote:
> I want to use s3a: urls in fetcher. I’m using dcos 1.7 which has hadoop 2.5 
> on its agents. This version has the necessary hadoop-aws and aws-sdk:
> 
> hadoop--afadb46fe64d0ee7ce23dbe769e44bfb0767a8b9]$ ls 
> usr/share/hadoop/tools/lib/ | grep aws
> aws-java-sdk-1.7.4.jar
> hadoop-aws-2.5.0-cdh5.3.3.jar
> 
> What config/scripts do I need to hack to get these guys on the classpath so 
> that "hadoop fs -copyToLocal” works?
> 
> Thanks,
> Jamie
> 

Reply via email to