Thanks, that is what I was looking for... Any Idea where I have to store and reference the corresponding hadoop-aws-2.6.0.jar ?:
java.io.IOException: No FileSystem for scheme: s3n 2015-07-20 8:33 GMT+02:00 Akhil Das <ak...@sigmoidanalytics.com>: > Not in the uri, but in the hadoop configuration you can specify it. > > <property> > <name>fs.s3a.endpoint</name> > <description>AWS S3 endpoint to connect to. An up-to-date list is > provided in the AWS Documentation: regions and endpoints. Without this > property, the standard region (s3.amazonaws.com) is assumed. > </description> > </property> > > > Thanks > Best Regards > > On Sun, Jul 19, 2015 at 9:13 PM, Schmirr Wurst <schmirrwu...@gmail.com> > wrote: >> >> I want to use pithos, were do I can specify that endpoint, is it >> possible in the url ? >> >> 2015-07-19 17:22 GMT+02:00 Akhil Das <ak...@sigmoidanalytics.com>: >> > Could you name the Storage service that you are using? Most of them >> > provides >> > a S3 like RestAPI endpoint for you to hit. >> > >> > Thanks >> > Best Regards >> > >> > On Fri, Jul 17, 2015 at 2:06 PM, Schmirr Wurst <schmirrwu...@gmail.com> >> > wrote: >> >> >> >> Hi, >> >> >> >> I wonder how to use S3 compatible Storage in Spark ? >> >> If I'm using s3n:// url schema, the it will point to amazon, is there >> >> a way I can specify the host somewhere ? >> >> >> >> --------------------------------------------------------------------- >> >> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org >> >> For additional commands, e-mail: user-h...@spark.apache.org >> >> >> > >> >> --------------------------------------------------------------------- >> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org >> For additional commands, e-mail: user-h...@spark.apache.org >> > --------------------------------------------------------------------- To unsubscribe, e-mail: user-unsubscr...@spark.apache.org For additional commands, e-mail: user-h...@spark.apache.org