Not in the uri, but in the hadoop configuration you can specify it. <property> <name>fs.s3a.endpoint</name> <description>AWS S3 endpoint to connect to. An up-to-date list is provided in the AWS Documentation: regions and endpoints. Without this property, the standard region (s3.amazonaws.com) is assumed. </description> </property>
Thanks Best Regards On Sun, Jul 19, 2015 at 9:13 PM, Schmirr Wurst <schmirrwu...@gmail.com> wrote: > I want to use pithos, were do I can specify that endpoint, is it > possible in the url ? > > 2015-07-19 17:22 GMT+02:00 Akhil Das <ak...@sigmoidanalytics.com>: > > Could you name the Storage service that you are using? Most of them > provides > > a S3 like RestAPI endpoint for you to hit. > > > > Thanks > > Best Regards > > > > On Fri, Jul 17, 2015 at 2:06 PM, Schmirr Wurst <schmirrwu...@gmail.com> > > wrote: > >> > >> Hi, > >> > >> I wonder how to use S3 compatible Storage in Spark ? > >> If I'm using s3n:// url schema, the it will point to amazon, is there > >> a way I can specify the host somewhere ? > >> > >> --------------------------------------------------------------------- > >> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org > >> For additional commands, e-mail: user-h...@spark.apache.org > >> > > > > --------------------------------------------------------------------- > To unsubscribe, e-mail: user-unsubscr...@spark.apache.org > For additional commands, e-mail: user-h...@spark.apache.org > >