[ https://issues.apache.org/jira/browse/SPARK-21077?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Sean Owen resolved SPARK-21077. ------------------------------- Resolution: Not A Problem > Cannot access public files over S3 protocol > ------------------------------------------- > > Key: SPARK-21077 > URL: https://issues.apache.org/jira/browse/SPARK-21077 > Project: Spark > Issue Type: Bug > Components: EC2 > Affects Versions: 2.1.0 > Environment: Spark 2.1.0 default installation. No existing hadoop, > using the one distributed with Spark. > Added in $SPARK_HOME/jars: > hadoop-aws-2.7.3.jar and aws-java-sdk-1.7.4.jar > Added endpoint configuration in $SPARK_HOME/conf/core-site.xml (I want to > access datasets hosted by organisation with CEPH; follows S3 protocols). > Ubuntu 14.04 x64. > Reporter: Ciprian Tomoiaga > > I am trying to access a dataset with public (anonymous) credentials via the > S3 (or S3a, s3n) protocol. > It fails with the error that no provider in chain can supply the credentials. > I asked our sysadmin to add some dummy credentials, and if I set them up (via > link or config) then I have access. > I tried setting the config : > {code:xml} > <property> > <name>fs.s3a.credentials.provider</name> > <value>org.apache.hadoop.fs.s3a.AnonymousAWSCredentialsProvider</value> > </property> > {code} > but it still doesn't work. > I suggested that it is a java-aws issue > [here|https://github.com/aws/aws-sdk-java/issues/1122#issuecomment-307814540], > but they said it is not. > Any hints on how to use public S3 files from Spark ? -- This message was sent by Atlassian JIRA (v6.4.14#64029) --------------------------------------------------------------------- To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org