[ 
https://issues.apache.org/jira/browse/SPARK-21077?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean Owen resolved SPARK-21077.
-------------------------------
    Resolution: Not A Problem

> Cannot access public files over S3 protocol
> -------------------------------------------
>
>                 Key: SPARK-21077
>                 URL: https://issues.apache.org/jira/browse/SPARK-21077
>             Project: Spark
>          Issue Type: Bug
>          Components: EC2
>    Affects Versions: 2.1.0
>         Environment: Spark 2.1.0 default installation. No existing hadoop, 
> using the one distributed with Spark.
> Added in $SPARK_HOME/jars:  
> hadoop-aws-2.7.3.jar and aws-java-sdk-1.7.4.jar
> Added endpoint configuration in $SPARK_HOME/conf/core-site.xml (I want to 
> access datasets hosted by organisation with CEPH; follows S3 protocols).
> Ubuntu 14.04 x64.
>            Reporter: Ciprian Tomoiaga
>
> I am trying to access a dataset with public (anonymous) credentials via the 
> S3 (or S3a, s3n) protocol. 
> It fails with the error that no provider in chain can supply the credentials.
> I asked our sysadmin to add some dummy credentials, and if I set them up (via 
> link or config) then I have access.
> I tried setting the config :
> {code:xml}
> <property>
>   <name>fs.s3a.credentials.provider</name>
>   <value>org.apache.hadoop.fs.s3a.AnonymousAWSCredentialsProvider</value>
> </property>
> {code}
> but it still doesn't work.
> I suggested that it is a java-aws issue 
> [here|https://github.com/aws/aws-sdk-java/issues/1122#issuecomment-307814540],
>  but they said it is not.
> Any hints on how to use public S3 files from Spark ?



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org

Reply via email to