Which version of Spark are you trying this with?

On Mon, Jan 27, 2014 at 7:59 PM, robin_up <robin...@gmail.com> wrote:

> Hi
>
> I try to run a small piece of code on Spark Steaming. It sets the s3 keys
> in
> sparkContext object and passed into a sparkStreaming object. However, I got
> the below error -- it seems StreamingContext did not use the hadoop config
> on work threads. It works ok if I run it in spark core (batch mode) without
> streaming.
>
> java.lang.IllegalArgumentException: AWS Access Key ID and Secret Access Key
> must be specified as the username or password (respectively) of a s3n URL,
> or by setting the fs.s3n.awsAccessKeyId or fs.s3n.awsSecretAccessKey
> properties (respectively).
>
>
> //my code:
>
> System.setProperty("spark.cleaner.ttl", "3600")
>         val spark_master = "spark://" + System.getenv("SPARK_MASTER_IP") +
> ":" + System.getenv("SPARK_MASTER_PORT")
>         val external_jars =
>
> Seq("target/scala-2.9.3/test_2.9.3-1.0.jar","/opt/json4s-core_2.9.3-3.2.2.jar","/opt/json4s-native_2.9.3-3.2.2.jar","/opt/json4s-ast_2.9.3-3.2.2.jar")
>
>         val sc = new SparkContext(spark_master, "test",
> System.getenv("SPARK_HOME"), external_jars)
>         sc.hadoopConfiguration.set("fs.s3n.awsAccessKeyId",
> System.getenv("ds_awsAccessKeyId"))
>         sc.hadoopConfiguration.set("fs.s3n.awsSecretAccessKey",
> System.getenv("ds_awsSecretAccessKey"))
>         val ssc = new StreamingContext(sc, Seconds(5))
>
>         val file =
> ssc.textFileStream("s3n://my-bucket/syslog-ng/2014-01-24/")
>
>
>
> -----
> -- Robin Li
> --
> View this message in context:
> http://apache-spark-user-list.1001560.n3.nabble.com/SparkStreaming-not-read-hadoop-configuration-from-its-sparkContext-on-Stand-Alone-mode-tp972.html
> Sent from the Apache Spark User List mailing list archive at Nabble.com.
>

Reply via email to