Hi, I am trying to read data from kafka and process it using spark. i have attached my source code , error log.
For integrating kafka,
i have added dependency in pom.xml
<dependency>
<groupId>org.apache.spark</groupId>
<artifactId>spark-streaming_2.10</artifactId>
<version>1.3.0</version>
</dependency>
<dependency>
<groupId>org.apache.spark</groupId>
<artifactId>spark-streaming-kafka_2.10</artifactId>
<version>1.3.0</version>
</dependency>
i have attached full error log.please check it why it is giving the error
. this class exits in my class path.
I am running spark and kafka locally using java class.
SparkConf conf = new SparkConf().setAppName("Spark Demo").setMaster(
"local[2]").set("spark.executor.memory", "1g");
I
[image: Inline image 2]
spark-error.log
Description: Binary data
kafka-spark.java
Description: Binary data
--------------------------------------------------------------------- To unsubscribe, e-mail: [email protected] For additional commands, e-mail: [email protected]
