Hi Tathagata, I am attaching a snapshot of my pom.xml. It would help immensely, if I can include max, and min values in my mapper phase.
The question is still open at : http://stackoverflow.com/questions/30902090/adding-max-and-min-in-spark-stream-in-java/30909796#30909796 I see that there is a bug report filed for a similar error as well: https://issues.apache.org/jira/browse/SPARK-3266 Please let me know, how I can get the same version of spark streaming in my assembly. I am using the following spark version: http://www.apache.org/dyn/closer.cgi/spark/spark-1.4.0/spark-1.4.0-bin-hadoop2.6.tgz .. no compilation, just an untar and use the spark-submit script in a local install. I still get the same error. Exception in thread "JobGenerator" java.lang.NoSuchMethodError: org.apache.spark.api.java.JavaPairRDD.max(Ljava/util/Comparator;)Lscala/Tuple2; <dependencies> <dependency> <!-- Spark dependency --> <groupId>org.apache.spark</groupId> <artifactId>spark-core_2.10</artifactId> <version>1.4.0</version> </dependency> <dependency> <!-- Spark Streaming dependency --> <groupId>org.apache.spark</groupId> <artifactId>spark-streaming_2.10</artifactId> <version>1.4.0</version> </dependency> Thanks Nipun On Thu, Jun 18, 2015 at 11:16 PM, Nipun Arora <nipunarora2...@gmail.com> wrote: > Hi Tathagata, > > When you say please mark spark-core and spark-streaming as dependencies > how do you mean? > I have installed the pre-build spark-1.4 for Hadoop 2.6 from spark > downloads. In my maven pom.xml, I am using version 1.4 as described. > > Please let me know how I can fix that? > > Thanks > Nipun > > On Thu, Jun 18, 2015 at 4:22 PM, Tathagata Das <t...@databricks.com> > wrote: > >> I think you may be including a different version of Spark Streaming in >> your assembly. Please mark spark-core nd spark-streaming as provided >> dependencies. Any installation of Spark will automatically provide Spark in >> the classpath so you do not have to bundle it. >> >> On Thu, Jun 18, 2015 at 8:44 AM, Nipun Arora <nipunarora2...@gmail.com> >> wrote: >> >>> Hi, >>> >>> I have the following piece of code, where I am trying to transform a >>> spark stream and add min and max to it of eachRDD. However, I get an error >>> saying max call does not exist, at run-time (compiles properly). I am using >>> spark-1.4 >>> >>> I have added the question to stackoverflow as well: >>> http://stackoverflow.com/questions/30902090/adding-max-and-min-in-spark-stream-in-java/30909796#30909796 >>> >>> Any help is greatly appreciated :) >>> >>> Thanks >>> Nipun >>> >>> JavaPairDStream<Tuple2<Long, Integer>, Tuple3<Integer,Long,Long>> >>> sortedtsStream = transformedMaxMintsStream.transformToPair(new Sort2()); >>> >>> sortedtsStream.foreach( >>> new Function<JavaPairRDD<Tuple2<Long, Integer>, Tuple3<Integer, >>> Long, Long>>, Void>() { >>> @Override >>> public Void call(JavaPairRDD<Tuple2<Long, Integer>, >>> Tuple3<Integer, Long, Long>> tuple2Tuple3JavaPairRDD) throws Exception { >>> List<Tuple2<Tuple2<Long, Integer>, >>> Tuple3<Integer,Long,Long>> >templist = tuple2Tuple3JavaPairRDD.collect(); >>> for(Tuple2<Tuple2<Long,Integer>, Tuple3<Integer,Long,Long>> >>> tuple :templist){ >>> >>> Date date = new Date(tuple._1._1); >>> int pattern = tuple._1._2; >>> int count = tuple._2._1(); >>> Date maxDate = new Date(tuple._2._2()); >>> Date minDate = new Date(tuple._2._2()); >>> System.out.println("TimeSlot: " + date.toString() + " >>> Pattern: " + pattern + " Count: " + count + " Max: " + maxDate.toString() + >>> " Min: " + minDate.toString()); >>> >>> } >>> return null; >>> } >>> } >>> ); >>> >>> Error: >>> >>> >>> 15/06/18 11:05:06 INFO BlockManagerInfo: Added input-0-1434639906000 in >>> memory on localhost:42829 (size: 464.0 KB, free: 264.9 MB)15/06/18 11:05:06 >>> INFO BlockGenerator: Pushed block input-0-1434639906000Exception in thread >>> "JobGenerator" java.lang.NoSuchMethodError: >>> org.apache.spark.api.java.JavaPairRDD.max(Ljava/util/Comparator;)Lscala/Tuple2; >>> at >>> org.necla.ngla.spark_streaming.MinMax.call(Type4ViolationChecker.java:346) >>> at >>> org.necla.ngla.spark_streaming.MinMax.call(Type4ViolationChecker.java:340) >>> at >>> org.apache.spark.streaming.api.java.JavaDStreamLike$class.scalaTransform$3(JavaDStreamLike.scala:360) >>> at >>> org.apache.spark.streaming.api.java.JavaDStreamLike$$anonfun$transformToPair$1.apply(JavaDStreamLike.scala:361) >>> at >>> org.apache.spark.streaming.api.java.JavaDStreamLike$$anonfun$transformToPair$1.apply(JavaDStreamLike.scala:361) >>> at >>> org.apache.spark.streaming.dstream.DStream$$anonfun$transform$1$$anonf >>> >>> >> >