[jira] [Comment Edited] (SPARK-3274) Spark Streaming Java API reports java.lang.ClassCastException when calling collectAsMap on JavaPairDStream
[ https://issues.apache.org/jira/browse/SPARK-3274?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14151447#comment-14151447 ] Pulkit Bhuwalka edited comment on SPARK-3274 at 9/29/14 7:27 AM: - [~sowen] - you are right. I was making the mistake of reading the sequence file as String instead of Text. Adding toString fixed the problem. Thanks a lot for your help. was (Author: pulkit.bosc...@gmail.com): [~sowen] - you are right. I was making the mistake of reading the sequence file as String instead of text. Addind toString fixed the problem. Thanks a lot for your help. > Spark Streaming Java API reports java.lang.ClassCastException when calling > collectAsMap on JavaPairDStream > -- > > Key: SPARK-3274 > URL: https://issues.apache.org/jira/browse/SPARK-3274 > Project: Spark > Issue Type: Bug > Components: Java API >Affects Versions: 1.0.2 >Reporter: Jack Hu > > Reproduce code: > scontext > .socketTextStream("localhost", 1) > .mapToPair(new PairFunction(){ > public Tuple2 call(String arg0) > throws Exception { > return new Tuple2("1", arg0); > } > }) > .foreachRDD(new Function2, Time, > Void>() { > public Void call(JavaPairRDD v1, Time > v2) throws Exception { > System.out.println(v2.toString() + ": " + > v1.collectAsMap().toString()); > return null; > } > }); > Exception: > java.lang.ClassCastException: [Ljava.lang.Object; cannot be cast to > [Lscala.Tupl > e2; > at > org.apache.spark.rdd.PairRDDFunctions.collectAsMap(PairRDDFunctions.s > cala:447) > at > org.apache.spark.api.java.JavaPairRDD.collectAsMap(JavaPairRDD.scala: > 464) > at tuk.usecase.failedcall.FailedCall$1.call(FailedCall.java:90) > at tuk.usecase.failedcall.FailedCall$1.call(FailedCall.java:88) > at > org.apache.spark.streaming.api.java.JavaDStreamLike$$anonfun$foreachR > DD$2.apply(JavaDStreamLike.scala:282) > at > org.apache.spark.streaming.api.java.JavaDStreamLike$$anonfun$foreachR > DD$2.apply(JavaDStreamLike.scala:282) > at > org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1.apply$mc > V$sp(ForEachDStream.scala:41) > at > org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1.apply(Fo > rEachDStream.scala:40) > at > org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1.apply(Fo > rEachDStream.scala:40) > at scala.util.Try$.apply(Try.scala:161) > at org.apache.spark.streaming.scheduler.Job.run(Job.scala:32) > at > org.apache.spark.streaming.scheduler.JobScheduler$JobHandler.run(JobS -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Commented] (SPARK-3274) Spark Streaming Java API reports java.lang.ClassCastException when calling collectAsMap on JavaPairDStream
[ https://issues.apache.org/jira/browse/SPARK-3274?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14151447#comment-14151447 ] Pulkit Bhuwalka commented on SPARK-3274: [~sowen] - you are right. I was making the mistake of reading the sequence file as String instead of text. Addind toString fixed the problem. Thanks a lot for your help. > Spark Streaming Java API reports java.lang.ClassCastException when calling > collectAsMap on JavaPairDStream > -- > > Key: SPARK-3274 > URL: https://issues.apache.org/jira/browse/SPARK-3274 > Project: Spark > Issue Type: Bug > Components: Java API >Affects Versions: 1.0.2 >Reporter: Jack Hu > > Reproduce code: > scontext > .socketTextStream("localhost", 1) > .mapToPair(new PairFunction(){ > public Tuple2 call(String arg0) > throws Exception { > return new Tuple2("1", arg0); > } > }) > .foreachRDD(new Function2, Time, > Void>() { > public Void call(JavaPairRDD v1, Time > v2) throws Exception { > System.out.println(v2.toString() + ": " + > v1.collectAsMap().toString()); > return null; > } > }); > Exception: > java.lang.ClassCastException: [Ljava.lang.Object; cannot be cast to > [Lscala.Tupl > e2; > at > org.apache.spark.rdd.PairRDDFunctions.collectAsMap(PairRDDFunctions.s > cala:447) > at > org.apache.spark.api.java.JavaPairRDD.collectAsMap(JavaPairRDD.scala: > 464) > at tuk.usecase.failedcall.FailedCall$1.call(FailedCall.java:90) > at tuk.usecase.failedcall.FailedCall$1.call(FailedCall.java:88) > at > org.apache.spark.streaming.api.java.JavaDStreamLike$$anonfun$foreachR > DD$2.apply(JavaDStreamLike.scala:282) > at > org.apache.spark.streaming.api.java.JavaDStreamLike$$anonfun$foreachR > DD$2.apply(JavaDStreamLike.scala:282) > at > org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1.apply$mc > V$sp(ForEachDStream.scala:41) > at > org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1.apply(Fo > rEachDStream.scala:40) > at > org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1.apply(Fo > rEachDStream.scala:40) > at scala.util.Try$.apply(Try.scala:161) > at org.apache.spark.streaming.scheduler.Job.run(Job.scala:32) > at > org.apache.spark.streaming.scheduler.JobScheduler$JobHandler.run(JobS -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Commented] (SPARK-3274) Spark Streaming Java API reports java.lang.ClassCastException when calling collectAsMap on JavaPairDStream
[ https://issues.apache.org/jira/browse/SPARK-3274?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14151233#comment-14151233 ] Pulkit Bhuwalka commented on SPARK-3274: SparkConf sparkConf = new SparkConf().setAppName("Page Rank").setMaster("local[4]"); JavaSparkContext context = new JavaSparkContext(sparkConf); JavaPairRDD transformedLinkMap = context.sequenceFile(pageRankOptions.getFileLocation(), String.class, String.class, 1) .mapToPair(new PairFunction, String, String>() { @Override public Tuple2 call(Tuple2 urlAndLinks) throws Exception { //return new Tuple2(urlAndLinks._1(), urlAndLinks._2()); return new Tuple2( urlAndLinks._1(), new LinkDetails(1.0, new LinkParser().parse(urlAndLinks._2())).toString() ); } }); When I use the commented line above, which simply returns the strings, it works. However, when I use the code after that with LinkDetails which simply parses the string into an object, the code fails with a ClassCastException. java.lang.ClassCastException: org.apache.hadoop.io.Text cannot be cast to java.lang.String at io.pulkit.cmu.acc.project1.phase2.PageRankSparkJob$1.call(PageRankSparkJob.java:28) at io.pulkit.cmu.acc.project1.phase2.PageRankSparkJob$1.call(PageRankSparkJob.java:24) at org.apache.spark.api.java.JavaPairRDD$$anonfun$pairFunToScalaFun$1.apply(JavaPairRDD.scala:926) at org.apache.spark.api.java.JavaPairRDD$$anonfun$pairFunToScalaFun$1.apply(JavaPairRDD.scala:926) at scala.collection.Iterator$$anon$11.next(Iterator.scala:328) at org.apache.spark.util.Utils$.getIteratorSize(Utils.scala:1167) at org.apache.spark.rdd.RDD$$anonfun$count$1.apply(RDD.scala:904) at org.apache.spark.rdd.RDD$$anonfun$count$1.apply(RDD.scala:904) at org.apache.spark.SparkContext$$anonfun$runJob$4.apply(SparkContext.scala:1121) at org.apache.spark.SparkContext$$anonfun$runJob$4.apply(SparkContext.scala:1121) at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:62) at org.apache.spark.scheduler.Task.run(Task.scala:54) at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:177) at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:895) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:918) at java.lang.Thread.run(Thread.java:695) I looked at the other link mentioned. However, the pull request link on that does not work and it is marked as resolved in 0.9. However, I'm using 1.1.0. Thanks a lot. > Spark Streaming Java API reports java.lang.ClassCastException when calling > collectAsMap on JavaPairDStream > -- > > Key: SPARK-3274 > URL: https://issues.apache.org/jira/browse/SPARK-3274 > Project: Spark > Issue Type: Bug > Components: Java API >Affects Versions: 1.0.2 >Reporter: Jack Hu > > Reproduce code: > scontext > .socketTextStream("localhost", 1) > .mapToPair(new PairFunction(){ > public Tuple2 call(String arg0) > throws Exception { > return new Tuple2("1", arg0); > } > }) > .foreachRDD(new Function2, Time, > Void>() { > public Void call(JavaPairRDD v1, Time > v2) throws Exception { > System.out.println(v2.toString() + ": " + > v1.collectAsMap().toString()); > return null; > } > }); > Exception: > java.lang.ClassCastException: [Ljava.lang.Object; cannot be cast to > [Lscala.Tupl > e2; > at > org.apache.spark.rdd.PairRDDFunctions.collectAsMap(PairRDDFunctions.s > cala:447) > at > org.apache.spark.api.java.JavaPairRDD.collectAsMap(JavaPairRDD.scala: > 464) > at tuk.usecase.failedcall.FailedCall$1.call(FailedCall.java:90) > at tuk.usecase.failedcall.FailedCall$1.call(FailedCall.java:88) > at > org.apache.spark.streaming.api.java.JavaDStreamLike$$anonfun$foreachR > DD$2.apply(JavaDStreamLike.scala:282) > at > org.apache.spark.streaming.api.java.JavaDStreamLike$$anonfun$foreachR > DD$2.apply(JavaDStreamLike.scala:282) > at > org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1.apply$mc > V$sp(ForEachDStream.scala:41) > at > org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$
[jira] [Comment Edited] (SPARK-1040) Collect as Map throws a casting exception when run on a JavaPairRDD object
[ https://issues.apache.org/jira/browse/SPARK-1040?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14151231#comment-14151231 ] Pulkit Bhuwalka edited comment on SPARK-1040 at 9/28/14 9:36 PM: - The github link above is not functional anymore. was (Author: pulkit.bosc...@gmail.com): The github link above is not funcitonal anymore. > Collect as Map throws a casting exception when run on a JavaPairRDD object > -- > > Key: SPARK-1040 > URL: https://issues.apache.org/jira/browse/SPARK-1040 > Project: Spark > Issue Type: Bug > Components: Java API >Affects Versions: 0.9.0 >Reporter: Kevin Mader >Assignee: Josh Rosen >Priority: Minor > Fix For: 0.9.1 > > > The error that arises > {code} > Exception in thread "main" java.lang.ClassCastException: [Ljava.lang.Object; > cannot be cast to [Lscala.Tuple2; > at > org.apache.spark.rdd.PairRDDFunctions.collectAsMap(PairRDDFunctions.scala:427) > at > org.apache.spark.api.java.JavaPairRDD.collectAsMap(JavaPairRDD.scala:409) > {code} > The code being executed > {code:java} > public static String ImageSummary(final JavaPairRDD inImg) { > final Set keyList=inImg.collectAsMap().keySet(); > for(Integer cVal: keyList) outString+=cVal+","; > return outString; > } > {code} > The line 426-427 from PairRDDFunctions.scala > {code:java} > def collectAsMap(): Map[K, V] = { > val data = self.toArray() > {code} -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Commented] (SPARK-1040) Collect as Map throws a casting exception when run on a JavaPairRDD object
[ https://issues.apache.org/jira/browse/SPARK-1040?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14151231#comment-14151231 ] Pulkit Bhuwalka commented on SPARK-1040: The github link above is not funcitonal anymore. > Collect as Map throws a casting exception when run on a JavaPairRDD object > -- > > Key: SPARK-1040 > URL: https://issues.apache.org/jira/browse/SPARK-1040 > Project: Spark > Issue Type: Bug > Components: Java API >Affects Versions: 0.9.0 >Reporter: Kevin Mader >Assignee: Josh Rosen >Priority: Minor > Fix For: 0.9.1 > > > The error that arises > {code} > Exception in thread "main" java.lang.ClassCastException: [Ljava.lang.Object; > cannot be cast to [Lscala.Tuple2; > at > org.apache.spark.rdd.PairRDDFunctions.collectAsMap(PairRDDFunctions.scala:427) > at > org.apache.spark.api.java.JavaPairRDD.collectAsMap(JavaPairRDD.scala:409) > {code} > The code being executed > {code:java} > public static String ImageSummary(final JavaPairRDD inImg) { > final Set keyList=inImg.collectAsMap().keySet(); > for(Integer cVal: keyList) outString+=cVal+","; > return outString; > } > {code} > The line 426-427 from PairRDDFunctions.scala > {code:java} > def collectAsMap(): Map[K, V] = { > val data = self.toArray() > {code} -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org