I tried the patch; it resolves this issue. Thanks.

On Wed, Sep 23, 2015 at 11:23 AM Josh Mahonin <jmaho...@interset.com> wrote:

> I've got a patch attached to the ticket that I think should fix your
> issue.
>
> If you're able to try it out and let us know how it goes, it'd be much
> appreciated.
>
> From: Babar Tareen
> Reply-To: "user@phoenix.apache.org"
> Date: Wednesday, September 23, 2015 at 1:14 PM
> To: "user@phoenix.apache.org"
> Subject: Re: Spark Plugin Exception - java.lang.ClassCastException:
> org.apache.spark.sql.catalyst.expressions.GenericMutableRow cannot be cast
> to org.apache.spark.sql.Row
>
> I have filed PHOENIX-2287
> <https://issues.apache.org/jira/browse/PHOENIX-2287> for this. And the
> code works fine with Spark 1.4.1.
>
> Thanks
>
> On Wed, Sep 23, 2015 at 6:06 AM Josh Mahonin <jmaho...@interset.com>
> wrote:
>
>> Hi Babar,
>>
>> Can you file a JIRA for this? I suspect this is something to do with the
>> Spark 1.5 data frame API data structures, perhaps they've gone and changed
>> them again!
>>
>> Can you try with previous Spark versions to see if there's a difference?
>> Also, you may have luck interfacing with the RDDs directly instead of the
>> data frames.
>>
>> Thanks!
>>
>> Josh
>>
>> From: Babar Tareen
>> Reply-To: "user@phoenix.apache.org"
>> Date: Tuesday, September 22, 2015 at 5:47 PM
>> To: "user@phoenix.apache.org"
>> Subject: Spark Plugin Exception - java.lang.ClassCastException:
>> org.apache.spark.sql.catalyst.expressions.GenericMutableRow cannot be cast
>> to org.apache.spark.sql.Row
>>
>> Hi,
>>
>> I am trying to run the spark plugin DataFrame sample code available here (
>> https://phoenix.apache.org/phoenix_spark.html) and getting following
>> exception.  I am running the code against hbase-1.1.1, spark 1.5.0 and
>> phoenix  4.5.2. HBase is running in standalone mode, locally on OS X.  Any
>> ideas what might be causing this exception?
>>
>>
>> java.lang.ClassCastException:
>> org.apache.spark.sql.catalyst.expressions.GenericMutableRow cannot be cast
>> to org.apache.spark.sql.Row
>>     at
>> org.apache.spark.sql.SQLContext$$anonfun$7.apply(SQLContext.scala:439)
>> ~[spark-sql_2.11-1.5.0.jar:1.5.0]
>>     at scala.collection.Iterator$$anon$11.next(Iterator.scala:363)
>> ~[scala-library-2.11.4.jar:na]
>>     at scala.collection.Iterator$$anon$11.next(Iterator.scala:363)
>> ~[scala-library-2.11.4.jar:na]
>>     at scala.collection.Iterator$$anon$11.next(Iterator.scala:363)
>> ~[scala-library-2.11.4.jar:na]
>>     at
>> org.apache.spark.sql.execution.aggregate.TungstenAggregationIterator.processInputs(TungstenAggregationIterator.scala:366)
>> ~[spark-sql_2.11-1.5.0.jar:1.5.0]
>>     at
>> org.apache.spark.sql.execution.aggregate.TungstenAggregationIterator.start(TungstenAggregationIterator.scala:622)
>> ~[spark-sql_2.11-1.5.0.jar:1.5.0]
>>     at
>> org.apache.spark.sql.execution.aggregate.TungstenAggregate$$anonfun$doExecute$
>> 1.org$apache$spark$sql$execution$aggregate$TungstenAggregate$$anonfun$$executePartition$1(TungstenAggregate.scala:110)
>> ~[spark-sql_2.11-1.5.0.jar:1.5.0]
>>     at
>> org.apache.spark.sql.execution.aggregate.TungstenAggregate$$anonfun$doExecute$1$$anonfun$2.apply(TungstenAggregate.scala:119)
>> ~[spark-sql_2.11-1.5.0.jar:1.5.0]
>>     at
>> org.apache.spark.sql.execution.aggregate.TungstenAggregate$$anonfun$doExecute$1$$anonfun$2.apply(TungstenAggregate.scala:119)
>> ~[spark-sql_2.11-1.5.0.jar:1.5.0]
>>     at
>> org.apache.spark.rdd.MapPartitionsWithPreparationRDD.compute(MapPartitionsWithPreparationRDD.scala:64)
>> ~[spark-core_2.11-1.5.0.jar:1.5.0]
>>     at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:297)
>> ~[spark-core_2.11-1.5.0.jar:1.5.0]
>>     at org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
>> ~[spark-core_2.11-1.5.0.jar:1.5.0]
>>     at
>> org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
>> ~[spark-core_2.11-1.5.0.jar:1.5.0]
>>     at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:297)
>> ~[spark-core_2.11-1.5.0.jar:1.5.0]
>>     at org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
>> ~[spark-core_2.11-1.5.0.jar:1.5.0]
>>     at
>> org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:73)
>> ~[spark-core_2.11-1.5.0.jar:1.5.0]
>>     at
>> org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)
>> ~[spark-core_2.11-1.5.0.jar:1.5.0]
>>     at org.apache.spark.scheduler.Task.run(Task.scala:88)
>> ~[spark-core_2.11-1.5.0.jar:1.5.0]
>>     at
>> org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
>> ~[spark-core_2.11-1.5.0.jar:1.5.0]
>>     at
>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
>> [na:1.8.0_45]
>>     at
>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
>> [na:1.8.0_45]
>>     at java.lang.Thread.run(Thread.java:745) [na:1.8.0_45]
>>
>> Thanks,
>> Babar
>>
>

Reply via email to