Hello Modi,

Thanks for the response.

i am running the code via spark-submit command, and i have included
following jars to spark classpath. still getting exception.

hoenix-4.8.0-HBase-1.1-client.jar
phoenix-spark-4.8.0-HBase-1.1.jar
phoenix-core-4.8.0-HBase-1.1.jar

On Mon, Mar 27, 2017 at 10:30 PM, Dhaval Modi <dhavalmod...@gmail.com>
wrote:

> Hi Sateesh,
>
> If you are running from spark shell, then please include Phoenix spark jar
> in classpath.
>
> Kindly refer to url that Sandeep provide.
>
>
> Regards,
> Dhaval
>
>
> On Mar 27, 2017 21:20, "Sateesh Karuturi" <sateesh.karutu...@gmail.com>
> wrote:
>
> Thanks Sandeep for your response.
>
> This is the exception what i am getting:
>
> org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in 
> stage 3.0 failed 4 times, most recent failure: Lost task 0.3 in stage 3.0 
> (TID 411, ip-xxxxx-xx-xxx.ap-southeast-1.compute.internal): 
> java.lang.RuntimeException: java.sql.SQLException: No suitable driver found 
> for jdbc:phoenix:localhost:2181:/hbase-unsecure;
>         at 
> org.apache.phoenix.mapreduce.PhoenixOutputFormat.getRecordWriter(PhoenixOutputFormat.java:58)
>         at 
> org.apache.spark.rdd.PairRDDFunctions$anonfun$saveAsNewAPIHadoopDataset$1$anonfun$12.apply(PairRDDFunctions.scala:1030)
>         at 
> org.apache.spark.rdd.PairRDDFunctions$anonfun$saveAsNewAPIHadoopDataset$1$anonfun$12.apply(PairRDDFunctions.scala:1014)
>         at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:66)
>         at org.apache.spark.scheduler.Task.run(Task.scala:88)
>         at 
> org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
>         at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>         at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>
>
> On Mon, Mar 27, 2017 at 8:17 PM, Sandeep Nemuri <nhsande...@gmail.com>
> wrote:
>
>> What is the error you are seeing ?
>>
>> Ref: https://phoenix.apache.org/phoenix_spark.html
>>
>> df.write \
>>   .format("org.apache.phoenix.spark") \
>>   .mode("overwrite") \
>>   .option("table", "TABLE1") \
>>   .option("zkUrl", "localhost:2181") \
>>   .save()
>>
>>
>>
>> On Mon, Mar 27, 2017 at 10:19 AM, Sateesh Karuturi <
>> sateesh.karutu...@gmail.com> wrote:
>>
>>> Please anyone help me out how to write dataframe to phoenix in java?
>>>
>>> here is my code:
>>>
>>> pos_offer_new_join.write().format("org.apache.phoenix.spark"
>>> ).mode(SaveMode.Overwrite)
>>>
>>>                         .options(ImmutableMap.of("driver",
>>> "org.apache.phoenix.jdbc.PhoenixDriver","zkUrl",
>>>
>>>                         "jdbc:phoenix:localhost:2181","table","RESULT"))
>>>
>>>                         .save();
>>>
>>>
>>> but i am not able to write data to phoenix.
>>>
>>>
>>> Thanks.
>>>
>>>
>>>
>>
>>
>> --
>> *  Regards*
>> *  Sandeep Nemuri*
>>
>
>
>

Reply via email to