[ 
https://issues.apache.org/jira/browse/PHOENIX-2632?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15120348#comment-15120348
 ] 

Josh Elser commented on PHOENIX-2632:
-------------------------------------

[~rgelhau], how do you see something like this getting included into Phoenix? A 
new Maven module that can sit downstream of phoenix-spark, maybe produce some 
uber-jar to make classpath stuff easier from the Phoenix side?

Any possibility to add some end-to-end tests? Such tests would be nice to have 
to help catch future breakages as they happen instead of realizing after a 
release when someone goes to use it.

In general, any tooling that can help get your data into Phoenix seems like a 
valuable addition to me. There are many ways to hammer that nail, but this 
seems like it would be a reasonably general-purpose one to provide.

> Easier Hive->Phoenix data movement
> ----------------------------------
>
>                 Key: PHOENIX-2632
>                 URL: https://issues.apache.org/jira/browse/PHOENIX-2632
>             Project: Phoenix
>          Issue Type: Improvement
>            Reporter: Randy Gelhausen
>
> Moving tables or query results from Hive into Phoenix today requires error 
> prone manual schema re-definition inside HBase storage handler properties. 
> Since Hive and Phoenix support near equivalent types, it should be easier for 
> users to pick a Hive table and load it (or derived query results) from it.
> I'm posting this to open design discussion, but also submit my own project 
> https://github.com/randerzander/HiveToPhoenix for consideration as an early 
> solution. It creates a Spark DataFrame from a Hive query, uses Phoenix JDBC 
> to "create if not exists" a Phoenix equivalent table, and uses the 
> phoenix-spark artifact to store the DataFrame into Phoenix.
> I'm eager to get feedback if this is interesting/useful to the Phoenix 
> community.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Reply via email to