[
https://issues.apache.org/jira/browse/PHOENIX-2632?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15120500#comment-15120500
]
Josh Mahonin commented on PHOENIX-2632:
---------------------------------------
This looks pretty neat [~rgelhau]
I bet there's a way to take your 'CREATE TABLE IF NOT EXISTS' functionality
could be wrapped into the existing Spark DataFrame code, and be made to use for
the SaveMode.Ignore option [1]. Right now it only supports SaveMode.Overwrite,
which assumes the table is setup already.
Once that's in, I think the Hive->Phoenix functionality becomes a documentation
exercise: show to to setup the Hive table as a DataFrame, then invoke
df.save("org.apache.phoenix.spark"...) on it.
[1] http://spark.apache.org/docs/latest/sql-programming-guide.html
> Easier Hive->Phoenix data movement
> ----------------------------------
>
> Key: PHOENIX-2632
> URL: https://issues.apache.org/jira/browse/PHOENIX-2632
> Project: Phoenix
> Issue Type: Improvement
> Reporter: Randy Gelhausen
>
> Moving tables or query results from Hive into Phoenix today requires error
> prone manual schema re-definition inside HBase storage handler properties.
> Since Hive and Phoenix support near equivalent types, it should be easier for
> users to pick a Hive table and load it (or derived query results) from it.
> I'm posting this to open design discussion, but also submit my own project
> https://github.com/randerzander/HiveToPhoenix for consideration as an early
> solution. It creates a Spark DataFrame from a Hive query, uses Phoenix JDBC
> to "create if not exists" a Phoenix equivalent table, and uses the
> phoenix-spark artifact to store the DataFrame into Phoenix.
> I'm eager to get feedback if this is interesting/useful to the Phoenix
> community.
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)