[
https://issues.apache.org/jira/browse/SQOOP-1393?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14092544#comment-14092544
]
Richard commented on SQOOP-1393:
--------------------------------
There are 2 solutions:
1. Use Kite to generate hive table, which is similar with function of import
into hdfs as Parquet file (SQOOP-1390). Set the Kite dataset uri like
"dataset:hive?dataset=test&hdfs:host=server-360&hdfs:port:8020" only to
generate hive table directly.
2. use Sqoop to generate hive table, which is similar with function of import
into hive as text file. There are 3 steps. First, import into hdfs as Parquet
file (SQOOP-1390). Second, use Sqoop to create table in hive, like text file.
Third, move Parquet file to hive warehouse (/user/hive/warehouse).
> Import data from database to Hive as Parquet files
> --------------------------------------------------
>
> Key: SQOOP-1393
> URL: https://issues.apache.org/jira/browse/SQOOP-1393
> Project: Sqoop
> Issue Type: Sub-task
> Components: tools
> Reporter: Qian Xu
> Assignee: Richard
>
> Import data to Hive as Parquet file can be separated into two steps:
> 1. Import an individual table from an RDBMS to HDFS as a set of Parquet files.
> 2. Import the data into Hive by generating and executing a CREATE TABLE
> statement to define the data's layout in Hive with Parquet format table
--
This message was sent by Atlassian JIRA
(v6.2#6252)