[ https://issues.apache.org/jira/browse/PHOENIX-918?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13964566#comment-13964566 ]
Gabriel Reid commented on PHOENIX-918: -------------------------------------- [~ndimiduk] thanks for pointing out those Hive tickets, I wasn't aware of them. And yes, definitely agreed, bulk loading by HFiles is the "right" way to do it as far as I'm concerned. [~jamestaylor] from your summary of the differences between HCatalog and Phoenix metadata, I get the feeling that you're talking about a more involved use case than I'm thinking of. The use case that I was thinking of was as follows: # User has a table in Hive # User creates a table (possibly the same table definition, or a subset) in Phoenix # User runs a MR-based import to move data from the Hive table into Phoenix, via HFile creation in the same way as the CSV loader. This involves specifying which columns from the HCatalog are to be mapped to which columns in Phoenix. Am I correct in assuming that you had a more automated workflow in mind? > Support importing directly from ORC formatted HDFS data > ------------------------------------------------------- > > Key: PHOENIX-918 > URL: https://issues.apache.org/jira/browse/PHOENIX-918 > Project: Phoenix > Issue Type: Bug > Reporter: James Taylor > > We currently have a good way to import from CSV, but we should also add the > ability to import from HDFS ORC files, as this would likely be common if > folks have Hive data they'd like to import. > [~enis], [~ndimiduk], [~devaraj] - Does this make sense, or is there a > better, existing way? Any takers on implementing it? -- This message was sent by Atlassian JIRA (v6.2#6252)