[
https://issues.apache.org/jira/browse/HIVE-3844?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13893934#comment-13893934
]
Venki Korukanti commented on HIVE-3844:
---------------------------------------
Updated the patch to use Pattern/Matcher to identify the format of timestamp. I
thought about not using the BigDecimal, but the output has floating point
errors.
Reg: Should we really be allowing numeric formats to be read as timestamp?
I have seen multiple customers that have logs with different timestamp formats
and they want to create one table with timestamp schema and read all sources of
data.
> Unix timestamps don't seem to be read correctly from HDFS as Timestamp column
> -----------------------------------------------------------------------------
>
> Key: HIVE-3844
> URL: https://issues.apache.org/jira/browse/HIVE-3844
> Project: Hive
> Issue Type: Bug
> Components: Serializers/Deserializers
> Affects Versions: 0.8.0
> Reporter: Mark Grover
> Assignee: Venki Korukanti
> Attachments: HIVE-3844.1.patch.txt
>
>
> Serega Shepak pointed out that something like
> {code}
> select cast(date_occurrence as timestamp) from xvlr_data limit 10
> {code}
> where date_occurrence has BIGINT type (timestamp in milliseconds) works. But
> it doesn't work if the declared type is TIMESTAMP on column. The data in the
> date_occurence column in unix timestamp in millis.
--
This message was sent by Atlassian JIRA
(v6.1.5#6160)