Hi,
I Need to check below points before concluding on it. If you already have
information on this, you can provide me.

1. About hive and spark default format; some place they mention upto 9
decimal precision. you mentioned 3 decimal precision.
so, which file of hive and spark has this default value ?
2. Why current test cases are not failing when we compare query results
from hive and carbon for timestamp column ?
3. Also after we change it, how many testcase need to modify? [because the
validation value may not match] ?

*As this value is configurable.* I am *neutral* about the changes proposed
If the effort is high.

Thanks,
Ajantha

On Tue, Jul 14, 2020 at 2:05 PM haomarch <marchp...@126.com> wrote:

> Spark's default TimeStampFormat  is yyyy-mm-dd hh:mm:ss.SSS
> CarbonData shall keep consistent with Spark.
>
>
>
> --
> Sent from:
> http://apache-carbondata-dev-mailing-list-archive.1130556.n5.nabble.com/
>

Reply via email to