[jira] [Updated] (SPARK-32021) make_interval does not accept seconds >100
[ https://issues.apache.org/jira/browse/SPARK-32021?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Dongjoon Hyun updated SPARK-32021: -- Fix Version/s: (was: 3.0.0) 3.0.1 > make_interval does not accept seconds >100 > -- > > Key: SPARK-32021 > URL: https://issues.apache.org/jira/browse/SPARK-32021 > Project: Spark > Issue Type: Bug > Components: SQL >Affects Versions: 3.0.0 >Reporter: Juliusz Sompolski >Assignee: Maxim Gekk >Priority: Major > Fix For: 3.0.1, 3.1.0 > > > In make_interval(years, months, weeks, days, hours, mins, secs), secs are > defined as Decimal(8, 6), which turns into null if the value of the > expression overflows 100 seconds. > Larger seconds values should be allowed. > This has been reported by Simba, who wants to use make_interval to implement > translation for TIMESTAMP_ADD ODBC function in Spark 3.0. > ODBC {fn TIMESTAMPADD(SECOND, integer_exp, timestamp} fails when integer_exp > returns seconds values >= 100. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Updated] (SPARK-32021) make_interval does not accept seconds >100
[ https://issues.apache.org/jira/browse/SPARK-32021?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Dongjoon Hyun updated SPARK-32021: -- Fix Version/s: 3.0.0 > make_interval does not accept seconds >100 > -- > > Key: SPARK-32021 > URL: https://issues.apache.org/jira/browse/SPARK-32021 > Project: Spark > Issue Type: Bug > Components: SQL >Affects Versions: 3.0.0 >Reporter: Juliusz Sompolski >Assignee: Maxim Gekk >Priority: Major > Fix For: 3.0.0, 3.1.0 > > > In make_interval(years, months, weeks, days, hours, mins, secs), secs are > defined as Decimal(8, 6), which turns into null if the value of the > expression overflows 100 seconds. > Larger seconds values should be allowed. > This has been reported by Simba, who wants to use make_interval to implement > translation for TIMESTAMP_ADD ODBC function in Spark 3.0. > ODBC {fn TIMESTAMPADD(SECOND, integer_exp, timestamp} fails when integer_exp > returns seconds values >= 100. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org
[jira] [Updated] (SPARK-32021) make_interval does not accept seconds >100
[ https://issues.apache.org/jira/browse/SPARK-32021?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Juliusz Sompolski updated SPARK-32021: -- Description: In make_interval(years, months, weeks, days, hours, mins, secs), secs are defined as Decimal(8, 6), which turns into null if the value of the expression overflows 100 seconds. Larger seconds values should be allowed. This has been reported by Simba, who wants to use make_interval to implement translation for TIMESTAMP_ADD ODBC function in Spark 3.0. ODBC {fn TIMESTAMPADD(SECOND, integer_exp, timestamp} fails when integer_exp returns seconds values >= 100. was: In make_interval(years, months, weeks, days, hours, mins, secs), secs are defined as Decimal(8, 6), which turns into null if the value of the expression overflows 100 seconds. Larger seconds values should be allowed. This has been reported by Simba, who wants to use make_interval to implement translation for TIMESTAMP_ADD ODBC function in Spark 3.0. > make_interval does not accept seconds >100 > -- > > Key: SPARK-32021 > URL: https://issues.apache.org/jira/browse/SPARK-32021 > Project: Spark > Issue Type: Bug > Components: SQL >Affects Versions: 3.0.0 >Reporter: Juliusz Sompolski >Priority: Major > > In make_interval(years, months, weeks, days, hours, mins, secs), secs are > defined as Decimal(8, 6), which turns into null if the value of the > expression overflows 100 seconds. > Larger seconds values should be allowed. > This has been reported by Simba, who wants to use make_interval to implement > translation for TIMESTAMP_ADD ODBC function in Spark 3.0. > ODBC {fn TIMESTAMPADD(SECOND, integer_exp, timestamp} fails when integer_exp > returns seconds values >= 100. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org