[jira] [Assigned] (SPARK-35107) Parse unit-to-unit interval literals to ANSI intervals

2021-04-16 Thread Apache Spark (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-35107?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Apache Spark reassigned SPARK-35107:


Assignee: Max Gekk  (was: Apache Spark)

> Parse unit-to-unit interval literals to ANSI intervals
> --
>
> Key: SPARK-35107
> URL: https://issues.apache.org/jira/browse/SPARK-35107
> Project: Spark
>  Issue Type: Sub-task
>  Components: SQL
>Affects Versions: 3.2.0
>Reporter: Max Gekk
>Assignee: Max Gekk
>Priority: Major
>
> Parse unit-to-unit intervals like INTERVAL '1-1' YEAR TO MONTH to either 
> YearMonthIntervalType or DayTimeIntervalType by default. But when 
> spark.sql.legacy.interval.enabled is set to true, parse them to 
> CalendarInterval.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Assigned] (SPARK-35107) Parse unit-to-unit interval literals to ANSI intervals

2021-04-16 Thread Apache Spark (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-35107?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Apache Spark reassigned SPARK-35107:


Assignee: Apache Spark  (was: Max Gekk)

> Parse unit-to-unit interval literals to ANSI intervals
> --
>
> Key: SPARK-35107
> URL: https://issues.apache.org/jira/browse/SPARK-35107
> Project: Spark
>  Issue Type: Sub-task
>  Components: SQL
>Affects Versions: 3.2.0
>Reporter: Max Gekk
>Assignee: Apache Spark
>Priority: Major
>
> Parse unit-to-unit intervals like INTERVAL '1-1' YEAR TO MONTH to either 
> YearMonthIntervalType or DayTimeIntervalType by default. But when 
> spark.sql.legacy.interval.enabled is set to true, parse them to 
> CalendarInterval.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org