[jira] [Created] (SPARK-33557) spark.storage.blockManagerSlaveTimeoutMs default value does not follow spark.network.timeout value when the latter was changed

2020-11-25 Thread Ohad (Jira)
Ohad created SPARK-33557:


 Summary: spark.storage.blockManagerSlaveTimeoutMs default value 
does not follow spark.network.timeout value when the latter was changed
 Key: SPARK-33557
 URL: https://issues.apache.org/jira/browse/SPARK-33557
 Project: Spark
  Issue Type: Bug
  Components: Spark Core
Affects Versions: 3.0.1, 3.0.0
Reporter: Ohad


According to the documentation "spark.network.timeout" is the default timeout 
for "spark.storage.blockManagerSlaveTimeoutMs" which implies that when the user 
sets "spark.network.timeout"  the effective value of 
"spark.storage.blockManagerSlaveTimeoutMs" should also be changed if it was not 
specifically changed.

However this is not the case since the default value of 
"spark.storage.blockManagerSlaveTimeoutMs" is always the default value of 
"spark.network.timeout" (120s)

 

"spark.storage.blockManagerSlaveTimeoutMs" is defined in the package object of 
"org.apache.spark.internal.config" as follows:
{code:java}
private[spark] val STORAGE_BLOCKMANAGER_SLAVE_TIMEOUT =
  ConfigBuilder("spark.storage.blockManagerSlaveTimeoutMs")
.version("0.7.0")
.timeConf(TimeUnit.MILLISECONDS)
.createWithDefaultString(Network.NETWORK_TIMEOUT.defaultValueString)
{code}
So it seems like the its default value is indeed "fixed" to 
"spark.network.timeout" default value.

 

 

 

 

 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-33556) Add array_to_vector function for dataframe column

2020-11-25 Thread Apache Spark (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-33556?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17238673#comment-17238673
 ] 

Apache Spark commented on SPARK-33556:
--

User 'WeichenXu123' has created a pull request for this issue:
https://github.com/apache/spark/pull/30498

> Add array_to_vector function for dataframe column
> -
>
> Key: SPARK-33556
> URL: https://issues.apache.org/jira/browse/SPARK-33556
> Project: Spark
>  Issue Type: New Feature
>  Components: ML
>Affects Versions: 3.1.0
>Reporter: Weichen Xu
>Priority: Major
>
> Add array_to_vector function for dataframe column.
> Input column type: Array
> Output column type: apache.spark.ml.linalg.Vector



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Assigned] (SPARK-33556) Add array_to_vector function for dataframe column

2020-11-25 Thread Apache Spark (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-33556?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Apache Spark reassigned SPARK-33556:


Assignee: Apache Spark

> Add array_to_vector function for dataframe column
> -
>
> Key: SPARK-33556
> URL: https://issues.apache.org/jira/browse/SPARK-33556
> Project: Spark
>  Issue Type: New Feature
>  Components: ML
>Affects Versions: 3.1.0
>Reporter: Weichen Xu
>Assignee: Apache Spark
>Priority: Major
>
> Add array_to_vector function for dataframe column.
> Input column type: Array
> Output column type: apache.spark.ml.linalg.Vector



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Assigned] (SPARK-33556) Add array_to_vector function for dataframe column

2020-11-25 Thread Apache Spark (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-33556?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Apache Spark reassigned SPARK-33556:


Assignee: (was: Apache Spark)

> Add array_to_vector function for dataframe column
> -
>
> Key: SPARK-33556
> URL: https://issues.apache.org/jira/browse/SPARK-33556
> Project: Spark
>  Issue Type: New Feature
>  Components: ML
>Affects Versions: 3.1.0
>Reporter: Weichen Xu
>Priority: Major
>
> Add array_to_vector function for dataframe column.
> Input column type: Array
> Output column type: apache.spark.ml.linalg.Vector



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-33556) Add array_to_vector function for dataframe column

2020-11-25 Thread Apache Spark (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-33556?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17238672#comment-17238672
 ] 

Apache Spark commented on SPARK-33556:
--

User 'WeichenXu123' has created a pull request for this issue:
https://github.com/apache/spark/pull/30498

> Add array_to_vector function for dataframe column
> -
>
> Key: SPARK-33556
> URL: https://issues.apache.org/jira/browse/SPARK-33556
> Project: Spark
>  Issue Type: New Feature
>  Components: ML
>Affects Versions: 3.1.0
>Reporter: Weichen Xu
>Priority: Major
>
> Add array_to_vector function for dataframe column.
> Input column type: Array
> Output column type: apache.spark.ml.linalg.Vector



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Created] (SPARK-33556) Add array_to_vector function for dataframe column

2020-11-25 Thread Weichen Xu (Jira)
Weichen Xu created SPARK-33556:
--

 Summary: Add array_to_vector function for dataframe column
 Key: SPARK-33556
 URL: https://issues.apache.org/jira/browse/SPARK-33556
 Project: Spark
  Issue Type: New Feature
  Components: ML
Affects Versions: 3.1.0
Reporter: Weichen Xu


Add array_to_vector function for dataframe column.

Input column type: Array

Output column type: apache.spark.ml.linalg.Vector



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-28064) Order by does not accept a call to rank()

2020-11-25 Thread Dylan Guedes (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-28064?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17238662#comment-17238662
 ] 

Dylan Guedes commented on SPARK-28064:
--

Sorry, my only intention was to help to map the differences between PostgreSQL 
and SparkSQL APIs.

> Order by does not accept a call to rank()
> -
>
> Key: SPARK-28064
> URL: https://issues.apache.org/jira/browse/SPARK-28064
> Project: Spark
>  Issue Type: Sub-task
>  Components: SQL
>Affects Versions: 3.1.0
>Reporter: Dylan Guedes
>Priority: Major
>
> Currently in Spark, we can't use a call to `rank()` in a order by; we need to 
> first rename the rank column to, for instance, `r` and then, use `order by 
> r`. For example:
>  This does not work:
> {code:sql}
>  SELECT depname, empno, salary, rank() OVER w FROM empsalary WINDOW w AS 
> (PARTITION BY depname ORDER BY salary) ORDER BY rank() OVER w;
> {code}
> However, this one does:
> {code:sql}
>  SELECT depname, empno, salary, rank() OVER w as r FROM empsalary WINDOW w AS 
> (PARTITION BY depname ORDER BY salary) ORDER BY r;
> {code}
> By the way, I took this one from Postgres behavior: postgres accept both ways.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-28648) Adds support to `groups` unit type in window clauses

2020-11-25 Thread jiaan.geng (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-28648?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17238654#comment-17238654
 ] 

jiaan.geng commented on SPARK-28648:


it seems groups only exists in PgSQL

> Adds support to `groups` unit type in window clauses
> 
>
> Key: SPARK-28648
> URL: https://issues.apache.org/jira/browse/SPARK-28648
> Project: Spark
>  Issue Type: Sub-task
>  Components: SQL
>Affects Versions: 3.0.0
>Reporter: Dylan Guedes
>Priority: Major
>
> Spark currently support the two most common window functions unit types: rows 
> and ranges. However, in PgSQL a new type was added: `groups`. 
> According to [this 
> source|https://blog.jooq.org/2018/07/05/postgresql-11s-support-for-sql-standard-groups-and-exclude-window-function-clauses/],
>  the difference is:
> """ROWS counts the exact number of rows in the frame.
> RANGE performs logical windowing where we don’t count the number of rows, but 
> look for a value offset.
> GROUPS counts all groups of tied rows within the window."""



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-28645) Throw an error on window redefinition

2020-11-25 Thread jiaan.geng (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-28645?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17238648#comment-17238648
 ] 

jiaan.geng commented on SPARK-28645:


I'm working on!

> Throw an error on window redefinition
> -
>
> Key: SPARK-28645
> URL: https://issues.apache.org/jira/browse/SPARK-28645
> Project: Spark
>  Issue Type: Sub-task
>  Components: SQL
>Affects Versions: 3.0.0
>Reporter: Dylan Guedes
>Priority: Major
>
> Currently in Spark one could redefine a window. For instance:
> {code:sql}select count(*) OVER w FROM tenk1 WINDOW w AS (ORDER BY unique1), w 
> AS (ORDER BY unique1);{code}
> The window `w` is defined two times. In PgSQL, on the other hand, a thrown 
> will happen:
> {code:sql}ERROR:  window "w" is already defined{code}
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Resolved] (SPARK-28508) Support for range frame+row frame in the same query

2020-11-25 Thread jiaan.geng (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-28508?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

jiaan.geng resolved SPARK-28508.

Resolution: Invalid

> Support for range frame+row frame in the same query
> ---
>
> Key: SPARK-28508
> URL: https://issues.apache.org/jira/browse/SPARK-28508
> Project: Spark
>  Issue Type: Sub-task
>  Components: SQL
>Affects Versions: 3.1.0
>Reporter: Dylan Guedes
>Priority: Major
>
> Currently, looks like some queries does not works if both, a range frame and 
> a row frame are given. However, PgSQL is able to handle them:
> {code:sql}
> select last(salary) over(order by enroll_date range between 1 preceding and 1 
> following), lag(salary) over(order by enroll_date range between 1 preceding 
> and 1 following),
> salary, enroll_date from empsalary;
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-28508) Support for range frame+row frame in the same query

2020-11-25 Thread jiaan.geng (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-28508?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17238638#comment-17238638
 ] 

jiaan.geng commented on SPARK-28508:


Spark has supported this feature.

> Support for range frame+row frame in the same query
> ---
>
> Key: SPARK-28508
> URL: https://issues.apache.org/jira/browse/SPARK-28508
> Project: Spark
>  Issue Type: Sub-task
>  Components: SQL
>Affects Versions: 3.1.0
>Reporter: Dylan Guedes
>Priority: Major
>
> Currently, looks like some queries does not works if both, a range frame and 
> a row frame are given. However, PgSQL is able to handle them:
> {code:sql}
> select last(salary) over(order by enroll_date range between 1 preceding and 1 
> following), lag(salary) over(order by enroll_date range between 1 preceding 
> and 1 following),
> salary, enroll_date from empsalary;
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Assigned] (SPARK-33540) Subexpression elimination for interpreted predicate

2020-11-25 Thread Apache Spark (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-33540?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Apache Spark reassigned SPARK-33540:


Assignee: L. C. Hsieh  (was: Apache Spark)

> Subexpression elimination for interpreted predicate
> ---
>
> Key: SPARK-33540
> URL: https://issues.apache.org/jira/browse/SPARK-33540
> Project: Spark
>  Issue Type: New Feature
>  Components: SQL
>Affects Versions: 3.1.0
>Reporter: L. C. Hsieh
>Assignee: L. C. Hsieh
>Priority: Major
>
> We can support subexpression elimination for interpreted predicate.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Assigned] (SPARK-33540) Subexpression elimination for interpreted predicate

2020-11-25 Thread Apache Spark (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-33540?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Apache Spark reassigned SPARK-33540:


Assignee: Apache Spark  (was: L. C. Hsieh)

> Subexpression elimination for interpreted predicate
> ---
>
> Key: SPARK-33540
> URL: https://issues.apache.org/jira/browse/SPARK-33540
> Project: Spark
>  Issue Type: New Feature
>  Components: SQL
>Affects Versions: 3.1.0
>Reporter: L. C. Hsieh
>Assignee: Apache Spark
>Priority: Major
>
> We can support subexpression elimination for interpreted predicate.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-33540) Subexpression elimination for interpreted predicate

2020-11-25 Thread Apache Spark (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-33540?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17238626#comment-17238626
 ] 

Apache Spark commented on SPARK-33540:
--

User 'viirya' has created a pull request for this issue:
https://github.com/apache/spark/pull/30497

> Subexpression elimination for interpreted predicate
> ---
>
> Key: SPARK-33540
> URL: https://issues.apache.org/jira/browse/SPARK-33540
> Project: Spark
>  Issue Type: New Feature
>  Components: SQL
>Affects Versions: 3.1.0
>Reporter: L. C. Hsieh
>Assignee: L. C. Hsieh
>Priority: Major
>
> We can support subexpression elimination for interpreted predicate.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-28064) Order by does not accept a call to rank()

2020-11-25 Thread jiaan.geng (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-28064?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17238625#comment-17238625
 ] 

jiaan.geng commented on SPARK-28064:


Yeah! But it seems nothing useful.

> Order by does not accept a call to rank()
> -
>
> Key: SPARK-28064
> URL: https://issues.apache.org/jira/browse/SPARK-28064
> Project: Spark
>  Issue Type: Sub-task
>  Components: SQL
>Affects Versions: 3.1.0
>Reporter: Dylan Guedes
>Priority: Major
>
> Currently in Spark, we can't use a call to `rank()` in a order by; we need to 
> first rename the rank column to, for instance, `r` and then, use `order by 
> r`. For example:
>  This does not work:
> {code:sql}
>  SELECT depname, empno, salary, rank() OVER w FROM empsalary WINDOW w AS 
> (PARTITION BY depname ORDER BY salary) ORDER BY rank() OVER w;
> {code}
> However, this one does:
> {code:sql}
>  SELECT depname, empno, salary, rank() OVER w as r FROM empsalary WINDOW w AS 
> (PARTITION BY depname ORDER BY salary) ORDER BY r;
> {code}
> By the way, I took this one from Postgres behavior: postgres accept both ways.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Updated] (SPARK-32223) Support adding a user provided config map.

2020-11-25 Thread Prashant Sharma (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-32223?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Prashant Sharma updated SPARK-32223:

Description: 
One of the challenge with this is, spark.properties is not user provided and is 
calculated based on certain factors. So a user provided config map, cannot be 
used as is to mount as SPARK_CONF_DIR, so it will have to be somehow augmented 
with the correct spark.properties.

Q, Do we support update to config map properties for an already running job?
Ans: No, since the spark.properties is calculated at the time of job 
submission, it cannot be updated on the fly and it is not supported by Spark at 
the moment for all the configuration values.

Q. What are the usecases, where supplying SPARK_CONF_DIR via a config map helps?
One of the use case, I can think of is programmatically submitting a `spark on 
k8s` job - e.g. spark as a service on a cloud deployment may find this feature 
useful.



  was:
One of the challenge with this is, spark.properties is not user provided and is 
calculated based on certain factors. So a user provided config map, cannot be 
used as is to mount as SPARK_CONF_DIR, so it will have to be somehow augmented 
with the correct spark.properties.

Q, Do we support update to config map properties for an already running job?
Ans: No, since the spark.properties is calculated at the time of job 
submission, it cannot be updated on the fly and it is not supported by Spark at 
the moment for all the configuration values.

Q. What are the usecases, where supplying SPARK_CONF_DIR via a config map helps?
One of the use case, I can think of is programmatically submitting a spark on 
k8s job - e.g. spark as a service on cloud deployment may find this feature 
useful.




> Support adding a user provided config map.
> --
>
> Key: SPARK-32223
> URL: https://issues.apache.org/jira/browse/SPARK-32223
> Project: Spark
>  Issue Type: Sub-task
>  Components: Kubernetes
>Affects Versions: 3.1.0
>Reporter: Prashant Sharma
>Priority: Major
>
> One of the challenge with this is, spark.properties is not user provided and 
> is calculated based on certain factors. So a user provided config map, cannot 
> be used as is to mount as SPARK_CONF_DIR, so it will have to be somehow 
> augmented with the correct spark.properties.
> Q, Do we support update to config map properties for an already running job?
> Ans: No, since the spark.properties is calculated at the time of job 
> submission, it cannot be updated on the fly and it is not supported by Spark 
> at the moment for all the configuration values.
> Q. What are the usecases, where supplying SPARK_CONF_DIR via a config map 
> helps?
> One of the use case, I can think of is programmatically submitting a `spark 
> on k8s` job - e.g. spark as a service on a cloud deployment may find this 
> feature useful.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-32223) Support adding a user provided config map.

2020-11-25 Thread Prashant Sharma (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-32223?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17238617#comment-17238617
 ] 

Prashant Sharma commented on SPARK-32223:
-

Hi [~dongjoon],

Do you think this is useful ? Any other thoughts on this? 

Thanks ! 

> Support adding a user provided config map.
> --
>
> Key: SPARK-32223
> URL: https://issues.apache.org/jira/browse/SPARK-32223
> Project: Spark
>  Issue Type: Sub-task
>  Components: Kubernetes
>Affects Versions: 3.1.0
>Reporter: Prashant Sharma
>Priority: Major
>
> One of the challenge with this is, spark.properties is not user provided and 
> is calculated based on certain factors. So a user provided config map, cannot 
> be used as is to mount as SPARK_CONF_DIR, so it will have to be somehow 
> augmented with the correct spark.properties.
> Q, Do we support update to config map properties for an already running job?
> Ans: No, since the spark.properties is calculated at the time of job 
> submission, it cannot be updated on the fly and it is not supported by Spark 
> at the moment for all the configuration values.
> Q. What are the usecases, where supplying SPARK_CONF_DIR via a config map 
> helps?
> One of the use case, I can think of is programmatically submitting a spark on 
> k8s job - e.g. spark as a service on cloud deployment may find this feature 
> useful.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Updated] (SPARK-32223) Support adding a user provided config map.

2020-11-25 Thread Prashant Sharma (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-32223?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Prashant Sharma updated SPARK-32223:

Description: 
One of the challenge with this is, spark.properties is not user provided and is 
calculated based on certain factors. So a user provided config map, cannot be 
used as is to mount as SPARK_CONF_DIR, so it will have to be somehow augmented 
with the correct spark.properties.

Q, Do we support update to config map properties for an already running job?
Ans: No, since the spark.properties is calculated at the time of job 
submission, it cannot be updated on the fly and it is not supported by Spark at 
the moment for all the configuration values.

Q. What are the usecases, where supplying SPARK_CONF_DIR via a config map helps?
One of the use case, I can think of is programmatically submitting a spark on 
k8s job - e.g. spark as a service on cloud deployment may find this feature 
useful.



  was:The semantics of this will be discussed and added soon.


> Support adding a user provided config map.
> --
>
> Key: SPARK-32223
> URL: https://issues.apache.org/jira/browse/SPARK-32223
> Project: Spark
>  Issue Type: Sub-task
>  Components: Kubernetes
>Affects Versions: 3.1.0
>Reporter: Prashant Sharma
>Priority: Major
>
> One of the challenge with this is, spark.properties is not user provided and 
> is calculated based on certain factors. So a user provided config map, cannot 
> be used as is to mount as SPARK_CONF_DIR, so it will have to be somehow 
> augmented with the correct spark.properties.
> Q, Do we support update to config map properties for an already running job?
> Ans: No, since the spark.properties is calculated at the time of job 
> submission, it cannot be updated on the fly and it is not supported by Spark 
> at the moment for all the configuration values.
> Q. What are the usecases, where supplying SPARK_CONF_DIR via a config map 
> helps?
> One of the use case, I can think of is programmatically submitting a spark on 
> k8s job - e.g. spark as a service on cloud deployment may find this feature 
> useful.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-33429) Support drop column in spark also like in postgresql

2020-11-25 Thread jiaan.geng (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-33429?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17238615#comment-17238615
 ] 

jiaan.geng commented on SPARK-33429:


It seems Spark has supported

{code:java}
| ALTER TABLE multipartIdentifier
DROP (COLUMN | COLUMNS)
'(' columns=multipartIdentifierList ')'
#dropTableColumns
| ALTER TABLE multipartIdentifier
DROP (COLUMN | COLUMNS) columns=multipartIdentifierList
#dropTableColumns
{code}


> Support drop column in spark also like in postgresql
> 
>
> Key: SPARK-33429
> URL: https://issues.apache.org/jira/browse/SPARK-33429
> Project: Spark
>  Issue Type: Sub-task
>  Components: SQL
>Affects Versions: 2.4.5
>Reporter: jobit mathew
>Priority: Minor
>
> Support alter table with drop column in spark also like in postgresql.
>  
> [https://www.postgresql.org/docs/12/sql-altertable.html]
> alter table tablename DROP COLUMN/columns [ IF EXISTS ]
> columnname
>  
> {code:java}
> spark-sql> drop database if exists hivemetastoretest cascade;
> Time taken: 1.067 seconds
> spark-sql> create database hivemetastoretest;
> Time taken: 0.326 seconds
> spark-sql> use hivemetastoretest;
> Time taken: 0.053 seconds
> spark-sql> create table jobit4 using parquet as select 2.5;
> Time taken: 5.058 seconds
> spark-sql> alter table jobit4 add columns(name string);
> Time taken: 1.194 seconds
> spark-sql> alter table jobit4 drop columns(name);
> Error in query:
> mismatched input 'columns' expecting \{'PARTITION', 'IF'}(line 1, pos 25)
> == SQL ==
>  alter table jobit4 drop columns(name)
> -^^^
> spark-sql> alter table jobit4 drop columns name;
> Error in query:
> mismatched input 'columns' expecting \{'PARTITION', 'IF'}(line 1, pos 25)
> == SQL ==
>  alter table jobit4 drop columns name
> -^^^
> spark-sql> [
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Updated] (SPARK-33527) Extend the function of decode so as consistent with mainstream databases

2020-11-25 Thread jiaan.geng (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-33527?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

jiaan.geng updated SPARK-33527:
---
Issue Type: New Feature  (was: Bug)

> Extend the function of decode so as consistent with mainstream databases
> 
>
> Key: SPARK-33527
> URL: https://issues.apache.org/jira/browse/SPARK-33527
> Project: Spark
>  Issue Type: New Feature
>  Components: SQL
>Affects Versions: 3.1.0
>Reporter: jiaan.geng
>Priority: Major
>
> In Spark, decode(bin, charset) - Decodes the first argument using the second 
> argument character set.
> Unfortunately this is NOT what any other SQL vendor understands DECODE to do.
> DECODE generally is a short hand for a simple case expression:
> {code:java}
> SELECT DECODE(c1, 1, 'Hello', 2, 'World', '!') FROM (VALUES (1), (2), (3)) AS 
> T(c1)
> => 
> (Hello),
> (World)
> (!)
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Resolved] (SPARK-33549) Remove configuration spark.sql.legacy.allowCastNumericToTimestamp

2020-11-25 Thread Wenchen Fan (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-33549?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Wenchen Fan resolved SPARK-33549.
-
Fix Version/s: 3.1.0
   Resolution: Fixed

Issue resolved by pull request 30493
[https://github.com/apache/spark/pull/30493]

> Remove configuration spark.sql.legacy.allowCastNumericToTimestamp
> -
>
> Key: SPARK-33549
> URL: https://issues.apache.org/jira/browse/SPARK-33549
> Project: Spark
>  Issue Type: Improvement
>  Components: SQL
>Affects Versions: 3.1.0
>Reporter: Gengliang Wang
>Assignee: Gengliang Wang
>Priority: Minor
> Fix For: 3.1.0
>
>
> In the current master branch, there is a new configuration 
> `spark.sql.legacy.allowCastNumericToTimestamp` which controls whether to cast 
> Numeric types to Timestamp or not. The default value is true.
> After https://github.com/apache/spark/pull/30260, the type conversion between 
> Timestamp type and Numeric type is disallowed in ANSI mode. So, we don't need 
> to a separate configuration `spark.sql.legacy.allowCastNumericToTimestamp` 
> for disallowing the conversion.
> We should remove the configuration.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Created] (SPARK-33555) Accessibility: spark history server page, the download button has insufficient 2.2:1 contrast .

2020-11-25 Thread danliu (Jira)
danliu created SPARK-33555:
--

 Summary: Accessibility: spark history server page, the download 
button has insufficient 2.2:1 contrast .
 Key: SPARK-33555
 URL: https://issues.apache.org/jira/browse/SPARK-33555
 Project: Spark
  Issue Type: Bug
  Components: Graph
Affects Versions: 2.4.5
Reporter: danliu


In spark history server page, the download button has insufficient 2.2:1 
contrast .Visual impaired user will find difficulty to see the control or 
element if that control or element have insufficient color contrast as user 
won't be able to perform functionality over such control.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Created] (SPARK-33554) Accessibility: insufficient color contrast

2020-11-25 Thread danliu (Jira)
danliu created SPARK-33554:
--

 Summary: Accessibility: insufficient color contrast 
 Key: SPARK-33554
 URL: https://issues.apache.org/jira/browse/SPARK-33554
 Project: Spark
  Issue Type: Bug
  Components: Graph
Affects Versions: 2.4.5
Reporter: danliu


Visual impaired user will find difficulty to see the control or element in 
spark history server page if that control or element have insufficient color 
contrast as user won't be able to perform functionality over such control.
 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Created] (SPARK-33553) Accessibility: "Spark 2.4.5" image link is not defined

2020-11-25 Thread danliu (Jira)
danliu created SPARK-33553:
--

 Summary: Accessibility: "Spark 2.4.5" image link is not defined 
 Key: SPARK-33553
 URL: https://issues.apache.org/jira/browse/SPARK-33553
 Project: Spark
  Issue Type: Bug
  Components: Graph
Affects Versions: 2.4.5
Reporter: danliu


In spark history server page, screen reader user will get confuse with the 
purpose of the image link of the "Spark 2.4.5" and can not access the image 
link efficiently if alt of the image is not defined.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Created] (SPARK-33552) Accessibility: in spark history server page, The element does not have a lang attribute, user depending on Assistive Technology get affected as lang attribute.

2020-11-25 Thread danliu (Jira)
danliu created SPARK-33552:
--

 Summary: Accessibility: in spark history server page, The  
element does not have a lang attribute, user depending on Assistive Technology 
get affected as lang attribute.
 Key: SPARK-33552
 URL: https://issues.apache.org/jira/browse/SPARK-33552
 Project: Spark
  Issue Type: Bug
  Components: Graph
Affects Versions: 3.0.1
Reporter: danliu


in spark history server page, The  elements do not have a lang attribute, 
user depending on Assistive Technology get affected as lang attribute.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Assigned] (SPARK-33547) Doc Type Construct Literal usage

2020-11-25 Thread Apache Spark (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-33547?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Apache Spark reassigned SPARK-33547:


Assignee: Apache Spark

> Doc Type Construct Literal usage
> 
>
> Key: SPARK-33547
> URL: https://issues.apache.org/jira/browse/SPARK-33547
> Project: Spark
>  Issue Type: Improvement
>  Components: Documentation
>Affects Versions: 3.1.0
>Reporter: angerszhu
>Assignee: Apache Spark
>Priority: Major
>
> Add Doc about type construct literal in    
> [https://spark.apache.org/docs/3.0.1/sql-ref-literals.html]



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Assigned] (SPARK-33547) Doc Type Construct Literal usage

2020-11-25 Thread Apache Spark (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-33547?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Apache Spark reassigned SPARK-33547:


Assignee: (was: Apache Spark)

> Doc Type Construct Literal usage
> 
>
> Key: SPARK-33547
> URL: https://issues.apache.org/jira/browse/SPARK-33547
> Project: Spark
>  Issue Type: Improvement
>  Components: Documentation
>Affects Versions: 3.1.0
>Reporter: angerszhu
>Priority: Major
>
> Add Doc about type construct literal in    
> [https://spark.apache.org/docs/3.0.1/sql-ref-literals.html]



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Commented] (SPARK-33547) Doc Type Construct Literal usage

2020-11-25 Thread Apache Spark (Jira)


[ 
https://issues.apache.org/jira/browse/SPARK-33547?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17238588#comment-17238588
 ] 

Apache Spark commented on SPARK-33547:
--

User 'AngersZh' has created a pull request for this issue:
https://github.com/apache/spark/pull/30496

> Doc Type Construct Literal usage
> 
>
> Key: SPARK-33547
> URL: https://issues.apache.org/jira/browse/SPARK-33547
> Project: Spark
>  Issue Type: Improvement
>  Components: Documentation
>Affects Versions: 3.1.0
>Reporter: angerszhu
>Priority: Major
>
> Add Doc about type construct literal in    
> [https://spark.apache.org/docs/3.0.1/sql-ref-literals.html]



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



<    1   2