[jira] [Updated] (SPARK-32640) Spark 3.1 log(NaN) returns null instead of NaN

2020-08-20 Thread Takeshi Yamamuro (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-32640?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Takeshi Yamamuro updated SPARK-32640:
-
Labels: correctness  (was: )

> Spark 3.1 log(NaN) returns null instead of NaN
> --
>
> Key: SPARK-32640
> URL: https://issues.apache.org/jira/browse/SPARK-32640
> Project: Spark
>  Issue Type: Bug
>  Components: SQL
>Affects Versions: 3.1.0
>Reporter: Thomas Graves
>Assignee: Wenchen Fan
>Priority: Major
>  Labels: correctness
> Fix For: 3.1.0
>
>
> I was testing Spark 3.1.0 and I noticed that if you take the log(NaN) it now 
> returns a null whereas in Spark 3.0 it returned a NaN.  I'm not an expert in 
> this but I thought NaN was correct.
> Spark 3.1.0 Example:
> >>> df.selectExpr(["value", "log1p(value)"]).show()
> +--+-+
> |    value|  LOG1P(value)|
> +--+-+
> |-3.4028235E38|  null|
> |3.4028235E38|88.72283906194683|
> |  0.0|   0.0|
> | -0.0|  -0.0|
> |  1.0|0.6931471805599453|
> | -1.0|  null|
> |  NaN|  null|
> +--+-+
>  
> Spark 3.0.0 example:
>  
> +-+--+
> | value| LOG1P(value)|
> +-+--+
> |-3.4028235E38| null|
> | 3.4028235E38| 88.72283906194683|
> | 0.0| 0.0|
> | -0.0| -0.0|
> | 1.0|0.6931471805599453|
> | -1.0| null|
> | NaN| NaN|
> +-+--+
>  
> Note it also does the same for log1p, log2, log10



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Updated] (SPARK-32640) Spark 3.1 log(NaN) returns null instead of NaN

2020-08-20 Thread Takeshi Yamamuro (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-32640?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Takeshi Yamamuro updated SPARK-32640:
-
Labels: correction  (was: )

> Spark 3.1 log(NaN) returns null instead of NaN
> --
>
> Key: SPARK-32640
> URL: https://issues.apache.org/jira/browse/SPARK-32640
> Project: Spark
>  Issue Type: Bug
>  Components: SQL
>Affects Versions: 3.1.0
>Reporter: Thomas Graves
>Assignee: Wenchen Fan
>Priority: Major
>  Labels: correction
> Fix For: 3.1.0
>
>
> I was testing Spark 3.1.0 and I noticed that if you take the log(NaN) it now 
> returns a null whereas in Spark 3.0 it returned a NaN.  I'm not an expert in 
> this but I thought NaN was correct.
> Spark 3.1.0 Example:
> >>> df.selectExpr(["value", "log1p(value)"]).show()
> +--+-+
> |    value|  LOG1P(value)|
> +--+-+
> |-3.4028235E38|  null|
> |3.4028235E38|88.72283906194683|
> |  0.0|   0.0|
> | -0.0|  -0.0|
> |  1.0|0.6931471805599453|
> | -1.0|  null|
> |  NaN|  null|
> +--+-+
>  
> Spark 3.0.0 example:
>  
> +-+--+
> | value| LOG1P(value)|
> +-+--+
> |-3.4028235E38| null|
> | 3.4028235E38| 88.72283906194683|
> | 0.0| 0.0|
> | -0.0| -0.0|
> | 1.0|0.6931471805599453|
> | -1.0| null|
> | NaN| NaN|
> +-+--+
>  
> Note it also does the same for log1p, log2, log10



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Updated] (SPARK-32640) Spark 3.1 log(NaN) returns null instead of NaN

2020-08-20 Thread Takeshi Yamamuro (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-32640?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Takeshi Yamamuro updated SPARK-32640:
-
Labels:   (was: correction)

> Spark 3.1 log(NaN) returns null instead of NaN
> --
>
> Key: SPARK-32640
> URL: https://issues.apache.org/jira/browse/SPARK-32640
> Project: Spark
>  Issue Type: Bug
>  Components: SQL
>Affects Versions: 3.1.0
>Reporter: Thomas Graves
>Assignee: Wenchen Fan
>Priority: Major
> Fix For: 3.1.0
>
>
> I was testing Spark 3.1.0 and I noticed that if you take the log(NaN) it now 
> returns a null whereas in Spark 3.0 it returned a NaN.  I'm not an expert in 
> this but I thought NaN was correct.
> Spark 3.1.0 Example:
> >>> df.selectExpr(["value", "log1p(value)"]).show()
> +--+-+
> |    value|  LOG1P(value)|
> +--+-+
> |-3.4028235E38|  null|
> |3.4028235E38|88.72283906194683|
> |  0.0|   0.0|
> | -0.0|  -0.0|
> |  1.0|0.6931471805599453|
> | -1.0|  null|
> |  NaN|  null|
> +--+-+
>  
> Spark 3.0.0 example:
>  
> +-+--+
> | value| LOG1P(value)|
> +-+--+
> |-3.4028235E38| null|
> | 3.4028235E38| 88.72283906194683|
> | 0.0| 0.0|
> | -0.0| -0.0|
> | 1.0|0.6931471805599453|
> | -1.0| null|
> | NaN| NaN|
> +-+--+
>  
> Note it also does the same for log1p, log2, log10



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org



[jira] [Updated] (SPARK-32640) Spark 3.1 log(NaN) returns null instead of NaN

2020-08-17 Thread Thomas Graves (Jira)


 [ 
https://issues.apache.org/jira/browse/SPARK-32640?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Thomas Graves updated SPARK-32640:
--
Description: 
I was testing Spark 3.1.0 and I noticed that if you take the log(NaN) it now 
returns a null whereas in Spark 3.0 it returned a NaN.  I'm not an expert in 
this but I thought NaN was correct.

Spark 3.1.0 Example:

>>> df.selectExpr(["value", "log1p(value)"]).show()

+--+-+
|    value|  LOG1P(value)|

+--+-+
|-3.4028235E38|  null|
|3.4028235E38|88.72283906194683|
|  0.0|   0.0|
| -0.0|  -0.0|
|  1.0|0.6931471805599453|
| -1.0|  null|
|  NaN|  null|

+--+-+

 

Spark 3.0.0 example:

 

+-+--+
| value| LOG1P(value)|
+-+--+
|-3.4028235E38| null|
| 3.4028235E38| 88.72283906194683|
| 0.0| 0.0|
| -0.0| -0.0|
| 1.0|0.6931471805599453|
| -1.0| null|
| NaN| NaN|
+-+--+

 

Note it also does the same for log1p, log2, log10

  was:
I was testing Spark 3.1.0 and I noticed that if you take the log(NaN) it now 
returns a null whereas in Spark 3.0 it returned a NaN.  I'm not an expert in 
this but I thought NaN was correct.

Spark 3.1.0 Example:

>>> df.selectExpr(["value", "log1p(value)"]).show()

+-+--+

|    value|  LOG1P(value)|

+-+--+

|-3.4028235E38|  null|

| 3.4028235E38| 88.72283906194683|

|  0.0|   0.0|

| -0.0|  -0.0|

|  1.0|0.6931471805599453|

| -1.0|  null|

|  NaN|  null|

+-+--+

 

Spark 3.0.0 example:

+-+-+
| value| LOG(E(), value)|
+-+-+
|-3.4028235E38| null|
| 3.4028235E38|88.72283906194683|
| 0.0| null|
| -0.0| null|
| 1.0| 0.0|
| -1.0| null|
| NaN| NaN|
+-+-+

 

Note it also does the same for log1p, log2, log10


> Spark 3.1 log(NaN) returns null instead of NaN
> --
>
> Key: SPARK-32640
> URL: https://issues.apache.org/jira/browse/SPARK-32640
> Project: Spark
>  Issue Type: Bug
>  Components: SQL
>Affects Versions: 3.1.0
>Reporter: Thomas Graves
>Priority: Major
>
> I was testing Spark 3.1.0 and I noticed that if you take the log(NaN) it now 
> returns a null whereas in Spark 3.0 it returned a NaN.  I'm not an expert in 
> this but I thought NaN was correct.
> Spark 3.1.0 Example:
> >>> df.selectExpr(["value", "log1p(value)"]).show()
> +--+-+
> |    value|  LOG1P(value)|
> +--+-+
> |-3.4028235E38|  null|
> |3.4028235E38|88.72283906194683|
> |  0.0|   0.0|
> | -0.0|  -0.0|
> |  1.0|0.6931471805599453|
> | -1.0|  null|
> |  NaN|  null|
> +--+-+
>  
> Spark 3.0.0 example:
>  
> +-+--+
> | value| LOG1P(value)|
> +-+--+
> |-3.4028235E38| null|
> | 3.4028235E38| 88.72283906194683|
> | 0.0| 0.0|
> | -0.0| -0.0|
> | 1.0|0.6931471805599453|
> | -1.0| null|
> | NaN| NaN|
> +-+--+
>  
> Note it also does the same for log1p, log2, log10



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org