[ 
https://issues.apache.org/jira/browse/SPARK-12491?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15072999#comment-15072999
 ] 

Herman van Hovell commented on SPARK-12491:
-------------------------------------------

The logical plans look fine.

I have defined the class in a separate project, create an assembly, {{gm.jar}}, 
from that, and finally start the the spark-shell using the following command: 
{{./spark-shell --master local\[\*\] --driver-memory 10G --jars gm.jar}}

Unfortunately I still cannot reproduce your problem (I'll attach the project).

Are you doing this in local mode or cluster mode? What kind of cluster are you 
using? What deps are in jars you pass to spark?



> UDAF result differs in SQL if alias is used
> -------------------------------------------
>
>                 Key: SPARK-12491
>                 URL: https://issues.apache.org/jira/browse/SPARK-12491
>             Project: Spark
>          Issue Type: Bug
>          Components: SQL
>    Affects Versions: 1.5.2
>            Reporter: Tristan
>
> Using the GeometricMean UDAF example 
> (https://databricks.com/blog/2015/09/16/spark-1-5-dataframe-api-highlights-datetimestring-handling-time-intervals-and-udafs.html),
>  I found the following discrepancy in results:
> {code}
> scala> sqlContext.sql("select group_id, gm(id) from simple group by 
> group_id").show()
> +--------+---+
> |group_id|_c1|
> +--------+---+
> |       0|0.0|
> |       1|0.0|
> |       2|0.0|
> +--------+---+
> scala> sqlContext.sql("select group_id, gm(id) as GeometricMean from simple 
> group by group_id").show()
> +--------+-----------------+
> |group_id|    GeometricMean|
> +--------+-----------------+
> |       0|8.981385496571725|
> |       1|7.301716979342118|
> |       2|7.706253151292568|
> +--------+-----------------+
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org

Reply via email to