[ 
https://issues.apache.org/jira/browse/SPARK-11009?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Saif Addin Ellafi updated SPARK-11009:
--------------------------------------
    Environment: Standalone cluster mode. No hadoop/hive is present in the 
environment (no hive-site.xml), only using HiveContext. Spark build as with 
hadoop 2.6.0. Default spark configuration variables. cluster has 4 nodes, but 
happens with n nodes as well.  (was: Standalone cluster mode
No hadoop/hive is present in the environment (no hive-site.xml), only using 
HiveContext. Spark build as with hadoop 2.6.0.
Default spark configuration variables.
cluster has 4 nodes, but happens with n nodes as well.)

> RowNumber in HiveContext returns negative values in cluster mode
> ----------------------------------------------------------------
>
>                 Key: SPARK-11009
>                 URL: https://issues.apache.org/jira/browse/SPARK-11009
>             Project: Spark
>          Issue Type: Bug
>          Components: Spark Core
>    Affects Versions: 1.5.1
>         Environment: Standalone cluster mode. No hadoop/hive is present in 
> the environment (no hive-site.xml), only using HiveContext. Spark build as 
> with hadoop 2.6.0. Default spark configuration variables. cluster has 4 
> nodes, but happens with n nodes as well.
>            Reporter: Saif Addin Ellafi
>
> This issue happens when submitting the job into a standalone cluster. Have 
> not tried YARN or MESOS. Repartition df into 1 piece or default parallelism=1 
> does not fix the issue. Also tried having only one node in the cluster, with 
> same result. Other shuffle configuration changes do not alter the results 
> either.
> The issue does NOT happen in --master local[*].
>         val ws = Window.
>             partitionBy("client_id").
>             orderBy("date")
>  
>         val nm = "repeatMe"
>         df.select(df.col("*"), rowNumber().over(ws).as(nm))
>  
>         
> df.filter(df("repeatMe").isNotNull).orderBy("repeatMe").take(50).foreach(println(_))
>  
> --->
>  
> Long, DateType, Int
> [219483904822,2006-06-01,-1863462909]
> [219483904822,2006-09-01,-1863462909]
> [219483904822,2007-01-01,-1863462909]
> [219483904822,2007-08-01,-1863462909]
> [219483904822,2007-07-01,-1863462909]
> [192489238423,2007-07-01,-1863462774]
> [192489238423,2007-02-01,-1863462774]
> [192489238423,2006-11-01,-1863462774]
> [192489238423,2006-08-01,-1863462774]
> [192489238423,2007-08-01,-1863462774]
> [192489238423,2006-09-01,-1863462774]
> [192489238423,2007-03-01,-1863462774]
> [192489238423,2006-10-01,-1863462774]
> [192489238423,2007-05-01,-1863462774]
> [192489238423,2006-06-01,-1863462774]
> [192489238423,2006-12-01,-1863462774]



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org

Reply via email to