[ 
https://issues.apache.org/jira/browse/SPARK-20520?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16033918#comment-16033918
 ] 

Dongjoon Hyun edited comment on SPARK-20520 at 6/1/17 11:46 PM:
----------------------------------------------------------------

Hi, [~felixcheung].
-Is this still targeting 2.2.0?-
Oh, I see. It should be tested on RC.


was (Author: dongjoon):
Hi, [~felixcheung].
Is this still targeting 2.2.0?

> R streaming tests failed on Windows
> -----------------------------------
>
>                 Key: SPARK-20520
>                 URL: https://issues.apache.org/jira/browse/SPARK-20520
>             Project: Spark
>          Issue Type: Bug
>          Components: SparkR
>            Reporter: Felix Cheung
>            Assignee: Felix Cheung
>            Priority: Critical
>
> Running R CMD check on SparkR 2.2 RC1 packages 
> {code}
> Failed 
> -------------------------------------------------------------------------
> 1. Failure: read.stream, write.stream, awaitTermination, stopQuery 
> (@test_streaming.R#56) 
> head(sql("SELECT count(*) FROM people"))[[1]] not equal to 3.
> 1/1 mismatches
> [1] 0 - 3 == -3
> 2. Failure: read.stream, write.stream, awaitTermination, stopQuery 
> (@test_streaming.R#60) 
> head(sql("SELECT count(*) FROM people"))[[1]] not equal to 6.
> 1/1 mismatches
> [1] 3 - 6 == -3
> 3. Failure: print from explain, lastProgress, status, isActive 
> (@test_streaming.R#75) 
> any(grepl("\"description\" : \"MemorySink\"", 
> capture.output(lastProgress(q)))) isn't true.
> 4. Failure: Stream other format (@test_streaming.R#95) 
> -------------------------
> head(sql("SELECT count(*) FROM people3"))[[1]] not equal to 3.
> 1/1 mismatches
> [1] 0 - 3 == -3
> 5. Failure: Stream other format (@test_streaming.R#98) 
> -------------------------
> any(...) isn't true.
> {code}
> Need to investigate



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org

Reply via email to