----- Original Message -----

> dev/run-tests fails two tests (1 Hive, 1 Kafka Streaming) for me
> locally on 1.1.0-rc3. Does anyone else see that? It may be my env.
> Although I still see the Hive failure on Debian too:
> 
> [info] - SET commands semantics for a HiveContext *** FAILED ***
> [info]   Expected Array("spark.sql.key.usedfortestonly=test.val.0",
> "spark.sql.key.usedfortestonlyspark.sql.key.usedfortestonly=test.val.0test.val.0"),
> but got
> Array("spark.sql.key.usedfortestonlyspark.sql.key.usedfortestonly=test.val.0test.val.0",
> "spark.sql.key.usedfortestonly=test.val.0") (HiveQuerySuite.scala:541)

I've seen this error before.  (In particular, I've seen it on my OS X machine 
using Oracle JDK 8 but not on Fedora using OpenJDK.)  I've also seen similar 
errors in topic branches (but not on master) that seem to indicate that tests 
depend on sets of pairs arriving from Hive in a particular order; it seems that 
this isn't a safe assumption.

I just submitted a (trivial) PR to fix this spurious failure:  
https://github.com/apache/spark/pull/2220


best,
wb

---------------------------------------------------------------------
To unsubscribe, e-mail: dev-unsubscr...@spark.apache.org
For additional commands, e-mail: dev-h...@spark.apache.org

Reply via email to