Github user vanzin commented on the issue:

    https://github.com/apache/spark/pull/21588
  
    The main thing is that this change is changing test coverage based on the 
Hadoop version. So that means that we're effectively changing supported 
versions of Hive here, and we should do all the necessary changes to let people 
know of that. That includes deciding whether disabling those tests is the right 
thing, or whether we should make them work.
    
    The error message you saw seems familiar to you but I'm pretty sure it 
would be very cryptic to someone who's not familiar with this problem. (Why is 
Hive complaining about a Hadoop version if I'm running Spark?)
    
    The Hive 2.1 suite you're disabling is also pretty important to keep 
working, since it tests behavior that changed from Spark's built in version of 
Hive, and most probably is similar in newer versions.
    
    We should be looking at what it means to support Hadoop 3, and answer that 
before we go hacking and disabling things just to get tests to pass.
    
    I also really don't see the point of this before we fix the Hive fork...


---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org

Reply via email to