I am following the Hive build instructions here: https://cwiki.apache.org/confluence/display/Hive/GettingStarted#GettingStarted-InstallationandConfiguration .
I noticed that Hive development seems to be using an old version of Derby: 10.10.2.0. Is there some defect in the most recent Derby version (10.11.1.1) which prevents Hive from upgrading to 10.11.1.1? The only Hive-tagged Derby bug which I can find is https://issues.apache.org/jira/browse/DERBY-6358. That issue doesn't seem to be version-specific and it mentions a resolved Hive issue: https://issues.apache.org/jira/browse/HIVE-8739. Staying with 10.10.2.0 makes sense if you need to run on some ancient JVMs: Java SE 5 or Java ME CDC/Foundation Profile 1.1. Hadoop, however, requires at least Java 6 according to https://wiki.apache.org/hadoop/HadoopJavaVersions. Note that the Derby community expects to release version 10.12.1.1 soon: https://wiki.apache.org/db-derby/DerbyTenTwelveOneRelease. This might be a good opportunity for Hive to upgrade to a more capable version of Derby. I mention this because the Derby version used by Hive ends up on the classpath used by downstream projects (like Spark). That makes it awkward for downstream projects to use more current Derby versions. Do you know of any reason that downstream projects shouldn't override the Derby version currently preferred by Hive? Thanks, -Rick