Whilst we¹re on this topic, I¹d be interested to see if you get hive failures. I¹m trying to build on a Mac using HDP and seem to be getting failures related to Parquet. I¹ll know for sure once I get in tomorrow and confirm with engineering, but this is likely because the version of Hive is 0.12.0, and Parquet is only supported in Hive 0.13 (HDP is 0.13)
Any idea on what it would take to bump the Hive version up to the latest? Regards, - SteveN On 7/27/14, 19:39, "Stephen Boesch" <java...@gmail.com> wrote: > OK i'll do it after confirming all the tests run > > >2014-07-27 19:36 GMT-07:00 Reynold Xin <r...@databricks.com>: > >> Would you like to submit a pull request? All doc source code are in the >> docs folder. Cheers. >> >> >> >> On Sun, Jul 27, 2014 at 7:35 PM, Stephen Boesch <java...@gmail.com> >>wrote: >> >> > i Reynold, >> > thanks for responding here. Yes I had looked at the building with >>maven >> > page in the past. I have not noticed that the "package" step must >> happen >> > *before *the test. I had assumed it were a corequisite -as seen in my >> > command line. >> > >> > So the following sequence appears to work fine (so far so good - well >> past >> > when the prior attempts failed): >> > >> > >> > mvn -Pyarn -Phadoop-2.3 -DskipTests -Phive clean package >> > mvn -Pyarn -Phadoop-2.3 -Phive test >> > >> > AFA documentation, yes adding another sentence to that same "Building >> with >> > Maven" page would likely be helpful to future generations. >> > >> > >> > 2014-07-27 19:10 GMT-07:00 Reynold Xin <r...@databricks.com>: >> > >> > > To run through all the tests you'd need to create the assembly jar >> first. >> > > >> > > >> > > I've seen this asked a few times. Maybe we should make it more >>obvious. >> > > >> > > >> > > >> > > http://spark.apache.org/docs/latest/building-with-maven.html >> > > >> > > Spark Tests in Maven >> > > >> > > Tests are run by default via the ScalaTest Maven plugin >> > > >><http://www.scalatest.org/user_guide/using_the_scalatest_maven_plugin >> >. >> > > Some of the require Spark to be packaged first, so always run mvn >> package >> > > with -DskipTests the first time. You can then run the tests with >>mvn >> > > -Dhadoop.version=... test. >> > > >> > > The ScalaTest plugin also supports running only a specific test >>suite >> as >> > > follows: >> > > >> > > mvn -Dhadoop.version=... >> -DwildcardSuites=org.apache.spark.repl.ReplSuite >> > > test >> > > >> > > >> > > >> > > >> > > >> > > On Sun, Jul 27, 2014 at 7:07 PM, Stephen Boesch <java...@gmail.com> >> > wrote: >> > > >> > > > I have pulled latest from github this afternoon. There are many >> many >> > > > errors: >> > > > >> > > > <source_home>/assembly/target/scala-2.10: No such file or >>directory >> > > > >> > > > This causes many tests to fail. >> > > > >> > > > Here is the command line I am running >> > > > >> > > > mvn -Pyarn -Phadoop-2.3 -Phive package test >> > > > >> > > >> > >> -- CONFIDENTIALITY NOTICE NOTICE: This message is intended for the use of the individual or entity to which it is addressed and may contain information that is confidential, privileged and exempt from disclosure under applicable law. If the reader of this message is not the intended recipient, you are hereby notified that any printing, copying, dissemination, distribution, disclosure or forwarding of this communication is strictly prohibited. If you have received this communication in error, please contact the sender immediately and delete it from your system. Thank You.