Oh, and you'll also need to add a dependency on "spark-sql_2.10".
On Thu, Apr 24, 2014 at 10:13 AM, Michael Armbrust <mich...@databricks.com>wrote: > Yeah, you'll need to run `sbt publish-local` to push the jars to your > local maven repository (~/.m2) and then depend on version 1.0.0-SNAPSHOT. > > > On Thu, Apr 24, 2014 at 9:58 AM, diplomatic Guru <diplomaticg...@gmail.com > > wrote: > >> It's a simple application based on the "People" example. >> >> I'm using Maven for building and below is the pom.xml. Perhaps, I need to >> change the version? >> >> <project> >> <groupId>Uthay.Test.App</groupId> >> <artifactId>test-app</artifactId> >> <modelVersion>4.0.0</modelVersion> >> <name>TestApp</name> >> <packaging>jar</packaging> >> <version>1.0</version> >> >> <repositories> >> <repository> >> <id>Akka repository</id> >> <url>http://repo.akka.io/releases</url> >> </repository> >> </repositories> >> >> <dependencies> >> <dependency> <!-- Spark dependency --> >> <groupId>org.apache.spark</groupId> >> <artifactId>spark-core_2.10</artifactId> >> <version>0.9.1</version> >> </dependency> >> </dependencies> >> </project> >> >> >> >> On 24 April 2014 17:47, Michael Armbrust <mich...@databricks.com> wrote: >> >>> You shouldn't need to set SPARK_HIVE=true unless you want to use the >>> JavaHiveContext. You should be able to access >>> org.apache.spark.sql.api.java.JavaSQLContext with the default build. >>> >>> How are you building your application? >>> >>> Michael >>> >>> >>> On Thu, Apr 24, 2014 at 9:17 AM, Andrew Or <and...@databricks.com>wrote: >>> >>>> Did you build it with SPARK_HIVE=true? >>>> >>>> >>>> On Thu, Apr 24, 2014 at 7:00 AM, diplomatic Guru < >>>> diplomaticg...@gmail.com> wrote: >>>> >>>>> Hi Matei, >>>>> >>>>> I checked out the git repository and built it. However, I'm still >>>>> getting below error. It couldn't find those SQL packages. Please advice. >>>>> >>>>> package org.apache.spark.sql.api.java does not exist >>>>> [ERROR] >>>>> /home/VirtualBoxImages.com/Documents/projects/errCount/src/main/java/errorCount/TransDriverSQL.java:[49,8] >>>>> cannot find symbol >>>>> [ERROR] symbol : class JavaSchemaRDD >>>>> >>>>> Kind regards, >>>>> >>>>> Raj. >>>>> >>>>> >>>>> >>>>> On 23 April 2014 22:09, Matei Zaharia <matei.zaha...@gmail.com> wrote: >>>>> >>>>>> It's currently in the master branch, on >>>>>> https://github.com/apache/spark. You can check that out from git, >>>>>> build it with sbt/sbt assembly, and then try it out. We're also going to >>>>>> post some release candidates soon that will be pre-built. >>>>>> >>>>>> Matei >>>>>> >>>>>> On Apr 23, 2014, at 1:30 PM, diplomatic Guru < >>>>>> diplomaticg...@gmail.com> wrote: >>>>>> >>>>>> > Hello Team, >>>>>> > >>>>>> > I'm new to SPARK and just came across SPARK SQL, which appears to >>>>>> be interesting but not sure how I could get it. >>>>>> > >>>>>> > I know it's an Alpha version but not sure if its available for >>>>>> community yet. >>>>>> > >>>>>> > Many thanks. >>>>>> > >>>>>> > Raj. >>>>>> >>>>>> >>>>> >>>> >>> >> >