It worked!! Many thanks for your brilliant support.
On 24 April 2014 18:20, diplomatic Guru <diplomaticg...@gmail.com> wrote: > Many thanks for your prompt reply. I'll try your suggestions and will get > back to you. > > > > > On 24 April 2014 18:17, Michael Armbrust <mich...@databricks.com> wrote: > >> Oh, and you'll also need to add a dependency on "spark-sql_2.10". >> >> >> On Thu, Apr 24, 2014 at 10:13 AM, Michael Armbrust < >> mich...@databricks.com> wrote: >> >>> Yeah, you'll need to run `sbt publish-local` to push the jars to your >>> local maven repository (~/.m2) and then depend on version 1.0.0-SNAPSHOT. >>> >>> >>> On Thu, Apr 24, 2014 at 9:58 AM, diplomatic Guru < >>> diplomaticg...@gmail.com> wrote: >>> >>>> It's a simple application based on the "People" example. >>>> >>>> I'm using Maven for building and below is the pom.xml. Perhaps, I need >>>> to change the version? >>>> >>>> <project> >>>> <groupId>Uthay.Test.App</groupId> >>>> <artifactId>test-app</artifactId> >>>> <modelVersion>4.0.0</modelVersion> >>>> <name>TestApp</name> >>>> <packaging>jar</packaging> >>>> <version>1.0</version> >>>> >>>> <repositories> >>>> <repository> >>>> <id>Akka repository</id> >>>> <url>http://repo.akka.io/releases</url> >>>> </repository> >>>> </repositories> >>>> >>>> <dependencies> >>>> <dependency> <!-- Spark dependency --> >>>> <groupId>org.apache.spark</groupId> >>>> <artifactId>spark-core_2.10</artifactId> >>>> <version>0.9.1</version> >>>> </dependency> >>>> </dependencies> >>>> </project> >>>> >>>> >>>> >>>> On 24 April 2014 17:47, Michael Armbrust <mich...@databricks.com>wrote: >>>> >>>>> You shouldn't need to set SPARK_HIVE=true unless you want to use the >>>>> JavaHiveContext. You should be able to access >>>>> org.apache.spark.sql.api.java.JavaSQLContext with the default build. >>>>> >>>>> How are you building your application? >>>>> >>>>> Michael >>>>> >>>>> >>>>> On Thu, Apr 24, 2014 at 9:17 AM, Andrew Or <and...@databricks.com>wrote: >>>>> >>>>>> Did you build it with SPARK_HIVE=true? >>>>>> >>>>>> >>>>>> On Thu, Apr 24, 2014 at 7:00 AM, diplomatic Guru < >>>>>> diplomaticg...@gmail.com> wrote: >>>>>> >>>>>>> Hi Matei, >>>>>>> >>>>>>> I checked out the git repository and built it. However, I'm still >>>>>>> getting below error. It couldn't find those SQL packages. Please advice. >>>>>>> >>>>>>> package org.apache.spark.sql.api.java does not exist >>>>>>> [ERROR] >>>>>>> /home/VirtualBoxImages.com/Documents/projects/errCount/src/main/java/errorCount/TransDriverSQL.java:[49,8] >>>>>>> cannot find symbol >>>>>>> [ERROR] symbol : class JavaSchemaRDD >>>>>>> >>>>>>> Kind regards, >>>>>>> >>>>>>> Raj. >>>>>>> >>>>>>> >>>>>>> >>>>>>> On 23 April 2014 22:09, Matei Zaharia <matei.zaha...@gmail.com>wrote: >>>>>>> >>>>>>>> It’s currently in the master branch, on >>>>>>>> https://github.com/apache/spark. You can check that out from git, >>>>>>>> build it with sbt/sbt assembly, and then try it out. We’re also going >>>>>>>> to >>>>>>>> post some release candidates soon that will be pre-built. >>>>>>>> >>>>>>>> Matei >>>>>>>> >>>>>>>> On Apr 23, 2014, at 1:30 PM, diplomatic Guru < >>>>>>>> diplomaticg...@gmail.com> wrote: >>>>>>>> >>>>>>>> > Hello Team, >>>>>>>> > >>>>>>>> > I'm new to SPARK and just came across SPARK SQL, which appears to >>>>>>>> be interesting but not sure how I could get it. >>>>>>>> > >>>>>>>> > I know it's an Alpha version but not sure if its available for >>>>>>>> community yet. >>>>>>>> > >>>>>>>> > Many thanks. >>>>>>>> > >>>>>>>> > Raj. >>>>>>>> >>>>>>>> >>>>>>> >>>>>> >>>>> >>>> >>> >> >