Looks like you're depending on Spark 0.9.1, which doesn't have Spark SQL.
Assuming you've downloaded Spark, just run 'mvn install' to publish Spark
locally, and depend on Spark version 1.0.0-SNAPSHOT.


On Thu, Apr 24, 2014 at 9:58 AM, diplomatic Guru
<diplomaticg...@gmail.com>wrote:

> It's a simple application based on the "People" example.
>
> I'm using Maven for building and below is the pom.xml. Perhaps, I need to
> change the version?
>
> <project>
>   <groupId>Uthay.Test.App</groupId>
>     <artifactId>test-app</artifactId>
>       <modelVersion>4.0.0</modelVersion>
>         <name>TestApp</name>
>           <packaging>jar</packaging>
>             <version>1.0</version>
>
>         <repositories>
>             <repository>
>                  <id>Akka repository</id>
>                  <url>http://repo.akka.io/releases</url>
>             </repository>
>         </repositories>
>
>         <dependencies>
>            <dependency> <!-- Spark dependency -->
>                 <groupId>org.apache.spark</groupId>
>                 <artifactId>spark-core_2.10</artifactId>
>                 <version>0.9.1</version>
>            </dependency>
>         </dependencies>
> </project>
>
>
>
> On 24 April 2014 17:47, Michael Armbrust <mich...@databricks.com> wrote:
>
>> You shouldn't need to set SPARK_HIVE=true unless you want to use the
>> JavaHiveContext.  You should be able to access
>> org.apache.spark.sql.api.java.JavaSQLContext with the default build.
>>
>> How are you building your application?
>>
>> Michael
>>
>>
>> On Thu, Apr 24, 2014 at 9:17 AM, Andrew Or <and...@databricks.com> wrote:
>>
>>> Did you build it with SPARK_HIVE=true?
>>>
>>>
>>> On Thu, Apr 24, 2014 at 7:00 AM, diplomatic Guru <
>>> diplomaticg...@gmail.com> wrote:
>>>
>>>> Hi Matei,
>>>>
>>>> I checked out the git repository and built it. However, I'm still
>>>> getting below error. It couldn't find those SQL packages. Please advice.
>>>>
>>>> package org.apache.spark.sql.api.java does not exist
>>>> [ERROR]
>>>> /home/VirtualBoxImages.com/Documents/projects/errCount/src/main/java/errorCount/TransDriverSQL.java:[49,8]
>>>> cannot find symbol
>>>> [ERROR] symbol  : class JavaSchemaRDD
>>>>
>>>> Kind regards,
>>>>
>>>> Raj.
>>>>
>>>>
>>>>
>>>> On 23 April 2014 22:09, Matei Zaharia <matei.zaha...@gmail.com> wrote:
>>>>
>>>>> It’s currently in the master branch, on
>>>>> https://github.com/apache/spark. You can check that out from git,
>>>>> build it with sbt/sbt assembly, and then try it out. We’re also going to
>>>>> post some release candidates soon that will be pre-built.
>>>>>
>>>>> Matei
>>>>>
>>>>> On Apr 23, 2014, at 1:30 PM, diplomatic Guru <diplomaticg...@gmail.com>
>>>>> wrote:
>>>>>
>>>>> > Hello Team,
>>>>> >
>>>>> > I'm new to SPARK and just came across SPARK SQL, which appears to be
>>>>> interesting but not sure how I could get it.
>>>>> >
>>>>> > I know it's an Alpha version but not sure if its available for
>>>>> community yet.
>>>>> >
>>>>> > Many thanks.
>>>>> >
>>>>> > Raj.
>>>>>
>>>>>
>>>>
>>>
>>
>

Reply via email to