I am really new to this but what should I look into maven logs?

I have tried mvn package -X -e

SHould I show the full trace?



On Mon, May 4, 2015 at 6:54 PM, Driesprong, Fokko <fo...@driesprong.frl>
wrote:

> Hi Saurabh,
>
> Did you check the log of maven?
>
> 2015-05-04 15:17 GMT+02:00 Saurabh Gupta <saurabh.gu...@semusi.com>:
>
>> HI,
>>
>> I am trying to build a example code given at
>>
>> https://spark.apache.org/docs/latest/sql-programming-guide.html#interoperating-with-rdds
>>
>> code is:
>>
>> // Import factory methods provided by DataType.import 
>> org.apache.spark.sql.types.DataType;// Import StructType and 
>> StructFieldimport org.apache.spark.sql.types.StructType;import 
>> org.apache.spark.sql.types.StructField;// Import Row.import 
>> org.apache.spark.sql.Row;
>> // sc is an existing JavaSparkContext.SQLContext sqlContext = new 
>> org.apache.spark.sql.SQLContext(sc);
>> // Load a text file and convert each line to a JavaBean.JavaRDD<String> 
>> people = sc.textFile("examples/src/main/resources/people.txt");
>> // The schema is encoded in a stringString schemaString = "name age";
>> // Generate the schema based on the string of schemaList<StructField> fields 
>> = new ArrayList<StructField>();for (String fieldName: schemaString.split(" 
>> ")) {
>>   fields.add(DataType.createStructField(fieldName, DataType.StringType, 
>> true));}StructType schema = DataType.createStructType(fields);
>> // Convert records of the RDD (people) to Rows.JavaRDD<Row> rowRDD = 
>> people.map(
>>   new Function<String, Row>() {
>>     public Row call(String record) throws Exception {
>>       String[] fields = record.split(",");
>>       return Row.create(fields[0], fields[1].trim());
>>     }
>>   });
>> // Apply the schema to the RDD.DataFrame peopleDataFrame = 
>> sqlContext.createDataFrame(rowRDD, schema);
>> // Register the DataFrame as a 
>> table.peopleDataFrame.registerTempTable("people");
>> // SQL can be run over RDDs that have been registered as tables.DataFrame 
>> results = sqlContext.sql("SELECT name FROM people");
>> // The results of SQL queries are DataFrames and support all the normal RDD 
>> operations.// The columns of a row in the result can be accessed by 
>> ordinal.List<String> names = results.map(new Function<Row, String>() {
>>   public String call(Row row) {
>>     return "Name: " + row.getString(0);
>>   }}).collect();
>>
>> my pom file looks like:
>>
>> <dependencies>
>>         <dependency>
>>             <groupId>org.apache.spark</groupId>
>>             <artifactId>spark-core_2.10</artifactId>
>>             <version>1.3.1</version>
>>         </dependency>
>>         <dependency>
>>             <groupId>org.apache.spark</groupId>
>>             <artifactId>spark-sql_2.10</artifactId>
>>             <version>1.3.1</version>
>>         </dependency>
>> <dependency>
>>         <groupId>org.apache.hbase</groupId>
>>         <artifactId>hbase</artifactId>
>>         <version>0.94.0</version>
>> </dependency>
>>
>> When I try to mvn package I am getting this issue:
>> cannot find symbol
>> [ERROR] symbol:   variable StringType
>> [ERROR] location: class org.apache.spark.sql.types.DataType
>>
>> I have gone through
>> https://spark.apache.org/docs/latest/api/java/org/apache/spark/sql/types/StringType.html
>>
>> What is missing here?
>>
>>
>

Reply via email to