Re: calling individual columns from spark temporary table

2016-03-24 Thread Michael Armbrust
df.filter(col("paid") > "").select(col("name1").as("newName"), ...) On Wed, Mar 23, 2016 at 6:17 PM, Ashok Kumar wrote: > Thank you again > > For > > val r = df.filter(col("paid") > "").map(x => > (x.getString(0),x.getString(1).) > > Can you give an example of column

Re: calling individual columns from spark temporary table

2016-03-23 Thread Ashok Kumar
Thank you again For val r = df.filter(col("paid") > "").map(x => (x.getString(0),x.getString(1).) Can you give an example of column expression please like df.filter(col("paid") > "").col("firstcolumn").getString   ? On Thursday, 24 March 2016, 0:45, Michael Armbrust

Re: calling individual columns from spark temporary table

2016-03-23 Thread Michael Armbrust
You can only use as on a Column expression, not inside of a lambda function. The reason is the lambda function is compiled into opaque bytecode that Spark SQL is not able to see. We just blindly execute it. However, there are a couple of ways to name the columns that come out of a map. Either

Re: calling individual columns from spark temporary table

2016-03-23 Thread Ashok Kumar
thank you sir sql("select `_1` as firstcolumn from items") is there anyway one can keep the csv column names using databricks when mapping val r = df.filter(col("paid") > "").map(x => (x.getString(0),x.getString(1).) can I call example  x.getString(0).as.(firstcolumn) in above when mapping

Re: calling individual columns from spark temporary table

2016-03-23 Thread Michael Armbrust
You probably need to use `backticks` to escape `_1` since I don't think that its a valid SQL identifier. On Wed, Mar 23, 2016 at 5:10 PM, Ashok Kumar wrote: > Gurus, > > If I register a temporary table as below > > r.toDF > res58: org.apache.spark.sql.DataFrame =

calling individual columns from spark temporary table

2016-03-23 Thread Ashok Kumar
Gurus, If I register a temporary table as below  r.toDFres58: org.apache.spark.sql.DataFrame = [_1: string, _2: string, _3: double, _4: double, _5: double] r.toDF.registerTempTable("items") sql("select * from items")res60: org.apache.spark.sql.DataFrame = [_1: string, _2: string, _3: double, _4: