Hi,
using dataframes you can use SQL, and SQL has an option of JOIN, BETWEEN,
IN and LIKE OPERATIONS. Why would someone use a dataframe and then use them
as RDD's? :)
Regards,
Gourav Sengupta
On Thu, Mar 3, 2016 at 4:28 PM, Sumedh Wale wrote:
> On Thursday 03 March 2016
On Thursday 03 March 2016 09:15 PM,
Gourav Sengupta wrote:
Hi,
why not read the table into a dataframe directly using SPARK
CSV package. You are trying to solve the problem the
Hi,
why not read the table into a dataframe directly using SPARK CSV package.
You are trying to solve the problem the round about way.
Regards,
Gourav Sengupta
On Thu, Mar 3, 2016 at 12:33 PM, Sumedh Wale wrote:
> On Thursday 03 March 2016 11:03 AM, Angel Angel wrote:
>
On Thursday 03 March 2016 11:03 AM,
Angel Angel wrote:
Hello Sir/Madam,
I am writing one application using spark sql.
i made the vary big table using the following command
val
Have you seen the thread 'Filter on a column having multiple values' where
Michael gave this example ?
https://databricks-prod-cloudfront.cloud.databricks.com/public/4027ec902e239c93eaaa8714f173bcfc/1023043053387187/107522969592/2840265927289860/2388bac36e.html
FYI
On Wed, Mar 2, 2016 at
Hello Sir/Madam,
I am writing one application using spark sql.
i made the vary big table using the following command
*val dfCustomers1 =
sc.textFile("/root/Desktop/database.txt").map(_.split(",")).map(p =>
Customer1(p(0),p(1).trim.toInt, p(2).trim.toInt, p(3)))toDF*
Now i want to search the