<http://www.amazon.com/Big-Data-Analytics-Spark-Practitioners/dp/1484209656/>
>
>
>
> *From:* Ashok Kumar [mailto:ashok34...@yahoo.com.INVALID]
> *Sent:* Friday, February 26, 2016 9:41 AM
> *To:* User
> *Subject:* Clarification on RDD
>
>
>
> Hi,
>
>
>
> Spark
-project-dist/hadoop-hdfs/HdfsDesign.html
Mohammed
Author: Big Data Analytics with
Spark<http://www.amazon.com/Big-Data-Analytics-Spark-Practitioners/dp/1484209656/>
From: Ashok Kumar [mailto:ashok34...@yahoo.com.INVALID]
Sent: Friday, February 26, 2016 9:41 AM
To: User
Subject: Clarifi
Hi,
Spark doco says
Spark’s primary abstraction is a distributed collection of items called a
Resilient Distributed Dataset (RDD). RDDs can be created from Hadoop
InputFormats (such as HDFS files) or by transforming other RDDs
example:
val textFile = sc.textFile("README.md")
my question is