Sorry I forgot to ask how can I use spark context here ? I have hdfs directory path of the files, as well as the name node of hdfs cluster.
Thanks for your help. On Mon, Nov 21, 2016 at 9:45 PM, Raghav <raghavas...@gmail.com> wrote: > Hi > > I am extremely new to Spark. I have to read a file form HDFS, and get it > in memory in RDD format. > > I have a Java class as follows: > > class Person { > private long UUID; > private String FirstName; > private String LastName; > private String zip; > > // public methods > } > > The file in HDFS is as follows: > > UUID. FirstName LastName Zip > 7462 John Doll 06903 > 5231 Brad Finley 32820 > > > Can someone point me how to get a JavaRDD<Person> object by reading the > file in HDFS ? > > Thanks. > > -- > Raghav > -- Raghav