Yes.. you should use maprfs:// I personally haven't used pyspark, I just used scala shell or standalone with MapR.
I think you need to set classpath right, adding jar like /opt/mapr/hadoop/hadoop-0.20.2/lib/hadoop-0.20.2-dev-core.jar to the classpath in the classpath. Sungwook On Wed, Oct 1, 2014 at 4:09 PM, Addanki, Santosh Kumar < santosh.kumar.adda...@sap.com> wrote: > Hi > > > > We would like to do this in PySpark Environment > > > > i.e something like > > > > test = sc.textFile("maprfs:///user/root/test") or > > test = sc.textFile("hdfs:///user/root/test") or > > > > > > Currently when we try > > test = sc.textFile("maprfs:///user/root/test") > > > > It throws the error > > No File-System for scheme: maprfs > > > > > > Best Regards > > Santosh > > > > > > > > *From:* Vladimir Rodionov [mailto:vrodio...@splicemachine.com] > *Sent:* Wednesday, October 01, 2014 3:59 PM > *To:* Addanki, Santosh Kumar > *Cc:* user@spark.apache.org > *Subject:* Re: Spark And Mapr > > > > There is doc on MapR: > > > > http://doc.mapr.com/display/MapR/Accessing+MapR-FS+in+Java+Applications > > > > -Vladimir Rodionov > > > > On Wed, Oct 1, 2014 at 3:00 PM, Addanki, Santosh Kumar < > santosh.kumar.adda...@sap.com> wrote: > > Hi > > > > We were using Horton 2.4.1 as our Hadoop distribution and now switched to > MapR > > > > Previously to read a text file we would use : > > > > test = sc.textFile(\"hdfs://10.48.101.111:8020/user/hdfs/test\ > <http://10.48.101.111:8020/user/hdfs/test%5C>")" > > > > > > What would be the equivalent of the same for Mapr. > > > > Best Regards > > Santosh > > >