It should just work in PySpark, the same way it does in Java / Scala apps.

Matei

On Oct 1, 2014, at 4:12 PM, Sungwook Yoon <sy...@maprtech.com> wrote:

> 
> Yes.. you should use maprfs://
> 
> I personally haven't used pyspark, I just used scala shell or standalone with 
> MapR.
> 
> I think you need to set classpath right, adding jar like
>  /opt/mapr/hadoop/hadoop-0.20.2/lib/hadoop-0.20.2-dev-core.jar to the 
> classpath
> in the classpath.
> 
> Sungwook
> 
> On Wed, Oct 1, 2014 at 4:09 PM, Addanki, Santosh Kumar 
> <santosh.kumar.adda...@sap.com> wrote:
> Hi
> 
>  
> 
> We would like to do this in PySpark Environment
> 
>  
> 
> i.e something like
> 
>  
> 
> test = sc.textFile("maprfs:///user/root/test") or
> 
> test = sc.textFile("hdfs:///user/root/test") or
> 
>  
> 
>  
> 
> Currently when we try
> 
> test = sc.textFile("maprfs:///user/root/test")
> 
>  
> 
> It throws the error
> 
> No File-System for scheme: maprfs
> 
>  
> 
>  
> 
> Best Regards
> 
> Santosh
> 
>  
> 
>  
> 
>  
> 
> From: Vladimir Rodionov [mailto:vrodio...@splicemachine.com] 
> Sent: Wednesday, October 01, 2014 3:59 PM
> To: Addanki, Santosh Kumar
> Cc: user@spark.apache.org
> Subject: Re: Spark And Mapr
> 
>  
> 
> There is doc on MapR:
> 
>  
> 
> http://doc.mapr.com/display/MapR/Accessing+MapR-FS+in+Java+Applications
> 
>  
> 
> -Vladimir Rodionov
> 
>  
> 
> On Wed, Oct 1, 2014 at 3:00 PM, Addanki, Santosh Kumar 
> <santosh.kumar.adda...@sap.com> wrote:
> 
> Hi
> 
>  
> 
> We were using Horton 2.4.1 as our Hadoop distribution and now switched to MapR
> 
>  
> 
> Previously to read a text file  we would use :
> 
>  
> 
> test = sc.textFile(\"hdfs://10.48.101.111:8020/user/hdfs/test\")"
> 
>  
> 
>  
> 
> What would be the equivalent of the same for Mapr.
> 
>  
> 
> Best Regards
> 
> Santosh
> 
>  
> 
> 

Reply via email to