finally understand it  

solved  

--  
Nan Zhu
School of Computer Science,
McGill University




On Sunday, December 15, 2013 at 1:43 AM, Nan Zhu wrote:

> Hi, all
>  
> I’m trying to run Spark on EC2 and using S3 as the data storage service,  
>  
> I set fs.default.name (http://fs.default.name) to 
> s3://myaccessid:mysecreteid@bucketid, and I tried to load a local file with 
> textFile
>  
> I found that Spark still tries to find http://mymasterip:9000
>  
> I also tried to load a file stored in s3, the same thing
>  
> Did I misunderstand something?
>  
> I once setup hadoop cluster in ec2 using s3 to store data, it’s 
> straightforward that I only need to set fs.default.name 
> (http://fs.default.name)
>  
> I assume that spark uses hadoop file interfaces to be able to interact with 
> S3, so there should be no difference?
>  
> Best,
>  
> Nan  

Reply via email to