Hi, all

I’m trying to run Spark on EC2 and using S3 as the data storage service,  

I set fs.default.name to s3://myaccessid:mysecreteid@bucketid, and I tried to 
load a local file with textFile

I found that Spark still tries to find http://mymasterip:9000

I also tried to load a file stored in s3, the same thing

Did I misunderstand something?

I once setup hadoop cluster in ec2 using s3 to store data, it’s straightforward 
that I only need to set fs.default.name

I assume that spark uses hadoop file interfaces to be able to interact with S3, 
so there should be no difference?

Best,

Nan

Reply via email to