Thanx Dean, i am giving unique output path and in every time i also delete the directory before i run the job.
2015-08-10 15:30 GMT+03:00 Dean Wampler <deanwamp...@gmail.com>: > Following Hadoop conventions, Spark won't overwrite an existing directory. > You need to provide a unique output path every time you run the program, or > delete or rename the target directory before you run the job. > > dean > > Dean Wampler, Ph.D. > Author: Programming Scala, 2nd Edition > <http://shop.oreilly.com/product/0636920033073.do> (O'Reilly) > Typesafe <http://typesafe.com> > @deanwampler <http://twitter.com/deanwampler> > http://polyglotprogramming.com > > On Mon, Aug 10, 2015 at 7:08 AM, Yasemin Kaya <godo...@gmail.com> wrote: > >> Hi, >> >> I have EC2 cluster, and am using spark 1.3, yarn and HDFS . When i submit >> at local there is no problem , but i run at cluster, saveAsTextFile doesn't >> work."*It says me User class threw exception: Output directory >> hdfs://172.31.42.10:54310/...../weblogReadResult >> <http://172.31.42.10:54310/...../weblogReadResult> already exists*" >> >> Is there anyone can help me about this issue ? >> >> Best, >> yasemin >> >> >> >> -- >> hiç ender hiç >> > > -- hiç ender hiç