Kill job gracefully?

2017-04-10 Thread Jane Chen
Using Apache Hadoop 2.6. I’m looking for a way to gracefully abort a Hadoop job so that the proper cleanups, such as RecordReader.close and RecordWriter.close are still called. I found if I use job.killJob(), the tasks are all terminated without the cleanups. Thanks in advance for any advice

how to unsubscribe ?

2017-04-10 Thread Pengcheng Fang
how to unsubscribe ?

Re: Anatomy of read in hdfs

2017-04-10 Thread Sidharth Kumar
Thanks Philippe but your answers raised another sets of questions to me .please help me to understand it 1) If we read anatomy of hdfs read in hadoop definitive guide it says data queue is consumed by streamer. So, can you just tell me that will there be only one streamer in a cluster which consume

Re: Anatomy of read in hdfs

2017-04-10 Thread Philippe Kernévez
On Mon, Apr 10, 2017 at 11:46 AM, Sidharth Kumar < sidharthkumar2...@gmail.com> wrote: > Thanks Philippe, > > I am looking for answer only restricted to HDFS. Because we can do read > and write operations from CLI using commands like "*hadoop fs > -copyfromlocal /(local disk location) /(hdfs path)

Re: Anatomy of read in hdfs

2017-04-10 Thread Sidharth Kumar
Thanks Philippe, I am looking for answer only restricted to HDFS. Because we can do read and write operations from CLI using commands like "*hadoop fs -copyfromlocal /(local disk location) /(hdfs path)" *and read using "*hadoop fs -text /(hdfs file)" *as well. So my question are 1) when I write d

Re: Anatomy of read in hdfs

2017-04-10 Thread Philippe Kernévez
Hi Sidharth, As it has been explained, HDFS is not just a file system. It's a part of the Hadoop platform. To take advantage of HDFS you have to understand how Hadoop storage (HDFS) AND Yarn processing (say MapReduce) work all together to implements jobs and parallel processing. That says that you