Using Apache Hadoop 2.6. I’m looking for a way to gracefully abort a Hadoop
job so that the proper cleanups, such as RecordReader.close and
RecordWriter.close are still called. I found if I use job.killJob(), the tasks
are all terminated without the cleanups.
Thanks in advance for any advice
how to unsubscribe ?
Thanks Philippe but your answers raised another sets of questions to me
.please help me to understand it
1) If we read anatomy of hdfs read in hadoop definitive guide it says data
queue is consumed by streamer. So, can you just tell me that will there be
only one streamer in a cluster which consume
On Mon, Apr 10, 2017 at 11:46 AM, Sidharth Kumar <
sidharthkumar2...@gmail.com> wrote:
> Thanks Philippe,
>
> I am looking for answer only restricted to HDFS. Because we can do read
> and write operations from CLI using commands like "*hadoop fs
> -copyfromlocal /(local disk location) /(hdfs path)
Thanks Philippe,
I am looking for answer only restricted to HDFS. Because we can do read and
write operations from CLI using commands like "*hadoop fs -copyfromlocal
/(local disk location) /(hdfs path)" *and read using "*hadoop fs -text
/(hdfs file)" *as well.
So my question are
1) when I write d
Hi Sidharth,
As it has been explained, HDFS is not just a file system. It's a part of
the Hadoop platform. To take advantage of HDFS you have to understand how
Hadoop storage (HDFS) AND Yarn processing (say MapReduce) work all together
to implements jobs and parallel processing.
That says that you