Hi Olivier, Thanks for the reply.
I had two concerns : 1. As I mentioned, I want to un install the entire cluster which means the 9 nodes where there are datanode directories, nn configs. and previously loaded data(less in size), missing blocks(!) etc. should be cleaned i.e I get back the 9 machines which can now be used for fresh installation of a cluster 2. The reset thing will I guess only clear the metadata but the HDFS and other components will remain unchanged - I am not sure if this will solve the problems that I am facing with my existing cluster Regards ! On Thu, Aug 6, 2015 at 10:12 AM, Olivier Renault <[email protected]> wrote: > Log on your ambari-server bring it down and reset it > > $ ambari-server stop > $ ambari-server reset > > If you are using postgresql – installed and configured by ambari server, > you should be able to restart. If you are using MySQL or Oracle, you will > need to drop / re-create the database manually. > > $ ambari-server setup > …. > > Good luck, > Olivier > > From: Kaliyug Antagonist > Reply-To: "[email protected]" > Date: Thursday, 6 August 2015 10:07 > To: "[email protected]" > Subject: Clean un-installation via Ambari > > I had installed HDP-2.2.4.2-2 using Ambari Version 2.0.0. > > There have been several issues in the cluster due to misconfiguration in > the datanode directories and so on. Now I get several alerts and any MR > that I execute fails with an error like this : > > 15/06/01 13:53:44 INFO mapreduce.Job: Job job_1431689151537_0003 running > in uber mode : false > 15/06/01 13:53:44 INFO mapreduce.Job: map 0% reduce 0% > 15/06/01 13:53:47 INFO mapreduce.Job: Task Id : > attempt_1431689151537_0003_m_000000_1000, Status : FAILED > java.io.FileNotFoundException: File > /opt/dev/sdb/hadoop/yarn/local/filecache does not exist > > 15/06/01 13:53:51 INFO mapreduce.Job: Task Id : > attempt_1431689151537_0003_m_000000_1001, Status : FAILED > java.io.FileNotFoundException: File > /opt/dev/sdd/hadoop/yarn/local/filecache does not exist > > 15/06/01 13:53:55 INFO mapreduce.Job: Task Id : > attempt_1431689151537_0003_m_000000_1002, Status : FAILED > java.io.FileNotFoundException: File > /opt/dev/sdh/hadoop/yarn/local/filecache does not exist > > > I wish to clean un install the cluster and reinstall it, it is ok even if > Ambari needs to be uninstalled and reinstalled. > > How can I do it ? >
