Failed to rollback from hadoop-2.4.1 to hadoop 2.2.0

2014-09-18 Thread sam liu
Hi Expert, Below are my steps and is it a hadoop bug or did I miss any thing? Thanks! Step: [A] Upgrade 1. Install Hadoop 2.2.0 cluster 2. Stop Hadoop services 3. Replace 2.2.0 binaries with 2.4.1 binaries 4. Start datanodes: $HADOOP_HOME/sbin/hadoop-daemon.sh start datanode 5. Start namenode

AW: ClassCastException on running map-reduce jobs + tests on Windows (mongo-hadoop)

2014-09-18 Thread Blanca Hernandez
Thanks, I made the changes and everything works fine!! Many thanks!! Now I am having problems converting BSONWritable to BSONObject and vice versa. Is there an automatic way to make it? Or should I write myself a parse? And regarding the tests on windows, any experience? Thanks again!! Best

RE: hbase use error

2014-09-18 Thread QiXiangming
Dear Yu, this is a snippet of log , thank you for diagnosis! it seems the upstream socket cracks. but i can't find valuable clue from other datanode's log! -- 2014-09-17 11:02:30,058 INFO

Re: Failed to rollback from hadoop-2.4.1 to hadoop 2.2.0

2014-09-18 Thread Susheel Kumar Gadalay
You have to upgrade both name node and data node. Better issue start-dfs.sh -upgrade. Check whether current and previous directories are present in both dfs.namenode.name.dir and dfs.datanode.data.dir directory. On 9/18/14, sam liu samliuhad...@gmail.com wrote: Hi Expert, Below are my steps

Re: core-site.xml not found while installing CDH 5 with YARN on a Single Linux Node in Pseudo-distributed mode

2014-09-18 Thread Ravindra
Please check if /etc/hadoop/conf/ exists. If it exists then export enviroment variable HADOOP_CONF_DIR set to this path. On Thu, Sep 18, 2014 at 2:46 PM, Vandana kumari kvandana1...@gmail.com wrote: Hello all I am manually Installing CDH 5 with YARN on a Single Linux Node in

Re: Failed to rollback from hadoop-2.4.1 to hadoop 2.2.0

2014-09-18 Thread sam liu
Thanks for your comment! I can upgrade from 2.2.0 to 2.4.1 using command 'start-dfs.sh -upgrade', however failed to rollback from 2.4.1 to 2.2.0 using command 'start-dfs.sh -rollback': the namenode always stays on safe mode(awaiting reported blocks (0/315)). Why? 2014-09-18 1:51 GMT-07:00

Re: core-site.xml not found while installing CDH 5 with YARN on a Single Linux Node in Pseudo-distributed mode

2014-09-18 Thread Vandana kumari
In the installation process /etc/hadoop/conf.pseudo/ directory was made which contains all the hdfs files: core-site.xml, hdfs-site.xml, mapred-site.xml, yarn-site.xml On Thu, Sep 18, 2014 at 2:53 PM, Ravindra ravin.i...@gmail.com wrote: Please check if /etc/hadoop/conf/ exists. If it exists

Re: core-site.xml not found while installing CDH 5 with YARN on a Single Linux Node in Pseudo-distributed mode

2014-09-18 Thread Ravindra
Please check if /etc/hadoop/conf/ is present and is a symbolic link to some other directory that doesn't exists. Ideally this should be take care by installer, are you sure that you didn't have an already existing hadoop setup on that machine? Regards, Ravindra On Thu, Sep 18, 2014 at 3:12 PM,

Re: core-site.xml not found while installing CDH 5 with YARN on a Single Linux Node in Pseudo-distributed mode

2014-09-18 Thread Vandana kumari
/etc/hadoop/conf/ is not present. Earlier i tried to install hadoop by apache hadoop and cdh4 installation too, but i had uninstalled both, still unable to figure out the error On Thu, Sep 18, 2014 at 3:41 PM, Ravindra ravin.i...@gmail.com wrote: Please check if /etc/hadoop/conf/ is present

Re: core-site.xml not found while installing CDH 5 with YARN on a Single Linux Node in Pseudo-distributed mode

2014-09-18 Thread Ritesh Kumar Singh
just for check, what's the output for: whereis hadoop and ls /etc/hadoop/ On Thu, Sep 18, 2014 at 3:46 PM, Vandana kumari kvandana1...@gmail.com wrote: /etc/hadoop/conf/ is not present. Earlier i tried to install hadoop by apache hadoop and cdh4 installation too, but i had uninstalled

Re: Failed to rollback from hadoop-2.4.1 to hadoop 2.2.0

2014-09-18 Thread Susheel Kumar Gadalay
What is the o/p of command hdfs dfsadmin -upgradeProgress status If it says upgrade is complete then you can do some sanity check by hdfs fsck. Stop the servers by stop-dfs.sh and then rollback by command start-dfs.sh -rollback On 9/18/14, sam liu samliuhad...@gmail.com wrote: Thanks for your

Re: core-site.xml not found while installing CDH 5 with YARN on a Single Linux Node in Pseudo-distributed mode

2014-09-18 Thread Ravindra
A quick fix is to run this command ln -s /etc/hadoop/conf.pseudo /etc/hadoop/conf On Thu, Sep 18, 2014 at 3:46 PM, Vandana kumari kvandana1...@gmail.com wrote: /etc/hadoop/conf/ is not present. Earlier i tried to install hadoop by apache hadoop and cdh4 installation too, but i had

Re: ClassCastException on running map-reduce jobs + tests on Windows (mongo-hadoop)

2014-09-18 Thread Shahab Yunus
You will have to convert BSONWritable to BSONObject yourself. You can abstract this parsing in a separate class/object model and reuse it but as far as I understand, objects being serialized or deserialized have to be Writable (conforming to the interface that Hadoop defines and Comparable if

Re: YARN ResourceManager Hostname in HA configuration

2014-09-18 Thread Matt Narrell
Correct. This is for client communication to the cluster. I’m talking about “yarn.resourcemanager.hostname” which is still used to construct the composite UI, it seems. mn On Sep 17, 2014, at 11:02 PM, Susheel Kumar Gadalay skgada...@gmail.com wrote: I observed in Yarn Cluster, you set

AW: ClassCastException on running map-reduce jobs + tests on Windows (mongo-hadoop)

2014-09-18 Thread Blanca Hernandez
Thanks a millio for your support! Von: Shahab Yunus [mailto:shahab.yu...@gmail.com] Gesendet: Donnerstag, 18. September 2014 13:40 An: user@hadoop.apache.org Betreff: Re: ClassCastException on running map-reduce jobs + tests on Windows (mongo-hadoop) You will have to convert BSONWritable to

Re: core-site.xml not found while installing CDH 5 with YARN on a Single Linux Node in Pseudo-distributed mode

2014-09-18 Thread Vandana kumari
No its not working ravindra and showing following error: SLF4J: Failed to load class org.slf4j.impl.StaticLoggerBinder. SLF4J: Defaulting to no-operation (NOP) logger implementation SLF4J: See http://www.slf4j.org/codes.html#StaticLoggerBinder for further details. ls: Call From belief/127.0.0.1

Re: Enable Debug logging for a job

2014-09-18 Thread Siddhi Mehta
Yes I crossed checked the final setting on the Job run on RM Configuration tab was set to DEBUG. Thanks, Siddhi On Fri, Sep 12, 2014 at 11:24 AM, Jakub Stransky stransky...@gmail.com wrote: Hi Siddhi, I did this today for Hadoop 2.2.0 with difference that I was setting level to WARN and on

RE: Issue with Hadoop/Kerberos security as client

2014-09-18 Thread John Lilley
Follow-up: problem seem to be solved. The issue is purely a JNI issue and actually doesn't have anything to do with security per se. Calling vm-AttachCurrentThread() sets the context class loader to the bootstrap loader, which is insufficient for ServiceLoader to function properly. Any C++