@Jean the log which i got at the start of running the hadoop jar: Maybe you can spot something
11:51:44,431 INFO ZooKeeper:100 - Client environment:java.library.path=/usr/lib/hadoop/lib/native 11:51:44,432 INFO ZooKeeper:100 - Client environment:java.io.tmpdir=/tmp 11:51:44,432 INFO ZooKeeper:100 - Client environment:java.compiler=<NA> 11:51:44,432 INFO ZooKeeper:100 - Client environment:os.name=Linux 11:51:44,432 INFO ZooKeeper:100 - Client environment:os.arch=amd64 11:51:44,432 INFO ZooKeeper:100 - Client environment:os.version=3.2.0-23-virtual 11:51:44,432 INFO ZooKeeper:100 - Client environment:user.name=root 11:51:44,433 INFO ZooKeeper:100 - Client environment:user.home=/root 11:51:44,433 INFO ZooKeeper:100 - Client environment:user.dir=/home/ubuntu/pasudhee/ActionDataInterpret 11:51:44,437 INFO ZooKeeper:438 - Initiating client connection, connectString=localhost:2181 sessionTimeout=180000 watcher=hconnection 11:51:44,493 INFO ClientCnxn:966 - Opening socket connection to server localhost/127.0.0.1:2181. Will not attempt to authenticate using SASL (Unable to locate a login configuration) 11:51:44,500 INFO RecoverableZooKeeper:104 - The identifier of this process is [email protected] 11:51:44,513 INFO ClientCnxn:849 - Socket connection established to localhost/127.0.0.1:2181, initiating session 11:51:44,532 INFO ClientCnxn:1207 - Session establishment complete on server localhost/127.0.0.1:2181, sessionid = 0x13ff1cff71bb167, negotiated timeout = 60000 11:51:44,743 INFO ZooKeeper:438 - Initiating client connection, connectString=localhost:2181 sessionTimeout=180000 watcher=hconnection 11:51:44,747 INFO ClientCnxn:966 - Opening socket connection to server localhost/127.0.0.1:2181. Will not attempt to authenticate using SASL (Unable to locate a login configuration) 11:51:44,747 INFO ClientCnxn:849 - Socket connection established to localhost/127.0.0.1:2181, initiating session 11:51:44,747 INFO RecoverableZooKeeper:104 - The identifier of this process is [email protected] 11:51:44,749 INFO ClientCnxn:1207 - Session establishment complete on server localhost/127.0.0.1:2181, sessionid = 0x13ff1cff71bb168, negotiated timeout = 60000 11:51:44,803 WARN Configuration:824 - hadoop.native.lib is deprecated. Instead, use io.native.lib.available 11:51:45,051 INFO HConnectionManager$HConnectionImplementation:1789 - Closed zookeeper sessionid=0x13ff1cff71bb168 11:51:45,054 INFO ZooKeeper:684 - Session: 0x13ff1cff71bb168 closed 11:51:45,054 INFO ClientCnxn:509 - EventThread shut down 11:51:45,057 INFO ZooKeeper:438 - Initiating client connection, connectString=localhost:2181 sessionTimeout=180000 watcher=hconnection 11:51:45,059 INFO ClientCnxn:966 - Opening socket connection to server localhost/127.0.0.1:2181. Will not attempt to authenticate using SASL (Unable to locate a login configuration) 11:51:45,060 INFO ClientCnxn:849 - Socket connection established to localhost/127.0.0.1:2181, initiating session 11:51:45,061 INFO ClientCnxn:1207 - Session establishment complete on server localhost/127.0.0.1:2181, sessionid = 0x13ff1cff71bb169, negotiated timeout = 60000 11:51:45,065 INFO RecoverableZooKeeper:104 - The identifier of this process is [email protected] 11:51:45,135 INFO ZooKeeper:438 - Initiating client connection, connectString=10.34.187.170:2181 sessionTimeout=180000 watcher=hconnection 11:51:45,137 INFO ClientCnxn:966 - Opening socket connection to server ip-10-34-187-170.eu-west-1.compute.internal/10.34.187.170:2181. Will not attempt to authenticate using SASL (Unable to locate a login configuration) 11:51:45,138 INFO ClientCnxn:849 - Socket connection established to ip-10-34-187-170.eu-west-1.compute.internal/10.34.187.170:2181, initiating session 11:51:45,138 INFO RecoverableZooKeeper:104 - The identifier of this process is [email protected] 11:51:45,140 INFO ClientCnxn:1207 - Session establishment complete on server ip-10-34-187-170.eu-west-1.compute.internal/10.34.187.170:2181, sessionid = 0x13ff1cff71bb16a, negotiated timeout = 60000 11:51:45,173 INFO ZooKeeper:438 - Initiating client connection, connectString=10.34.187.170:2181 sessionTimeout=180000 watcher=catalogtracker-on-org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation@7444f787 11:51:45,176 INFO ClientCnxn:966 - Opening socket connection to server ip-10-34-187-170.eu-west-1.compute.internal/10.34.187.170:2181. Will not attempt to authenticate using SASL (Unable to locate a login configuration) 11:51:45,176 INFO ClientCnxn:849 - Socket connection established to ip-10-34-187-170.eu-west-1.compute.internal/10.34.187.170:2181, initiating session 11:51:45,178 INFO ClientCnxn:1207 - Session establishment complete on server ip-10-34-187-170.eu-west-1.compute.internal/10.34.187.170:2181, sessionid = 0x13ff1cff71bb16b, negotiated timeout = 60000 11:51:45,180 INFO RecoverableZooKeeper:104 - The identifier of this process is [email protected] 11:51:45,211 INFO ZooKeeper:684 - Session: 0x13ff1cff71bb16b closed 11:51:45,211 INFO ClientCnxn:509 - EventThread shut down 11:51:45,218 INFO ZooKeeper:438 - Initiating client connection, connectString=10.34.187.170:2181 sessionTimeout=180000 watcher=catalogtracker-on-org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation@7444f787 11:51:45,220 INFO ClientCnxn:966 - Opening socket connection to server ip-10-34-187-170.eu-west-1.compute.internal/10.34.187.170:2181. Will not attempt to authenticate using SASL (Unable to locate a login configuration) On Wed, Aug 21, 2013 at 6:23 PM, Pavan Sudheendra <[email protected]>wrote: > Yes .. The zookeeper server is also 10.34.187.170 .. > > > On Wed, Aug 21, 2013 at 6:21 PM, Jean-Marc Spaggiari < > [email protected]> wrote: > >> Are you able to connect to your ZK server shell and list the nodes? >> >> 2013/8/21 Pavan Sudheendra <[email protected]> >> >> > Yes.. I can do everything.. But i do not want my Hadoop Namenode to >> report >> > logs like this.. Also, it says >> > >> > KeeperException, re-throwing exception >> > org.apache.zookeeper.KeeperException$ConnectionLossException: >> > KeeperErrorCode = ConnectionLoss for /hbase/ >> > >> > >> > >> > >> > >> > On Wed, Aug 21, 2013 at 6:14 PM, Jean-Marc Spaggiari < >> > [email protected]> wrote: >> > >> > > Sound correct. You are able to start the shell and scan the few first >> > line >> > > of the tables, right? >> > > >> > > 2013/8/21 Pavan Sudheendra <[email protected]> >> > > >> > > > This is my hbase-site.xml file if it helps: >> > > > >> > > > <?xml version="1.0" encoding="UTF-8"?> >> > > > >> > > > <!--Autogenerated by Cloudera CM on 2013-07-09T09:26:49.841Z--> >> > > > <configuration> >> > > > <property> >> > > > <name>hbase.rootdir</name> >> > > > >> > > > >> > > > >> > > >> > >> <value>hdfs://ip-10-34-187-170.eu-west-1.compute.internal:8020/hbase</value> >> > > > </property> >> > > > <property> >> > > > <name>hbase.client.write.buffer</name> >> > > > <value>2097152</value> >> > > > </property> >> > > > <property> >> > > > <name>hbase.client.pause</name> >> > > > <value>1000</value> >> > > > </property> >> > > > <property> >> > > > <name>hbase.client.retries.number</name> >> > > > <value>10</value> >> > > > </property> >> > > > <property> >> > > > <name>hbase.client.scanner.caching</name> >> > > > <value>1</value> >> > > > </property> >> > > > <property> >> > > > <name>hbase.client.keyvalue.maxsize</name> >> > > > <value>10485760</value> >> > > > </property> >> > > > <property> >> > > > <name>hbase.rpc.timeout</name> >> > > > <value>60000</value> >> > > > </property> >> > > > <property> >> > > > <name>hbase.security.authentication</name> >> > > > <value>simple</value> >> > > > </property> >> > > > <property> >> > > > <name>zookeeper.session.timeout</name> >> > > > <value>60000</value> >> > > > </property> >> > > > <property> >> > > > <name>zookeeper.znode.parent</name> >> > > > <value>/hbase</value> >> > > > </property> >> > > > <property> >> > > > <name>zookeeper.znode.rootserver</name> >> > > > <value>root-region-server</value> >> > > > </property> >> > > > <property> >> > > > <name>hbase.zookeeper.quorum</name> >> > > > <value>ip-10-34-187-170.eu-west-1.compute.internal</value> >> > > > </property> >> > > > <property> >> > > > <name>hbase.zookeeper.property.clientPort</name> >> > > > <value>2181</value> >> > > > </property> >> > > > </configuration> >> > > > >> > > > >> > > > >> > > > On Wed, Aug 21, 2013 at 6:09 PM, Jean-Marc Spaggiari < >> > > > [email protected]> wrote: >> > > > >> > > > > Hum. >> > > > > >> > > > > Things seems to be correct there. >> > > > > >> > > > > Can you try something simple like: >> > > > > >> > > > > Configuration config = HBaseConfiguration.create(); >> > > > > config.set("hbase.zookeeper.quorum", >> "ip-10-34-187-170"); >> > > > > HTable table = new HTable (config, >> > > > > Bytes.toBytes("TABLE_NAME")); >> > > > > >> > > > > And see if it works? >> > > > > >> > > > > JM >> > > > > >> > > > > 2013/8/21 Pavan Sudheendra <[email protected]> >> > > > > >> > > > > > Is this a zookeeper specific error or something? >> > > > > > >> > > > > > >> > > > > > On Wed, Aug 21, 2013 at 6:06 PM, Pavan Sudheendra < >> > > [email protected] >> > > > > > >wrote: >> > > > > > >> > > > > > > Hi Jean, >> > > > > > > >> > > > > > > ubuntu@ip-10-34-187-170:~$ cat /etc/hostname >> > > > > > > ip-10-34-187-170 >> > > > > > > ubuntu@ip-10-34-187-170:~$ hostname >> > > > > > > ip-10-34-187-170 >> > > > > > > >> > > > > > > >> > > > > > > >> > > > > > > On Wed, Aug 21, 2013 at 6:01 PM, Jean-Marc Spaggiari < >> > > > > > > [email protected]> wrote: >> > > > > > > >> > > > > > >> And what about: >> > > > > > >> # cat /etc/hostname >> > > > > > >> >> > > > > > >> and >> > > > > > >> # hostname >> > > > > > >> >> > > > > > >> ? >> > > > > > >> >> > > > > > >> 2013/8/21 Pavan Sudheendra <[email protected]> >> > > > > > >> >> > > > > > >> > Sure.. >> > > > > > >> > /etc/hosts file: >> > > > > > >> > >> > > > > > >> > 127.0.0.1 localhost >> > > > > > >> > 10.34.187.170 ip-10-34-187-170 >> > > > > > >> > # The following lines are desirable for IPv6 capable hosts >> > > > > > >> > ::1 ip6-localhost ip6-loopback >> > > > > > >> > fe00::0 ip6-localnet >> > > > > > >> > ff00::0 ip6-mcastprefix >> > > > > > >> > ff02::1 ip6-allnodes >> > > > > > >> > ff02::2 ip6-allrouters >> > > > > > >> > ff02::3 ip6-allhosts >> > > > > > >> > >> > > > > > >> > Configuration conf = HBaseConfiguration.create(); >> > > > > > >> > conf.set("hbase.zookeeper.quorum", "10.34.187.170"); >> > > > > > >> > conf.set("hbase.zookeeper.property.clientPort","2181"); >> > > > > > >> > conf.set("hbase.master","10.34.187.170"); >> > > > > > >> > Job job = new Job(conf, ViewersTable); >> > > > > > >> > >> > > > > > >> > I'm trying to process table data which has 19 million >> rows..It >> > > > runs >> > > > > > fine >> > > > > > >> > for a while although i don't see the map percent completion >> > > change >> > > > > > from >> > > > > > >> 0% >> > > > > > >> > .. After a while it says >> > > > > > >> > >> > > > > > >> > Task attempt_201304161625_0028_m_000000_0 failed to report >> > > status >> > > > > for >> > > > > > >> > 600 seconds. Killing! >> > > > > > >> > >> > > > > > >> > >> > > > > > >> > >> > > > > > >> > >> > > > > > >> > >> > > > > > >> > On Wed, Aug 21, 2013 at 5:52 PM, Jean-Marc Spaggiari < >> > > > > > >> > [email protected]> wrote: >> > > > > > >> > >> > > > > > >> > > Can you past you host file here again with the >> modification >> > > you >> > > > > have >> > > > > > >> > done? >> > > > > > >> > > >> > > > > > >> > > Also, can you share a big more of you code? What are you >> > doing >> > > > > with >> > > > > > >> the >> > > > > > >> > > config object after, how do you create your table object, >> > > etc.? >> > > > > > >> > > >> > > > > > >> > > Thanks, >> > > > > > >> > > >> > > > > > >> > > JM >> > > > > > >> > > >> > > > > > >> > > 2013/8/21 Pavan Sudheendra <[email protected]> >> > > > > > >> > > >> > > > > > >> > > > @Jean tried your method didn't work.. >> > > > > > >> > > > >> > > > > > >> > > > 2013-08-21 12:17:10,908 INFO >> > > org.apache.zookeeper.ClientCnxn: >> > > > > > >> Opening >> > > > > > >> > > > socket connection to server localhost/127.0.0.1:2181. >> > Will >> > > > not >> > > > > > >> attempt >> > > > > > >> > > to >> > > > > > >> > > > authenticate using SASL (Unable to locate a login >> > > > configuration) >> > > > > > >> > > > 2013-08-21 12:17:10,908 WARN >> > > org.apache.zookeeper.ClientCnxn: >> > > > > > >> Session >> > > > > > >> > 0x0 >> > > > > > >> > > > for server null, unexpected error, closing socket >> > connection >> > > > and >> > > > > > >> > > attempting >> > > > > > >> > > > reconnect >> > > > > > >> > > > java.net.ConnectException: Connection refused >> > > > > > >> > > > at sun.nio.ch.SocketChannelImpl.checkConnect(Native >> > > > Method) >> > > > > > >> > > > at >> > > > > > >> > > > >> > > > > > >> >> > > > >> sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:567) >> > > > > > >> > > > at >> > > > > > >> > > > >> > > > > > >> > > > >> > > > > > >> > > >> > > > > > >> > >> > > > > > >> >> > > > > > >> > > > > >> > > > >> > > >> > >> org.apache.zookeeper.ClientCnxnSocketNIO.doTransport(ClientCnxnSocketNIO.java:350) >> > > > > > >> > > > at >> > > > > > >> > > >> > > > > >> org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1068) >> > > > > > >> > > > 2013-08-21 12:17:11,009 WARN >> > > > > > >> > > > org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper: >> > > > Possibly >> > > > > > >> > > transient >> > > > > > >> > > > ZooKeeper exception: >> > > > > > >> > > > >> > > org.apache.zookeeper.KeeperException$ConnectionLossException: >> > > > > > >> > > > KeeperErrorCode = ConnectionLoss for /hbase >> > > > > > >> > > > 2013-08-21 12:17:11,009 INFO >> > > > > > >> org.apache.hadoop.hbase.util.RetryCounter: >> > > > > > >> > > > Sleeping 8000ms before retry #3...\ >> > > > > > >> > > > >> > > > > > >> > > > Any tips? >> > > > > > >> > > > >> > > > > > >> > > > >> > > > > > >> > > > >> > > > > > >> > > > On Wed, Aug 21, 2013 at 5:15 PM, Jean-Marc Spaggiari < >> > > > > > >> > > > [email protected]> wrote: >> > > > > > >> > > > >> > > > > > >> > > > > Hi Pavan, >> > > > > > >> > > > > >> > > > > > >> > > > > I don't think Cloudera Manager assign the address to >> > your >> > > > > > >> computer. >> > > > > > >> > > When >> > > > > > >> > > > CM >> > > > > > >> > > > > is down, your computer still have an IP, and even if >> you >> > > > > > >> un-install >> > > > > > >> > CM, >> > > > > > >> > > > you >> > > > > > >> > > > > will still have an IP assigned to your computer. >> > > > > > >> > > > > >> > > > > > >> > > > > If you have not configured anything there, then you >> most >> > > > > > probably >> > > > > > >> > have >> > > > > > >> > > a >> > > > > > >> > > > > DHCP. Just give a try to what I told you on the other >> > > > message. >> > > > > > >> > > > > >> > > > > > >> > > > > JM >> > > > > > >> > > > > >> > > > > > >> > > > > 2013/8/21 Pavan Sudheendra <[email protected]> >> > > > > > >> > > > > >> > > > > > >> > > > > > @Manoj i have set hbase.zookeeper.quorum in my M-R >> > > > > > application.. >> > > > > > >> > > > > > >> > > > > > >> > > > > > @Jean The cloudera manager picks up the ip address >> > > > > > >> automatically.. >> > > > > > >> > > > > > >> > > > > > >> > > > > > >> > > > > > >> > > > > > On Wed, Aug 21, 2013 at 5:07 PM, manoj p < >> > > > [email protected] >> > > > > > >> > > > > > >> > wrote: >> > > > > > >> > > > > > >> > > > > > >> > > > > > > Can you try passing the argument >> > > > > > >> > > > -Dhbase.zookeeper.quorum=10.34.187.170 >> > > > > > >> > > > > > > while running the program >> > > > > > >> > > > > > > >> > > > > > >> > > > > > > If this does'nt work either please check if >> > HBASE_HOME >> > > > and >> > > > > > >> > > > > HBASE_CONF_DIR >> > > > > > >> > > > > > > is set correctly. >> > > > > > >> > > > > > > >> > > > > > >> > > > > > > BR/Manoj >> > > > > > >> > > > > > > >> > > > > > >> > > > > > > >> > > > > > >> > > > > > > On Wed, Aug 21, 2013 at 4:48 PM, Pavan >> Sudheendra < >> > > > > > >> > > > [email protected] >> > > > > > >> > > > > > > >wrote: >> > > > > > >> > > > > > > >> > > > > > >> > > > > > > > Yes. My /etc/hosts have the correct mapping to >> > > > localhost >> > > > > > >> > > > > > > > >> > > > > > >> > > > > > > > 127.0.0.1 localhost >> > > > > > >> > > > > > > > >> > > > > > >> > > > > > > > # The following lines are desirable for IPv6 >> > capable >> > > > > hosts >> > > > > > >> > > > > > > > ::1 ip6-localhost ip6-loopback >> > > > > > >> > > > > > > > fe00::0 ip6-localnet >> > > > > > >> > > > > > > > ff00::0 ip6-mcastprefix >> > > > > > >> > > > > > > > ff02::1 ip6-allnodes >> > > > > > >> > > > > > > > ff02::2 ip6-allrouters >> > > > > > >> > > > > > > > >> > > > > > >> > > > > > > > I've added the HBase jars to the Hadoop >> Classpath >> > as >> > > > > well. >> > > > > > >> Not >> > > > > > >> > > sure >> > > > > > >> > > > > > why.. >> > > > > > >> > > > > > > > I'm running this on a 6 node cloudera cluster >> > which >> > > > > > consist >> > > > > > >> of >> > > > > > >> > 1 >> > > > > > >> > > > > > > > jobtrackers and 5 tasktrackers.. >> > > > > > >> > > > > > > > >> > > > > > >> > > > > > > > After a while all my map jobs fail.. Completely >> > > > baffled >> > > > > > >> because >> > > > > > >> > > the >> > > > > > >> > > > > map >> > > > > > >> > > > > > > > tasks were doing the required tasks.. >> > > > > > >> > > > > > > > >> > > > > > >> > > > > > > > >> > > > > > >> > > > > > > > >> > > > > > >> > > > > > > > On Wed, Aug 21, 2013 at 4:45 PM, manoj p < >> > > > > > [email protected] >> > > > > > >> > >> > > > > > >> > > > wrote: >> > > > > > >> > > > > > > > >> > > > > > >> > > > > > > > > For your code to run, please ensure if you >> use >> > the >> > > > > > correct >> > > > > > >> > > > > > HBase/Hadoop >> > > > > > >> > > > > > > > jar >> > > > > > >> > > > > > > > > versions while compiling your program. >> > > > > > >> > > > > > > > > >> > > > > > >> > > > > > > > > BR/Manoj >> > > > > > >> > > > > > > > > >> > > > > > >> > > > > > > > > >> > > > > > >> > > > > > > > > On Wed, Aug 21, 2013 at 4:38 PM, manoj p < >> > > > > > >> [email protected]> >> > > > > > >> > > > > wrote: >> > > > > > >> > > > > > > > > >> > > > > > >> > > > > > > > > > Check your /etc/hosts file if you have the >> > > correct >> > > > > > >> mapping >> > > > > > >> > to >> > > > > > >> > > > > > > localhost >> > > > > > >> > > > > > > > > > for 127.0.0.1. Also ensure that if you have >> > > > > > >> > > > > hbase.zookeeper.quorum >> > > > > > >> > > > > > in >> > > > > > >> > > > > > > > > your >> > > > > > >> > > > > > > > > > configuration and also check if HBase >> > classpath >> > > is >> > > > > > >> appended >> > > > > > >> > > to >> > > > > > >> > > > > > Hadoop >> > > > > > >> > > > > > > > > > classpath. >> > > > > > >> > > > > > > > > > >> > > > > > >> > > > > > > > > > >> > > > > > >> > > > > > > > > > BR/Manoj >> > > > > > >> > > > > > > > > > >> > > > > > >> > > > > > > > > > >> > > > > > >> > > > > > > > > > On Wed, Aug 21, 2013 at 4:10 PM, Pavan >> > > Sudheendra >> > > > < >> > > > > > >> > > > > > > [email protected] >> > > > > > >> > > > > > > > > >wrote: >> > > > > > >> > > > > > > > > > >> > > > > > >> > > > > > > > > >> Hadoop Namenode reports the following >> error >> > > which >> > > > > is >> > > > > > >> > > unusual : >> > > > > > >> > > > > > > > > >> >> > > > > > >> > > > > > > > > >> >> > > > > > >> > > > > > > > > >> 013-08-21 09:21:12,328 INFO >> > > > > > >> > org.apache.zookeeper.ClientCnxn: >> > > > > > >> > > > > > Opening >> > > > > > >> > > > > > > > > >> socket >> > > > > > >> > > > > > > > > >> connection to server localhost/ >> > 127.0.0.1:2181. >> > > > > Will >> > > > > > >> not >> > > > > > >> > > > attempt >> > > > > > >> > > > > > to >> > > > > > >> > > > > > > > > >> authenticate using SASL (Unable to locate >> a >> > > login >> > > > > > >> > > > configuration) >> > > > > > >> > > > > > > > > >> java.net.ConnectException: Connection >> refused >> > > > > > >> > > > > > > > > >> at >> > > > > > sun.nio.ch.SocketChannelImpl.checkConnect(Native >> > > > > > >> > > > Method) >> > > > > > >> > > > > > > > > >> at >> > > > > > >> > > > > > > > > >> >> > > > > > >> > > > > > > >> > > > > > >> > > > >> > > > > > >> >> > > > >> sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:567) >> > > > > > >> > > > > > > > > >> at >> > > > > > >> > > > > > > > > >> >> > > > > > >> > > > > > > > > >> >> > > > > > >> > > > > > > > > >> > > > > > >> > > > > > > > >> > > > > > >> > > > > > > >> > > > > > >> > > > > > >> > > > > > >> > > > > >> > > > > > >> > > > >> > > > > > >> > > >> > > > > > >> > >> > > > > > >> >> > > > > > >> > > > > >> > > > >> > > >> > >> org.apache.zookeeper.ClientCnxnSocketNIO.doTransport(ClientCnxnSocketNIO.java:350) >> > > > > > >> > > > > > > > > >> at >> > > > > > >> > > > > > > > > >> >> > > > > > >> > > > > > >> > > > > > >> > >> > > > org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1068) >> > > > > > >> > > > > > > > > >> 2013-08-21 09:33:11,033 WARN >> > > > > > >> > > > > > > > > >> >> > > > > > org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper: >> > > > > > >> > > > Possibly >> > > > > > >> > > > > > > > > transient >> > > > > > >> > > > > > > > > >> ZooKeeper exception: >> > > > > > >> > > > > > > > > >> >> > > > > > >> > > >> > org.apache.zookeeper.KeeperException$ConnectionLossException: >> > > > > > >> > > > > > > > > >> KeeperErrorCode = ConnectionLoss for >> /hbase >> > > > > > >> > > > > > > > > >> 2013-08-21 09:33:11,033 INFO >> > > > > > >> > > > > > > > org.apache.hadoop.hbase.util.RetryCounter: >> > > > > > >> > > > > > > > > >> Sleeping 8000ms before retry #3... >> > > > > > >> > > > > > > > > >> 2013-08-21 09:33:11,043 WARN >> > > > > > >> > org.apache.hadoop.mapred.Task: >> > > > > > >> > > > > Parent >> > > > > > >> > > > > > > > died. >> > > > > > >> > > > > > > > > >> Exiting >> attempt_201307181246_0548_m_000022_2 >> > > > > > >> > > > > > > > > >> >> > > > > > >> > > > > > > > > >> >> > > > > > >> > > > > > > > > >> Because i have specified the address in >> the >> > > java >> > > > > file >> > > > > > >> > > > > > > > > >> Configuration conf = >> > > > > HBaseConfiguration.create(); >> > > > > > >> > > > > > > > > >> conf.set("hbase.zookeeper.quorum", >> > > > > > >> "10.34.187.170"); >> > > > > > >> > > > > > > > > >> >> > > > > > >> > conf.set("hbase.zookeeper.property.clientPort","2181"); >> > > > > > >> > > > > > > > > >> >> conf.set("hbase.master","10.34.187.170"); >> > > > > > >> > > > > > > > > >> >> > > > > > >> > > > > > > > > >> >> > > > > > >> > > > > > > > > >> >> > > > > > >> > > > > > > > > >> All my map tasks fail like this! Please >> > help.. >> > > > I'm >> > > > > > on a >> > > > > > >> > > > timebomb >> > > > > > >> > > > > > > > > >> -- >> > > > > > >> > > > > > > > > >> Regards- >> > > > > > >> > > > > > > > > >> Pavan >> > > > > > >> > > > > > > > > >> >> > > > > > >> > > > > > > > > > >> > > > > > >> > > > > > > > > > >> > > > > > >> > > > > > > > > >> > > > > > >> > > > > > > > >> > > > > > >> > > > > > > > >> > > > > > >> > > > > > > > >> > > > > > >> > > > > > > > -- >> > > > > > >> > > > > > > > Regards- >> > > > > > >> > > > > > > > Pavan >> > > > > > >> > > > > > > > >> > > > > > >> > > > > > > >> > > > > > >> > > > > > >> > > > > > >> > > > > > >> > > > > > >> > > > > > >> > > > > > >> > > > > > -- >> > > > > > >> > > > > > Regards- >> > > > > > >> > > > > > Pavan >> > > > > > >> > > > > > >> > > > > > >> > > > > >> > > > > > >> > > > >> > > > > > >> > > > >> > > > > > >> > > > >> > > > > > >> > > > -- >> > > > > > >> > > > Regards- >> > > > > > >> > > > Pavan >> > > > > > >> > > > >> > > > > > >> > > >> > > > > > >> > >> > > > > > >> > >> > > > > > >> > >> > > > > > >> > -- >> > > > > > >> > Regards- >> > > > > > >> > Pavan >> > > > > > >> > >> > > > > > >> >> > > > > > > >> > > > > > > >> > > > > > > >> > > > > > > -- >> > > > > > > Regards- >> > > > > > > Pavan >> > > > > > > >> > > > > > >> > > > > > >> > > > > > >> > > > > > -- >> > > > > > Regards- >> > > > > > Pavan >> > > > > > >> > > > > >> > > > >> > > > >> > > > >> > > > -- >> > > > Regards- >> > > > Pavan >> > > > >> > > >> > >> > >> > >> > -- >> > Regards- >> > Pavan >> > >> > > > > -- > Regards- > Pavan > -- Regards- Pavan
