Sure.. /etc/hosts file: 127.0.0.1 localhost 10.34.187.170 ip-10-34-187-170 # The following lines are desirable for IPv6 capable hosts ::1 ip6-localhost ip6-loopback fe00::0 ip6-localnet ff00::0 ip6-mcastprefix ff02::1 ip6-allnodes ff02::2 ip6-allrouters ff02::3 ip6-allhosts
Configuration conf = HBaseConfiguration.create(); conf.set("hbase.zookeeper.quorum", "10.34.187.170"); conf.set("hbase.zookeeper.property.clientPort","2181"); conf.set("hbase.master","10.34.187.170"); Job job = new Job(conf, ViewersTable); I'm trying to process table data which has 19 million rows..It runs fine for a while although i don't see the map percent completion change from 0% .. After a while it says Task attempt_201304161625_0028_m_000000_0 failed to report status for 600 seconds. Killing! On Wed, Aug 21, 2013 at 5:52 PM, Jean-Marc Spaggiari < jean-m...@spaggiari.org> wrote: > Can you past you host file here again with the modification you have done? > > Also, can you share a big more of you code? What are you doing with the > config object after, how do you create your table object, etc.? > > Thanks, > > JM > > 2013/8/21 Pavan Sudheendra <pavan0...@gmail.com> > > > @Jean tried your method didn't work.. > > > > 2013-08-21 12:17:10,908 INFO org.apache.zookeeper.ClientCnxn: Opening > > socket connection to server localhost/127.0.0.1:2181. Will not attempt > to > > authenticate using SASL (Unable to locate a login configuration) > > 2013-08-21 12:17:10,908 WARN org.apache.zookeeper.ClientCnxn: Session 0x0 > > for server null, unexpected error, closing socket connection and > attempting > > reconnect > > java.net.ConnectException: Connection refused > > at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method) > > at > > sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:567) > > at > > > > > org.apache.zookeeper.ClientCnxnSocketNIO.doTransport(ClientCnxnSocketNIO.java:350) > > at > org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1068) > > 2013-08-21 12:17:11,009 WARN > > org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper: Possibly > transient > > ZooKeeper exception: > > org.apache.zookeeper.KeeperException$ConnectionLossException: > > KeeperErrorCode = ConnectionLoss for /hbase > > 2013-08-21 12:17:11,009 INFO org.apache.hadoop.hbase.util.RetryCounter: > > Sleeping 8000ms before retry #3...\ > > > > Any tips? > > > > > > > > On Wed, Aug 21, 2013 at 5:15 PM, Jean-Marc Spaggiari < > > jean-m...@spaggiari.org> wrote: > > > > > Hi Pavan, > > > > > > I don't think Cloudera Manager assign the address to your computer. > When > > CM > > > is down, your computer still have an IP, and even if you un-install CM, > > you > > > will still have an IP assigned to your computer. > > > > > > If you have not configured anything there, then you most probably have > a > > > DHCP. Just give a try to what I told you on the other message. > > > > > > JM > > > > > > 2013/8/21 Pavan Sudheendra <pavan0...@gmail.com> > > > > > > > @Manoj i have set hbase.zookeeper.quorum in my M-R application.. > > > > > > > > @Jean The cloudera manager picks up the ip address automatically.. > > > > > > > > > > > > On Wed, Aug 21, 2013 at 5:07 PM, manoj p <eors...@gmail.com> wrote: > > > > > > > > > Can you try passing the argument > > -Dhbase.zookeeper.quorum=10.34.187.170 > > > > > while running the program > > > > > > > > > > If this does'nt work either please check if HBASE_HOME and > > > HBASE_CONF_DIR > > > > > is set correctly. > > > > > > > > > > BR/Manoj > > > > > > > > > > > > > > > On Wed, Aug 21, 2013 at 4:48 PM, Pavan Sudheendra < > > pavan0...@gmail.com > > > > > >wrote: > > > > > > > > > > > Yes. My /etc/hosts have the correct mapping to localhost > > > > > > > > > > > > 127.0.0.1 localhost > > > > > > > > > > > > # The following lines are desirable for IPv6 capable hosts > > > > > > ::1 ip6-localhost ip6-loopback > > > > > > fe00::0 ip6-localnet > > > > > > ff00::0 ip6-mcastprefix > > > > > > ff02::1 ip6-allnodes > > > > > > ff02::2 ip6-allrouters > > > > > > > > > > > > I've added the HBase jars to the Hadoop Classpath as well. Not > sure > > > > why.. > > > > > > I'm running this on a 6 node cloudera cluster which consist of 1 > > > > > > jobtrackers and 5 tasktrackers.. > > > > > > > > > > > > After a while all my map jobs fail.. Completely baffled because > the > > > map > > > > > > tasks were doing the required tasks.. > > > > > > > > > > > > > > > > > > > > > > > > On Wed, Aug 21, 2013 at 4:45 PM, manoj p <eors...@gmail.com> > > wrote: > > > > > > > > > > > > > For your code to run, please ensure if you use the correct > > > > HBase/Hadoop > > > > > > jar > > > > > > > versions while compiling your program. > > > > > > > > > > > > > > BR/Manoj > > > > > > > > > > > > > > > > > > > > > On Wed, Aug 21, 2013 at 4:38 PM, manoj p <eors...@gmail.com> > > > wrote: > > > > > > > > > > > > > > > Check your /etc/hosts file if you have the correct mapping to > > > > > localhost > > > > > > > > for 127.0.0.1. Also ensure that if you have > > > hbase.zookeeper.quorum > > > > in > > > > > > > your > > > > > > > > configuration and also check if HBase classpath is appended > to > > > > Hadoop > > > > > > > > classpath. > > > > > > > > > > > > > > > > > > > > > > > > BR/Manoj > > > > > > > > > > > > > > > > > > > > > > > > On Wed, Aug 21, 2013 at 4:10 PM, Pavan Sudheendra < > > > > > pavan0...@gmail.com > > > > > > > >wrote: > > > > > > > > > > > > > > > >> Hadoop Namenode reports the following error which is > unusual : > > > > > > > >> > > > > > > > >> > > > > > > > >> 013-08-21 09:21:12,328 INFO org.apache.zookeeper.ClientCnxn: > > > > Opening > > > > > > > >> socket > > > > > > > >> connection to server localhost/127.0.0.1:2181. Will not > > attempt > > > > to > > > > > > > >> authenticate using SASL (Unable to locate a login > > configuration) > > > > > > > >> java.net.ConnectException: Connection refused > > > > > > > >> at sun.nio.ch.SocketChannelImpl.checkConnect(Native > > Method) > > > > > > > >> at > > > > > > > >> > > > > > > > sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:567) > > > > > > > >> at > > > > > > > >> > > > > > > > >> > > > > > > > > > > > > > > > > > > > > > > > > > > > > org.apache.zookeeper.ClientCnxnSocketNIO.doTransport(ClientCnxnSocketNIO.java:350) > > > > > > > >> at > > > > > > > >> > > > > org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1068) > > > > > > > >> 2013-08-21 09:33:11,033 WARN > > > > > > > >> org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper: > > Possibly > > > > > > > transient > > > > > > > >> ZooKeeper exception: > > > > > > > >> > org.apache.zookeeper.KeeperException$ConnectionLossException: > > > > > > > >> KeeperErrorCode = ConnectionLoss for /hbase > > > > > > > >> 2013-08-21 09:33:11,033 INFO > > > > > > org.apache.hadoop.hbase.util.RetryCounter: > > > > > > > >> Sleeping 8000ms before retry #3... > > > > > > > >> 2013-08-21 09:33:11,043 WARN org.apache.hadoop.mapred.Task: > > > Parent > > > > > > died. > > > > > > > >> Exiting attempt_201307181246_0548_m_000022_2 > > > > > > > >> > > > > > > > >> > > > > > > > >> Because i have specified the address in the java file > > > > > > > >> Configuration conf = HBaseConfiguration.create(); > > > > > > > >> conf.set("hbase.zookeeper.quorum", "10.34.187.170"); > > > > > > > >> conf.set("hbase.zookeeper.property.clientPort","2181"); > > > > > > > >> conf.set("hbase.master","10.34.187.170"); > > > > > > > >> > > > > > > > >> > > > > > > > >> > > > > > > > >> All my map tasks fail like this! Please help.. I'm on a > > timebomb > > > > > > > >> -- > > > > > > > >> Regards- > > > > > > > >> Pavan > > > > > > > >> > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > -- > > > > > > Regards- > > > > > > Pavan > > > > > > > > > > > > > > > > > > > > > > > > > > > -- > > > > Regards- > > > > Pavan > > > > > > > > > > > > > > > -- > > Regards- > > Pavan > > > -- Regards- Pavan