How can i be sure previous failures are shut down, there is no more effects?
With what other application might hbase started? 2012/11/9 Eric Charles <[email protected]> > It builds/tests fine on the Apache CI [1] and on my laptop (at least last > week, for now I have an unstable local trunk, so I can not confirm). > > Can you double check if you don't have any hadoop/hbase/zookeeper process > running (wherever because you have launched them or because of a previous > test failure) and try again (numerous time if needed). > > Thx, Eric > > [1] https://builds.apache.org/**view/G-L/view/James/job/** > mailbox/1109/org.apache.james$**apache-james-mailbox-hbase/<https://builds.apache.org/view/G-L/view/James/job/mailbox/1109/org.apache.james$apache-james-mailbox-hbase/> > > > > > On 09/11/2012 16:10, Merve Temizer wrote: > >> os: Ubuntu 12.04 lts >> >> i use a jdk by specifying on command >> >> sudo JAVA_HOME /home/merve/dev/jdk/jdk1.7.0_**03 mvn clean compile >> install >> >> "mvn --v" outputs >> >> Apache Maven 3.0.4 (r1232337; 2012-01-17 10:44:56+0200) >> Maven home: /usr/local/apache-maven-3.0.4 >> Java version: 1.7.0_03, vendor: Oracle Corporation >> Java home: /usr/lib/jvm/jdk1.7.0_03/jre >> Default locale: en_US, platform encoding: UTF-8 >> OS name: "linux", version: "3.0.0-16-generic", arch: "i386", family: >> "unix" >> >> thanks very much for time. >> >> 2012/11/9 Eric Charles <[email protected]> >> >> Hi Merve, >>> Can you send env details: os and jdk version? >>> >>> Thx, Eric >>> >>> >>> On 09/11/2012 13:59, Merve Temizer wrote: >>> >>> Hello, >>>> >>>> I had problems on hbase by building tag 3.0 beta, >>>> >>>> then i checked out trunk from >>>> >>>> http://svn.apache.org/repos/****asf/james/server/trunk<http://svn.apache.org/repos/**asf/james/server/trunk> >>>> <http://**svn.apache.org/repos/asf/**james/server/trunk<http://svn.apache.org/repos/asf/james/server/trunk> >>>> > >>>> >>>> >>>> I am using maven 3 to "mvn clean compile install" >>>> >>>> Below is the console output: >>>> >>>> Running org.apache.james.domainlist.****hbase.HBaseDomainListTest >>>> >>>> 2012-11-09 15:47:07,044 [main] WARN >>>> org.apache.hadoop.metrics2.****impl.MetricsSystemImpl - Source name >>>> ugi >>>> >>>> already exists! >>>> 2012-11-09 15:47:07,201 [main] WARN >>>> org.apache.hadoop.metrics2.****impl.MetricsSystemImpl - Source name >>>> ugi >>>> >>>> already exists! >>>> Starting DataNode 0 with dfs.data.dir: >>>> /home/merve/dev/source/james2/****trunk/data-hbase/target/**test-** >>>> data/b84d7dd5-91ad-4082-b02f-****411efb69948b/dfscluster_** >>>> d7a93b05-d4b6-4248-80c6-****7f509a9a3446/dfs/data/data1,/**** >>>> home/merve/dev/source/james2/****trunk/data-hbase/target/test-**** >>>> data/b84d7dd5-91ad-4082-b02f-****411efb69948b/dfscluster_** >>>> d7a93b05-d4b6-4248-80c6-****7f509a9a3446/dfs/data/data2 >>>> >>>> 2012-11-09 15:47:07,877 [main] WARN >>>> org.apache.hadoop.metrics2.****impl.MetricsSystemImpl - NameNode >>>> metrics >>>> >>>> system already initialized! >>>> 2012-11-09 15:47:07,877 [main] WARN >>>> org.apache.hadoop.metrics2.****impl.MetricsSystemImpl - Source name >>>> ugi >>>> >>>> already exists! >>>> 2012-11-09 15:47:08,233 [main] WARN >>>> org.apache.hadoop.metrics2.****impl.MetricsSystemImpl - Source name >>>> jvm >>>> >>>> already exists! >>>> 2012-11-09 15:47:09,454 [IPC Server handler 2 on 28077] WARN >>>> org.apache.hadoop.hdfs.server.****namenode.FSNamesystem - Not able >>>> to >>>> >>>> place >>>> enough replicas, still in need of 1 >>>> 2012-11-09 15:47:09,455 [IPC Server handler 2 on 28077] ERROR >>>> org.apache.hadoop.security.****UserGroupInformation - >>>> >>>> PriviledgedActionException as:root cause:java.io.IOException: File >>>> /user/root/hbase.version could only be replicated to 0 nodes, instead >>>> of 1 >>>> 2012-11-09 15:47:09,456 [Thread-44] WARN org.apache.hadoop.hdfs.** >>>> DFSClient >>>> - DataStreamer Exception: org.apache.hadoop.ipc.****RemoteException: >>>> >>>> java.io.IOException: File /user/root/hbase.version could only be >>>> replicated >>>> to 0 nodes, instead of 1 >>>> at >>>> org.apache.hadoop.hdfs.server.****namenode.FSNamesystem.** >>>> getAdditionalBlock(****FSNamesystem.java:1556) >>>> at >>>> org.apache.hadoop.hdfs.server.****namenode.NameNode.addBlock(**** >>>> NameNode.java:696) >>>> at sun.reflect.****NativeMethodAccessorImpl.****invoke0(Native Method) >>>> at >>>> sun.reflect.****NativeMethodAccessorImpl.****invoke(** >>>> NativeMethodAccessorImpl.java:****57) >>>> at >>>> sun.reflect.****DelegatingMethodAccessorImpl.****invoke(** >>>> DelegatingMethodAccessorImpl.****java:43) >>>> at java.lang.reflect.Method.****invoke(Method.java:601) >>>> at org.apache.hadoop.ipc.RPC$****Server.call(RPC.java:563) >>>> at org.apache.hadoop.ipc.Server$****Handler$1.run(Server.java:****1388) >>>> at org.apache.hadoop.ipc.Server$****Handler$1.run(Server.java:****1384) >>>> at java.security.****AccessController.doPrivileged(****Native Method) >>>> at javax.security.auth.Subject.****doAs(Subject.java:415) >>>> at >>>> org.apache.hadoop.security.****UserGroupInformation.doAs(** >>>> UserGroupInformation.java:****1093) >>>> at org.apache.hadoop.ipc.Server$****Handler.run(Server.java:1382) >>>> >>>> at org.apache.hadoop.ipc.Client.****call(Client.java:1066) >>>> at org.apache.hadoop.ipc.RPC$****Invoker.invoke(RPC.java:225) >>>> at $Proxy8.addBlock(Unknown Source) >>>> at sun.reflect.****NativeMethodAccessorImpl.****invoke0(Native Method) >>>> at >>>> sun.reflect.****NativeMethodAccessorImpl.****invoke(** >>>> NativeMethodAccessorImpl.java:****57) >>>> at >>>> sun.reflect.****DelegatingMethodAccessorImpl.****invoke(** >>>> DelegatingMethodAccessorImpl.****java:43) >>>> at java.lang.reflect.Method.****invoke(Method.java:601) >>>> at >>>> org.apache.hadoop.io.retry.****RetryInvocationHandler.**** >>>> invokeMethod(** >>>> RetryInvocationHandler.java:****82) >>>> at >>>> org.apache.hadoop.io.retry.****RetryInvocationHandler.invoke(**** >>>> RetryInvocationHandler.java:****59) >>>> at $Proxy8.addBlock(Unknown Source) >>>> at >>>> org.apache.hadoop.hdfs.****DFSClient$DFSOutputStream.** >>>> locateFollowingBlock(****DFSClient.java:3507) >>>> at >>>> org.apache.hadoop.hdfs.****DFSClient$DFSOutputStream.** >>>> nextBlockOutputStream(****DFSClient.java:3370) >>>> at >>>> org.apache.hadoop.hdfs.****DFSClient$DFSOutputStream.** >>>> access$2700(DFSClient.java:****2586) >>>> at >>>> org.apache.hadoop.hdfs.****DFSClient$DFSOutputStream$** >>>> DataStreamer.run(DFSClient.****java:2826) >>>> >>>> 2012-11-09 15:47:09,460 [Thread-44] WARN org.apache.hadoop.hdfs.** >>>> >>>> DFSClient >>>> - Error Recovery for block null bad datanode[0] nodes == null >>>> 2012-11-09 15:47:09,460 [Thread-44] WARN org.apache.hadoop.hdfs.** >>>> >>>> DFSClient >>>> - Could not get block locations. Source file "/user/root/hbase.version" >>>> - >>>> Aborting... >>>> 2012-11-09 15:47:09,462 [IPC Server handler 3 on 28077] WARN >>>> org.apache.hadoop.hdfs.****StateChange - DIR* NameSystem.startFile: >>>> >>>> failed to >>>> create file /user/root/hbase.version for DFSClient_1286943058 on client >>>> 127.0.0.1 because current leaseholder is trying to recreate file. >>>> 2012-11-09 15:47:09,463 [IPC Server handler 3 on 28077] ERROR >>>> org.apache.hadoop.security.****UserGroupInformation - >>>> PriviledgedActionException as:root >>>> cause:org.apache.hadoop.hdfs.****protocol.**** >>>> AlreadyBeingCreatedException: >>>> >>>> failed >>>> to create file /user/root/hbase.version for DFSClient_1286943058 on >>>> client >>>> 127.0.0.1 because current leaseholder is trying to recreate file. >>>> 2012-11-09 15:48:09,469 [IPC Server handler 6 on 28077] WARN >>>> org.apache.hadoop.hdfs.****StateChange - DIR* NameSystem.startFile: >>>> >>>> failed to >>>> create file /user/root/hbase.version for DFSClient_1286943058 on client >>>> 127.0.0.1 because current leaseholder is trying to recreate file. >>>> 2012-11-09 15:48:09,469 [IPC Server handler 6 on 28077] ERROR >>>> org.apache.hadoop.security.****UserGroupInformation - >>>> PriviledgedActionException as:root >>>> cause:org.apache.hadoop.hdfs.****protocol.**** >>>> AlreadyBeingCreatedException: >>>> >>>> failed >>>> to create file /user/root/hbase.version for DFSClient_1286943058 on >>>> client >>>> 127.0.0.1 because current leaseholder is trying to recreate file. >>>> 2012-11-09 15:49:09,477 [IPC Server handler 8 on 28077] WARN >>>> org.apache.hadoop.hdfs.****StateChange - DIR* NameSystem.startFile: >>>> >>>> failed to >>>> create file /user/root/hbase.version for DFSClient_1286943058 on client >>>> 127.0.0.1 because current leaseholder is trying to recreate file. >>>> 2012-11-09 15:49:09,477 [IPC Server handler 8 on 28077] ERROR >>>> org.apache.hadoop.security.****UserGroupInformation - >>>> PriviledgedActionException as:root >>>> cause:org.apache.hadoop.hdfs.****protocol.**** >>>> AlreadyBeingCreatedException: >>>> >>>> failed >>>> to create file /user/root/hbase.version for DFSClient_1286943058 on >>>> client >>>> 127.0.0.1 because current leaseholder is trying to recreate file. >>>> 2012-11-09 15:50:09,486 [IPC Server handler 1 on 28077] WARN >>>> org.apache.hadoop.hdfs.****StateChange - DIR* NameSystem.startFile: >>>> >>>> failed to >>>> create file /user/root/hbase.version for DFSClient_1286943058 on client >>>> 127.0.0.1 because current leaseholder is trying to recreate file. >>>> 2012-11-09 15:50:09,487 [IPC Server handler 1 on 28077] ERROR >>>> org.apache.hadoop.security.****UserGroupInformation - >>>> PriviledgedActionException as:root >>>> cause:org.apache.hadoop.hdfs.****protocol.**** >>>> AlreadyBeingCreatedException: >>>> >>>> failed >>>> to create file /user/root/hbase.version for DFSClient_1286943058 on >>>> client >>>> 127.0.0.1 because current leaseholder is trying to recreate file. >>>> 2012-11-09 15:51:09,495 [IPC Server handler 4 on 28077] WARN >>>> org.apache.hadoop.hdfs.****StateChange - DIR* NameSystem.startFile: >>>> >>>> failed to >>>> create file /user/root/hbase.version for DFSClient_1286943058 on client >>>> 127.0.0.1 because current leaseholder is trying to recreate file. >>>> 2012-11-09 15:51:09,496 [IPC Server handler 4 on 28077] ERROR >>>> org.apache.hadoop.security.****UserGroupInformation - >>>> PriviledgedActionException as:root >>>> cause:org.apache.hadoop.hdfs.****protocol.**** >>>> AlreadyBeingCreatedException: >>>> >>>> failed >>>> to create file /user/root/hbase.version for DFSClient_1286943058 on >>>> client >>>> 127.0.0.1 because current leaseholder is trying to recreate file. >>>> 2012-11-09 15:52:09,515 [IPC Server handler 7 on 28077] WARN >>>> org.apache.hadoop.hdfs.****StateChange - DIR* NameSystem.startFile: >>>> >>>> failed to >>>> create file /user/root/hbase.version for DFSClient_1286943058 on client >>>> 127.0.0.1 because current leaseholder is trying to recreate file. >>>> 2012-11-09 15:52:09,515 [IPC Server handler 7 on 28077] ERROR >>>> org.apache.hadoop.security.****UserGroupInformation - >>>> PriviledgedActionException as:root >>>> cause:org.apache.hadoop.hdfs.****protocol.**** >>>> AlreadyBeingCreatedException: >>>> >>>> failed >>>> to create file /user/root/hbase.version for DFSClient_1286943058 on >>>> client >>>> 127.0.0.1 because current leaseholder is trying to recreate file. >>>> 2012-11-09 15:52:09,521 [IPC Server handler 8 on 28077] WARN >>>> org.apache.hadoop.hdfs.****StateChange - DIR* NameSystem.startFile: >>>> >>>> failed to >>>> create file /user/root/hbase.version for DFSClient_1286943058 on client >>>> 127.0.0.1 because current leaseholder is trying to recreate file. >>>> 2012-11-09 15:52:09,522 [IPC Server handler 8 on 28077] ERROR >>>> org.apache.hadoop.security.****UserGroupInformation - >>>> PriviledgedActionException as:root >>>> cause:org.apache.hadoop.hdfs.****protocol.**** >>>> AlreadyBeingCreatedException: >>>> >>>> failed >>>> to create file /user/root/hbase.version for DFSClient_1286943058 on >>>> client >>>> 127.0.0.1 because current leaseholder is trying to recreate file. >>>> 2012-11-09 15:53:09,534 [IPC Server handler 1 on 28077] WARN >>>> org.apache.hadoop.hdfs.****StateChange - DIR* NameSystem.startFile: >>>> >>>> failed to >>>> create file /user/root/hbase.version for DFSClient_1286943058 on client >>>> 127.0.0.1 because current leaseholder is trying to recreate file. >>>> 2012-11-09 15:53:09,534 [IPC Server handler 1 on 28077] ERROR >>>> org.apache.hadoop.security.****UserGroupInformation - >>>> PriviledgedActionException as:root >>>> cause:org.apache.hadoop.hdfs.****protocol.**** >>>> AlreadyBeingCreatedException: >>>> >>>> failed >>>> to create file /user/root/hbase.version for DFSClient_1286943058 on >>>> client >>>> 127.0.0.1 because current leaseholder is trying to recreate file. >>>> 2012-11-09 15:54:09,547 [IPC Server handler 4 on 28077] WARN >>>> org.apache.hadoop.hdfs.****StateChange - DIR* NameSystem.startFile: >>>> >>>> failed to >>>> create file /user/root/hbase.version for DFSClient_1286943058 on client >>>> 127.0.0.1 because current leaseholder is trying to recreate file. >>>> >>>> >>>> ------------------------------****----------------------------** >>> --**--------- >>> To unsubscribe, e-mail: server-dev-unsubscribe@james.****apache.org< >>> server-dev-**[email protected]<[email protected]> >>> > >>> For additional commands, e-mail: [email protected].****org< >>> server-dev-help@james.**apache.org <[email protected]>> >>> >>> >>> >> > ------------------------------**------------------------------**--------- > To unsubscribe, e-mail: > server-dev-unsubscribe@james.**apache.org<[email protected]> > For additional commands, e-mail: > [email protected].**org<[email protected]> > >
