How can i be sure previous failures are shut down, there is no more effects?

With what other application might hbase started?


2012/11/9 Eric Charles <[email protected]>

> It builds/tests fine on the Apache CI [1] and on my laptop (at least last
> week, for now I have an unstable local trunk, so I can not confirm).
>
> Can you double check if you don't have any hadoop/hbase/zookeeper process
> running (wherever because you have launched them or because of a previous
> test failure) and try again (numerous time if needed).
>
> Thx, Eric
>
> [1] https://builds.apache.org/**view/G-L/view/James/job/**
> mailbox/1109/org.apache.james$**apache-james-mailbox-hbase/<https://builds.apache.org/view/G-L/view/James/job/mailbox/1109/org.apache.james$apache-james-mailbox-hbase/>
>
>
>
>
> On 09/11/2012 16:10, Merve Temizer wrote:
>
>> os: Ubuntu 12.04 lts
>>
>> i use a jdk by specifying on command
>>
>> sudo JAVA_HOME /home/merve/dev/jdk/jdk1.7.0_**03 mvn clean compile
>> install
>>
>> "mvn --v" outputs
>>
>> Apache Maven 3.0.4 (r1232337; 2012-01-17 10:44:56+0200)
>> Maven home: /usr/local/apache-maven-3.0.4
>> Java version: 1.7.0_03, vendor: Oracle Corporation
>> Java home: /usr/lib/jvm/jdk1.7.0_03/jre
>> Default locale: en_US, platform encoding: UTF-8
>> OS name: "linux", version: "3.0.0-16-generic", arch: "i386", family:
>> "unix"
>>
>> thanks very much for time.
>>
>> 2012/11/9 Eric Charles <[email protected]>
>>
>>  Hi Merve,
>>> Can you send env details: os and jdk version?
>>>
>>> Thx, Eric
>>>
>>>
>>> On 09/11/2012 13:59, Merve Temizer wrote:
>>>
>>>  Hello,
>>>>
>>>> I had problems on hbase by building tag 3.0 beta,
>>>>
>>>> then i checked out trunk from
>>>>
>>>> http://svn.apache.org/repos/****asf/james/server/trunk<http://svn.apache.org/repos/**asf/james/server/trunk>
>>>> <http://**svn.apache.org/repos/asf/**james/server/trunk<http://svn.apache.org/repos/asf/james/server/trunk>
>>>> >
>>>>
>>>>
>>>> I am using maven 3 to "mvn clean compile install"
>>>>
>>>> Below is the console output:
>>>>
>>>> Running org.apache.james.domainlist.****hbase.HBaseDomainListTest
>>>>
>>>> 2012-11-09 15:47:07,044 [main] WARN
>>>>    org.apache.hadoop.metrics2.****impl.MetricsSystemImpl - Source name
>>>> ugi
>>>>
>>>> already exists!
>>>> 2012-11-09 15:47:07,201 [main] WARN
>>>>    org.apache.hadoop.metrics2.****impl.MetricsSystemImpl - Source name
>>>> ugi
>>>>
>>>> already exists!
>>>> Starting DataNode 0 with dfs.data.dir:
>>>> /home/merve/dev/source/james2/****trunk/data-hbase/target/**test-**
>>>> data/b84d7dd5-91ad-4082-b02f-****411efb69948b/dfscluster_**
>>>> d7a93b05-d4b6-4248-80c6-****7f509a9a3446/dfs/data/data1,/****
>>>> home/merve/dev/source/james2/****trunk/data-hbase/target/test-****
>>>> data/b84d7dd5-91ad-4082-b02f-****411efb69948b/dfscluster_**
>>>> d7a93b05-d4b6-4248-80c6-****7f509a9a3446/dfs/data/data2
>>>>
>>>> 2012-11-09 15:47:07,877 [main] WARN
>>>>    org.apache.hadoop.metrics2.****impl.MetricsSystemImpl - NameNode
>>>> metrics
>>>>
>>>> system already initialized!
>>>> 2012-11-09 15:47:07,877 [main] WARN
>>>>    org.apache.hadoop.metrics2.****impl.MetricsSystemImpl - Source name
>>>> ugi
>>>>
>>>> already exists!
>>>> 2012-11-09 15:47:08,233 [main] WARN
>>>>    org.apache.hadoop.metrics2.****impl.MetricsSystemImpl - Source name
>>>> jvm
>>>>
>>>> already exists!
>>>> 2012-11-09 15:47:09,454 [IPC Server handler 2 on 28077] WARN
>>>>    org.apache.hadoop.hdfs.server.****namenode.FSNamesystem - Not able
>>>> to
>>>>
>>>> place
>>>> enough replicas, still in need of 1
>>>> 2012-11-09 15:47:09,455 [IPC Server handler 2 on 28077] ERROR
>>>> org.apache.hadoop.security.****UserGroupInformation -
>>>>
>>>> PriviledgedActionException as:root cause:java.io.IOException: File
>>>> /user/root/hbase.version could only be replicated to 0 nodes, instead
>>>> of 1
>>>> 2012-11-09 15:47:09,456 [Thread-44] WARN  org.apache.hadoop.hdfs.**
>>>> DFSClient
>>>> - DataStreamer Exception: org.apache.hadoop.ipc.****RemoteException:
>>>>
>>>> java.io.IOException: File /user/root/hbase.version could only be
>>>> replicated
>>>> to 0 nodes, instead of 1
>>>> at
>>>> org.apache.hadoop.hdfs.server.****namenode.FSNamesystem.**
>>>> getAdditionalBlock(****FSNamesystem.java:1556)
>>>> at
>>>> org.apache.hadoop.hdfs.server.****namenode.NameNode.addBlock(****
>>>> NameNode.java:696)
>>>> at sun.reflect.****NativeMethodAccessorImpl.****invoke0(Native Method)
>>>> at
>>>> sun.reflect.****NativeMethodAccessorImpl.****invoke(**
>>>> NativeMethodAccessorImpl.java:****57)
>>>> at
>>>> sun.reflect.****DelegatingMethodAccessorImpl.****invoke(**
>>>> DelegatingMethodAccessorImpl.****java:43)
>>>> at java.lang.reflect.Method.****invoke(Method.java:601)
>>>> at org.apache.hadoop.ipc.RPC$****Server.call(RPC.java:563)
>>>> at org.apache.hadoop.ipc.Server$****Handler$1.run(Server.java:****1388)
>>>> at org.apache.hadoop.ipc.Server$****Handler$1.run(Server.java:****1384)
>>>> at java.security.****AccessController.doPrivileged(****Native Method)
>>>> at javax.security.auth.Subject.****doAs(Subject.java:415)
>>>> at
>>>> org.apache.hadoop.security.****UserGroupInformation.doAs(**
>>>> UserGroupInformation.java:****1093)
>>>> at org.apache.hadoop.ipc.Server$****Handler.run(Server.java:1382)
>>>>
>>>> at org.apache.hadoop.ipc.Client.****call(Client.java:1066)
>>>> at org.apache.hadoop.ipc.RPC$****Invoker.invoke(RPC.java:225)
>>>> at $Proxy8.addBlock(Unknown Source)
>>>> at sun.reflect.****NativeMethodAccessorImpl.****invoke0(Native Method)
>>>> at
>>>> sun.reflect.****NativeMethodAccessorImpl.****invoke(**
>>>> NativeMethodAccessorImpl.java:****57)
>>>> at
>>>> sun.reflect.****DelegatingMethodAccessorImpl.****invoke(**
>>>> DelegatingMethodAccessorImpl.****java:43)
>>>> at java.lang.reflect.Method.****invoke(Method.java:601)
>>>> at
>>>> org.apache.hadoop.io.retry.****RetryInvocationHandler.****
>>>> invokeMethod(**
>>>> RetryInvocationHandler.java:****82)
>>>> at
>>>> org.apache.hadoop.io.retry.****RetryInvocationHandler.invoke(****
>>>> RetryInvocationHandler.java:****59)
>>>> at $Proxy8.addBlock(Unknown Source)
>>>> at
>>>> org.apache.hadoop.hdfs.****DFSClient$DFSOutputStream.**
>>>> locateFollowingBlock(****DFSClient.java:3507)
>>>> at
>>>> org.apache.hadoop.hdfs.****DFSClient$DFSOutputStream.**
>>>> nextBlockOutputStream(****DFSClient.java:3370)
>>>> at
>>>> org.apache.hadoop.hdfs.****DFSClient$DFSOutputStream.**
>>>> access$2700(DFSClient.java:****2586)
>>>> at
>>>> org.apache.hadoop.hdfs.****DFSClient$DFSOutputStream$**
>>>> DataStreamer.run(DFSClient.****java:2826)
>>>>
>>>> 2012-11-09 15:47:09,460 [Thread-44] WARN  org.apache.hadoop.hdfs.**
>>>>
>>>> DFSClient
>>>> - Error Recovery for block null bad datanode[0] nodes == null
>>>> 2012-11-09 15:47:09,460 [Thread-44] WARN  org.apache.hadoop.hdfs.**
>>>>
>>>> DFSClient
>>>> - Could not get block locations. Source file "/user/root/hbase.version"
>>>> -
>>>> Aborting...
>>>> 2012-11-09 15:47:09,462 [IPC Server handler 3 on 28077] WARN
>>>>    org.apache.hadoop.hdfs.****StateChange - DIR* NameSystem.startFile:
>>>>
>>>> failed to
>>>> create file /user/root/hbase.version for DFSClient_1286943058 on client
>>>> 127.0.0.1 because current leaseholder is trying to recreate file.
>>>> 2012-11-09 15:47:09,463 [IPC Server handler 3 on 28077] ERROR
>>>> org.apache.hadoop.security.****UserGroupInformation -
>>>> PriviledgedActionException as:root
>>>> cause:org.apache.hadoop.hdfs.****protocol.****
>>>> AlreadyBeingCreatedException:
>>>>
>>>> failed
>>>> to create file /user/root/hbase.version for DFSClient_1286943058 on
>>>> client
>>>> 127.0.0.1 because current leaseholder is trying to recreate file.
>>>> 2012-11-09 15:48:09,469 [IPC Server handler 6 on 28077] WARN
>>>>    org.apache.hadoop.hdfs.****StateChange - DIR* NameSystem.startFile:
>>>>
>>>> failed to
>>>> create file /user/root/hbase.version for DFSClient_1286943058 on client
>>>> 127.0.0.1 because current leaseholder is trying to recreate file.
>>>> 2012-11-09 15:48:09,469 [IPC Server handler 6 on 28077] ERROR
>>>> org.apache.hadoop.security.****UserGroupInformation -
>>>> PriviledgedActionException as:root
>>>> cause:org.apache.hadoop.hdfs.****protocol.****
>>>> AlreadyBeingCreatedException:
>>>>
>>>> failed
>>>> to create file /user/root/hbase.version for DFSClient_1286943058 on
>>>> client
>>>> 127.0.0.1 because current leaseholder is trying to recreate file.
>>>> 2012-11-09 15:49:09,477 [IPC Server handler 8 on 28077] WARN
>>>>    org.apache.hadoop.hdfs.****StateChange - DIR* NameSystem.startFile:
>>>>
>>>> failed to
>>>> create file /user/root/hbase.version for DFSClient_1286943058 on client
>>>> 127.0.0.1 because current leaseholder is trying to recreate file.
>>>> 2012-11-09 15:49:09,477 [IPC Server handler 8 on 28077] ERROR
>>>> org.apache.hadoop.security.****UserGroupInformation -
>>>> PriviledgedActionException as:root
>>>> cause:org.apache.hadoop.hdfs.****protocol.****
>>>> AlreadyBeingCreatedException:
>>>>
>>>> failed
>>>> to create file /user/root/hbase.version for DFSClient_1286943058 on
>>>> client
>>>> 127.0.0.1 because current leaseholder is trying to recreate file.
>>>> 2012-11-09 15:50:09,486 [IPC Server handler 1 on 28077] WARN
>>>>    org.apache.hadoop.hdfs.****StateChange - DIR* NameSystem.startFile:
>>>>
>>>> failed to
>>>> create file /user/root/hbase.version for DFSClient_1286943058 on client
>>>> 127.0.0.1 because current leaseholder is trying to recreate file.
>>>> 2012-11-09 15:50:09,487 [IPC Server handler 1 on 28077] ERROR
>>>> org.apache.hadoop.security.****UserGroupInformation -
>>>> PriviledgedActionException as:root
>>>> cause:org.apache.hadoop.hdfs.****protocol.****
>>>> AlreadyBeingCreatedException:
>>>>
>>>> failed
>>>> to create file /user/root/hbase.version for DFSClient_1286943058 on
>>>> client
>>>> 127.0.0.1 because current leaseholder is trying to recreate file.
>>>> 2012-11-09 15:51:09,495 [IPC Server handler 4 on 28077] WARN
>>>>    org.apache.hadoop.hdfs.****StateChange - DIR* NameSystem.startFile:
>>>>
>>>> failed to
>>>> create file /user/root/hbase.version for DFSClient_1286943058 on client
>>>> 127.0.0.1 because current leaseholder is trying to recreate file.
>>>> 2012-11-09 15:51:09,496 [IPC Server handler 4 on 28077] ERROR
>>>> org.apache.hadoop.security.****UserGroupInformation -
>>>> PriviledgedActionException as:root
>>>> cause:org.apache.hadoop.hdfs.****protocol.****
>>>> AlreadyBeingCreatedException:
>>>>
>>>> failed
>>>> to create file /user/root/hbase.version for DFSClient_1286943058 on
>>>> client
>>>> 127.0.0.1 because current leaseholder is trying to recreate file.
>>>> 2012-11-09 15:52:09,515 [IPC Server handler 7 on 28077] WARN
>>>>    org.apache.hadoop.hdfs.****StateChange - DIR* NameSystem.startFile:
>>>>
>>>> failed to
>>>> create file /user/root/hbase.version for DFSClient_1286943058 on client
>>>> 127.0.0.1 because current leaseholder is trying to recreate file.
>>>> 2012-11-09 15:52:09,515 [IPC Server handler 7 on 28077] ERROR
>>>> org.apache.hadoop.security.****UserGroupInformation -
>>>> PriviledgedActionException as:root
>>>> cause:org.apache.hadoop.hdfs.****protocol.****
>>>> AlreadyBeingCreatedException:
>>>>
>>>> failed
>>>> to create file /user/root/hbase.version for DFSClient_1286943058 on
>>>> client
>>>> 127.0.0.1 because current leaseholder is trying to recreate file.
>>>> 2012-11-09 15:52:09,521 [IPC Server handler 8 on 28077] WARN
>>>>    org.apache.hadoop.hdfs.****StateChange - DIR* NameSystem.startFile:
>>>>
>>>> failed to
>>>> create file /user/root/hbase.version for DFSClient_1286943058 on client
>>>> 127.0.0.1 because current leaseholder is trying to recreate file.
>>>> 2012-11-09 15:52:09,522 [IPC Server handler 8 on 28077] ERROR
>>>> org.apache.hadoop.security.****UserGroupInformation -
>>>> PriviledgedActionException as:root
>>>> cause:org.apache.hadoop.hdfs.****protocol.****
>>>> AlreadyBeingCreatedException:
>>>>
>>>> failed
>>>> to create file /user/root/hbase.version for DFSClient_1286943058 on
>>>> client
>>>> 127.0.0.1 because current leaseholder is trying to recreate file.
>>>> 2012-11-09 15:53:09,534 [IPC Server handler 1 on 28077] WARN
>>>>    org.apache.hadoop.hdfs.****StateChange - DIR* NameSystem.startFile:
>>>>
>>>> failed to
>>>> create file /user/root/hbase.version for DFSClient_1286943058 on client
>>>> 127.0.0.1 because current leaseholder is trying to recreate file.
>>>> 2012-11-09 15:53:09,534 [IPC Server handler 1 on 28077] ERROR
>>>> org.apache.hadoop.security.****UserGroupInformation -
>>>> PriviledgedActionException as:root
>>>> cause:org.apache.hadoop.hdfs.****protocol.****
>>>> AlreadyBeingCreatedException:
>>>>
>>>> failed
>>>> to create file /user/root/hbase.version for DFSClient_1286943058 on
>>>> client
>>>> 127.0.0.1 because current leaseholder is trying to recreate file.
>>>> 2012-11-09 15:54:09,547 [IPC Server handler 4 on 28077] WARN
>>>>    org.apache.hadoop.hdfs.****StateChange - DIR* NameSystem.startFile:
>>>>
>>>> failed to
>>>> create file /user/root/hbase.version for DFSClient_1286943058 on client
>>>> 127.0.0.1 because current leaseholder is trying to recreate file.
>>>>
>>>>
>>>>  ------------------------------****----------------------------**
>>> --**---------
>>> To unsubscribe, e-mail: server-dev-unsubscribe@james.****apache.org<
>>> server-dev-**[email protected]<[email protected]>
>>> >
>>> For additional commands, e-mail: [email protected].****org<
>>> server-dev-help@james.**apache.org <[email protected]>>
>>>
>>>
>>>
>>
> ------------------------------**------------------------------**---------
> To unsubscribe, e-mail: 
> server-dev-unsubscribe@james.**apache.org<[email protected]>
> For additional commands, e-mail: 
> [email protected].**org<[email protected]>
>
>

Reply via email to