Thanks, i built without the test.

2012/11/11 Eric Charles <[email protected]>

> Well, this is probably not the cause, but as the test launches a
> minicluster, if this one is not correctly shutted down, the next step may
> fail. In the log you pasted, nothing indicates that the cause of the
> failure is such a remaining process, but did you paste the complete log?
>
> If you want to build without the test, add -DskipTests
>
> Thx, Eric
>
>
>
> On 11/11/2012 17:41, Merve Temizer wrote:
>
>> How can i be sure previous failures are shut down, there is no more
>> effects?
>>
>> With what other application might hbase started?
>>
>>
>> 2012/11/9 Eric Charles <[email protected]>
>>
>>  It builds/tests fine on the Apache CI [1] and on my laptop (at least last
>>> week, for now I have an unstable local trunk, so I can not confirm).
>>>
>>> Can you double check if you don't have any hadoop/hbase/zookeeper process
>>> running (wherever because you have launched them or because of a previous
>>> test failure) and try again (numerous time if needed).
>>>
>>> Thx, Eric
>>>
>>> [1] 
>>> https://builds.apache.org/****view/G-L/view/James/job/**<https://builds.apache.org/**view/G-L/view/James/job/**>
>>> mailbox/1109/org.apache.james$****apache-james-mailbox-hbase/<**
>>> https://builds.apache.org/**view/G-L/view/James/job/**
>>> mailbox/1109/org.apache.james$**apache-james-mailbox-hbase/<https://builds.apache.org/view/G-L/view/James/job/mailbox/1109/org.apache.james$apache-james-mailbox-hbase/>
>>> >
>>>
>>>
>>>
>>>
>>>
>>> On 09/11/2012 16:10, Merve Temizer wrote:
>>>
>>>  os: Ubuntu 12.04 lts
>>>>
>>>> i use a jdk by specifying on command
>>>>
>>>> sudo JAVA_HOME /home/merve/dev/jdk/jdk1.7.0_****03 mvn clean compile
>>>>
>>>> install
>>>>
>>>> "mvn --v" outputs
>>>>
>>>> Apache Maven 3.0.4 (r1232337; 2012-01-17 10:44:56+0200)
>>>> Maven home: /usr/local/apache-maven-3.0.4
>>>> Java version: 1.7.0_03, vendor: Oracle Corporation
>>>> Java home: /usr/lib/jvm/jdk1.7.0_03/jre
>>>> Default locale: en_US, platform encoding: UTF-8
>>>> OS name: "linux", version: "3.0.0-16-generic", arch: "i386", family:
>>>> "unix"
>>>>
>>>> thanks very much for time.
>>>>
>>>> 2012/11/9 Eric Charles <[email protected]>
>>>>
>>>>   Hi Merve,
>>>>
>>>>> Can you send env details: os and jdk version?
>>>>>
>>>>> Thx, Eric
>>>>>
>>>>>
>>>>> On 09/11/2012 13:59, Merve Temizer wrote:
>>>>>
>>>>>   Hello,
>>>>>
>>>>>>
>>>>>> I had problems on hbase by building tag 3.0 beta,
>>>>>>
>>>>>> then i checked out trunk from
>>>>>>
>>>>>> http://svn.apache.org/repos/******asf/james/server/trunk<http://svn.apache.org/repos/****asf/james/server/trunk>
>>>>>> <http:**//svn.apache.org/repos/**asf/**james/server/trunk<http://svn.apache.org/repos/**asf/james/server/trunk>
>>>>>> >
>>>>>> <http://**svn.apache.org/**repos/asf/**james/server/trunk<http://svn.apache.org/repos/asf/**james/server/trunk>
>>>>>> **<http://svn.apache.org/repos/**asf/james/server/trunk<http://svn.apache.org/repos/asf/james/server/trunk>
>>>>>> >
>>>>>>
>>>>>>
>>>>>>>
>>>>>>
>>>>>> I am using maven 3 to "mvn clean compile install"
>>>>>>
>>>>>> Below is the console output:
>>>>>>
>>>>>> Running org.apache.james.domainlist.******hbase.HBaseDomainListTest
>>>>>>
>>>>>>
>>>>>> 2012-11-09 15:47:07,044 [main] WARN
>>>>>>     org.apache.hadoop.metrics2.******impl.MetricsSystemImpl - Source
>>>>>> name
>>>>>>
>>>>>> ugi
>>>>>>
>>>>>> already exists!
>>>>>> 2012-11-09 15:47:07,201 [main] WARN
>>>>>>     org.apache.hadoop.metrics2.******impl.MetricsSystemImpl - Source
>>>>>> name
>>>>>>
>>>>>> ugi
>>>>>>
>>>>>> already exists!
>>>>>> Starting DataNode 0 with dfs.data.dir:
>>>>>> /home/merve/dev/source/james2/******trunk/data-hbase/target/****
>>>>>> test-**
>>>>>> data/b84d7dd5-91ad-4082-b02f-******411efb69948b/dfscluster_**
>>>>>> d7a93b05-d4b6-4248-80c6-******7f509a9a3446/dfs/data/data1,/******
>>>>>> home/merve/dev/source/james2/******trunk/data-hbase/target/**
>>>>>> test-****
>>>>>> data/b84d7dd5-91ad-4082-b02f-******411efb69948b/dfscluster_**
>>>>>> d7a93b05-d4b6-4248-80c6-******7f509a9a3446/dfs/data/data2
>>>>>>
>>>>>>
>>>>>> 2012-11-09 15:47:07,877 [main] WARN
>>>>>>     org.apache.hadoop.metrics2.******impl.MetricsSystemImpl -
>>>>>> NameNode
>>>>>>
>>>>>> metrics
>>>>>>
>>>>>> system already initialized!
>>>>>> 2012-11-09 15:47:07,877 [main] WARN
>>>>>>     org.apache.hadoop.metrics2.******impl.MetricsSystemImpl - Source
>>>>>> name
>>>>>>
>>>>>> ugi
>>>>>>
>>>>>> already exists!
>>>>>> 2012-11-09 15:47:08,233 [main] WARN
>>>>>>     org.apache.hadoop.metrics2.******impl.MetricsSystemImpl - Source
>>>>>> name
>>>>>>
>>>>>> jvm
>>>>>>
>>>>>> already exists!
>>>>>> 2012-11-09 15:47:09,454 [IPC Server handler 2 on 28077] WARN
>>>>>>     org.apache.hadoop.hdfs.server.******namenode.FSNamesystem - Not
>>>>>> able
>>>>>>
>>>>>> to
>>>>>>
>>>>>> place
>>>>>> enough replicas, still in need of 1
>>>>>> 2012-11-09 15:47:09,455 [IPC Server handler 2 on 28077] ERROR
>>>>>> org.apache.hadoop.security.******UserGroupInformation -
>>>>>>
>>>>>>
>>>>>> PriviledgedActionException as:root cause:java.io.IOException: File
>>>>>> /user/root/hbase.version could only be replicated to 0 nodes, instead
>>>>>> of 1
>>>>>> 2012-11-09 15:47:09,456 [Thread-44] WARN  org.apache.hadoop.hdfs.**
>>>>>> DFSClient
>>>>>> - DataStreamer Exception: org.apache.hadoop.ipc.******
>>>>>> RemoteException:
>>>>>>
>>>>>>
>>>>>> java.io.IOException: File /user/root/hbase.version could only be
>>>>>> replicated
>>>>>> to 0 nodes, instead of 1
>>>>>> at
>>>>>> org.apache.hadoop.hdfs.server.******namenode.FSNamesystem.**
>>>>>> getAdditionalBlock(******FSNamesystem.java:1556)
>>>>>> at
>>>>>> org.apache.hadoop.hdfs.server.******namenode.NameNode.**addBlock(****
>>>>>> NameNode.java:696)
>>>>>> at sun.reflect.******NativeMethodAccessorImpl.******invoke0(Native
>>>>>> Method)
>>>>>> at
>>>>>> sun.reflect.******NativeMethodAccessorImpl.******invoke(**
>>>>>> NativeMethodAccessorImpl.java:******57)
>>>>>> at
>>>>>> sun.reflect.******DelegatingMethodAccessorImpl.******invoke(**
>>>>>> DelegatingMethodAccessorImpl.******java:43)
>>>>>> at java.lang.reflect.Method.******invoke(Method.java:601)
>>>>>> at org.apache.hadoop.ipc.RPC$******Server.call(RPC.java:563)
>>>>>> at org.apache.hadoop.ipc.Server$******Handler$1.run(Server.java:***
>>>>>> ***1388)
>>>>>> at org.apache.hadoop.ipc.Server$******Handler$1.run(Server.java:***
>>>>>> ***1384)
>>>>>> at java.security.******AccessController.doPrivileged(******Native
>>>>>> Method)
>>>>>> at javax.security.auth.Subject.******doAs(Subject.java:415)
>>>>>> at
>>>>>> org.apache.hadoop.security.******UserGroupInformation.doAs(**
>>>>>> UserGroupInformation.java:******1093)
>>>>>> at org.apache.hadoop.ipc.Server$******Handler.run(Server.java:**1382)
>>>>>>
>>>>>> at org.apache.hadoop.ipc.Client.******call(Client.java:1066)
>>>>>> at org.apache.hadoop.ipc.RPC$******Invoker.invoke(RPC.java:225)
>>>>>> at $Proxy8.addBlock(Unknown Source)
>>>>>> at sun.reflect.******NativeMethodAccessorImpl.******invoke0(Native
>>>>>> Method)
>>>>>> at
>>>>>> sun.reflect.******NativeMethodAccessorImpl.******invoke(**
>>>>>> NativeMethodAccessorImpl.java:******57)
>>>>>> at
>>>>>> sun.reflect.******DelegatingMethodAccessorImpl.******invoke(**
>>>>>> DelegatingMethodAccessorImpl.******java:43)
>>>>>> at java.lang.reflect.Method.******invoke(Method.java:601)
>>>>>> at
>>>>>> org.apache.hadoop.io.retry.******RetryInvocationHandler.****
>>>>>> invokeMethod(**
>>>>>> RetryInvocationHandler.java:******82)
>>>>>> at
>>>>>> org.apache.hadoop.io.retry.******RetryInvocationHandler.**invoke(****
>>>>>> RetryInvocationHandler.java:******59)
>>>>>> at $Proxy8.addBlock(Unknown Source)
>>>>>> at
>>>>>> org.apache.hadoop.hdfs.******DFSClient$DFSOutputStream.**
>>>>>> locateFollowingBlock(******DFSClient.java:3507)
>>>>>> at
>>>>>> org.apache.hadoop.hdfs.******DFSClient$DFSOutputStream.**
>>>>>> nextBlockOutputStream(******DFSClient.java:3370)
>>>>>> at
>>>>>> org.apache.hadoop.hdfs.******DFSClient$DFSOutputStream.**
>>>>>> access$2700(DFSClient.java:******2586)
>>>>>> at
>>>>>> org.apache.hadoop.hdfs.******DFSClient$DFSOutputStream$**
>>>>>> DataStreamer.run(DFSClient.******java:2826)
>>>>>>
>>>>>>
>>>>>> 2012-11-09 15:47:09,460 [Thread-44] WARN  org.apache.hadoop.hdfs.**
>>>>>>
>>>>>> DFSClient
>>>>>> - Error Recovery for block null bad datanode[0] nodes == null
>>>>>> 2012-11-09 15:47:09,460 [Thread-44] WARN  org.apache.hadoop.hdfs.**
>>>>>>
>>>>>> DFSClient
>>>>>> - Could not get block locations. Source file
>>>>>> "/user/root/hbase.version"
>>>>>> -
>>>>>> Aborting...
>>>>>> 2012-11-09 15:47:09,462 [IPC Server handler 3 on 28077] WARN
>>>>>>     org.apache.hadoop.hdfs.******StateChange - DIR*
>>>>>> NameSystem.startFile:
>>>>>>
>>>>>>
>>>>>> failed to
>>>>>> create file /user/root/hbase.version for DFSClient_1286943058 on
>>>>>> client
>>>>>> 127.0.0.1 because current leaseholder is trying to recreate file.
>>>>>> 2012-11-09 15:47:09,463 [IPC Server handler 3 on 28077] ERROR
>>>>>> org.apache.hadoop.security.******UserGroupInformation -
>>>>>> PriviledgedActionException as:root
>>>>>> cause:org.apache.hadoop.hdfs.******protocol.****
>>>>>>
>>>>>> AlreadyBeingCreatedException:
>>>>>>
>>>>>> failed
>>>>>> to create file /user/root/hbase.version for DFSClient_1286943058 on
>>>>>> client
>>>>>> 127.0.0.1 because current leaseholder is trying to recreate file.
>>>>>> 2012-11-09 15:48:09,469 [IPC Server handler 6 on 28077] WARN
>>>>>>     org.apache.hadoop.hdfs.******StateChange - DIR*
>>>>>> NameSystem.startFile:
>>>>>>
>>>>>>
>>>>>> failed to
>>>>>> create file /user/root/hbase.version for DFSClient_1286943058 on
>>>>>> client
>>>>>> 127.0.0.1 because current leaseholder is trying to recreate file.
>>>>>> 2012-11-09 15:48:09,469 [IPC Server handler 6 on 28077] ERROR
>>>>>> org.apache.hadoop.security.******UserGroupInformation -
>>>>>> PriviledgedActionException as:root
>>>>>> cause:org.apache.hadoop.hdfs.******protocol.****
>>>>>>
>>>>>> AlreadyBeingCreatedException:
>>>>>>
>>>>>> failed
>>>>>> to create file /user/root/hbase.version for DFSClient_1286943058 on
>>>>>> client
>>>>>> 127.0.0.1 because current leaseholder is trying to recreate file.
>>>>>> 2012-11-09 15:49:09,477 [IPC Server handler 8 on 28077] WARN
>>>>>>     org.apache.hadoop.hdfs.******StateChange - DIR*
>>>>>> NameSystem.startFile:
>>>>>>
>>>>>>
>>>>>> failed to
>>>>>> create file /user/root/hbase.version for DFSClient_1286943058 on
>>>>>> client
>>>>>> 127.0.0.1 because current leaseholder is trying to recreate file.
>>>>>> 2012-11-09 15:49:09,477 [IPC Server handler 8 on 28077] ERROR
>>>>>> org.apache.hadoop.security.******UserGroupInformation -
>>>>>> PriviledgedActionException as:root
>>>>>> cause:org.apache.hadoop.hdfs.******protocol.****
>>>>>>
>>>>>> AlreadyBeingCreatedException:
>>>>>>
>>>>>> failed
>>>>>> to create file /user/root/hbase.version for DFSClient_1286943058 on
>>>>>> client
>>>>>> 127.0.0.1 because current leaseholder is trying to recreate file.
>>>>>> 2012-11-09 15:50:09,486 [IPC Server handler 1 on 28077] WARN
>>>>>>     org.apache.hadoop.hdfs.******StateChange - DIR*
>>>>>> NameSystem.startFile:
>>>>>>
>>>>>>
>>>>>> failed to
>>>>>> create file /user/root/hbase.version for DFSClient_1286943058 on
>>>>>> client
>>>>>> 127.0.0.1 because current leaseholder is trying to recreate file.
>>>>>> 2012-11-09 15:50:09,487 [IPC Server handler 1 on 28077] ERROR
>>>>>> org.apache.hadoop.security.******UserGroupInformation -
>>>>>> PriviledgedActionException as:root
>>>>>> cause:org.apache.hadoop.hdfs.******protocol.****
>>>>>>
>>>>>> AlreadyBeingCreatedException:
>>>>>>
>>>>>> failed
>>>>>> to create file /user/root/hbase.version for DFSClient_1286943058 on
>>>>>> client
>>>>>> 127.0.0.1 because current leaseholder is trying to recreate file.
>>>>>> 2012-11-09 15:51:09,495 [IPC Server handler 4 on 28077] WARN
>>>>>>     org.apache.hadoop.hdfs.******StateChange - DIR*
>>>>>> NameSystem.startFile:
>>>>>>
>>>>>>
>>>>>> failed to
>>>>>> create file /user/root/hbase.version for DFSClient_1286943058 on
>>>>>> client
>>>>>> 127.0.0.1 because current leaseholder is trying to recreate file.
>>>>>> 2012-11-09 15:51:09,496 [IPC Server handler 4 on 28077] ERROR
>>>>>> org.apache.hadoop.security.******UserGroupInformation -
>>>>>> PriviledgedActionException as:root
>>>>>> cause:org.apache.hadoop.hdfs.******protocol.****
>>>>>>
>>>>>> AlreadyBeingCreatedException:
>>>>>>
>>>>>> failed
>>>>>> to create file /user/root/hbase.version for DFSClient_1286943058 on
>>>>>> client
>>>>>> 127.0.0.1 because current leaseholder is trying to recreate file.
>>>>>> 2012-11-09 15:52:09,515 [IPC Server handler 7 on 28077] WARN
>>>>>>     org.apache.hadoop.hdfs.******StateChange - DIR*
>>>>>> NameSystem.startFile:
>>>>>>
>>>>>>
>>>>>> failed to
>>>>>> create file /user/root/hbase.version for DFSClient_1286943058 on
>>>>>> client
>>>>>> 127.0.0.1 because current leaseholder is trying to recreate file.
>>>>>> 2012-11-09 15:52:09,515 [IPC Server handler 7 on 28077] ERROR
>>>>>> org.apache.hadoop.security.******UserGroupInformation -
>>>>>> PriviledgedActionException as:root
>>>>>> cause:org.apache.hadoop.hdfs.******protocol.****
>>>>>>
>>>>>> AlreadyBeingCreatedException:
>>>>>>
>>>>>> failed
>>>>>> to create file /user/root/hbase.version for DFSClient_1286943058 on
>>>>>> client
>>>>>> 127.0.0.1 because current leaseholder is trying to recreate file.
>>>>>> 2012-11-09 15:52:09,521 [IPC Server handler 8 on 28077] WARN
>>>>>>     org.apache.hadoop.hdfs.******StateChange - DIR*
>>>>>> NameSystem.startFile:
>>>>>>
>>>>>>
>>>>>> failed to
>>>>>> create file /user/root/hbase.version for DFSClient_1286943058 on
>>>>>> client
>>>>>> 127.0.0.1 because current leaseholder is trying to recreate file.
>>>>>> 2012-11-09 15:52:09,522 [IPC Server handler 8 on 28077] ERROR
>>>>>> org.apache.hadoop.security.******UserGroupInformation -
>>>>>> PriviledgedActionException as:root
>>>>>> cause:org.apache.hadoop.hdfs.******protocol.****
>>>>>>
>>>>>> AlreadyBeingCreatedException:
>>>>>>
>>>>>> failed
>>>>>> to create file /user/root/hbase.version for DFSClient_1286943058 on
>>>>>> client
>>>>>> 127.0.0.1 because current leaseholder is trying to recreate file.
>>>>>> 2012-11-09 15:53:09,534 [IPC Server handler 1 on 28077] WARN
>>>>>>     org.apache.hadoop.hdfs.******StateChange - DIR*
>>>>>> NameSystem.startFile:
>>>>>>
>>>>>>
>>>>>> failed to
>>>>>> create file /user/root/hbase.version for DFSClient_1286943058 on
>>>>>> client
>>>>>> 127.0.0.1 because current leaseholder is trying to recreate file.
>>>>>> 2012-11-09 15:53:09,534 [IPC Server handler 1 on 28077] ERROR
>>>>>> org.apache.hadoop.security.******UserGroupInformation -
>>>>>> PriviledgedActionException as:root
>>>>>> cause:org.apache.hadoop.hdfs.******protocol.****
>>>>>>
>>>>>> AlreadyBeingCreatedException:
>>>>>>
>>>>>> failed
>>>>>> to create file /user/root/hbase.version for DFSClient_1286943058 on
>>>>>> client
>>>>>> 127.0.0.1 because current leaseholder is trying to recreate file.
>>>>>> 2012-11-09 15:54:09,547 [IPC Server handler 4 on 28077] WARN
>>>>>>     org.apache.hadoop.hdfs.******StateChange - DIR*
>>>>>> NameSystem.startFile:
>>>>>>
>>>>>>
>>>>>> failed to
>>>>>> create file /user/root/hbase.version for DFSClient_1286943058 on
>>>>>> client
>>>>>> 127.0.0.1 because current leaseholder is trying to recreate file.
>>>>>>
>>>>>>
>>>>>>   ------------------------------******--------------------------**
>>>>>> --**
>>>>>>
>>>>> --**---------
>>>>> To unsubscribe, e-mail: server-dev-unsubscribe@james.******apache.org<
>>>>> server-dev-**unsubscribe@**james.apache.org<[email protected]>
>>>>> <server-dev-**[email protected]<[email protected]>
>>>>> >
>>>>>
>>>>>>
>>>>>>  For additional commands, e-mail: [email protected].***
>>>>> ***org<
>>>>> server-dev-help@james.**apache**.org <http://apache.org> <
>>>>> [email protected].**org <[email protected]>
>>>>> >>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>  ------------------------------****----------------------------**
>>> --**---------
>>> To unsubscribe, e-mail: server-dev-unsubscribe@james.****apache.org<
>>> server-dev-**[email protected]<[email protected]>
>>> >
>>> For additional commands, e-mail: [email protected].****org<
>>> server-dev-help@james.**apache.org <[email protected]>>
>>>
>>>
>>>
>>
> ------------------------------**------------------------------**---------
> To unsubscribe, e-mail: 
> server-dev-unsubscribe@james.**apache.org<[email protected]>
> For additional commands, e-mail: 
> [email protected].**org<[email protected]>
>
>

Reply via email to