Daniilchik commented on PR #7328:
URL: https://github.com/apache/ozone/pull/7328#issuecomment-2422002010

   2024-10-18 09:45:42,616 [main] ERROR volume.VolumeUsage: Reserved space 
should be configured in a pair, but current value is 15GB
   2024-10-18 09:45:42,616 [main] ERROR ozone.HddsDatanodeService: Exception in 
HddsDatanodeService.
   org.apache.hadoop.hdds.conf.ConfigurationException: Reserved space should be 
configured in a pair
           at 
org.apache.hadoop.ozone.container.common.volume.VolumeUsage.getReserved(VolumeUsage.java:226)
           at 
org.apache.hadoop.ozone.container.common.volume.VolumeUsage.<init>(VolumeUsage.java:58)
           at 
org.apache.hadoop.ozone.container.common.volume.VolumeInfo.<init>(VolumeInfo.java:152)
           at 
org.apache.hadoop.ozone.container.common.volume.VolumeInfo.<init>(VolumeInfo.java:91)
           at 
org.apache.hadoop.ozone.container.common.volume.VolumeInfo$Builder.build(VolumeInfo.java:126)
           at 
org.apache.hadoop.ozone.container.common.volume.StorageVolume.<init>(StorageVolume.java:151)
           at 
org.apache.hadoop.ozone.container.common.volume.HddsVolume.<init>(HddsVolume.java:118)
           at 
org.apache.hadoop.ozone.container.common.volume.HddsVolume.<init>(HddsVolume.java:72)
           at 
org.apache.hadoop.ozone.container.common.volume.HddsVolume$Builder.build(HddsVolume.java:113)
           at 
org.apache.hadoop.ozone.container.common.volume.HddsVolumeFactory.createVolume(HddsVolumeFactory.java:48)
           at 
org.apache.hadoop.ozone.container.common.volume.MutableVolumeSet.initializeVolumeSet(MutableVolumeSet.java:170)
           at 
org.apache.hadoop.ozone.container.common.volume.MutableVolumeSet.<init>(MutableVolumeSet.java:132)
           at 
org.apache.hadoop.ozone.container.common.volume.MutableVolumeSet.<init>(MutableVolumeSet.java:96)
           at 
org.apache.hadoop.ozone.container.ozoneimpl.OzoneContainer.<init>(OzoneContainer.java:155)
           at 
org.apache.hadoop.ozone.container.common.statemachine.DatanodeStateMachine.<init>(DatanodeStateMachine.java:183)
           at 
org.apache.hadoop.ozone.HddsDatanodeService.start(HddsDatanodeService.java:291)
           at 
org.apache.hadoop.ozone.HddsDatanodeService.start(HddsDatanodeService.java:209)
           at 
org.apache.hadoop.ozone.HddsDatanodeService.call(HddsDatanodeService.java:177)
           at 
org.apache.hadoop.ozone.HddsDatanodeService.call(HddsDatanodeService.java:95)
           at picocli.CommandLine.executeUserObject(CommandLine.java:2041)
           at picocli.CommandLine.access$1500(CommandLine.java:148)
           at 
picocli.CommandLine$RunLast.executeUserObjectOfLastSubcommandWithSameParent(CommandLine.java:2461)
           at picocli.CommandLine$RunLast.handle(CommandLine.java:2453)
           at picocli.CommandLine$RunLast.handle(CommandLine.java:2415)
           at 
picocli.CommandLine$AbstractParseResultHandler.execute(CommandLine.java:2273)
           at picocli.CommandLine$RunLast.execute(CommandLine.java:2417)
           at picocli.CommandLine.execute(CommandLine.java:2170)
           at org.apache.hadoop.hdds.cli.GenericCli.execute(GenericCli.java:100)
           at org.apache.hadoop.hdds.cli.GenericCli.run(GenericCli.java:91)
           at 
org.apache.hadoop.ozone.HddsDatanodeService.main(HddsDatanodeService.java:159)
   2024-10-18 09:45:42,624 [shutdown-hook-0] INFO ozone.HddsDatanodeService: 
SHUTDOWN_MSG: 
   /************************************************************
   SHUTDOWN_MSG: Shutting down HddsDatanodeService at 60cb06476235/172.18.0.4
   ************************************************************/


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]


---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to