This is an automated email from the ASF dual-hosted git repository.
adoroszlai pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/ozone.git
The following commit(s) were added to refs/heads/master by this push:
new 3ee57e6052 HDDS-12813. Replace calls to deprecated RandomUtils methods
(#8302)
3ee57e6052 is described below
commit 3ee57e60521347a6b6900f8b6e66050415084f03
Author: Chia-Chuan Yu <[email protected]>
AuthorDate: Sat Apr 19 21:30:52 2025 +0800
HDDS-12813. Replace calls to deprecated RandomUtils methods (#8302)
---
.../hadoop/hdds/scm/storage/TestBlockInputStream.java | 2 +-
.../scm/storage/TestBlockOutputStreamCorrectness.java | 2 +-
.../io/TestECBlockReconstructedStripeInputStream.java | 2 +-
.../scm/protocolPB/ContainerCommandResponseBuilders.java | 2 +-
.../hadoop/ozone/common/TestChecksumByteBuffer.java | 3 ++-
.../ozone/common/TestChecksumImplsComputeSameValues.java | 4 ++--
.../common/report/ContainerReportPublisher.java | 3 +--
.../container/common/report/PipelineReportPublisher.java | 3 +--
.../common/impl/TestContainerDeletionChoosingPolicy.java | 2 +-
.../ozone/container/common/impl/TestHddsDispatcher.java | 4 ++--
.../container/replication/TestGrpcContainerUploader.java | 2 +-
.../hadoop/hdds/fs/TestCachingSpaceUsageSource.java | 2 +-
.../hdds/security/symmetric/TestManagedSecretKey.java | 6 +++---
.../security/token/TestOzoneBlockTokenIdentifier.java | 2 +-
.../hadoop/hdds/server/http/TestHttpServer2Metrics.java | 8 ++++----
.../java/org/apache/hadoop/hdds/scm/HddsTestUtils.java | 6 +++---
.../hadoop/hdds/scm/block/TestDeletedBlockLog.java | 8 ++++----
.../algorithms/TestSCMContainerPlacementRackAware.java | 2 +-
.../checksum/TestReplicatedBlockChecksumComputer.java | 4 ++--
.../org/apache/hadoop/ozone/MiniOzoneChaosCluster.java | 6 +++---
.../org/apache/hadoop/ozone/failure/FailureManager.java | 4 ++--
.../ozone/loadgenerators/AgedDirLoadGenerator.java | 2 +-
.../hadoop/ozone/loadgenerators/AgedLoadGenerator.java | 4 ++--
.../apache/hadoop/ozone/loadgenerators/DataBuffer.java | 2 +-
.../ozone/loadgenerators/FilesystemLoadGenerator.java | 2 +-
.../apache/hadoop/ozone/loadgenerators/LoadBucket.java | 2 +-
.../hadoop/ozone/loadgenerators/LoadExecutors.java | 2 +-
.../ozone/loadgenerators/NestedDirLoadGenerator.java | 2 +-
.../ozone/loadgenerators/RandomDirLoadGenerator.java | 2 +-
.../hadoop/ozone/loadgenerators/RandomLoadGenerator.java | 2 +-
.../ozone/loadgenerators/ReadOnlyLoadGenerator.java | 2 +-
.../hadoop/fs/contract/AbstractContractSeekTest.java | 4 ++--
.../fs/ozone/AbstractRootedOzoneFileSystemTest.java | 4 ++--
.../apache/hadoop/hdds/scm/TestContainerOperations.java | 2 +-
.../hdds/scm/TestGetCommittedBlockLengthAndPutKey.java | 4 ++--
.../hadoop/hdds/scm/TestStorageContainerManager.java | 4 ++--
.../org/apache/hadoop/hdds/scm/TestWatchForCommit.java | 6 ++++--
.../apache/hadoop/hdds/scm/TestXceiverClientGrpc.java | 6 ++++--
.../container/TestContainerStateManagerIntegration.java | 2 +-
.../metrics/TestSCMContainerManagerMetrics.java | 2 +-
.../hadoop/hdds/scm/pipeline/TestLeaderChoosePolicy.java | 2 +-
.../hadoop/hdds/scm/storage/TestContainerCommandsEC.java | 6 +++---
.../utils/db/managed/TestRocksObjectLeakDetector.java | 2 +-
.../java/org/apache/hadoop/ozone/TestBlockTokens.java | 2 +-
.../org/apache/hadoop/ozone/TestMultipartObjectGet.java | 2 +-
.../hadoop/ozone/client/rpc/OzoneRpcClientTests.java | 8 ++++----
.../hadoop/ozone/client/rpc/TestBlockOutputStream.java | 12 ++++++------
.../client/rpc/TestBlockOutputStreamWithFailures.java | 16 ++++++++--------
.../rpc/TestOzoneClientMultipartUploadWithFSO.java | 2 +-
.../rpc/TestOzoneRpcClientWithKeyLatestVersion.java | 6 +++---
.../apache/hadoop/ozone/client/rpc/TestReadRetries.java | 2 +-
.../container/server/TestSecureContainerServer.java | 2 +-
.../volume/TestDatanodeHddsVolumeFailureDetection.java | 2 +-
.../hadoop/ozone/om/TestObjectStoreWithLegacyFS.java | 2 +-
.../hadoop/ozone/shell/TestDeletedBlocksTxnShell.java | 4 ++--
.../org/apache/hadoop/ozone/om/TestKeyManagerUnit.java | 4 ++--
.../om/service/TestMultipartUploadCleanupService.java | 6 +++---
.../ozone/om/service/TestOpenKeyCleanupService.java | 10 +++++-----
.../hadoop/ozone/security/TestOzoneTokenIdentifier.java | 2 +-
.../ozone/security/acl/TestOzoneNativeAuthorizer.java | 4 ++--
.../apache/hadoop/ozone/security/acl/TestParentAcl.java | 4 ++--
.../apache/hadoop/fs/ozone/TestOzoneFSInputStream.java | 2 +-
.../org/apache/hadoop/ozone/recon/TestReconUtils.java | 4 ++--
.../hadoop/ozone/freon/DatanodeSimulationState.java | 2 +-
.../org/apache/hadoop/ozone/freon/DatanodeSimulator.java | 4 ++--
.../ozone/freon/OzoneClientKeyReadWriteListOps.java | 2 +-
66 files changed, 125 insertions(+), 122 deletions(-)
diff --git
a/hadoop-hdds/client/src/test/java/org/apache/hadoop/hdds/scm/storage/TestBlockInputStream.java
b/hadoop-hdds/client/src/test/java/org/apache/hadoop/hdds/scm/storage/TestBlockInputStream.java
index ab612bcaea..2e9a84cad4 100644
---
a/hadoop-hdds/client/src/test/java/org/apache/hadoop/hdds/scm/storage/TestBlockInputStream.java
+++
b/hadoop-hdds/client/src/test/java/org/apache/hadoop/hdds/scm/storage/TestBlockInputStream.java
@@ -193,7 +193,7 @@ public void testSeek() throws Exception {
// Seek to random positions between 0 and the block size.
for (int i = 0; i < 10; i++) {
- pos = RandomUtils.nextInt(0, blockSize);
+ pos = RandomUtils.secure().randomInt(0, blockSize);
seekAndVerify(pos);
}
}
diff --git
a/hadoop-hdds/client/src/test/java/org/apache/hadoop/hdds/scm/storage/TestBlockOutputStreamCorrectness.java
b/hadoop-hdds/client/src/test/java/org/apache/hadoop/hdds/scm/storage/TestBlockOutputStreamCorrectness.java
index 7724c50283..a6f589e9cc 100644
---
a/hadoop-hdds/client/src/test/java/org/apache/hadoop/hdds/scm/storage/TestBlockOutputStreamCorrectness.java
+++
b/hadoop-hdds/client/src/test/java/org/apache/hadoop/hdds/scm/storage/TestBlockOutputStreamCorrectness.java
@@ -69,7 +69,7 @@
class TestBlockOutputStreamCorrectness {
private static final int DATA_SIZE = 256 * (int) OzoneConsts.MB;
- private static final byte[] DATA = RandomUtils.nextBytes(DATA_SIZE);
+ private static final byte[] DATA =
RandomUtils.secure().randomBytes(DATA_SIZE);
@ParameterizedTest
@ValueSource(ints = { 1, 1024, 1024 * 1024 })
diff --git
a/hadoop-hdds/client/src/test/java/org/apache/hadoop/ozone/client/io/TestECBlockReconstructedStripeInputStream.java
b/hadoop-hdds/client/src/test/java/org/apache/hadoop/ozone/client/io/TestECBlockReconstructedStripeInputStream.java
index effef61301..2fed95bfa6 100644
---
a/hadoop-hdds/client/src/test/java/org/apache/hadoop/ozone/client/io/TestECBlockReconstructedStripeInputStream.java
+++
b/hadoop-hdds/client/src/test/java/org/apache/hadoop/ozone/client/io/TestECBlockReconstructedStripeInputStream.java
@@ -647,7 +647,7 @@ public void testSeekToPartialOffsetFails() {
}
private Integer getRandomStreamIndex(Set<Integer> set) {
- return set.stream().skip(RandomUtils.nextInt(0, set.size()))
+ return set.stream().skip(RandomUtils.secure().randomInt(0, set.size()))
.findFirst().orElse(null);
}
diff --git
a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/protocolPB/ContainerCommandResponseBuilders.java
b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/protocolPB/ContainerCommandResponseBuilders.java
index 6285e2caf7..9ce971c081 100644
---
a/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/protocolPB/ContainerCommandResponseBuilders.java
+++
b/hadoop-hdds/common/src/main/java/org/apache/hadoop/hdds/scm/protocolPB/ContainerCommandResponseBuilders.java
@@ -362,7 +362,7 @@ public static ContainerCommandResponseProto getEchoResponse(
ContainerProtos.EchoResponseProto.Builder echo =
ContainerProtos.EchoResponseProto
.newBuilder()
-
.setPayload(UnsafeByteOperations.unsafeWrap(RandomUtils.nextBytes(responsePayload)));
+
.setPayload(UnsafeByteOperations.unsafeWrap(RandomUtils.secure().randomBytes(responsePayload)));
return getSuccessResponseBuilder(msg)
.setEcho(echo)
diff --git
a/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/common/TestChecksumByteBuffer.java
b/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/common/TestChecksumByteBuffer.java
index 48ed716d2e..6151d71da5 100644
---
a/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/common/TestChecksumByteBuffer.java
+++
b/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/common/TestChecksumByteBuffer.java
@@ -79,7 +79,8 @@ void testCorrectness() {
final int len = 1 << 10;
for (int i = 0; i < 1000; i++) {
- checkBytes(RandomUtils.nextBytes(len), RandomUtils.nextInt(0, len));
+ checkBytes(RandomUtils.secure().randomBytes(len),
+ RandomUtils.secure().randomInt(0, len));
}
}
diff --git
a/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/common/TestChecksumImplsComputeSameValues.java
b/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/common/TestChecksumImplsComputeSameValues.java
index 96f1c52c22..fed48f63ff 100644
---
a/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/common/TestChecksumImplsComputeSameValues.java
+++
b/hadoop-hdds/common/src/test/java/org/apache/hadoop/ozone/common/TestChecksumImplsComputeSameValues.java
@@ -42,7 +42,7 @@ public class TestChecksumImplsComputeSameValues {
@Test
public void testCRC32ImplsMatch() {
data.clear();
- data.put(RandomUtils.nextBytes(data.remaining()));
+ data.put(RandomUtils.secure().randomBytes(data.remaining()));
for (int bpc : bytesPerChecksum) {
List<ChecksumByteBuffer> impls = new ArrayList<>();
impls.add(new PureJavaCrc32ByteBuffer());
@@ -58,7 +58,7 @@ public void testCRC32ImplsMatch() {
@Test
public void testCRC32CImplsMatch() {
data.clear();
- data.put(RandomUtils.nextBytes(data.remaining()));
+ data.put(RandomUtils.secure().randomBytes(data.remaining()));
for (int bpc : bytesPerChecksum) {
List<ChecksumByteBuffer> impls = new ArrayList<>();
impls.add(new PureJavaCrc32CByteBuffer());
diff --git
a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/report/ContainerReportPublisher.java
b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/report/ContainerReportPublisher.java
index 26c0c69466..6926543b7c 100644
---
a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/report/ContainerReportPublisher.java
+++
b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/report/ContainerReportPublisher.java
@@ -69,9 +69,8 @@ protected long getReportFrequency() {
return containerReportInterval + getRandomReportDelay();
}
- @SuppressWarnings("java:S2245") // no need for secure random
private long getRandomReportDelay() {
- return RandomUtils.nextLong(0, containerReportInterval);
+ return RandomUtils.secure().randomLong(0, containerReportInterval);
}
@Override
diff --git
a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/report/PipelineReportPublisher.java
b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/report/PipelineReportPublisher.java
index 5e6049843b..3064b74210 100644
---
a/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/report/PipelineReportPublisher.java
+++
b/hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/common/report/PipelineReportPublisher.java
@@ -59,9 +59,8 @@ protected long getReportFrequency() {
return pipelineReportInterval + getRandomReportDelay();
}
- @SuppressWarnings("java:S2245") // no need for secure random
private long getRandomReportDelay() {
- return RandomUtils.nextLong(0, pipelineReportInterval);
+ return RandomUtils.secure().randomLong(0, pipelineReportInterval);
}
@Override
diff --git
a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/common/impl/TestContainerDeletionChoosingPolicy.java
b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/common/impl/TestContainerDeletionChoosingPolicy.java
index bb6aadae84..bd2e1237dd 100644
---
a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/common/impl/TestContainerDeletionChoosingPolicy.java
+++
b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/common/impl/TestContainerDeletionChoosingPolicy.java
@@ -151,7 +151,7 @@ public void
testTopNOrderedChoosingPolicy(ContainerLayoutVersion layout)
List<Integer> numberOfBlocks = new ArrayList<Integer>();
// create [numContainers + 1] containers
for (int i = 0; i <= numContainers; i++) {
- long containerId = RandomUtils.nextLong();
+ long containerId = RandomUtils.secure().randomLong();
KeyValueContainerData data =
new KeyValueContainerData(containerId,
layout,
diff --git
a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/common/impl/TestHddsDispatcher.java
b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/common/impl/TestHddsDispatcher.java
index 5a2e362f14..6afcadb809 100644
---
a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/common/impl/TestHddsDispatcher.java
+++
b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/common/impl/TestHddsDispatcher.java
@@ -609,7 +609,7 @@ static ChecksumData checksum(ByteString data) {
private ContainerCommandRequestProto getWriteChunkRequest0(
String datanodeId, Long containerId, Long localId, int chunkNum) {
final int lenOfBytes = 32;
- ByteString chunkData = ByteString.copyFrom(RandomUtils.nextBytes(32));
+ ByteString chunkData =
ByteString.copyFrom(RandomUtils.secure().randomBytes(32));
ContainerProtos.ChunkInfo chunk = ContainerProtos.ChunkInfo
.newBuilder()
@@ -638,7 +638,7 @@ private ContainerCommandRequestProto getWriteChunkRequest0(
}
static ContainerCommandRequestProto newPutSmallFile(Long containerId, Long
localId) {
- ByteString chunkData = ByteString.copyFrom(RandomUtils.nextBytes(32));
+ ByteString chunkData =
ByteString.copyFrom(RandomUtils.secure().randomBytes(32));
return newPutSmallFile(new BlockID(containerId, localId), chunkData);
}
diff --git
a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/replication/TestGrpcContainerUploader.java
b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/replication/TestGrpcContainerUploader.java
index cd3587d351..b8df5c18e8 100644
---
a/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/replication/TestGrpcContainerUploader.java
+++
b/hadoop-hdds/container-service/src/test/java/org/apache/hadoop/ozone/container/replication/TestGrpcContainerUploader.java
@@ -89,7 +89,7 @@ public void onNext(SendContainerRequest value) {
// WHEN
OutputStream out = startUpload(subject, callback);
- out.write(RandomUtils.nextBytes(4));
+ out.write(RandomUtils.secure().randomBytes(4));
out.close();
// THEN
diff --git
a/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/fs/TestCachingSpaceUsageSource.java
b/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/fs/TestCachingSpaceUsageSource.java
index d63f23578c..db0a2c89c9 100644
---
a/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/fs/TestCachingSpaceUsageSource.java
+++
b/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/fs/TestCachingSpaceUsageSource.java
@@ -194,7 +194,7 @@ private static long missingInitialValue() {
}
private static long validInitialValue() {
- return RandomUtils.nextLong(1, 100);
+ return RandomUtils.secure().randomLong(1, 100);
}
private static Builder paramsBuilder(AtomicLong savedValue) {
diff --git
a/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/security/symmetric/TestManagedSecretKey.java
b/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/security/symmetric/TestManagedSecretKey.java
index cd5bf80b2b..386c8afb49 100644
---
a/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/security/symmetric/TestManagedSecretKey.java
+++
b/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/security/symmetric/TestManagedSecretKey.java
@@ -39,7 +39,7 @@ public class TestManagedSecretKey {
@Test
public void testSignAndVerifySuccess() throws Exception {
// Data can be signed and verified by same key.
- byte[] data = RandomUtils.nextBytes(100);
+ byte[] data = RandomUtils.secure().randomBytes(100);
ManagedSecretKey secretKey = generateHmac(now(), ofDays(1));
byte[] signature = secretKey.sign(data);
assertTrue(secretKey.isValidSignature(data, signature));
@@ -62,10 +62,10 @@ public void testSignAndVerifySuccess() throws Exception {
@Test
public void testVerifyFailure() throws Exception {
- byte[] data = RandomUtils.nextBytes(100);
+ byte[] data = RandomUtils.secure().randomBytes(100);
ManagedSecretKey secretKey = generateHmac(now(), ofDays(1));
// random signature is not valid.
- assertFalse(secretKey.isValidSignature(data, RandomUtils.nextBytes(100)));
+ assertFalse(secretKey.isValidSignature(data,
RandomUtils.secure().randomBytes(100)));
// Data sign by one key can't be verified by another key.
byte[] signature = secretKey.sign(data);
diff --git
a/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/security/token/TestOzoneBlockTokenIdentifier.java
b/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/security/token/TestOzoneBlockTokenIdentifier.java
index 4199af1068..a04df4f618 100644
---
a/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/security/token/TestOzoneBlockTokenIdentifier.java
+++
b/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/security/token/TestOzoneBlockTokenIdentifier.java
@@ -65,7 +65,7 @@ public void testSignToken() {
// Verify an invalid signed OzoneMaster Token with Ozone Master.
assertFalse(secretKey.isValidSignature(tokenId.getBytes(),
- RandomUtils.nextBytes(128)));
+ RandomUtils.secure().randomBytes(128)));
}
@Test
diff --git
a/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/server/http/TestHttpServer2Metrics.java
b/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/server/http/TestHttpServer2Metrics.java
index 14e0a57bb5..07e842068d 100644
---
a/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/server/http/TestHttpServer2Metrics.java
+++
b/hadoop-hdds/framework/src/test/java/org/apache/hadoop/hdds/server/http/TestHttpServer2Metrics.java
@@ -56,10 +56,10 @@ public void setup() {
@Test
public void testMetrics() {
// crate mock metrics
- int threadCount = RandomUtils.nextInt();
- int maxThreadCount = RandomUtils.nextInt();
- int idleThreadCount = RandomUtils.nextInt();
- int threadQueueWaitingTaskCount = RandomUtils.nextInt();
+ int threadCount = RandomUtils.secure().randomInt();
+ int maxThreadCount = RandomUtils.secure().randomInt();
+ int idleThreadCount = RandomUtils.secure().randomInt();
+ int threadQueueWaitingTaskCount = RandomUtils.secure().randomInt();
String name = "s3g";
when(threadPool.getThreads()).thenReturn(threadCount);
diff --git
a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/HddsTestUtils.java
b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/HddsTestUtils.java
index a3e948a5e9..4029b49172 100644
---
a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/HddsTestUtils.java
+++
b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/HddsTestUtils.java
@@ -623,7 +623,7 @@ public static StorageContainerManager
getScm(OzoneConfiguration conf,
private static ContainerInfo.Builder getDefaultContainerInfoBuilder(
final HddsProtos.LifeCycleState state) {
return new ContainerInfo.Builder()
- .setContainerID(RandomUtils.nextLong())
+ .setContainerID(RandomUtils.secure().randomLong())
.setReplicationConfig(
RatisReplicationConfig
.getInstance(ReplicationFactor.THREE))
@@ -816,7 +816,7 @@ public static List<ContainerInfo> getContainerInfo(int
numContainers) {
for (int i = 0; i < numContainers; i++) {
ContainerInfo.Builder builder = new ContainerInfo.Builder();
containerInfoList.add(builder
- .setContainerID(RandomUtils.nextLong())
+ .setContainerID(RandomUtils.secure().randomLong())
.setReplicationConfig(ratisReplicationConfig)
.build());
}
@@ -837,7 +837,7 @@ public static List<ContainerInfo> getECContainerInfo(int
numContainers, int data
for (int i = 0; i < numContainers; i++) {
ContainerInfo.Builder builder = new ContainerInfo.Builder();
containerInfoList.add(builder
- .setContainerID(RandomUtils.nextLong())
+ .setContainerID(RandomUtils.secure().randomLong())
.setOwner("test-owner")
.setPipelineID(PipelineID.randomId())
.setReplicationConfig(eCReplicationConfig)
diff --git
a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/block/TestDeletedBlockLog.java
b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/block/TestDeletedBlockLog.java
index 23583d03b5..f37fa1ac10 100644
---
a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/block/TestDeletedBlockLog.java
+++
b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/block/TestDeletedBlockLog.java
@@ -205,8 +205,8 @@ private Map<Long, List<Long>> generateData(int dataSize)
throws IOException {
private Map<Long, List<Long>> generateData(int dataSize,
HddsProtos.LifeCycleState state) throws IOException {
Map<Long, List<Long>> blockMap = new HashMap<>();
- int continerIDBase = RandomUtils.nextInt(0, 100);
- int localIDBase = RandomUtils.nextInt(0, 1000);
+ int continerIDBase = RandomUtils.secure().randomInt(0, 100);
+ int localIDBase = RandomUtils.secure().randomInt(0, 1000);
for (int i = 0; i < dataSize; i++) {
long containerID = continerIDBase + i;
updateContainerMetadata(containerID, state);
@@ -752,7 +752,7 @@ public void testRandomOperateTransactions() throws
Exception {
List<Long> txIDs;
// Randomly add/get/commit/increase transactions.
for (int i = 0; i < 100; i++) {
- int state = RandomUtils.nextInt(0, 4);
+ int state = RandomUtils.secure().randomInt(0, 4);
if (state == 0) {
addTransactions(generateData(10), true);
added += 10;
@@ -851,7 +851,7 @@ public void testDeletedBlockTransactions()
// add two transactions for same container
containerID = blocks.get(0).getContainerID();
Map<Long, List<Long>> deletedBlocksMap = new HashMap<>();
- long localId = RandomUtils.nextLong();
+ long localId = RandomUtils.secure().randomLong();
deletedBlocksMap.put(containerID, new LinkedList<>(
Collections.singletonList(localId)));
addTransactions(deletedBlocksMap, true);
diff --git
a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/placement/algorithms/TestSCMContainerPlacementRackAware.java
b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/placement/algorithms/TestSCMContainerPlacementRackAware.java
index c60a4f81c7..9497abfdee 100644
---
a/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/placement/algorithms/TestSCMContainerPlacementRackAware.java
+++
b/hadoop-hdds/server-scm/src/test/java/org/apache/hadoop/hdds/scm/container/placement/algorithms/TestSCMContainerPlacementRackAware.java
@@ -621,7 +621,7 @@ public void testOutOfServiceNodesNotSelected(int
datanodeCount) {
for (int i = 0; i < 10; i++) {
// Set a random DN to in_service and ensure it is always picked
- int index = RandomUtils.nextInt(0, dnInfos.size());
+ int index = RandomUtils.secure().randomInt(0, dnInfos.size());
dnInfos.get(index).setNodeStatus(NodeStatus.inServiceHealthy());
try {
List<DatanodeDetails> datanodeDetails =
diff --git
a/hadoop-ozone/client/src/test/java/org/apache/hadoop/ozone/client/checksum/TestReplicatedBlockChecksumComputer.java
b/hadoop-ozone/client/src/test/java/org/apache/hadoop/ozone/client/checksum/TestReplicatedBlockChecksumComputer.java
index 2cfb59dfba..bc232e8e05 100644
---
a/hadoop-ozone/client/src/test/java/org/apache/hadoop/ozone/client/checksum/TestReplicatedBlockChecksumComputer.java
+++
b/hadoop-ozone/client/src/test/java/org/apache/hadoop/ozone/client/checksum/TestReplicatedBlockChecksumComputer.java
@@ -39,7 +39,7 @@ public class TestReplicatedBlockChecksumComputer {
@Test
public void testComputeMd5Crc() throws IOException {
final int lenOfBytes = 32;
- byte[] randomChunkChecksum = RandomUtils.nextBytes(lenOfBytes);
+ byte[] randomChunkChecksum = RandomUtils.secure().randomBytes(lenOfBytes);
MD5Hash emptyBlockMD5 = MD5Hash.digest(randomChunkChecksum);
byte[] emptyBlockMD5Hash = emptyBlockMD5.getDigest();
@@ -54,7 +54,7 @@ public void testComputeMd5Crc() throws IOException {
@Test
public void testComputeCompositeCrc() throws IOException {
final int lenOfBytes = 32;
- byte[] randomChunkChecksum = RandomUtils.nextBytes(lenOfBytes);
+ byte[] randomChunkChecksum = RandomUtils.secure().randomBytes(lenOfBytes);
CrcComposer crcComposer =
CrcComposer.newCrcComposer(DataChecksum.Type.CRC32C, 4);
diff --git
a/hadoop-ozone/fault-injection-test/mini-chaos-tests/src/test/java/org/apache/hadoop/ozone/MiniOzoneChaosCluster.java
b/hadoop-ozone/fault-injection-test/mini-chaos-tests/src/test/java/org/apache/hadoop/ozone/MiniOzoneChaosCluster.java
index 17f135ebbf..555055a630 100644
---
a/hadoop-ozone/fault-injection-test/mini-chaos-tests/src/test/java/org/apache/hadoop/ozone/MiniOzoneChaosCluster.java
+++
b/hadoop-ozone/fault-injection-test/mini-chaos-tests/src/test/java/org/apache/hadoop/ozone/MiniOzoneChaosCluster.java
@@ -317,12 +317,12 @@ public boolean shouldStopOm() {
if (failedOmSet.size() >= numOzoneManagers / 2) {
return false;
}
- return RandomUtils.nextBoolean();
+ return RandomUtils.secure().randomBoolean();
}
// Datanode specific
private int getNumberOfDnToFail() {
- return RandomUtils.nextBoolean() ? 1 : 2;
+ return RandomUtils.secure().randomBoolean() ? 1 : 2;
}
public Set<DatanodeDetails> dnToFail() {
@@ -395,7 +395,7 @@ public boolean shouldStopScm() {
if (failedScmSet.size() >= numStorageContainerManagers / 2) {
return false;
}
- return RandomUtils.nextBoolean();
+ return RandomUtils.secure().randomBoolean();
}
}
diff --git
a/hadoop-ozone/fault-injection-test/mini-chaos-tests/src/test/java/org/apache/hadoop/ozone/failure/FailureManager.java
b/hadoop-ozone/fault-injection-test/mini-chaos-tests/src/test/java/org/apache/hadoop/ozone/failure/FailureManager.java
index 6d80fd362e..2a27708ef9 100644
---
a/hadoop-ozone/fault-injection-test/mini-chaos-tests/src/test/java/org/apache/hadoop/ozone/failure/FailureManager.java
+++
b/hadoop-ozone/fault-injection-test/mini-chaos-tests/src/test/java/org/apache/hadoop/ozone/failure/FailureManager.java
@@ -89,10 +89,10 @@ public void stop() throws Exception {
}
public static boolean isFastRestart() {
- return RandomUtils.nextBoolean();
+ return RandomUtils.secure().randomBoolean();
}
public static int getBoundedRandomIndex(int size) {
- return RandomUtils.nextInt(0, size);
+ return RandomUtils.secure().randomInt(0, size);
}
}
diff --git
a/hadoop-ozone/fault-injection-test/mini-chaos-tests/src/test/java/org/apache/hadoop/ozone/loadgenerators/AgedDirLoadGenerator.java
b/hadoop-ozone/fault-injection-test/mini-chaos-tests/src/test/java/org/apache/hadoop/ozone/loadgenerators/AgedDirLoadGenerator.java
index 19ca066db9..2187309839 100644
---
a/hadoop-ozone/fault-injection-test/mini-chaos-tests/src/test/java/org/apache/hadoop/ozone/loadgenerators/AgedDirLoadGenerator.java
+++
b/hadoop-ozone/fault-injection-test/mini-chaos-tests/src/test/java/org/apache/hadoop/ozone/loadgenerators/AgedDirLoadGenerator.java
@@ -33,7 +33,7 @@ public AgedDirLoadGenerator(DataBuffer dataBuffer, LoadBucket
fsBucket) {
@Override
public void generateLoad() throws Exception {
- int index = RandomUtils.nextInt(0, maxDirIndex);
+ int index = RandomUtils.secure().randomInt(0, maxDirIndex);
String keyName = getKeyName(index);
fsBucket.readDirectory(keyName);
}
diff --git
a/hadoop-ozone/fault-injection-test/mini-chaos-tests/src/test/java/org/apache/hadoop/ozone/loadgenerators/AgedLoadGenerator.java
b/hadoop-ozone/fault-injection-test/mini-chaos-tests/src/test/java/org/apache/hadoop/ozone/loadgenerators/AgedLoadGenerator.java
index ff101786c9..4551eedea4 100644
---
a/hadoop-ozone/fault-injection-test/mini-chaos-tests/src/test/java/org/apache/hadoop/ozone/loadgenerators/AgedLoadGenerator.java
+++
b/hadoop-ozone/fault-injection-test/mini-chaos-tests/src/test/java/org/apache/hadoop/ozone/loadgenerators/AgedLoadGenerator.java
@@ -45,7 +45,7 @@ public AgedLoadGenerator(DataBuffer data, LoadBucket
agedLoadBucket) {
@Override
public void generateLoad() throws Exception {
- if (RandomUtils.nextInt(0, 100) <= 10) {
+ if (RandomUtils.secure().randomInt(0, 100) <= 10) {
synchronized (agedFileAllocationIndex) {
int index = agedFileAllocationIndex.getAndIncrement();
ByteBuffer buffer = dataBuffer.getBuffer(index);
@@ -66,7 +66,7 @@ public void generateLoad() throws Exception {
private Optional<Integer> randomKeyToRead() {
int currentIndex = agedFileWrittenIndex.get();
return currentIndex != 0
- ? Optional.of(RandomUtils.nextInt(0, currentIndex))
+ ? Optional.of(RandomUtils.secure().randomInt(0, currentIndex))
: Optional.empty();
}
diff --git
a/hadoop-ozone/fault-injection-test/mini-chaos-tests/src/test/java/org/apache/hadoop/ozone/loadgenerators/DataBuffer.java
b/hadoop-ozone/fault-injection-test/mini-chaos-tests/src/test/java/org/apache/hadoop/ozone/loadgenerators/DataBuffer.java
index 1778481692..9c5019e698 100644
---
a/hadoop-ozone/fault-injection-test/mini-chaos-tests/src/test/java/org/apache/hadoop/ozone/loadgenerators/DataBuffer.java
+++
b/hadoop-ozone/fault-injection-test/mini-chaos-tests/src/test/java/org/apache/hadoop/ozone/loadgenerators/DataBuffer.java
@@ -39,7 +39,7 @@ public DataBuffer(int numBuffers) {
for (int i = 0; i < numBuffers; i++) {
int size = (int) StorageUnit.KB.toBytes(1 << i);
ByteBuffer buffer = ByteBuffer.allocate(size);
- buffer.put(RandomUtils.nextBytes(size));
+ buffer.put(RandomUtils.secure().randomBytes(size));
this.buffers.add(buffer);
}
// TODO: add buffers of sizes of prime numbers.
diff --git
a/hadoop-ozone/fault-injection-test/mini-chaos-tests/src/test/java/org/apache/hadoop/ozone/loadgenerators/FilesystemLoadGenerator.java
b/hadoop-ozone/fault-injection-test/mini-chaos-tests/src/test/java/org/apache/hadoop/ozone/loadgenerators/FilesystemLoadGenerator.java
index 5a3c6c300e..61ec463a5c 100644
---
a/hadoop-ozone/fault-injection-test/mini-chaos-tests/src/test/java/org/apache/hadoop/ozone/loadgenerators/FilesystemLoadGenerator.java
+++
b/hadoop-ozone/fault-injection-test/mini-chaos-tests/src/test/java/org/apache/hadoop/ozone/loadgenerators/FilesystemLoadGenerator.java
@@ -38,7 +38,7 @@ public FilesystemLoadGenerator(DataBuffer dataBuffer,
LoadBucket fsBucket) {
@Override
public void generateLoad() throws Exception {
- int index = RandomUtils.nextInt();
+ int index = RandomUtils.secure().randomInt();
ByteBuffer buffer = dataBuffer.getBuffer(index);
String keyName = getKeyName(index);
fsBucket.writeKey(true, buffer, keyName);
diff --git
a/hadoop-ozone/fault-injection-test/mini-chaos-tests/src/test/java/org/apache/hadoop/ozone/loadgenerators/LoadBucket.java
b/hadoop-ozone/fault-injection-test/mini-chaos-tests/src/test/java/org/apache/hadoop/ozone/loadgenerators/LoadBucket.java
index 6722586dce..4f0c1b9211 100644
---
a/hadoop-ozone/fault-injection-test/mini-chaos-tests/src/test/java/org/apache/hadoop/ozone/loadgenerators/LoadBucket.java
+++
b/hadoop-ozone/fault-injection-test/mini-chaos-tests/src/test/java/org/apache/hadoop/ozone/loadgenerators/LoadBucket.java
@@ -63,7 +63,7 @@ public LoadBucket(OzoneBucket bucket, OzoneConfiguration conf,
}
private boolean isFsOp() {
- return RandomUtils.nextBoolean();
+ return RandomUtils.secure().randomBoolean();
}
// Write ops.
diff --git
a/hadoop-ozone/fault-injection-test/mini-chaos-tests/src/test/java/org/apache/hadoop/ozone/loadgenerators/LoadExecutors.java
b/hadoop-ozone/fault-injection-test/mini-chaos-tests/src/test/java/org/apache/hadoop/ozone/loadgenerators/LoadExecutors.java
index a0b3e8fb8b..c7c8ccf9ec 100644
---
a/hadoop-ozone/fault-injection-test/mini-chaos-tests/src/test/java/org/apache/hadoop/ozone/loadgenerators/LoadExecutors.java
+++
b/hadoop-ozone/fault-injection-test/mini-chaos-tests/src/test/java/org/apache/hadoop/ozone/loadgenerators/LoadExecutors.java
@@ -57,7 +57,7 @@ private void load(long runTimeMillis) {
while (Time.monotonicNow() - startTime < runTimeMillis) {
LoadGenerator gen =
- generators.get(RandomUtils.nextInt(0, numGenerators));
+ generators.get(RandomUtils.secure().randomInt(0, numGenerators));
try {
gen.generateLoad();
diff --git
a/hadoop-ozone/fault-injection-test/mini-chaos-tests/src/test/java/org/apache/hadoop/ozone/loadgenerators/NestedDirLoadGenerator.java
b/hadoop-ozone/fault-injection-test/mini-chaos-tests/src/test/java/org/apache/hadoop/ozone/loadgenerators/NestedDirLoadGenerator.java
index 4ed69f13ab..f1a82719b6 100644
---
a/hadoop-ozone/fault-injection-test/mini-chaos-tests/src/test/java/org/apache/hadoop/ozone/loadgenerators/NestedDirLoadGenerator.java
+++
b/hadoop-ozone/fault-injection-test/mini-chaos-tests/src/test/java/org/apache/hadoop/ozone/loadgenerators/NestedDirLoadGenerator.java
@@ -42,7 +42,7 @@ private String createNewPath(int i, String s) {
@Override
public void generateLoad() throws Exception {
- int index = RandomUtils.nextInt(0, maxDirDepth);
+ int index = RandomUtils.secure().randomInt(0, maxDirDepth);
String str = this.pathMap.compute(index, this::createNewPath);
fsBucket.createDirectory(str);
fsBucket.readDirectory(str);
diff --git
a/hadoop-ozone/fault-injection-test/mini-chaos-tests/src/test/java/org/apache/hadoop/ozone/loadgenerators/RandomDirLoadGenerator.java
b/hadoop-ozone/fault-injection-test/mini-chaos-tests/src/test/java/org/apache/hadoop/ozone/loadgenerators/RandomDirLoadGenerator.java
index c81df6e22e..88ef4a60b3 100644
---
a/hadoop-ozone/fault-injection-test/mini-chaos-tests/src/test/java/org/apache/hadoop/ozone/loadgenerators/RandomDirLoadGenerator.java
+++
b/hadoop-ozone/fault-injection-test/mini-chaos-tests/src/test/java/org/apache/hadoop/ozone/loadgenerators/RandomDirLoadGenerator.java
@@ -31,7 +31,7 @@ public RandomDirLoadGenerator(DataBuffer dataBuffer,
LoadBucket fsBucket) {
@Override
public void generateLoad() throws Exception {
- int index = RandomUtils.nextInt();
+ int index = RandomUtils.secure().randomInt();
String keyName = getKeyName(index);
fsBucket.createDirectory(keyName);
fsBucket.readDirectory(keyName);
diff --git
a/hadoop-ozone/fault-injection-test/mini-chaos-tests/src/test/java/org/apache/hadoop/ozone/loadgenerators/RandomLoadGenerator.java
b/hadoop-ozone/fault-injection-test/mini-chaos-tests/src/test/java/org/apache/hadoop/ozone/loadgenerators/RandomLoadGenerator.java
index 91d5d23ee3..f9cda3b5f6 100644
---
a/hadoop-ozone/fault-injection-test/mini-chaos-tests/src/test/java/org/apache/hadoop/ozone/loadgenerators/RandomLoadGenerator.java
+++
b/hadoop-ozone/fault-injection-test/mini-chaos-tests/src/test/java/org/apache/hadoop/ozone/loadgenerators/RandomLoadGenerator.java
@@ -36,7 +36,7 @@ public RandomLoadGenerator(DataBuffer dataBuffer, LoadBucket
bucket) {
@Override
public void generateLoad() throws Exception {
- int index = RandomUtils.nextInt();
+ int index = RandomUtils.secure().randomInt();
ByteBuffer buffer = dataBuffer.getBuffer(index);
String keyName = getKeyName(index);
ozoneBucket.writeKey(buffer, keyName);
diff --git
a/hadoop-ozone/fault-injection-test/mini-chaos-tests/src/test/java/org/apache/hadoop/ozone/loadgenerators/ReadOnlyLoadGenerator.java
b/hadoop-ozone/fault-injection-test/mini-chaos-tests/src/test/java/org/apache/hadoop/ozone/loadgenerators/ReadOnlyLoadGenerator.java
index a991245a62..9d19f9f6c7 100644
---
a/hadoop-ozone/fault-injection-test/mini-chaos-tests/src/test/java/org/apache/hadoop/ozone/loadgenerators/ReadOnlyLoadGenerator.java
+++
b/hadoop-ozone/fault-injection-test/mini-chaos-tests/src/test/java/org/apache/hadoop/ozone/loadgenerators/ReadOnlyLoadGenerator.java
@@ -35,7 +35,7 @@ public ReadOnlyLoadGenerator(DataBuffer dataBuffer,
LoadBucket replBucket) {
@Override
public void generateLoad() throws Exception {
- int index = RandomUtils.nextInt(0, NUM_KEYS);
+ int index = RandomUtils.secure().randomInt(0, NUM_KEYS);
ByteBuffer buffer = dataBuffer.getBuffer(index);
String keyName = getKeyName(index);
replBucket.readKey(buffer, keyName);
diff --git
a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/fs/contract/AbstractContractSeekTest.java
b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/fs/contract/AbstractContractSeekTest.java
index d19475fa6a..ca1450382d 100644
---
a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/fs/contract/AbstractContractSeekTest.java
+++
b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/fs/contract/AbstractContractSeekTest.java
@@ -349,8 +349,8 @@ public void testRandomSeeks() throws Throwable {
int[] reads = new int[10];
try (FSDataInputStream stm = getFileSystem().open(randomSeekFile)) {
for (int i = 0; i < limit; i++) {
- int seekOff = RandomUtils.nextInt(0, buf.length);
- int toRead = RandomUtils.nextInt(0, Math.min(buf.length - seekOff,
32000));
+ int seekOff = RandomUtils.secure().randomInt(0, buf.length);
+ int toRead = RandomUtils.secure().randomInt(0, Math.min(buf.length -
seekOff, 32000));
seeks[i % seeks.length] = seekOff;
reads[i % reads.length] = toRead;
diff --git
a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/fs/ozone/AbstractRootedOzoneFileSystemTest.java
b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/fs/ozone/AbstractRootedOzoneFileSystemTest.java
index f8b9c26d91..70df69038d 100644
---
a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/fs/ozone/AbstractRootedOzoneFileSystemTest.java
+++
b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/fs/ozone/AbstractRootedOzoneFileSystemTest.java
@@ -1984,7 +1984,7 @@ private void checkInvalidPath(Path path) {
@Test
void testRenameFile() throws Exception {
- final String dir = "/dir" + RandomUtils.nextInt(0, 1000);
+ final String dir = "/dir" + RandomUtils.secure().randomInt(0, 1000);
Path dirPath = new Path(getBucketPath() + dir);
Path file1Source = new Path(getBucketPath() + dir
+ "/file1_Copy");
@@ -2010,7 +2010,7 @@ void testRenameFile() throws Exception {
*/
@Test
void testRenameFileToDir() throws Exception {
- final String dir = "/dir" + RandomUtils.nextInt(0, 1000);
+ final String dir = "/dir" + RandomUtils.secure().randomInt(0, 1000);
Path dirPath = new Path(getBucketPath() + dir);
getFs().mkdirs(dirPath);
diff --git
a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/TestContainerOperations.java
b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/TestContainerOperations.java
index b98aee7286..fcea426ad4 100644
---
a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/TestContainerOperations.java
+++
b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/TestContainerOperations.java
@@ -96,7 +96,7 @@ void testContainerStateMachineIdempotency() throws Exception {
// call create Container again
BlockID blockID = ContainerTestHelper.getTestBlockID(containerID);
byte[] data =
- RandomStringUtils.random(RandomUtils.nextInt(0, 1024)).getBytes(UTF_8);
+ RandomStringUtils.random(RandomUtils.secure().randomInt(0,
1024)).getBytes(UTF_8);
ContainerProtos.ContainerCommandRequestProto writeChunkRequest =
ContainerTestHelper
.getWriteChunkRequest(container.getPipeline(), blockID,
diff --git
a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/TestGetCommittedBlockLengthAndPutKey.java
b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/TestGetCommittedBlockLengthAndPutKey.java
index d1b52da2d5..6168f4c3ee 100644
---
a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/TestGetCommittedBlockLengthAndPutKey.java
+++
b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/TestGetCommittedBlockLengthAndPutKey.java
@@ -90,7 +90,7 @@ public void tesGetCommittedBlockLength() throws Exception {
BlockID blockID = ContainerTestHelper.getTestBlockID(containerID);
byte[] data =
- RandomStringUtils.random(RandomUtils.nextInt(1, 1024)).getBytes(UTF_8);
+ RandomStringUtils.random(RandomUtils.secure().randomInt(1,
1024)).getBytes(UTF_8);
ContainerProtos.ContainerCommandRequestProto writeChunkRequest =
ContainerTestHelper
.getWriteChunkRequest(container.getPipeline(), blockID,
@@ -154,7 +154,7 @@ public void tesPutKeyResposne() throws Exception {
BlockID blockID = ContainerTestHelper.getTestBlockID(containerID);
byte[] data =
- RandomStringUtils.random(RandomUtils.nextInt(1, 1024)).getBytes(UTF_8);
+ RandomStringUtils.random(RandomUtils.secure().randomInt(1,
1024)).getBytes(UTF_8);
ContainerProtos.ContainerCommandRequestProto writeChunkRequest =
ContainerTestHelper
.getWriteChunkRequest(container.getPipeline(), blockID,
diff --git
a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/TestStorageContainerManager.java
b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/TestStorageContainerManager.java
index c70173d19e..4df849e6e0 100644
---
a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/TestStorageContainerManager.java
+++
b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/TestStorageContainerManager.java
@@ -288,8 +288,8 @@ private void testBlockDeletionTransactions(MiniOzoneCluster
cluster) throws Exce
// Add 2 TXs per container.
Map<Long, List<Long>> deletedBlocks = new HashMap<>();
List<Long> blocks = new ArrayList<>();
- blocks.add(RandomUtils.nextLong());
- blocks.add(RandomUtils.nextLong());
+ blocks.add(RandomUtils.secure().randomLong());
+ blocks.add(RandomUtils.secure().randomLong());
deletedBlocks.put(containerID, blocks);
addTransactions(cluster.getStorageContainerManager(), delLog,
deletedBlocks);
diff --git
a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/TestWatchForCommit.java
b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/TestWatchForCommit.java
index 11b67acf31..f52083fa21 100644
---
a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/TestWatchForCommit.java
+++
b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/TestWatchForCommit.java
@@ -282,7 +282,8 @@ public void
testWatchForCommitForRetryfailure(RaftProtos.ReplicationLevel watchT
// as well as there is no logIndex generate in Ratis.
// The basic idea here is just to test if its throws an exception.
ExecutionException e = assertThrows(ExecutionException.class,
- () -> xceiverClient.watchForCommit(index + RandomUtils.nextInt(0,
100) + 10).get());
+ () -> xceiverClient.watchForCommit(index +
RandomUtils.secure().randomInt(0, 100) + 10)
+ .get());
// since the timeout value is quite long, the watch request will either
// fail with NotReplicated exceptio, RetryFailureException or
// RuntimeException
@@ -382,7 +383,8 @@ public void testWatchForCommitForGroupMismatchException()
throws Exception {
// as well as there is no logIndex generate in Ratis.
// The basic idea here is just to test if its throws an exception.
final Exception e = assertThrows(Exception.class,
- () -> xceiverClient.watchForCommit(reply.getLogIndex() +
RandomUtils.nextInt(0, 100) + 10).get());
+ () -> xceiverClient.watchForCommit(reply.getLogIndex() +
RandomUtils.secure().randomInt(0, 100) + 10)
+ .get());
assertInstanceOf(GroupMismatchException.class,
HddsClientUtils.checkForException(e));
} finally {
clientManager.releaseClient(xceiverClient, false);
diff --git
a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/TestXceiverClientGrpc.java
b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/TestXceiverClientGrpc.java
index 092d428a77..d6b69e3a03 100644
---
a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/TestXceiverClientGrpc.java
+++
b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/TestXceiverClientGrpc.java
@@ -189,9 +189,11 @@ public void testPrimaryReadFromNormalDatanode()
node -> assertEquals(NodeOperationalState.IN_SERVICE,
node.getPersistedOpState()));
randomPipeline.getNodes().get(
- RandomUtils.nextInt(0,
nodeCount)).setPersistedOpState(NodeOperationalState.IN_MAINTENANCE);
+ RandomUtils.secure().randomInt(0, nodeCount)).
+ setPersistedOpState(NodeOperationalState.IN_MAINTENANCE);
randomPipeline.getNodes().get(
- RandomUtils.nextInt(0,
nodeCount)).setPersistedOpState(NodeOperationalState.IN_MAINTENANCE);
+ RandomUtils.secure().randomInt(0, nodeCount)).
+ setPersistedOpState(NodeOperationalState.IN_MAINTENANCE);
try (XceiverClientGrpc client = new XceiverClientGrpc(randomPipeline,
conf) {
@Override
public XceiverClientReply sendCommandAsync(
diff --git
a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/container/TestContainerStateManagerIntegration.java
b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/container/TestContainerStateManagerIntegration.java
index f050edf935..629ae05270 100644
---
a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/container/TestContainerStateManagerIntegration.java
+++
b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/container/TestContainerStateManagerIntegration.java
@@ -326,7 +326,7 @@ public void testReplicaMap() throws Exception {
.setUuid(UUID.randomUUID()).build();
// Test 1: no replica's exist
- ContainerID containerID = ContainerID.valueOf(RandomUtils.nextLong());
+ ContainerID containerID =
ContainerID.valueOf(RandomUtils.secure().randomLong());
Set<ContainerReplica> replicaSet =
containerStateManager.getContainerReplicas(containerID);
assertNull(replicaSet);
diff --git
a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/container/metrics/TestSCMContainerManagerMetrics.java
b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/container/metrics/TestSCMContainerManagerMetrics.java
index 1acb081c10..300ea81e52 100644
---
a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/container/metrics/TestSCMContainerManagerMetrics.java
+++
b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/container/metrics/TestSCMContainerManagerMetrics.java
@@ -111,7 +111,7 @@ public void testContainerOpsMetrics() throws Exception {
assertThrows(ContainerNotFoundException.class, () ->
containerManager.deleteContainer(
- ContainerID.valueOf(RandomUtils.nextLong(10000, 20000))));
+ ContainerID.valueOf(RandomUtils.secure().randomLong(10000,
20000))));
// deleteContainer should fail, so it should have the old metric value.
metrics = getMetrics(SCMContainerManagerMetrics.class.getSimpleName());
assertEquals(getLongCounter("NumSuccessfulDeleteContainers",
diff --git
a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/pipeline/TestLeaderChoosePolicy.java
b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/pipeline/TestLeaderChoosePolicy.java
index 58863bb83b..3c19a039e1 100644
---
a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/pipeline/TestLeaderChoosePolicy.java
+++
b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/pipeline/TestLeaderChoosePolicy.java
@@ -182,7 +182,7 @@ public void testMinLeaderCountChoosePolicy() throws
Exception {
.getPipelines(RatisReplicationConfig.getInstance(
ReplicationFactor.THREE), Pipeline.PipelineState.OPEN);
- int destroyNum = RandomUtils.nextInt(0, pipelines.size());
+ int destroyNum = RandomUtils.secure().randomInt(0, pipelines.size());
for (int k = 0; k <= destroyNum; k++) {
pipelineManager.closePipeline(pipelines.get(k), false);
}
diff --git
a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/storage/TestContainerCommandsEC.java
b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/storage/TestContainerCommandsEC.java
index d79c312662..7606c38c56 100644
---
a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/storage/TestContainerCommandsEC.java
+++
b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/scm/storage/TestContainerCommandsEC.java
@@ -255,7 +255,7 @@ public void testOrphanBlock() throws Exception {
String keyName = UUID.randomUUID().toString();
try (OutputStream out = classBucket
.createKey(keyName, keyLen, repConfig, new HashMap<>())) {
- out.write(RandomUtils.nextBytes(keyLen));
+ out.write(RandomUtils.secure().randomBytes(keyLen));
}
long orphanContainerID = classBucket.getKey(keyName)
.getOzoneKeyLocations().get(0).getContainerID();
@@ -1023,8 +1023,8 @@ public static void prepareData(int[][] ranges) throws
Exception {
new ECReplicationConfig(EC_DATA, EC_PARITY, EC_CODEC, EC_CHUNK_SIZE);
values = new byte[ranges.length][];
for (int i = 0; i < ranges.length; i++) {
- int keySize = RandomUtils.nextInt(ranges[i][0], ranges[i][1]);
- values[i] = RandomUtils.nextBytes(keySize);
+ int keySize = RandomUtils.secure().randomInt(ranges[i][0], ranges[i][1]);
+ values[i] = RandomUtils.secure().randomBytes(keySize);
final String keyName = UUID.randomUUID().toString();
try (OutputStream out = classBucket
.createKey(keyName, values[i].length, repConfig, new HashMap<>())) {
diff --git
a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/utils/db/managed/TestRocksObjectLeakDetector.java
b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/utils/db/managed/TestRocksObjectLeakDetector.java
index b21dc8af69..167c8383f8 100644
---
a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/utils/db/managed/TestRocksObjectLeakDetector.java
+++
b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/hdds/utils/db/managed/TestRocksObjectLeakDetector.java
@@ -70,7 +70,7 @@ public void testLeakDetector() throws Exception {
testLeakDetector(() -> new ManagedLRUCache(1L));
testLeakDetector(ManagedOptions::new);
testLeakDetector(ManagedReadOptions::new);
- testLeakDetector(() -> new ManagedSlice(RandomUtils.nextBytes(10)));
+ testLeakDetector(() -> new
ManagedSlice(RandomUtils.secure().randomBytes(10)));
testLeakDetector(ManagedStatistics::new);
testLeakDetector(ManagedWriteBatch::new);
testLeakDetector(ManagedWriteOptions::new);
diff --git
a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/TestBlockTokens.java
b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/TestBlockTokens.java
index 752d2894da..62225911d6 100644
---
a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/TestBlockTokens.java
+++
b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/TestBlockTokens.java
@@ -250,7 +250,7 @@ public void blockTokenFailsOnWrongPassword() throws
Exception {
for (OmKeyLocationInfoGroup v : keyInfo.getKeyLocationVersions()) {
for (OmKeyLocationInfo l : v.getLocationList()) {
Token<OzoneBlockTokenIdentifier> token = l.getToken();
- byte[] randomPassword = RandomUtils.nextBytes(100);
+ byte[] randomPassword = RandomUtils.secure().randomBytes(100);
Token<OzoneBlockTokenIdentifier> override = new Token<>(
token.getIdentifier(), randomPassword,
token.getKind(), token.getService());
diff --git
a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/TestMultipartObjectGet.java
b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/TestMultipartObjectGet.java
index 9b1fcf0015..c9af9a4ffd 100644
---
a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/TestMultipartObjectGet.java
+++
b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/TestMultipartObjectGet.java
@@ -218,7 +218,7 @@ public void testMultipart() throws Exception {
private static String generateRandomContent(int sizeInMB) {
int bytesToGenerate = sizeInMB * 1024 * 1024;
- byte[] randomBytes = RandomUtils.nextBytes(bytesToGenerate);
+ byte[] randomBytes = RandomUtils.secure().randomBytes(bytesToGenerate);
return Base64.getEncoder().encodeToString(randomBytes);
}
}
diff --git
a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/OzoneRpcClientTests.java
b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/OzoneRpcClientTests.java
index 292723bfc6..6cc155b45a 100644
---
a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/OzoneRpcClientTests.java
+++
b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/OzoneRpcClientTests.java
@@ -1946,7 +1946,7 @@ public void testUsedBytesWithUploadPart() throws
IOException {
int blockSize = (int) ozoneManager.getConfiguration().getStorageSize(
OZONE_SCM_BLOCK_SIZE, OZONE_SCM_BLOCK_SIZE_DEFAULT, StorageUnit.BYTES);
String sampleData = Arrays.toString(generateData(blockSize + 100,
- (byte) RandomUtils.nextLong()));
+ (byte) RandomUtils.secure().randomLong()));
int valueLength = sampleData.getBytes(UTF_8).length;
store.createVolume(volumeName);
@@ -1983,7 +1983,7 @@ public void testValidateBlockLengthWithCommitKey() throws
IOException {
String volumeName = UUID.randomUUID().toString();
String bucketName = UUID.randomUUID().toString();
- String value = RandomStringUtils.random(RandomUtils.nextInt(1, 1024));
+ String value = RandomStringUtils.random(RandomUtils.secure().randomInt(1,
1024));
store.createVolume(volumeName);
OzoneVolume volume = store.getVolume(volumeName);
volume.createBucket(bucketName);
@@ -2097,7 +2097,7 @@ public void testPutKeyRatisThreeNodesParallel() throws
IOException,
for (int i = 0; i < 5; i++) {
String keyName = UUID.randomUUID().toString();
String data = Arrays.toString(generateData(5 * 1024 * 1024,
- (byte) RandomUtils.nextLong()));
+ (byte) RandomUtils.secure().randomLong()));
TestDataUtil.createKey(bucket, keyName,
ReplicationConfig.fromTypeAndFactor(RATIS, THREE),
data.getBytes(UTF_8));
@@ -3645,7 +3645,7 @@ void testCommitPartAfterCompleteUpload() throws Exception
{
// upload part 1.
byte[] data = generateData(5 * 1024 * 1024,
- (byte) RandomUtils.nextLong());
+ (byte) RandomUtils.secure().randomLong());
OzoneOutputStream ozoneOutputStream = bucket.createMultipartKey(keyName,
data.length, 1, uploadID);
ozoneOutputStream.write(data, 0, data.length);
diff --git
a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/TestBlockOutputStream.java
b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/TestBlockOutputStream.java
index a67dca9e8e..12eede0eb0 100644
---
a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/TestBlockOutputStream.java
+++
b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/TestBlockOutputStream.java
@@ -200,7 +200,7 @@ void testWriteLessThanChunkSize(boolean flushDelay, boolean
enablePiggybacking)
OzoneOutputStream key = createKey(client, keyName);
int dataLength = 50;
final int totalWriteLength = dataLength * 2;
- byte[] data1 = RandomUtils.nextBytes(dataLength);
+ byte[] data1 = RandomUtils.secure().randomBytes(dataLength);
key.write(data1);
KeyOutputStream keyOutputStream =
assertInstanceOf(KeyOutputStream.class, key.getOutputStream());
@@ -298,7 +298,7 @@ void testWriteExactlyFlushSize(boolean flushDelay, boolean
enablePiggybacking) t
OzoneOutputStream key = createKey(client, keyName);
// write data equal to 2 chunks
int dataLength = FLUSH_SIZE;
- byte[] data1 = RandomUtils.nextBytes(dataLength);
+ byte[] data1 = RandomUtils.secure().randomBytes(dataLength);
key.write(data1);
assertEquals(writeChunkCount + 2,
@@ -416,7 +416,7 @@ void testWriteMoreThanChunkSize(boolean flushDelay, boolean
enablePiggybacking)
OzoneOutputStream key = createKey(client, keyName);
// write data more than 1 chunk
int dataLength = CHUNK_SIZE + 50;
- byte[] data1 = RandomUtils.nextBytes(dataLength);
+ byte[] data1 = RandomUtils.secure().randomBytes(dataLength);
key.write(data1);
assertEquals(totalOpCount + 1, metrics.getTotalOpCount());
KeyOutputStream keyOutputStream =
@@ -505,7 +505,7 @@ void testWriteMoreThanFlushSize(boolean flushDelay, boolean
enablePiggybacking)
String keyName = getKeyName();
OzoneOutputStream key = createKey(client, keyName);
int dataLength = FLUSH_SIZE + 50;
- byte[] data1 = RandomUtils.nextBytes(dataLength);
+ byte[] data1 = RandomUtils.secure().randomBytes(dataLength);
key.write(data1);
assertEquals(totalOpCount + 3, metrics.getTotalOpCount());
@@ -594,7 +594,7 @@ void testWriteExactlyMaxFlushSize(boolean flushDelay,
boolean enablePiggybacking
String keyName = getKeyName();
OzoneOutputStream key = createKey(client, keyName);
int dataLength = MAX_FLUSH_SIZE;
- byte[] data1 = RandomUtils.nextBytes(dataLength);
+ byte[] data1 = RandomUtils.secure().randomBytes(dataLength);
key.write(data1);
KeyOutputStream keyOutputStream =
@@ -689,7 +689,7 @@ void testWriteMoreThanMaxFlushSize(boolean flushDelay,
boolean enablePiggybackin
OzoneOutputStream key = createKey(client, keyName);
int dataLength = MAX_FLUSH_SIZE + 50;
// write data more than 1 chunk
- byte[] data1 = RandomUtils.nextBytes(dataLength);
+ byte[] data1 = RandomUtils.secure().randomBytes(dataLength);
key.write(data1);
KeyOutputStream keyOutputStream =
assertInstanceOf(KeyOutputStream.class, key.getOutputStream());
diff --git
a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/TestBlockOutputStreamWithFailures.java
b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/TestBlockOutputStreamWithFailures.java
index 5450b830ef..a4cb32edd1 100644
---
a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/TestBlockOutputStreamWithFailures.java
+++
b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/TestBlockOutputStreamWithFailures.java
@@ -111,7 +111,7 @@ private void
testWatchForCommitWithCloseContainerException(OzoneClient client)
String keyName = getKeyName();
OzoneOutputStream key = createKey(client, keyName);
int dataLength = MAX_FLUSH_SIZE + CHUNK_SIZE;
- byte[] data1 = RandomUtils.nextBytes(dataLength);
+ byte[] data1 = RandomUtils.secure().randomBytes(dataLength);
key.write(data1);
KeyOutputStream keyOutputStream =
@@ -196,7 +196,7 @@ void testWatchForCommitDatanodeFailure(boolean flushDelay,
boolean enablePiggyba
String keyName = getKeyName();
OzoneOutputStream key = createKey(client, keyName);
int dataLength = MAX_FLUSH_SIZE + CHUNK_SIZE;
- byte[] data1 = RandomUtils.nextBytes(dataLength);
+ byte[] data1 = RandomUtils.secure().randomBytes(dataLength);
key.write(data1);
// since its hitting the full bufferCondition, it will call
watchForCommit
// and completes at least putBlock for first flushSize worth of data
@@ -280,7 +280,7 @@ void test2DatanodesFailure(boolean flushDelay, boolean
enablePiggybacking) throw
String keyName = getKeyName();
OzoneOutputStream key = createKey(client, keyName);
int dataLength = MAX_FLUSH_SIZE + CHUNK_SIZE;
- byte[] data1 = RandomUtils.nextBytes(dataLength);
+ byte[] data1 = RandomUtils.secure().randomBytes(dataLength);
key.write(data1);
// since its hitting the full bufferCondition, it will call
watchForCommit
// and completes atleast putBlock for first flushSize worth of data
@@ -376,7 +376,7 @@ private void testWriteMoreThanMaxFlushSize(OzoneClient
client)
String keyName = getKeyName();
OzoneOutputStream key = createKey(client, keyName);
int dataLength = MAX_FLUSH_SIZE + CHUNK_SIZE;
- byte[] data1 = RandomUtils.nextBytes(dataLength);
+ byte[] data1 = RandomUtils.secure().randomBytes(dataLength);
key.write(data1);
KeyOutputStream keyOutputStream =
@@ -433,7 +433,7 @@ private void testExceptionDuringClose(OzoneClient client)
throws Exception {
String keyName = getKeyName();
OzoneOutputStream key = createKey(client, keyName);
int dataLength = 167;
- byte[] data1 = RandomUtils.nextBytes(dataLength);
+ byte[] data1 = RandomUtils.secure().randomBytes(dataLength);
key.write(data1);
KeyOutputStream keyOutputStream =
@@ -498,7 +498,7 @@ private void
testWatchForCommitWithSingleNodeRatis(OzoneClient client)
OzoneOutputStream key =
createKey(client, keyName, 0, ReplicationFactor.ONE);
int dataLength = MAX_FLUSH_SIZE + CHUNK_SIZE;
- byte[] data1 = RandomUtils.nextBytes(dataLength);
+ byte[] data1 = RandomUtils.secure().randomBytes(dataLength);
key.write(data1);
KeyOutputStream keyOutputStream =
@@ -587,7 +587,7 @@ void testDatanodeFailureWithSingleNode(boolean flushDelay,
boolean enablePiggyba
OzoneOutputStream key =
createKey(client, keyName, 0, ReplicationFactor.ONE);
int dataLength = MAX_FLUSH_SIZE + CHUNK_SIZE;
- byte[] data1 = RandomUtils.nextBytes(dataLength);
+ byte[] data1 = RandomUtils.secure().randomBytes(dataLength);
key.write(data1);
// since its hitting the full bufferCondition, it will call
watchForCommit
// and completes at least putBlock for first flushSize worth of data
@@ -678,7 +678,7 @@ void testDatanodeFailureWithPreAllocation(boolean
flushDelay, boolean enablePigg
createKey(client, keyName, 3 * BLOCK_SIZE,
ReplicationFactor.ONE);
int dataLength = MAX_FLUSH_SIZE + CHUNK_SIZE;
- byte[] data1 = RandomUtils.nextBytes(dataLength);
+ byte[] data1 = RandomUtils.secure().randomBytes(dataLength);
key.write(data1);
// since its hitting the full bufferCondition, it will call
watchForCommit
// and completes at least putBlock for first flushSize worth of data
diff --git
a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/TestOzoneClientMultipartUploadWithFSO.java
b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/TestOzoneClientMultipartUploadWithFSO.java
index 8e6a74f838..f3cbbcaa77 100644
---
a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/TestOzoneClientMultipartUploadWithFSO.java
+++
b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/TestOzoneClientMultipartUploadWithFSO.java
@@ -444,7 +444,7 @@ public void testCommitPartAfterCompleteUpload() throws
Exception {
// upload part 1.
byte[] data = generateData(5 * 1024 * 1024,
- (byte) RandomUtils.nextLong());
+ (byte) RandomUtils.secure().randomLong());
OzoneOutputStream ozoneOutputStream = bucket.createMultipartKey(keyName,
data.length, 1, uploadID);
ozoneOutputStream.write(data, 0, data.length);
diff --git
a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/TestOzoneRpcClientWithKeyLatestVersion.java
b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/TestOzoneRpcClientWithKeyLatestVersion.java
index fa6bf2274c..b437aa7202 100644
---
a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/TestOzoneRpcClientWithKeyLatestVersion.java
+++
b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/TestOzoneRpcClientWithKeyLatestVersion.java
@@ -76,8 +76,8 @@ void testWithGetLatestVersion(boolean getLatestVersionOnly)
throws Exception {
OzoneBucket bucket = volume.getBucket(bucketName);
String keyName = UUID.randomUUID().toString();
- byte[] content = RandomUtils.nextBytes(128);
- int versions = RandomUtils.nextInt(2, 5);
+ byte[] content = RandomUtils.secure().randomBytes(128);
+ int versions = RandomUtils.secure().randomInt(2, 5);
createAndOverwriteKey(bucket, keyName, versions, content);
@@ -96,7 +96,7 @@ private void createAndOverwriteKey(OzoneBucket bucket, String
key,
int versions, byte[] content) throws IOException {
ReplicationConfig replication = RatisReplicationConfig.getInstance(THREE);
for (int i = 1; i < versions; i++) {
- writeKey(bucket, key, RandomUtils.nextBytes(content.length),
replication);
+ writeKey(bucket, key, RandomUtils.secure().randomBytes(content.length),
replication);
}
// overwrite it
writeKey(bucket, key, content, replication);
diff --git
a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/TestReadRetries.java
b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/TestReadRetries.java
index 06cd29529e..543acb089b 100644
---
a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/TestReadRetries.java
+++
b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/TestReadRetries.java
@@ -79,7 +79,7 @@ void testPutKeyAndGetKeyThreeNodes() throws Exception {
OzoneBucket bucket = volume.getBucket(bucketName);
String keyName = "a/b/c/" + UUID.randomUUID();
- byte[] content = RandomUtils.nextBytes(128);
+ byte[] content = RandomUtils.secure().randomBytes(128);
TestDataUtil.createKey(bucket, keyName,
RatisReplicationConfig.getInstance(THREE), content);
diff --git
a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/container/server/TestSecureContainerServer.java
b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/container/server/TestSecureContainerServer.java
index be048e65b3..ef88ee1c4b 100644
---
a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/container/server/TestSecureContainerServer.java
+++
b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/container/server/TestSecureContainerServer.java
@@ -263,7 +263,7 @@ private static void runTestClientServer(
Token<OzoneBlockTokenIdentifier> token =
blockTokenSecretManager.generateToken(blockID,
- EnumSet.allOf(AccessModeProto.class), RandomUtils.nextLong());
+ EnumSet.allOf(AccessModeProto.class),
RandomUtils.secure().randomLong());
String encodedToken = token.encodeToUrlString();
ContainerCommandRequestProto.Builder writeChunk =
diff --git
a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/dn/volume/TestDatanodeHddsVolumeFailureDetection.java
b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/dn/volume/TestDatanodeHddsVolumeFailureDetection.java
index 7d3eaffaec..5e22465825 100644
---
a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/dn/volume/TestDatanodeHddsVolumeFailureDetection.java
+++
b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/dn/volume/TestDatanodeHddsVolumeFailureDetection.java
@@ -277,7 +277,7 @@ private static MiniOzoneCluster newCluster(boolean schemaV3)
private static long createKey(OzoneBucket bucket, String key)
throws IOException {
- byte[] bytes = RandomUtils.nextBytes(KEY_SIZE);
+ byte[] bytes = RandomUtils.secure().randomBytes(KEY_SIZE);
RatisReplicationConfig replication =
RatisReplicationConfig.getInstance(ReplicationFactor.ONE);
TestDataUtil.createKey(bucket, key, replication, bytes);
diff --git
a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/TestObjectStoreWithLegacyFS.java
b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/TestObjectStoreWithLegacyFS.java
index 490fb2f42c..cf28bbab41 100644
---
a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/TestObjectStoreWithLegacyFS.java
+++
b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/om/TestObjectStoreWithLegacyFS.java
@@ -215,7 +215,7 @@ private OmMultipartUploadCompleteInfo
uploadMPUWithDirectoryExists(
String uploadID = omMultipartInfo.getUploadID();
// upload part 1.
- byte[] data = generateData(128, (byte) RandomUtils.nextLong());
+ byte[] data = generateData(128, (byte) RandomUtils.secure().randomLong());
OzoneOutputStream ozoneOutputStream = bucket.createMultipartKey(keyName,
data.length, 1, uploadID);
ozoneOutputStream.write(data, 0, data.length);
diff --git
a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/shell/TestDeletedBlocksTxnShell.java
b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/shell/TestDeletedBlocksTxnShell.java
index 54e9fdeff0..de1a0d360d 100644
---
a/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/shell/TestDeletedBlocksTxnShell.java
+++
b/hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/shell/TestDeletedBlocksTxnShell.java
@@ -120,8 +120,8 @@ public void shutdown() {
//<containerID, List<blockID>>
private Map<Long, List<Long>> generateData(int dataSize) throws Exception {
Map<Long, List<Long>> blockMap = new HashMap<>();
- int continerIDBase = RandomUtils.nextInt(0, 100);
- int localIDBase = RandomUtils.nextInt(0, 1000);
+ int continerIDBase = RandomUtils.secure().randomInt(0, 100);
+ int localIDBase = RandomUtils.secure().randomInt(0, 1000);
for (int i = 0; i < dataSize; i++) {
long containerID = continerIDBase + i;
updateContainerMetadata(containerID);
diff --git
a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/TestKeyManagerUnit.java
b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/TestKeyManagerUnit.java
index b4c8861476..6b0b710a83 100644
---
a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/TestKeyManagerUnit.java
+++
b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/TestKeyManagerUnit.java
@@ -472,7 +472,7 @@ private OmMultipartInfo addinitMultipartUploadToCache(
metadataManager.getMultipartInfoTable().addCacheEntry(
new CacheKey<>(metadataManager.getMultipartKey(volume, bucket, key,
uploadID)),
- CacheValue.get(RandomUtils.nextInt(), multipartKeyInfo));
+ CacheValue.get(RandomUtils.secure().randomInt(), multipartKeyInfo));
return new OmMultipartInfo(volume, bucket, key, uploadID);
}
@@ -480,7 +480,7 @@ private void abortMultipart(
String volume, String bucket, String key, String uploadID) {
metadataManager.getMultipartInfoTable().addCacheEntry(
new CacheKey<>(metadataManager.getMultipartKey(volume, bucket, key,
- uploadID)), CacheValue.get(RandomUtils.nextInt()));
+ uploadID)), CacheValue.get(RandomUtils.secure().randomInt()));
}
@Test
diff --git
a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/service/TestMultipartUploadCleanupService.java
b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/service/TestMultipartUploadCleanupService.java
index 6004fea29e..aaf3c267d9 100644
---
a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/service/TestMultipartUploadCleanupService.java
+++
b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/service/TestMultipartUploadCleanupService.java
@@ -161,16 +161,16 @@ private void createIncompleteMPUKeys(int mpuKeyCount,
String volume = UUID.randomUUID().toString();
String bucket = UUID.randomUUID().toString();
for (int x = 0; x < mpuKeyCount; x++) {
- if (RandomUtils.nextBoolean()) {
+ if (RandomUtils.secure().randomBoolean()) {
bucket = UUID.randomUUID().toString();
- if (RandomUtils.nextBoolean()) {
+ if (RandomUtils.secure().randomBoolean()) {
volume = UUID.randomUUID().toString();
}
}
String key = UUID.randomUUID().toString();
createVolumeAndBucket(volume, bucket, bucketLayout);
- final int numParts = RandomUtils.nextInt(0, 5);
+ final int numParts = RandomUtils.secure().randomInt(0, 5);
// Create the MPU key
createIncompleteMPUKey(volume, bucket, key, numParts);
}
diff --git
a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/service/TestOpenKeyCleanupService.java
b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/service/TestOpenKeyCleanupService.java
index e2da9e8216..97fa7db3c0 100644
---
a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/service/TestOpenKeyCleanupService.java
+++
b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/om/service/TestOpenKeyCleanupService.java
@@ -469,16 +469,16 @@ private void createOpenKeys(int keyCount, boolean hsync,
String volume = UUID.randomUUID().toString();
String bucket = UUID.randomUUID().toString();
for (int x = 0; x < keyCount; x++) {
- if (RandomUtils.nextBoolean()) {
+ if (RandomUtils.secure().randomBoolean()) {
bucket = UUID.randomUUID().toString();
- if (RandomUtils.nextBoolean()) {
+ if (RandomUtils.secure().randomBoolean()) {
volume = UUID.randomUUID().toString();
}
}
String key = withDir ? "dir1/dir2/" + UUID.randomUUID() :
UUID.randomUUID().toString();
createVolumeAndBucket(volume, bucket, bucketLayout);
- final int numBlocks = RandomUtils.nextInt(1, 3);
+ final int numBlocks = RandomUtils.secure().randomInt(1, 3);
// Create the key
createOpenKey(volume, bucket, key, numBlocks, hsync, recovery);
}
@@ -538,9 +538,9 @@ private void createIncompleteMPUKeys(int mpuKeyCount,
String volume = UUID.randomUUID().toString();
String bucket = UUID.randomUUID().toString();
for (int x = 0; x < mpuKeyCount; x++) {
- if (RandomUtils.nextBoolean()) {
+ if (RandomUtils.secure().randomBoolean()) {
bucket = UUID.randomUUID().toString();
- if (RandomUtils.nextBoolean()) {
+ if (RandomUtils.secure().randomBoolean()) {
volume = UUID.randomUUID().toString();
}
}
diff --git
a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/security/TestOzoneTokenIdentifier.java
b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/security/TestOzoneTokenIdentifier.java
index 85b1d24bd1..dade773b80 100644
---
a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/security/TestOzoneTokenIdentifier.java
+++
b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/security/TestOzoneTokenIdentifier.java
@@ -105,7 +105,7 @@ public void testSignToken(@TempDir Path baseDir) throws
GeneralSecurityException
new Text("rm"), new Text("client"));
tokenId.setOmCertSerialId("123");
LOG.info("Unsigned token {} is {}", tokenId,
- verifyTokenAsymmetric(tokenId, RandomUtils.nextBytes(128), cert));
+ verifyTokenAsymmetric(tokenId, RandomUtils.secure().randomBytes(128),
cert));
}
diff --git
a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/security/acl/TestOzoneNativeAuthorizer.java
b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/security/acl/TestOzoneNativeAuthorizer.java
index 5f9a1ad44d..48b978064e 100644
---
a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/security/acl/TestOzoneNativeAuthorizer.java
+++
b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/security/acl/TestOzoneNativeAuthorizer.java
@@ -122,7 +122,7 @@ public static Collection<Object[]> data() {
public void createAll(
String keyName, String prefixName, ACLType userRight,
ACLType groupRight, boolean expectedResult) throws IOException {
- int randomInt = RandomUtils.nextInt();
+ int randomInt = RandomUtils.secure().randomInt();
this.vol = "vol" + randomInt;
this.buck = "bucket" + randomInt;
this.key = keyName + randomInt;
@@ -427,7 +427,7 @@ private void resetAclsAndValidateAccess(
+ " name:" + (accessType == USER ? user : group));
// Randomize next type.
- int type = RandomUtils.nextInt(0, 3);
+ int type = RandomUtils.secure().randomInt(0, 3);
ACLIdentityType identityType = ACLIdentityType.values()[type];
// Add remaining acls one by one and then check access.
OzoneAcl addAcl = OzoneAcl.of(identityType,
diff --git
a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/security/acl/TestParentAcl.java
b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/security/acl/TestParentAcl.java
index df53684561..92bcc7b7a2 100644
---
a/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/security/acl/TestParentAcl.java
+++
b/hadoop-ozone/ozone-manager/src/test/java/org/apache/hadoop/ozone/security/acl/TestParentAcl.java
@@ -119,7 +119,7 @@ static void setup() throws Exception {
public void testKeyAcl()
throws IOException {
OzoneObj keyObj;
- int randomInt = RandomUtils.nextInt();
+ int randomInt = RandomUtils.secure().randomInt();
String vol = "vol" + randomInt;
String buck = "bucket" + randomInt;
String key = "key" + randomInt;
@@ -165,7 +165,7 @@ public void testKeyAcl()
public void testBucketAcl()
throws IOException {
OzoneObj bucketObj;
- int randomInt = RandomUtils.nextInt();
+ int randomInt = RandomUtils.secure().randomInt();
String vol = "vol" + randomInt;
String buck = "bucket" + randomInt;
diff --git
a/hadoop-ozone/ozonefs-common/src/test/java/org/apache/hadoop/fs/ozone/TestOzoneFSInputStream.java
b/hadoop-ozone/ozonefs-common/src/test/java/org/apache/hadoop/fs/ozone/TestOzoneFSInputStream.java
index beca218a52..86df63949b 100644
---
a/hadoop-ozone/ozonefs-common/src/test/java/org/apache/hadoop/fs/ozone/TestOzoneFSInputStream.java
+++
b/hadoop-ozone/ozonefs-common/src/test/java/org/apache/hadoop/fs/ozone/TestOzoneFSInputStream.java
@@ -78,7 +78,7 @@ private static void testReadToByteBuffer(
IntFunction<ByteBuffer> bufferConstructor,
int streamLength, int bufferCapacity,
int bufferPosition) throws IOException {
- final byte[] source = RandomUtils.nextBytes(streamLength);
+ final byte[] source = RandomUtils.secure().randomBytes(streamLength);
final InputStream input = new ByteArrayInputStream(source);
final OzoneFSInputStream subject = createTestSubject(input);
diff --git
a/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/TestReconUtils.java
b/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/TestReconUtils.java
index 07fea95819..5cfd5f26fd 100644
---
a/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/TestReconUtils.java
+++
b/hadoop-ozone/recon/src/test/java/org/apache/hadoop/ozone/recon/TestReconUtils.java
@@ -171,7 +171,7 @@ public void testNextClosestPowerIndexOfTwo() {
}
for (int i = 0; i < 10; i++) {
- assertNextClosestPowerIndexOfTwo(RandomUtils.nextLong());
+ assertNextClosestPowerIndexOfTwo(RandomUtils.secure().randomLong());
}
}
@@ -202,7 +202,7 @@ private static int oldNextClosestPowerIndexOfTwo(long
dataSize) {
private static ContainerInfo.Builder getDefaultContainerInfoBuilder(
final HddsProtos.LifeCycleState state) {
return new ContainerInfo.Builder()
- .setContainerID(RandomUtils.nextLong())
+ .setContainerID(RandomUtils.secure().randomLong())
.setReplicationConfig(
RatisReplicationConfig
.getInstance(HddsProtos.ReplicationFactor.THREE))
diff --git
a/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/freon/DatanodeSimulationState.java
b/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/freon/DatanodeSimulationState.java
index 9dde33c5e6..1e3bae684d 100644
---
a/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/freon/DatanodeSimulationState.java
+++
b/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/freon/DatanodeSimulationState.java
@@ -152,7 +152,7 @@ private void addContainerReport(InetSocketAddress endpoint,
// to avoid peaks.
if (state.nextFullContainerReport == Instant.MIN) {
state.nextFullContainerReport = Instant.now().plusMillis(
- RandomUtils.nextLong(1, fullContainerReportDurationMs));
+ RandomUtils.secure().randomLong(1, fullContainerReportDurationMs));
} else {
state.nextFullContainerReport = Instant.now()
.plusMillis(fullContainerReportDurationMs);
diff --git
a/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/freon/DatanodeSimulator.java
b/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/freon/DatanodeSimulator.java
index d1749f9a15..b31fde4cbd 100644
---
a/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/freon/DatanodeSimulator.java
+++
b/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/freon/DatanodeSimulator.java
@@ -345,7 +345,7 @@ private boolean startDatanode(DatanodeSimulationState dn)
long scmHeartbeatInterval = HddsServerUtil.getScmHeartbeatInterval(conf);
scmClients.forEach((endpoint, client) -> {
// Use random initial delay as a jitter to avoid peaks.
- long initialDelay = RandomUtils.nextLong(0, scmHeartbeatInterval);
+ long initialDelay = RandomUtils.secure().randomLong(0,
scmHeartbeatInterval);
Runnable runnable = () -> heartbeat(endpoint, client, dn);
heartbeatScheduler.scheduleAtFixedRate(runnable, initialDelay,
scmHeartbeatInterval, TimeUnit.MILLISECONDS);
@@ -353,7 +353,7 @@ private boolean startDatanode(DatanodeSimulationState dn)
long reconHeartbeatInterval =
HddsServerUtil.getReconHeartbeatInterval(conf);
- long initialDelay = RandomUtils.nextLong(0, reconHeartbeatInterval);
+ long initialDelay = RandomUtils.secure().randomLong(0,
reconHeartbeatInterval);
Runnable runnable = () -> heartbeat(reconAddress, reconClient, dn);
heartbeatScheduler.scheduleAtFixedRate(runnable, initialDelay,
reconHeartbeatInterval, TimeUnit.MILLISECONDS);
diff --git
a/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/freon/OzoneClientKeyReadWriteListOps.java
b/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/freon/OzoneClientKeyReadWriteListOps.java
index 602cf1ec07..2051aeec25 100644
---
a/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/freon/OzoneClientKeyReadWriteListOps.java
+++
b/hadoop-ozone/tools/src/main/java/org/apache/hadoop/ozone/freon/OzoneClientKeyReadWriteListOps.java
@@ -169,7 +169,7 @@ public Void call() throws Exception {
timer = getMetrics().timer("key-read-write-list");
if (objectSizeInBytes >= 0) {
- keyContent = RandomUtils.nextBytes(objectSizeInBytes);
+ keyContent = RandomUtils.secure().randomBytes(objectSizeInBytes);
}
if (kg == null) {
kg = new KeyGeneratorUtil();
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]