This is an automated email from the ASF dual-hosted git repository. ishan pushed a commit to branch ishan/cuvs-integration in repository https://gitbox.apache.org/repos/asf/solr.git
commit e53a039d804288fa7c31cb31d848cb28b211409e Author: noblepaul <[email protected]> AuthorDate: Thu Feb 27 14:36:17 2025 +1100 fixed some compilation errors --- .../solr/schema/TestICUCollationFieldUDVAS.java | 2 +- .../apache/solr/hdfs/cloud/MoveReplicaHdfsTest.java | 2 +- .../SharedFileSystemAutoReplicaFailoverTest.java | 2 +- .../apache/solr/hdfs/store/HdfsDirectoryTest.java | 20 ++++++++++---------- .../hdfs/store/blockcache/BlockDirectoryTest.java | 6 +++--- .../security/jwt/JWTAuthPluginIntegrationTest.java | 2 +- 6 files changed, 17 insertions(+), 17 deletions(-) diff --git a/solr/modules/analysis-extras/src/test/org/apache/solr/schema/TestICUCollationFieldUDVAS.java b/solr/modules/analysis-extras/src/test/org/apache/solr/schema/TestICUCollationFieldUDVAS.java index a57a45abee3..acfffe150e0 100644 --- a/solr/modules/analysis-extras/src/test/org/apache/solr/schema/TestICUCollationFieldUDVAS.java +++ b/solr/modules/analysis-extras/src/test/org/apache/solr/schema/TestICUCollationFieldUDVAS.java @@ -32,7 +32,7 @@ public class TestICUCollationFieldUDVAS extends SolrTestCaseJ4 { private static final String ICU_FIELD_UDVAS_PROPNAME = "tests.icu_collation_field.udvas"; private static final String ICU_TYPE_UDVAS_PROPNAME = "tests.icu_collation_fieldType.udvas"; private static final String TEST_LUCENE_MATCH_VERSION_PROPNAME = "tests.luceneMatchVersion"; - private static final Version WARN_CEILING = Version.LUCENE_8_12_0; + private static final Version WARN_CEILING = Version.LUCENE_9_0_0; private static String home; @BeforeClass diff --git a/solr/modules/hdfs/src/test/org/apache/solr/hdfs/cloud/MoveReplicaHdfsTest.java b/solr/modules/hdfs/src/test/org/apache/solr/hdfs/cloud/MoveReplicaHdfsTest.java index 3d504762802..cb4f2ad2bc2 100644 --- a/solr/modules/hdfs/src/test/org/apache/solr/hdfs/cloud/MoveReplicaHdfsTest.java +++ b/solr/modules/hdfs/src/test/org/apache/solr/hdfs/cloud/MoveReplicaHdfsTest.java @@ -82,7 +82,7 @@ public class MoveReplicaHdfsTest extends AbstractMoveReplicaTestBase { // 6-Sep-2018 // Commented 14-Oct-2018 @BadApple(bugUrl="https://issues.apache.org/jira/browse/SOLR-12028") // // added 20-Sep-2018 - @BadApple(bugUrl = "https://issues.apache.org/jira/browse/SOLR-12028") // 14-Oct-2018 + @AwaitsFix(bugUrl = "https://issues.apache.org/jira/browse/SOLR-12028") // 14-Oct-2018 public void testNormalFailedMove() throws Exception { inPlaceMove = false; super.testFailedMove(); diff --git a/solr/modules/hdfs/src/test/org/apache/solr/hdfs/cloud/SharedFileSystemAutoReplicaFailoverTest.java b/solr/modules/hdfs/src/test/org/apache/solr/hdfs/cloud/SharedFileSystemAutoReplicaFailoverTest.java index 68fd8715181..6be45c2d36b 100644 --- a/solr/modules/hdfs/src/test/org/apache/solr/hdfs/cloud/SharedFileSystemAutoReplicaFailoverTest.java +++ b/solr/modules/hdfs/src/test/org/apache/solr/hdfs/cloud/SharedFileSystemAutoReplicaFailoverTest.java @@ -77,7 +77,7 @@ import org.slf4j.LoggerFactory; BadHdfsThreadsFilter.class // hdfs currently leaks thread(s) }) @LogLevel("org.apache.solr.cloud.*=DEBUG") [email protected]( [email protected]( bugUrl = "https://issues.apache.org/jira/browse/SOLR-12028") // added 20-Jul-2018 public class SharedFileSystemAutoReplicaFailoverTest extends AbstractFullDistribZkTestBase { private static final Logger log = LoggerFactory.getLogger(MethodHandles.lookup().lookupClass()); diff --git a/solr/modules/hdfs/src/test/org/apache/solr/hdfs/store/HdfsDirectoryTest.java b/solr/modules/hdfs/src/test/org/apache/solr/hdfs/store/HdfsDirectoryTest.java index 0a0b9aad759..ac36ec5cbcf 100644 --- a/solr/modules/hdfs/src/test/org/apache/solr/hdfs/store/HdfsDirectoryTest.java +++ b/solr/modules/hdfs/src/test/org/apache/solr/hdfs/store/HdfsDirectoryTest.java @@ -112,11 +112,11 @@ public class HdfsDirectoryTest extends SolrTestCaseJ4 { directory.deleteFile(file); } - IndexOutput output = directory.createOutput("testing.test", new IOContext()); + IndexOutput output = directory.createOutput("testing.test", IOContext.DEFAULT); output.writeInt(12345); output.close(); - IndexInput input = directory.openInput("testing.test", new IOContext()); + IndexInput input = directory.openInput("testing.test", IOContext.DEFAULT); assertEquals(12345, input.readInt()); input.close(); @@ -126,7 +126,7 @@ public class HdfsDirectoryTest extends SolrTestCaseJ4 { assertEquals(4, directory.fileLength("testing.test")); - IndexInput input1 = directory.openInput("testing.test", new IOContext()); + IndexInput input1 = directory.openInput("testing.test", IOContext.DEFAULT); IndexInput input2 = input1.clone(); assertEquals(12345, input2.readInt()); @@ -147,13 +147,13 @@ public class HdfsDirectoryTest extends SolrTestCaseJ4 { directory.deleteFile(file); } - IndexOutput output = directory.createOutput("testing.test", new IOContext()); + IndexOutput output = directory.createOutput("testing.test", IOContext.DEFAULT); output.writeInt(12345); output.close(); directory.rename("testing.test", "testing.test.renamed"); assertFalse(slowFileExists(directory, "testing.test")); assertTrue(slowFileExists(directory, "testing.test.renamed")); - IndexInput input = directory.openInput("testing.test.renamed", new IOContext()); + IndexInput input = directory.openInput("testing.test.renamed", IOContext.DEFAULT); assertEquals(12345, input.readInt()); assertEquals(input.getFilePointer(), input.length()); input.close(); @@ -174,7 +174,7 @@ public class HdfsDirectoryTest extends SolrTestCaseJ4 { } private void testEof(String name, Directory directory, long length) throws IOException { - IndexInput input = directory.openInput(name, new IOContext()); + IndexInput input = directory.openInput(name, IOContext.DEFAULT); input.seek(length); expectThrows(Exception.class, input::readByte); } @@ -202,8 +202,8 @@ public class HdfsDirectoryTest extends SolrTestCaseJ4 { private void assertInputsEquals(String name, Directory fsDir, HdfsDirectory hdfs) throws IOException { int reads = random.nextInt(MAX_NUMBER_OF_READS); - IndexInput fsInput = fsDir.openInput(name, new IOContext()); - IndexInput hdfsInput = hdfs.openInput(name, new IOContext()); + IndexInput fsInput = fsDir.openInput(name, IOContext.DEFAULT); + IndexInput hdfsInput = hdfs.openInput(name, IOContext.DEFAULT); assertEquals(fsInput.length(), hdfsInput.length()); int fileLength = (int) fsInput.length(); for (int i = 0; i < reads; i++) { @@ -233,8 +233,8 @@ public class HdfsDirectoryTest extends SolrTestCaseJ4 { private void createFile(String name, Directory fsDir, HdfsDirectory hdfs) throws IOException { int writes = random.nextInt(MAX_NUMBER_OF_WRITES); int fileLength = random.nextInt(MAX_FILE_SIZE - MIN_FILE_SIZE) + MIN_FILE_SIZE; - IndexOutput fsOutput = fsDir.createOutput(name, new IOContext()); - IndexOutput hdfsOutput = hdfs.createOutput(name, new IOContext()); + IndexOutput fsOutput = fsDir.createOutput(name, IOContext.DEFAULT); + IndexOutput hdfsOutput = hdfs.createOutput(name, IOContext.DEFAULT); for (int i = 0; i < writes; i++) { byte[] buf = new byte diff --git a/solr/modules/hdfs/src/test/org/apache/solr/hdfs/store/blockcache/BlockDirectoryTest.java b/solr/modules/hdfs/src/test/org/apache/solr/hdfs/store/blockcache/BlockDirectoryTest.java index f9e0dba9ac0..4550e639729 100644 --- a/solr/modules/hdfs/src/test/org/apache/solr/hdfs/store/blockcache/BlockDirectoryTest.java +++ b/solr/modules/hdfs/src/test/org/apache/solr/hdfs/store/blockcache/BlockDirectoryTest.java @@ -150,7 +150,7 @@ public class BlockDirectoryTest extends SolrTestCaseJ4 { } private void testEof(String name, Directory directory, long length) throws IOException { - IndexInput input = directory.openInput(name, new IOContext()); + IndexInput input = directory.openInput(name, IOContext.DEFAULT); try { input.seek(length); try { @@ -191,8 +191,8 @@ public class BlockDirectoryTest extends SolrTestCaseJ4 { private void assertInputsEquals(String name, Directory fsDir, Directory hdfs) throws IOException { int reads = random.nextInt(MAX_NUMBER_OF_READS); - IndexInput fsInput = fsDir.openInput(name, new IOContext()); - IndexInput hdfsInput = hdfs.openInput(name, new IOContext()); + IndexInput fsInput = fsDir.openInput(name, IOContext.DEFAULT); + IndexInput hdfsInput = hdfs.openInput(name, IOContext.DEFAULT); assertEquals(fsInput.length(), hdfsInput.length()); int fileLength = (int) fsInput.length(); for (int i = 0; i < reads; i++) { diff --git a/solr/modules/jwt-auth/src/test/org/apache/solr/security/jwt/JWTAuthPluginIntegrationTest.java b/solr/modules/jwt-auth/src/test/org/apache/solr/security/jwt/JWTAuthPluginIntegrationTest.java index 3bb08460218..a617c50d1e5 100644 --- a/solr/modules/jwt-auth/src/test/org/apache/solr/security/jwt/JWTAuthPluginIntegrationTest.java +++ b/solr/modules/jwt-auth/src/test/org/apache/solr/security/jwt/JWTAuthPluginIntegrationTest.java @@ -136,7 +136,7 @@ public class JWTAuthPluginIntegrationTest extends SolrCloudAuthTestCase { } @Test - @BadApple(bugUrl = "https://issues.apache.org/jira/browse/SOLR-15484") + @AwaitsFix(bugUrl = "https://issues.apache.org/jira/browse/SOLR-15484") public void mockOAuth2Server() throws Exception { MiniSolrCloudCluster myCluster = configureClusterMockOauth(2, pemFilePath, 10000); String baseUrl = myCluster.getRandomJetty(random()).getBaseUrl().toString();
