OK will do. Mike
On Tue, Oct 6, 2009 at 10:23 AM, Uwe Schindler <u...@thetaphi.de> wrote: > Can you add this patch to backwards, too? I forgot, that some of the > backwards-changes also applied to BW, but for completeness, not sure, if a > tag is also needed. > > ----- > Uwe Schindler > H.-H.-Meier-Allee 63, D-28213 Bremen > http://www.thetaphi.de > eMail: u...@thetaphi.de > > >> -----Original Message----- >> From: mikemcc...@apache.org [mailto:mikemcc...@apache.org] >> Sent: Tuesday, October 06, 2009 4:13 PM >> To: java-comm...@lucene.apache.org >> Subject: svn commit: r822284 - >> /lucene/java/trunk/src/test/org/apache/lucene/index/TestBackwardsCompatibi >> lity.java >> >> Author: mikemccand >> Date: Tue Oct 6 14:12:46 2009 >> New Revision: 822284 >> >> URL: http://svn.apache.org/viewvc?rev=822284&view=rev >> Log: >> fix TestBackwardsCompability to not use deprecated APIs >> >> Modified: >> >> lucene/java/trunk/src/test/org/apache/lucene/index/TestBackwardsCompatibil >> ity.java >> >> Modified: >> lucene/java/trunk/src/test/org/apache/lucene/index/TestBackwardsCompatibil >> ity.java >> URL: >> http://svn.apache.org/viewvc/lucene/java/trunk/src/test/org/apache/lucene/ >> index/TestBackwardsCompatibility.java?rev=822284&r1=822283&r2=822284&view= >> diff >> ========================================================================== >> ==== >> --- >> lucene/java/trunk/src/test/org/apache/lucene/index/TestBackwardsCompatibil >> ity.java (original) >> +++ >> lucene/java/trunk/src/test/org/apache/lucene/index/TestBackwardsCompatibil >> ity.java Tue Oct 6 14:12:46 2009 >> @@ -158,11 +158,7 @@ >> for(int i=0;i<oldNames.length;i++) { >> String dirName = "src/test/org/apache/lucene/index/index." + >> oldNames[i]; >> unzip(dirName, oldNames[i]); >> - changeIndexNoAdds(oldNames[i], true); >> - rmDir(oldNames[i]); >> - >> - unzip(dirName, oldNames[i]); >> - changeIndexNoAdds(oldNames[i], false); >> + changeIndexNoAdds(oldNames[i]); >> rmDir(oldNames[i]); >> } >> } >> @@ -171,11 +167,7 @@ >> for(int i=0;i<oldNames.length;i++) { >> String dirName = "src/test/org/apache/lucene/index/index." + >> oldNames[i]; >> unzip(dirName, oldNames[i]); >> - changeIndexWithAdds(oldNames[i], true); >> - rmDir(oldNames[i]); >> - >> - unzip(dirName, oldNames[i]); >> - changeIndexWithAdds(oldNames[i], false); >> + changeIndexWithAdds(oldNames[i]); >> rmDir(oldNames[i]); >> } >> } >> @@ -196,7 +188,7 @@ >> dirName = fullDir(dirName); >> >> Directory dir = FSDirectory.open(new File(dirName)); >> - IndexSearcher searcher = new IndexSearcher(dir); >> + IndexSearcher searcher = new IndexSearcher(dir, true); >> IndexReader reader = searcher.getIndexReader(); >> >> _TestUtil.checkIndex(dir); >> @@ -267,14 +259,14 @@ >> >> /* Open pre-lockless index, add docs, do a delete & >> * setNorm, and search */ >> - public void changeIndexWithAdds(String dirName, boolean autoCommit) >> throws IOException { >> + public void changeIndexWithAdds(String dirName) throws IOException { >> String origDirName = dirName; >> dirName = fullDir(dirName); >> >> Directory dir = FSDirectory.open(new File(dirName)); >> >> // open writer >> - IndexWriter writer = new IndexWriter(dir, autoCommit, new >> WhitespaceAnalyzer(), false); >> + IndexWriter writer = new IndexWriter(dir, new WhitespaceAnalyzer(), >> false, IndexWriter.MaxFieldLength.UNLIMITED); >> >> // add 10 docs >> for(int i=0;i<10;i++) { >> @@ -288,11 +280,11 @@ >> } else { >> expected = 46; >> } >> - assertEquals("wrong doc count", expected, writer.docCount()); >> + assertEquals("wrong doc count", expected, writer.maxDoc()); >> writer.close(); >> >> // make sure searching sees right # hits >> - IndexSearcher searcher = new IndexSearcher(dir); >> + IndexSearcher searcher = new IndexSearcher(dir, true); >> ScoreDoc[] hits = searcher.search(new TermQuery(new Term("content", >> "aaa")), null, 1000).scoreDocs; >> Document d = searcher.doc(hits[0].doc); >> assertEquals("wrong first document", "21", d.get("id")); >> @@ -301,7 +293,7 @@ >> >> // make sure we can do delete & setNorm against this >> // pre-lockless segment: >> - IndexReader reader = IndexReader.open(dir); >> + IndexReader reader = IndexReader.open(dir, false); >> Term searchTerm = new Term("id", "6"); >> int delCount = reader.deleteDocuments(searchTerm); >> assertEquals("wrong delete count", 1, delCount); >> @@ -309,7 +301,7 @@ >> reader.close(); >> >> // make sure they "took": >> - searcher = new IndexSearcher(dir); >> + searcher = new IndexSearcher(dir, true); >> hits = searcher.search(new TermQuery(new Term("content", "aaa")), >> null, 1000).scoreDocs; >> assertEquals("wrong number of hits", 43, hits.length); >> d = searcher.doc(hits[0].doc); >> @@ -318,11 +310,11 @@ >> searcher.close(); >> >> // optimize >> - writer = new IndexWriter(dir, autoCommit, new WhitespaceAnalyzer(), >> false); >> + writer = new IndexWriter(dir, new WhitespaceAnalyzer(), false, >> IndexWriter.MaxFieldLength.UNLIMITED); >> writer.optimize(); >> writer.close(); >> >> - searcher = new IndexSearcher(dir); >> + searcher = new IndexSearcher(dir, true); >> hits = searcher.search(new TermQuery(new Term("content", "aaa")), >> null, 1000).scoreDocs; >> assertEquals("wrong number of hits", 43, hits.length); >> d = searcher.doc(hits[0].doc); >> @@ -335,14 +327,14 @@ >> >> /* Open pre-lockless index, add docs, do a delete & >> * setNorm, and search */ >> - public void changeIndexNoAdds(String dirName, boolean autoCommit) >> throws IOException { >> + public void changeIndexNoAdds(String dirName) throws IOException { >> >> dirName = fullDir(dirName); >> >> Directory dir = FSDirectory.open(new File(dirName)); >> >> // make sure searching sees right # hits >> - IndexSearcher searcher = new IndexSearcher(dir); >> + IndexSearcher searcher = new IndexSearcher(dir, true); >> ScoreDoc[] hits = searcher.search(new TermQuery(new Term("content", >> "aaa")), null, 1000).scoreDocs; >> assertEquals("wrong number of hits", 34, hits.length); >> Document d = searcher.doc(hits[0].doc); >> @@ -351,7 +343,7 @@ >> >> // make sure we can do a delete & setNorm against this >> // pre-lockless segment: >> - IndexReader reader = IndexReader.open(dir); >> + IndexReader reader = IndexReader.open(dir, false); >> Term searchTerm = new Term("id", "6"); >> int delCount = reader.deleteDocuments(searchTerm); >> assertEquals("wrong delete count", 1, delCount); >> @@ -359,7 +351,7 @@ >> reader.close(); >> >> // make sure they "took": >> - searcher = new IndexSearcher(dir); >> + searcher = new IndexSearcher(dir, true); >> hits = searcher.search(new TermQuery(new Term("content", "aaa")), >> null, 1000).scoreDocs; >> assertEquals("wrong number of hits", 33, hits.length); >> d = searcher.doc(hits[0].doc); >> @@ -368,11 +360,11 @@ >> searcher.close(); >> >> // optimize >> - IndexWriter writer = new IndexWriter(dir, autoCommit, new >> WhitespaceAnalyzer(), false); >> + IndexWriter writer = new IndexWriter(dir, new WhitespaceAnalyzer(), >> false, IndexWriter.MaxFieldLength.UNLIMITED); >> writer.optimize(); >> writer.close(); >> >> - searcher = new IndexSearcher(dir); >> + searcher = new IndexSearcher(dir, true); >> hits = searcher.search(new TermQuery(new Term("content", "aaa")), >> null, 1000).scoreDocs; >> assertEquals("wrong number of hits", 33, hits.length); >> d = searcher.doc(hits[0].doc); >> @@ -397,7 +389,7 @@ >> for(int i=0;i<35;i++) { >> addDoc(writer, i); >> } >> - assertEquals("wrong doc count", 35, writer.docCount()); >> + assertEquals("wrong doc count", 35, writer.maxDoc()); >> writer.close(); >> >> // open fresh writer so we get no prx file in the added segment >> @@ -408,7 +400,7 @@ >> writer.close(); >> >> // Delete one doc so we get a .del file: >> - IndexReader reader = IndexReader.open(dir); >> + IndexReader reader = IndexReader.open(dir, false); >> Term searchTerm = new Term("id", "7"); >> int delCount = reader.deleteDocuments(searchTerm); >> assertEquals("didn't delete the right number of documents", 1, >> delCount); >> @@ -422,71 +414,66 @@ >> >> public void testExactFileNames() throws IOException { >> >> - for(int pass=0;pass<2;pass++) { >> - >> - String outputDir = "lucene.backwardscompat0.index"; >> - rmDir(outputDir); >> + String outputDir = "lucene.backwardscompat0.index"; >> + rmDir(outputDir); >> >> - try { >> - Directory dir = FSDirectory.open(new File(fullDir(outputDir))); >> + try { >> + Directory dir = FSDirectory.open(new File(fullDir(outputDir))); >> >> - boolean autoCommit = 0 == pass; >> - >> - IndexWriter writer = new IndexWriter(dir, autoCommit, new >> WhitespaceAnalyzer(), true); >> - writer.setRAMBufferSizeMB(16.0); >> - for(int i=0;i<35;i++) { >> - addDoc(writer, i); >> - } >> - assertEquals("wrong doc count", 35, writer.docCount()); >> - writer.close(); >> + IndexWriter writer = new IndexWriter(dir, new WhitespaceAnalyzer(), >> true, IndexWriter.MaxFieldLength.UNLIMITED); >> + writer.setRAMBufferSizeMB(16.0); >> + for(int i=0;i<35;i++) { >> + addDoc(writer, i); >> + } >> + assertEquals("wrong doc count", 35, writer.maxDoc()); >> + writer.close(); >> >> - // Delete one doc so we get a .del file: >> - IndexReader reader = IndexReader.open(dir); >> - Term searchTerm = new Term("id", "7"); >> - int delCount = reader.deleteDocuments(searchTerm); >> - assertEquals("didn't delete the right number of documents", 1, >> delCount); >> - >> - // Set one norm so we get a .s0 file: >> - reader.setNorm(21, "content", (float) 1.5); >> - reader.close(); >> - >> - // The numbering of fields can vary depending on which >> - // JRE is in use. On some JREs we see content bound to >> - // field 0; on others, field 1. So, here we have to >> - // figure out which field number corresponds to >> - // "content", and then set our expected file names below >> - // accordingly: >> - CompoundFileReader cfsReader = new CompoundFileReader(dir, >> "_0.cfs"); >> - FieldInfos fieldInfos = new FieldInfos(cfsReader, "_0.fnm"); >> - int contentFieldIndex = -1; >> - for(int i=0;i<fieldInfos.size();i++) { >> - FieldInfo fi = fieldInfos.fieldInfo(i); >> - if (fi.name.equals("content")) { >> - contentFieldIndex = i; >> - break; >> - } >> + // Delete one doc so we get a .del file: >> + IndexReader reader = IndexReader.open(dir, false); >> + Term searchTerm = new Term("id", "7"); >> + int delCount = reader.deleteDocuments(searchTerm); >> + assertEquals("didn't delete the right number of documents", 1, >> delCount); >> + >> + // Set one norm so we get a .s0 file: >> + reader.setNorm(21, "content", (float) 1.5); >> + reader.close(); >> + >> + // The numbering of fields can vary depending on which >> + // JRE is in use. On some JREs we see content bound to >> + // field 0; on others, field 1. So, here we have to >> + // figure out which field number corresponds to >> + // "content", and then set our expected file names below >> + // accordingly: >> + CompoundFileReader cfsReader = new CompoundFileReader(dir, >> "_0.cfs"); >> + FieldInfos fieldInfos = new FieldInfos(cfsReader, "_0.fnm"); >> + int contentFieldIndex = -1; >> + for(int i=0;i<fieldInfos.size();i++) { >> + FieldInfo fi = fieldInfos.fieldInfo(i); >> + if (fi.name.equals("content")) { >> + contentFieldIndex = i; >> + break; >> } >> - cfsReader.close(); >> - assertTrue("could not locate the 'content' field number in the >> _2.cfs segment", contentFieldIndex != -1); >> + } >> + cfsReader.close(); >> + assertTrue("could not locate the 'content' field number in the >> _2.cfs segment", contentFieldIndex != -1); >> >> - // Now verify file names: >> - String[] expected; >> - expected = new String[] {"_0.cfs", >> - "_0_1.del", >> - "_0_1.s" + contentFieldIndex, >> - "segments_3", >> - "segments.gen"}; >> - >> - String[] actual = dir.listAll(); >> - Arrays.sort(expected); >> - Arrays.sort(actual); >> - if (!Arrays.equals(expected, actual)) { >> - fail("incorrect filenames in index: expected:\n " + >> asString(expected) + "\n actual:\n " + asString(actual)); >> - } >> - dir.close(); >> - } finally { >> - rmDir(outputDir); >> + // Now verify file names: >> + String[] expected; >> + expected = new String[] {"_0.cfs", >> + "_0_1.del", >> + "_0_1.s" + contentFieldIndex, >> + "segments_3", >> + "segments.gen"}; >> + >> + String[] actual = dir.listAll(); >> + Arrays.sort(expected); >> + Arrays.sort(actual); >> + if (!Arrays.equals(expected, actual)) { >> + fail("incorrect filenames in index: expected:\n " + >> asString(expected) + "\n actual:\n " + asString(actual)); >> } >> + dir.close(); >> + } finally { >> + rmDir(outputDir); >> } >> } >> >> @@ -509,17 +496,17 @@ >> doc.add(new Field("autf8", "Lu\uD834\uDD1Ece\uD834\uDD60ne \u0000 >> \u2620 ab\ud917\udc17cd", Field.Store.YES, Field.Index.ANALYZED, >> Field.TermVector.WITH_POSITIONS_OFFSETS)); >> doc.add(new Field("utf8", "Lu\uD834\uDD1Ece\uD834\uDD60ne \u0000 >> \u2620 ab\ud917\udc17cd", Field.Store.YES, Field.Index.ANALYZED, >> Field.TermVector.WITH_POSITIONS_OFFSETS)); >> doc.add(new Field("content2", "here is more content with aaa aaa >> aaa", Field.Store.YES, Field.Index.ANALYZED, >> Field.TermVector.WITH_POSITIONS_OFFSETS)); >> - doc.add(new Field("fie\u2C77ld", "field with non-ascii name", >> Field.Store.YES, Field.Index.TOKENIZED, >> Field.TermVector.WITH_POSITIONS_OFFSETS)); >> + doc.add(new Field("fie\u2C77ld", "field with non-ascii name", >> Field.Store.YES, Field.Index.ANALYZED, >> Field.TermVector.WITH_POSITIONS_OFFSETS)); >> writer.addDocument(doc); >> } >> >> private void addNoProxDoc(IndexWriter writer) throws IOException { >> Document doc = new Document(); >> Field f = new Field("content3", "aaa", Field.Store.YES, >> Field.Index.ANALYZED); >> - f.setOmitTf(true); >> + f.setOmitTermFreqAndPositions(true); >> doc.add(f); >> f = new Field("content4", "aaa", Field.Store.YES, Field.Index.NO); >> - f.setOmitTf(true); >> + f.setOmitTermFreqAndPositions(true); >> doc.add(f); >> writer.addDocument(doc); >> } >> > > > > --------------------------------------------------------------------- > To unsubscribe, e-mail: java-dev-unsubscr...@lucene.apache.org > For additional commands, e-mail: java-dev-h...@lucene.apache.org > > --------------------------------------------------------------------- To unsubscribe, e-mail: java-dev-unsubscr...@lucene.apache.org For additional commands, e-mail: java-dev-h...@lucene.apache.org