http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Hu/TestHungarianLightStemFilter.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Hu/TestHungarianLightStemFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Hu/TestHungarianLightStemFilter.cs index eeeef95..89ef132 100644 --- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Hu/TestHungarianLightStemFilter.cs +++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Hu/TestHungarianLightStemFilter.cs @@ -36,7 +36,7 @@ namespace Lucene.Net.Analysis.Hu { } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer source = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); return new TokenStreamComponents(source, new HungarianLightStemFilter(source)); @@ -71,7 +71,7 @@ namespace Lucene.Net.Analysis.Hu this.exclusionSet = exclusionSet; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer source = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); TokenStream sink = new SetKeywordMarkerFilter(source, exclusionSet); @@ -95,7 +95,7 @@ namespace Lucene.Net.Analysis.Hu this.outerInstance = outerInstance; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new KeywordTokenizer(reader); return new TokenStreamComponents(tokenizer, new HungarianLightStemFilter(tokenizer));
http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Hunspell/TestHunspellStemFilter.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Hunspell/TestHunspellStemFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Hunspell/TestHunspellStemFilter.cs index 4ece2fb..07d3541 100644 --- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Hunspell/TestHunspellStemFilter.cs +++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Hunspell/TestHunspellStemFilter.cs @@ -98,7 +98,7 @@ namespace Lucene.Net.Analysis.Hunspell this.outerInstance = outerInstance; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); return new TokenStreamComponents(tokenizer, new HunspellStemFilter(tokenizer, dictionary)); @@ -121,7 +121,7 @@ namespace Lucene.Net.Analysis.Hunspell this.outerInstance = outerInstance; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new KeywordTokenizer(reader); return new TokenStreamComponents(tokenizer, new HunspellStemFilter(tokenizer, dictionary)); @@ -158,7 +158,7 @@ namespace Lucene.Net.Analysis.Hunspell this.d = d; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new KeywordTokenizer(reader); return new TokenStreamComponents(tokenizer, new HunspellStemFilter(tokenizer, d)); http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Id/TestIndonesianStemmer.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Id/TestIndonesianStemmer.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Id/TestIndonesianStemmer.cs index 73857df..e55eff0 100644 --- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Id/TestIndonesianStemmer.cs +++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Id/TestIndonesianStemmer.cs @@ -35,7 +35,7 @@ namespace Lucene.Net.Analysis.Id { } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new KeywordTokenizer(reader); return new TokenStreamComponents(tokenizer, new IndonesianStemFilter(tokenizer)); @@ -129,7 +129,7 @@ namespace Lucene.Net.Analysis.Id { } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new KeywordTokenizer(reader); return new TokenStreamComponents(tokenizer, new IndonesianStemFilter(tokenizer, false)); @@ -171,7 +171,7 @@ namespace Lucene.Net.Analysis.Id this.outerInstance = outerInstance; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new KeywordTokenizer(reader); return new TokenStreamComponents(tokenizer, new IndonesianStemFilter(tokenizer)); http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/In/TestIndicNormalizer.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/In/TestIndicNormalizer.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/In/TestIndicNormalizer.cs index 8294ffe..47731bf 100644 --- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/In/TestIndicNormalizer.cs +++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/In/TestIndicNormalizer.cs @@ -65,7 +65,7 @@ namespace Lucene.Net.Analysis.In this.outerInstance = outerInstance; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new KeywordTokenizer(reader); return new TokenStreamComponents(tokenizer, new IndicNormalizationFilter(tokenizer)); http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/It/TestItalianLightStemFilter.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/It/TestItalianLightStemFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/It/TestItalianLightStemFilter.cs index d4f0fd7..1bd67dc 100644 --- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/It/TestItalianLightStemFilter.cs +++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/It/TestItalianLightStemFilter.cs @@ -34,7 +34,7 @@ namespace Lucene.Net.Analysis.It { } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer source = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); return new TokenStreamComponents(source, new ItalianLightStemFilter(source)); @@ -73,7 +73,7 @@ namespace Lucene.Net.Analysis.It this.outerInstance = outerInstance; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new KeywordTokenizer(reader); return new TokenStreamComponents(tokenizer, new ItalianLightStemFilter(tokenizer)); http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Lv/TestLatvianStemmer.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Lv/TestLatvianStemmer.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Lv/TestLatvianStemmer.cs index 52adc34..5192510 100644 --- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Lv/TestLatvianStemmer.cs +++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Lv/TestLatvianStemmer.cs @@ -34,7 +34,7 @@ namespace Lucene.Net.Analysis.Lv { } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); return new TokenStreamComponents(tokenizer, new LatvianStemFilter(tokenizer)); @@ -307,7 +307,7 @@ namespace Lucene.Net.Analysis.Lv this.outerInstance = outerInstance; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new KeywordTokenizer(reader); return new TokenStreamComponents(tokenizer, new LatvianStemFilter(tokenizer)); http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestASCIIFoldingFilter.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestASCIIFoldingFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestASCIIFoldingFilter.cs index 2b7cd61..d4d246f 100644 --- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestASCIIFoldingFilter.cs +++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestASCIIFoldingFilter.cs @@ -242,7 +242,7 @@ namespace Lucene.Net.Analysis.Miscellaneous this.outerInstance = outerInstance; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); return new TokenStreamComponents(tokenizer, new ASCIIFoldingFilter(tokenizer, Random().nextBoolean())); @@ -265,7 +265,7 @@ namespace Lucene.Net.Analysis.Miscellaneous this.outerInstance = outerInstance; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new KeywordTokenizer(reader); return new TokenStreamComponents(tokenizer, new ASCIIFoldingFilter(tokenizer, Random().nextBoolean())); http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestCapitalizationFilter.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestCapitalizationFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestCapitalizationFilter.cs index a95c131..781bdee 100644 --- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestCapitalizationFilter.cs +++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestCapitalizationFilter.cs @@ -109,7 +109,7 @@ namespace Lucene.Net.Analysis.Miscellaneous this.outerInstance = outerInstance; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); return new TokenStreamComponents(tokenizer, new CapitalizationFilter(tokenizer)); @@ -132,7 +132,7 @@ namespace Lucene.Net.Analysis.Miscellaneous this.outerInstance = outerInstance; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new KeywordTokenizer(reader); return new TokenStreamComponents(tokenizer, new CapitalizationFilter(tokenizer)); http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestCodepointCountFilter.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestCodepointCountFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestCodepointCountFilter.cs index e06097d..a53bab8 100644 --- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestCodepointCountFilter.cs +++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestCodepointCountFilter.cs @@ -50,7 +50,7 @@ namespace Lucene.Net.Analysis.Miscellaneous this.outerInstance = outerInstance; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new KeywordTokenizer(reader); return new TokenStreamComponents(tokenizer, new CodepointCountFilter(TEST_VERSION_CURRENT, tokenizer, 0, 5)); http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestHyphenatedWordsFilter.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestHyphenatedWordsFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestHyphenatedWordsFilter.cs index a04f8ed..a2819c4 100644 --- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestHyphenatedWordsFilter.cs +++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestHyphenatedWordsFilter.cs @@ -77,7 +77,7 @@ namespace Lucene.Net.Analysis.Miscellaneous this.outerInstance = outerInstance; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); return new TokenStreamComponents(tokenizer, new HyphenatedWordsFilter(tokenizer)); @@ -100,7 +100,7 @@ namespace Lucene.Net.Analysis.Miscellaneous this.outerInstance = outerInstance; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new KeywordTokenizer(reader); return new TokenStreamComponents(tokenizer, new HyphenatedWordsFilter(tokenizer)); http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestKeepWordFilter.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestKeepWordFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestKeepWordFilter.cs index f92e74f..c8148e2 100644 --- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestKeepWordFilter.cs +++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestKeepWordFilter.cs @@ -86,7 +86,7 @@ namespace Lucene.Net.Analysis.Miscellaneous this.words = words; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); TokenStream stream = new KeepWordFilter(TEST_VERSION_CURRENT, tokenizer, new CharArraySet(TEST_VERSION_CURRENT, words, true)); http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestLengthFilter.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestLengthFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestLengthFilter.cs index 3c55dc8..7180aed 100644 --- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestLengthFilter.cs +++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestLengthFilter.cs @@ -60,7 +60,7 @@ namespace Lucene.Net.Analysis.Miscellaneous this.outerInstance = outerInstance; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new KeywordTokenizer(reader); return new TokenStreamComponents(tokenizer, new LengthFilter(TEST_VERSION_CURRENT, tokenizer, 0, 5)); http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestLimitTokenPositionFilter.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestLimitTokenPositionFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestLimitTokenPositionFilter.cs index 3b4c014..c3afa22 100644 --- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestLimitTokenPositionFilter.cs +++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestLimitTokenPositionFilter.cs @@ -55,7 +55,7 @@ namespace Lucene.Net.Analysis.Miscellaneous this.consumeAll = consumeAll; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { MockTokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); // if we are consuming all tokens, we can use the checks, otherwise we can't http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestLucene47WordDelimiterFilter.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestLucene47WordDelimiterFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestLucene47WordDelimiterFilter.cs index 3d97c0d..48d76d7 100644 --- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestLucene47WordDelimiterFilter.cs +++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestLucene47WordDelimiterFilter.cs @@ -293,7 +293,7 @@ namespace Lucene.Net.Analysis.Miscellaneous this.protWords = protWords; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); return new TokenStreamComponents(tokenizer, new Lucene47WordDelimiterFilter(tokenizer, flags, protWords)); @@ -314,7 +314,7 @@ namespace Lucene.Net.Analysis.Miscellaneous this.protWords = protWords; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); return new TokenStreamComponents(tokenizer, new Lucene47WordDelimiterFilter(new LargePosIncTokenFilter(outerInstance, tokenizer), flags, protWords)); @@ -335,7 +335,7 @@ namespace Lucene.Net.Analysis.Miscellaneous this.protWords = protWords; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); StopFilter filter = new StopFilter(TEST_VERSION_CURRENT, tokenizer, StandardAnalyzer.STOP_WORDS_SET); @@ -381,7 +381,7 @@ namespace Lucene.Net.Analysis.Miscellaneous this.protectedWords = protectedWords; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); return new TokenStreamComponents(tokenizer, new Lucene47WordDelimiterFilter(tokenizer, flags, protectedWords)); @@ -425,7 +425,7 @@ namespace Lucene.Net.Analysis.Miscellaneous this.protectedWords = protectedWords; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new KeywordTokenizer(reader); return new TokenStreamComponents(tokenizer, new Lucene47WordDelimiterFilter(tokenizer, flags, protectedWords)); http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestPerFieldAnalyzerWrapper.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestPerFieldAnalyzerWrapper.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestPerFieldAnalyzerWrapper.cs index e3aa88e..7429429 100644 --- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestPerFieldAnalyzerWrapper.cs +++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestPerFieldAnalyzerWrapper.cs @@ -90,12 +90,12 @@ namespace Lucene.Net.Analysis.Miscellaneous this.outerInstance = outerInstance; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { return new TokenStreamComponents(new MockTokenizer(reader)); } - protected internal override TextReader InitReader(string fieldName, TextReader reader) + protected override TextReader InitReader(string fieldName, TextReader reader) { return new MockCharFilter(reader, 7); } http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestRemoveDuplicatesTokenFilter.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestRemoveDuplicatesTokenFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestRemoveDuplicatesTokenFilter.cs index 5eb577b..7634702 100644 --- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestRemoveDuplicatesTokenFilter.cs +++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestRemoveDuplicatesTokenFilter.cs @@ -171,7 +171,7 @@ namespace Lucene.Net.Analysis.Miscellaneous this.ignoreCase = ignoreCase; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.SIMPLE, true); TokenStream stream = new SynonymFilter(tokenizer, map, ignoreCase); @@ -195,7 +195,7 @@ namespace Lucene.Net.Analysis.Miscellaneous this.outerInstance = outerInstance; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new KeywordTokenizer(reader); return new TokenStreamComponents(tokenizer, new RemoveDuplicatesTokenFilter(tokenizer)); http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestScandinavianFoldingFilter.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestScandinavianFoldingFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestScandinavianFoldingFilter.cs index fcbbee9..9c1f248 100644 --- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestScandinavianFoldingFilter.cs +++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestScandinavianFoldingFilter.cs @@ -31,7 +31,7 @@ namespace Lucene.Net.Analysis.Miscellaneous { } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); TokenStream stream = new ScandinavianFoldingFilter(tokenizer); @@ -128,7 +128,7 @@ namespace Lucene.Net.Analysis.Miscellaneous this.outerInstance = outerInstance; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new KeywordTokenizer(reader); return new TokenStreamComponents(tokenizer, new ScandinavianFoldingFilter(tokenizer)); http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestScandinavianNormalizationFilter.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestScandinavianNormalizationFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestScandinavianNormalizationFilter.cs index 597d3f0..369d28d 100644 --- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestScandinavianNormalizationFilter.cs +++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestScandinavianNormalizationFilter.cs @@ -33,7 +33,7 @@ namespace Lucene.Net.Analysis.Miscellaneous { } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); TokenStream stream = new ScandinavianNormalizationFilter(tokenizer); @@ -128,7 +128,7 @@ namespace Lucene.Net.Analysis.Miscellaneous this.outerInstance = outerInstance; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new KeywordTokenizer(reader); return new TokenStreamComponents(tokenizer, new ScandinavianNormalizationFilter(tokenizer)); http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestTrimFilter.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestTrimFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestTrimFilter.cs index c61e5b1..5b0d900 100644 --- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestTrimFilter.cs +++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestTrimFilter.cs @@ -125,7 +125,7 @@ namespace Lucene.Net.Analysis.Miscellaneous this.outerInstance = outerInstance; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.KEYWORD, false); return new TokenStreamComponents(tokenizer, new TrimFilter(LuceneVersion.LUCENE_43, tokenizer, true)); @@ -141,7 +141,7 @@ namespace Lucene.Net.Analysis.Miscellaneous this.outerInstance = outerInstance; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.KEYWORD, false); return new TokenStreamComponents(tokenizer, new TrimFilter(TEST_VERSION_CURRENT, tokenizer, false)); @@ -164,7 +164,7 @@ namespace Lucene.Net.Analysis.Miscellaneous this.outerInstance = outerInstance; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new KeywordTokenizer(reader); bool updateOffsets = Random().nextBoolean(); http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestWordDelimiterFilter.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestWordDelimiterFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestWordDelimiterFilter.cs index 530682a..df9df2b 100644 --- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestWordDelimiterFilter.cs +++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestWordDelimiterFilter.cs @@ -287,7 +287,7 @@ namespace Lucene.Net.Analysis.Miscellaneous this.protWords = protWords; } - protected internal override TokenStreamComponents CreateComponents(string field, TextReader reader) + protected override TokenStreamComponents CreateComponents(string field, TextReader reader) { Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); return new TokenStreamComponents(tokenizer, new WordDelimiterFilter(TEST_VERSION_CURRENT, tokenizer, flags, protWords)); @@ -308,7 +308,7 @@ namespace Lucene.Net.Analysis.Miscellaneous this.protWords = protWords; } - protected internal override TokenStreamComponents CreateComponents(string field, TextReader reader) + protected override TokenStreamComponents CreateComponents(string field, TextReader reader) { Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); return new TokenStreamComponents(tokenizer, new WordDelimiterFilter(TEST_VERSION_CURRENT, new LargePosIncTokenFilter(outerInstance, tokenizer), flags, protWords)); @@ -329,7 +329,7 @@ namespace Lucene.Net.Analysis.Miscellaneous this.protWords = protWords; } - protected internal override TokenStreamComponents CreateComponents(string field, TextReader reader) + protected override TokenStreamComponents CreateComponents(string field, TextReader reader) { Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); StopFilter filter = new StopFilter(TEST_VERSION_CURRENT, tokenizer, StandardAnalyzer.STOP_WORDS_SET); @@ -369,7 +369,7 @@ namespace Lucene.Net.Analysis.Miscellaneous this.flags = flags; } - protected internal override TokenStreamComponents CreateComponents(string field, TextReader reader) + protected override TokenStreamComponents CreateComponents(string field, TextReader reader) { Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); return new TokenStreamComponents(tokenizer, new WordDelimiterFilter(TEST_VERSION_CURRENT, tokenizer, flags, null)); @@ -409,7 +409,7 @@ namespace Lucene.Net.Analysis.Miscellaneous this.flags = flags; } - protected internal override TokenStreamComponents CreateComponents(string field, TextReader reader) + protected override TokenStreamComponents CreateComponents(string field, TextReader reader) { Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); return new TokenStreamComponents(tokenizer, new WordDelimiterFilter(TEST_VERSION_CURRENT, tokenizer, flags, null)); @@ -453,7 +453,7 @@ namespace Lucene.Net.Analysis.Miscellaneous this.protectedWords = protectedWords; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); return new TokenStreamComponents(tokenizer, new WordDelimiterFilter(TEST_VERSION_CURRENT, tokenizer, flags, protectedWords)); @@ -499,7 +499,7 @@ namespace Lucene.Net.Analysis.Miscellaneous } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); return new TokenStreamComponents(tokenizer, new WordDelimiterFilter(TEST_VERSION_CURRENT, tokenizer, flags, protectedWords)); @@ -543,7 +543,7 @@ namespace Lucene.Net.Analysis.Miscellaneous this.protectedWords = protectedWords; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new KeywordTokenizer(reader); return new TokenStreamComponents(tokenizer, new WordDelimiterFilter(TEST_VERSION_CURRENT, tokenizer, flags, protectedWords)); http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/NGram/EdgeNGramTokenFilterTest.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/NGram/EdgeNGramTokenFilterTest.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/NGram/EdgeNGramTokenFilterTest.cs index ea6fbd7..bfdd72e 100644 --- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/NGram/EdgeNGramTokenFilterTest.cs +++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/NGram/EdgeNGramTokenFilterTest.cs @@ -240,7 +240,7 @@ namespace Lucene.Net.Analysis.NGram this.outerInstance = outerInstance; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); TokenFilter filters = new ASCIIFoldingFilter(tokenizer); @@ -283,7 +283,7 @@ namespace Lucene.Net.Analysis.NGram this.max = max; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); return new TokenStreamComponents(tokenizer, new EdgeNGramTokenFilter(TEST_VERSION_CURRENT, tokenizer, min, max)); @@ -299,7 +299,7 @@ namespace Lucene.Net.Analysis.NGram this.outerInstance = outerInstance; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); #pragma warning disable 612, 618 @@ -328,7 +328,7 @@ namespace Lucene.Net.Analysis.NGram this.outerInstance = outerInstance; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new KeywordTokenizer(reader); #pragma warning disable 612, 618 @@ -346,7 +346,7 @@ namespace Lucene.Net.Analysis.NGram this.outerInstance = outerInstance; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new KeywordTokenizer(reader); #pragma warning disable 612, 618 http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/NGram/EdgeNGramTokenizerTest.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/NGram/EdgeNGramTokenizerTest.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/NGram/EdgeNGramTokenizerTest.cs index 6cddafe..b9248fb 100644 --- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/NGram/EdgeNGramTokenizerTest.cs +++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/NGram/EdgeNGramTokenizerTest.cs @@ -164,7 +164,7 @@ namespace Lucene.Net.Analysis.NGram this.max = max; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, Reader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, Reader reader) { Tokenizer tokenizer = new EdgeNGramTokenizer(TEST_VERSION_CURRENT, reader, min, max); return new TokenStreamComponents(tokenizer, tokenizer); @@ -180,7 +180,7 @@ namespace Lucene.Net.Analysis.NGram this.outerInstance = outerInstance; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, Reader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, Reader reader) { #pragma warning disable 612, 618 Tokenizer tokenizer = new Lucene43EdgeNGramTokenizer(Version.LUCENE_43, reader, Lucene43EdgeNGramTokenizer.Side.BACK, 2, 4); http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/NGram/NGramTokenFilterTest.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/NGram/NGramTokenFilterTest.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/NGram/NGramTokenFilterTest.cs index e485fc0..5d9fee4 100644 --- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/NGram/NGramTokenFilterTest.cs +++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/NGram/NGramTokenFilterTest.cs @@ -142,7 +142,7 @@ namespace Lucene.Net.Analysis.NGram this.outerInstance = outerInstance; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); TokenFilter filters = new ASCIIFoldingFilter(tokenizer); @@ -179,7 +179,7 @@ namespace Lucene.Net.Analysis.NGram this.max = max; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); return new TokenStreamComponents(tokenizer, new NGramTokenFilter(TEST_VERSION_CURRENT, tokenizer, min, max)); @@ -203,7 +203,7 @@ namespace Lucene.Net.Analysis.NGram this.outerInstance = outerInstance; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new KeywordTokenizer(reader); return new TokenStreamComponents(tokenizer, new NGramTokenFilter(TEST_VERSION_CURRENT, tokenizer, 2, 15)); http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/NGram/NGramTokenizerTest.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/NGram/NGramTokenizerTest.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/NGram/NGramTokenizerTest.cs index 2fc1356..b8db76e 100644 --- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/NGram/NGramTokenizerTest.cs +++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/NGram/NGramTokenizerTest.cs @@ -132,7 +132,7 @@ namespace Lucene.Net.Analysis.NGram this.max = max; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new NGramTokenizer(TEST_VERSION_CURRENT, reader, min, max); return new TokenStreamComponents(tokenizer, tokenizer); http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/No/TestNorwegianLightStemFilter.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/No/TestNorwegianLightStemFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/No/TestNorwegianLightStemFilter.cs index cedb8c4..0b3e183 100644 --- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/No/TestNorwegianLightStemFilter.cs +++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/No/TestNorwegianLightStemFilter.cs @@ -37,7 +37,7 @@ namespace Lucene.Net.Analysis.No { } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer source = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); return new TokenStreamComponents(source, new NorwegianLightStemFilter(source, NorwegianStandard.BOKMAAL)); @@ -70,7 +70,7 @@ namespace Lucene.Net.Analysis.No this.outerInstance = outerInstance; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer source = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); return new TokenStreamComponents(source, new NorwegianLightStemFilter(source, NorwegianStandard.NYNORSK)); @@ -97,7 +97,7 @@ namespace Lucene.Net.Analysis.No this.exclusionSet = exclusionSet; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer source = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); TokenStream sink = new SetKeywordMarkerFilter(source, exclusionSet); @@ -130,7 +130,7 @@ namespace Lucene.Net.Analysis.No this.outerInstance = outerInstance; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new KeywordTokenizer(reader); return new TokenStreamComponents(tokenizer, new NorwegianLightStemFilter(tokenizer)); http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/No/TestNorwegianMinimalStemFilter.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/No/TestNorwegianMinimalStemFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/No/TestNorwegianMinimalStemFilter.cs index e1c6df7..7aecc9f 100644 --- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/No/TestNorwegianMinimalStemFilter.cs +++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/No/TestNorwegianMinimalStemFilter.cs @@ -37,7 +37,7 @@ namespace Lucene.Net.Analysis.No { } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer source = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); return new TokenStreamComponents(source, new NorwegianMinimalStemFilter(source, NorwegianStandard.BOKMAAL)); @@ -70,7 +70,7 @@ namespace Lucene.Net.Analysis.No this.outerInstance = outerInstance; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer source = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); return new TokenStreamComponents(source, new NorwegianMinimalStemFilter(source, NorwegianStandard.NYNORSK)); @@ -97,7 +97,7 @@ namespace Lucene.Net.Analysis.No this.exclusionSet = exclusionSet; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer source = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); TokenStream sink = new SetKeywordMarkerFilter(source, exclusionSet); @@ -130,7 +130,7 @@ namespace Lucene.Net.Analysis.No this.outerInstance = outerInstance; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new KeywordTokenizer(reader); return new TokenStreamComponents(tokenizer, new NorwegianMinimalStemFilter(tokenizer)); http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Path/TestPathHierarchyTokenizer.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Path/TestPathHierarchyTokenizer.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Path/TestPathHierarchyTokenizer.cs index 6663527..d926e6c 100644 --- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Path/TestPathHierarchyTokenizer.cs +++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Path/TestPathHierarchyTokenizer.cs @@ -168,7 +168,7 @@ namespace Lucene.Net.Analysis.Path this.outerInstance = outerInstance; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new PathHierarchyTokenizer(reader); return new TokenStreamComponents(tokenizer, tokenizer); @@ -194,7 +194,7 @@ namespace Lucene.Net.Analysis.Path this.outerInstance = outerInstance; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new PathHierarchyTokenizer(reader); return new TokenStreamComponents(tokenizer, tokenizer); http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Path/TestReversePathHierarchyTokenizer.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Path/TestReversePathHierarchyTokenizer.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Path/TestReversePathHierarchyTokenizer.cs index 9c3ae0a..21ff553 100644 --- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Path/TestReversePathHierarchyTokenizer.cs +++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Path/TestReversePathHierarchyTokenizer.cs @@ -137,7 +137,7 @@ namespace Lucene.Net.Analysis.Path this.outerInstance = outerInstance; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new ReversePathHierarchyTokenizer(reader); return new TokenStreamComponents(tokenizer, tokenizer); @@ -163,7 +163,7 @@ namespace Lucene.Net.Analysis.Path this.outerInstance = outerInstance; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new ReversePathHierarchyTokenizer(reader); return new TokenStreamComponents(tokenizer, tokenizer); http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pattern/TestPatternCaptureGroupTokenFilter.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pattern/TestPatternCaptureGroupTokenFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pattern/TestPatternCaptureGroupTokenFilter.cs index 6ca751e..c69cb84 100644 --- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pattern/TestPatternCaptureGroupTokenFilter.cs +++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pattern/TestPatternCaptureGroupTokenFilter.cs @@ -192,7 +192,7 @@ namespace Lucene.Net.Analysis.Pattern this.outerInstance = outerInstance; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); return new TokenStreamComponents(tokenizer, new PatternCaptureGroupTokenFilter(tokenizer, false, new Regex("((..)(..))", RegexOptions.Compiled))); http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pattern/TestPatternReplaceCharFilter.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pattern/TestPatternReplaceCharFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pattern/TestPatternReplaceCharFilter.cs index 5e1b3ea..4976c7b 100644 --- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pattern/TestPatternReplaceCharFilter.cs +++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pattern/TestPatternReplaceCharFilter.cs @@ -280,12 +280,12 @@ namespace Lucene.Net.Analysis.Pattern this.replacement = replacement; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); return new TokenStreamComponents(tokenizer, tokenizer); } - protected internal override TextReader InitReader(string fieldName, TextReader reader) + protected override TextReader InitReader(string fieldName, TextReader reader) { return new PatternReplaceCharFilter(p, replacement, reader); } http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pattern/TestPatternReplaceFilter.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pattern/TestPatternReplaceFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pattern/TestPatternReplaceFilter.cs index bae6dbc..4ebe0f9 100644 --- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pattern/TestPatternReplaceFilter.cs +++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pattern/TestPatternReplaceFilter.cs @@ -87,7 +87,7 @@ namespace Lucene.Net.Analysis.Pattern this.outerInstance = outerInstance; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); TokenStream filter = new PatternReplaceFilter(tokenizer, new Regex("a", RegexOptions.Compiled), "b", false); @@ -104,7 +104,7 @@ namespace Lucene.Net.Analysis.Pattern this.outerInstance = outerInstance; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); TokenStream filter = new PatternReplaceFilter(tokenizer, new Regex("a", RegexOptions.Compiled), "b", true); @@ -128,7 +128,7 @@ namespace Lucene.Net.Analysis.Pattern this.outerInstance = outerInstance; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new KeywordTokenizer(reader); return new TokenStreamComponents(tokenizer, new PatternReplaceFilter(tokenizer, new Regex("a", RegexOptions.Compiled), "b", true)); http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pattern/TestPatternTokenizer.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pattern/TestPatternTokenizer.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pattern/TestPatternTokenizer.cs index 98a4062..f3c0c47 100644 --- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pattern/TestPatternTokenizer.cs +++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pattern/TestPatternTokenizer.cs @@ -135,7 +135,7 @@ namespace Lucene.Net.Analysis.Pattern this.outerInstance = outerInstance; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new PatternTokenizer(reader, new Regex("a", RegexOptions.Compiled), -1); return new TokenStreamComponents(tokenizer); @@ -151,7 +151,7 @@ namespace Lucene.Net.Analysis.Pattern this.outerInstance = outerInstance; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new PatternTokenizer(reader, new Regex("a", RegexOptions.Compiled), 0); return new TokenStreamComponents(tokenizer); http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pt/TestPortugueseLightStemFilter.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pt/TestPortugueseLightStemFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pt/TestPortugueseLightStemFilter.cs index 77641f9..14871c5 100644 --- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pt/TestPortugueseLightStemFilter.cs +++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pt/TestPortugueseLightStemFilter.cs @@ -37,7 +37,7 @@ namespace Lucene.Net.Analysis.Pt { } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer source = new StandardTokenizer(TEST_VERSION_CURRENT, reader); TokenStream result = new LowerCaseFilter(TEST_VERSION_CURRENT, source); @@ -111,7 +111,7 @@ namespace Lucene.Net.Analysis.Pt this.exclusionSet = exclusionSet; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer source = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); TokenStream sink = new SetKeywordMarkerFilter(source, exclusionSet); @@ -143,7 +143,7 @@ namespace Lucene.Net.Analysis.Pt this.outerInstance = outerInstance; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new KeywordTokenizer(reader); return new TokenStreamComponents(tokenizer, new PortugueseLightStemFilter(tokenizer)); http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pt/TestPortugueseMinimalStemFilter.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pt/TestPortugueseMinimalStemFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pt/TestPortugueseMinimalStemFilter.cs index 65bdc33..10bdf1c 100644 --- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pt/TestPortugueseMinimalStemFilter.cs +++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pt/TestPortugueseMinimalStemFilter.cs @@ -37,7 +37,7 @@ namespace Lucene.Net.Analysis.Pt { } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer source = new StandardTokenizer(TEST_VERSION_CURRENT, reader); TokenStream result = new LowerCaseFilter(TEST_VERSION_CURRENT, source); @@ -83,7 +83,7 @@ namespace Lucene.Net.Analysis.Pt this.exclusionSet = exclusionSet; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer source = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); TokenStream sink = new SetKeywordMarkerFilter(source, exclusionSet); @@ -115,7 +115,7 @@ namespace Lucene.Net.Analysis.Pt this.outerInstance = outerInstance; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new KeywordTokenizer(reader); return new TokenStreamComponents(tokenizer, new PortugueseMinimalStemFilter(tokenizer)); http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pt/TestPortugueseStemFilter.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pt/TestPortugueseStemFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pt/TestPortugueseStemFilter.cs index 12acb52..7dca23e 100644 --- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pt/TestPortugueseStemFilter.cs +++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pt/TestPortugueseStemFilter.cs @@ -37,7 +37,7 @@ namespace Lucene.Net.Analysis.Pt { } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer source = new StandardTokenizer(TEST_VERSION_CURRENT, reader); TokenStream result = new LowerCaseFilter(TEST_VERSION_CURRENT, source); @@ -83,7 +83,7 @@ namespace Lucene.Net.Analysis.Pt this.exclusionSet = exclusionSet; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer source = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); TokenStream sink = new SetKeywordMarkerFilter(source, exclusionSet); @@ -115,7 +115,7 @@ namespace Lucene.Net.Analysis.Pt this.outerInstance = outerInstance; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new KeywordTokenizer(reader); return new TokenStreamComponents(tokenizer, new PortugueseStemFilter(tokenizer)); http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Reverse/TestReverseStringFilter.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Reverse/TestReverseStringFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Reverse/TestReverseStringFilter.cs index cafde23..7dd43ed 100644 --- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Reverse/TestReverseStringFilter.cs +++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Reverse/TestReverseStringFilter.cs @@ -125,7 +125,7 @@ namespace Lucene.Net.Analysis.Reverse this.outerInstance = outerInstance; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); return new TokenStreamComponents(tokenizer, new ReverseStringFilter(TEST_VERSION_CURRENT, tokenizer)); @@ -148,7 +148,7 @@ namespace Lucene.Net.Analysis.Reverse this.outerInstance = outerInstance; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new KeywordTokenizer(reader); return new TokenStreamComponents(tokenizer, new ReverseStringFilter(TEST_VERSION_CURRENT, tokenizer)); http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Ru/TestRussianLightStemFilter.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Ru/TestRussianLightStemFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Ru/TestRussianLightStemFilter.cs index 2ed12ae..a1cae8a 100644 --- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Ru/TestRussianLightStemFilter.cs +++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Ru/TestRussianLightStemFilter.cs @@ -36,7 +36,7 @@ namespace Lucene.Net.Analysis.Ru { } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer source = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); return new TokenStreamComponents(source, new RussianLightStemFilter(source)); @@ -71,7 +71,7 @@ namespace Lucene.Net.Analysis.Ru this.exclusionSet = exclusionSet; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer source = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); TokenStream sink = new SetKeywordMarkerFilter(source, exclusionSet); @@ -103,7 +103,7 @@ namespace Lucene.Net.Analysis.Ru this.outerInstance = outerInstance; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new KeywordTokenizer(reader); return new TokenStreamComponents(tokenizer, new RussianLightStemFilter(tokenizer)); http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Shingle/ShingleAnalyzerWrapperTest.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Shingle/ShingleAnalyzerWrapperTest.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Shingle/ShingleAnalyzerWrapperTest.cs index e5caf56..d9e4f7d 100644 --- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Shingle/ShingleAnalyzerWrapperTest.cs +++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Shingle/ShingleAnalyzerWrapperTest.cs @@ -233,7 +233,7 @@ namespace Lucene.Net.Analysis.Shingle this.outerInstance = outerInstance; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { CharArraySet stopSet = StopFilter.MakeStopSet(TEST_VERSION_CURRENT, "into"); Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Shingle/ShingleFilterTest.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Shingle/ShingleFilterTest.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Shingle/ShingleFilterTest.cs index 5e8a315..da2dc92 100644 --- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Shingle/ShingleFilterTest.cs +++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Shingle/ShingleFilterTest.cs @@ -523,7 +523,7 @@ namespace Lucene.Net.Analysis.Shingle this.outerInstance = outerInstance; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); return new TokenStreamComponents(tokenizer, new ShingleFilter(tokenizer)); @@ -549,7 +549,7 @@ namespace Lucene.Net.Analysis.Shingle this.outerInstance = outerInstance; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); return new TokenStreamComponents(tokenizer, new ShingleFilter(tokenizer)); @@ -572,7 +572,7 @@ namespace Lucene.Net.Analysis.Shingle this.outerInstance = outerInstance; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new KeywordTokenizer(reader); return new TokenStreamComponents(tokenizer, new ShingleFilter(tokenizer)); http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Snowball/TestSnowball.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Snowball/TestSnowball.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Snowball/TestSnowball.cs index e48f489..7ea9af1 100644 --- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Snowball/TestSnowball.cs +++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Snowball/TestSnowball.cs @@ -186,7 +186,7 @@ namespace Lucene.Net.Analysis.Snowball this.lang = lang; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new KeywordTokenizer(reader); return new TokenStreamComponents(tokenizer, new SnowballFilter(tokenizer, lang)); @@ -220,7 +220,7 @@ namespace Lucene.Net.Analysis.Snowball this.snowballLanguage = snowballLanguage; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer t = new MockTokenizer(reader); return new TokenStreamComponents(t, new SnowballFilter(t, snowballLanguage)); http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Snowball/TestSnowballVocab.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Snowball/TestSnowballVocab.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Snowball/TestSnowballVocab.cs index d86813a..a29e54b 100644 --- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Snowball/TestSnowballVocab.cs +++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Snowball/TestSnowballVocab.cs @@ -87,7 +87,7 @@ namespace Lucene.Net.Analysis.Snowball this.snowballLanguage = snowballLanguage; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer t = new KeywordTokenizer(reader); return new TokenStreamComponents(t, new SnowballFilter(t, snowballLanguage)); http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Sv/TestSwedishLightStemFilter.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Sv/TestSwedishLightStemFilter.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Sv/TestSwedishLightStemFilter.cs index 4b55f5f..4961365 100644 --- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Sv/TestSwedishLightStemFilter.cs +++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Sv/TestSwedishLightStemFilter.cs @@ -36,7 +36,7 @@ namespace Lucene.Net.Analysis.Sv { } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer source = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); return new TokenStreamComponents(source, new SwedishLightStemFilter(source)); @@ -71,7 +71,7 @@ namespace Lucene.Net.Analysis.Sv this.exclusionSet = exclusionSet; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer source = new MockTokenizer(reader, MockTokenizer.WHITESPACE, false); TokenStream sink = new SetKeywordMarkerFilter(source, exclusionSet); @@ -103,7 +103,7 @@ namespace Lucene.Net.Analysis.Sv this.outerInstance = outerInstance; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new KeywordTokenizer(reader); return new TokenStreamComponents(tokenizer, new SwedishLightStemFilter(tokenizer)); http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Synonym/TestSolrSynonymParser.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Synonym/TestSolrSynonymParser.cs b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Synonym/TestSolrSynonymParser.cs index b9f2a74..da25523 100644 --- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Synonym/TestSolrSynonymParser.cs +++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Synonym/TestSolrSynonymParser.cs @@ -63,7 +63,7 @@ namespace Lucene.Net.Analysis.Synonym this.map = map; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.WHITESPACE, true); return new TokenStreamComponents(tokenizer, new SynonymFilter(tokenizer, map, true)); @@ -150,7 +150,7 @@ namespace Lucene.Net.Analysis.Synonym this.map = map; } - protected internal override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) + protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader) { Tokenizer tokenizer = new MockTokenizer(reader, MockTokenizer.KEYWORD, false); return new TokenStreamComponents(tokenizer, new SynonymFilter(tokenizer, map, false));
