http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Hu/TestHungarianLightStemFilter.cs
----------------------------------------------------------------------
diff --git 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Hu/TestHungarianLightStemFilter.cs
 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Hu/TestHungarianLightStemFilter.cs
index eeeef95..89ef132 100644
--- 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Hu/TestHungarianLightStemFilter.cs
+++ 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Hu/TestHungarianLightStemFilter.cs
@@ -36,7 +36,7 @@ namespace Lucene.Net.Analysis.Hu
             {
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer source = new MockTokenizer(reader, 
MockTokenizer.WHITESPACE, false);
                 return new TokenStreamComponents(source, new 
HungarianLightStemFilter(source));
@@ -71,7 +71,7 @@ namespace Lucene.Net.Analysis.Hu
                 this.exclusionSet = exclusionSet;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer source = new MockTokenizer(reader, 
MockTokenizer.WHITESPACE, false);
                 TokenStream sink = new SetKeywordMarkerFilter(source, 
exclusionSet);
@@ -95,7 +95,7 @@ namespace Lucene.Net.Analysis.Hu
                 this.outerInstance = outerInstance;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new KeywordTokenizer(reader);
                 return new TokenStreamComponents(tokenizer, new 
HungarianLightStemFilter(tokenizer));

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Hunspell/TestHunspellStemFilter.cs
----------------------------------------------------------------------
diff --git 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Hunspell/TestHunspellStemFilter.cs
 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Hunspell/TestHunspellStemFilter.cs
index 4ece2fb..07d3541 100644
--- 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Hunspell/TestHunspellStemFilter.cs
+++ 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Hunspell/TestHunspellStemFilter.cs
@@ -98,7 +98,7 @@ namespace Lucene.Net.Analysis.Hunspell
                 this.outerInstance = outerInstance;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new MockTokenizer(reader, 
MockTokenizer.WHITESPACE, false);
                 return new TokenStreamComponents(tokenizer, new 
HunspellStemFilter(tokenizer, dictionary));
@@ -121,7 +121,7 @@ namespace Lucene.Net.Analysis.Hunspell
                 this.outerInstance = outerInstance;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new KeywordTokenizer(reader);
                 return new TokenStreamComponents(tokenizer, new 
HunspellStemFilter(tokenizer, dictionary));
@@ -158,7 +158,7 @@ namespace Lucene.Net.Analysis.Hunspell
                 this.d = d;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new KeywordTokenizer(reader);
                 return new TokenStreamComponents(tokenizer, new 
HunspellStemFilter(tokenizer, d));

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Id/TestIndonesianStemmer.cs
----------------------------------------------------------------------
diff --git 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Id/TestIndonesianStemmer.cs 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Id/TestIndonesianStemmer.cs
index 73857df..e55eff0 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Id/TestIndonesianStemmer.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Id/TestIndonesianStemmer.cs
@@ -35,7 +35,7 @@ namespace Lucene.Net.Analysis.Id
             {
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new KeywordTokenizer(reader);
                 return new TokenStreamComponents(tokenizer, new 
IndonesianStemFilter(tokenizer));
@@ -129,7 +129,7 @@ namespace Lucene.Net.Analysis.Id
             {
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new KeywordTokenizer(reader);
                 return new TokenStreamComponents(tokenizer, new 
IndonesianStemFilter(tokenizer, false));
@@ -171,7 +171,7 @@ namespace Lucene.Net.Analysis.Id
                 this.outerInstance = outerInstance;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new KeywordTokenizer(reader);
                 return new TokenStreamComponents(tokenizer, new 
IndonesianStemFilter(tokenizer));

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/In/TestIndicNormalizer.cs
----------------------------------------------------------------------
diff --git 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/In/TestIndicNormalizer.cs 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/In/TestIndicNormalizer.cs
index 8294ffe..47731bf 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/In/TestIndicNormalizer.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/In/TestIndicNormalizer.cs
@@ -65,7 +65,7 @@ namespace Lucene.Net.Analysis.In
                 this.outerInstance = outerInstance;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new KeywordTokenizer(reader);
                 return new TokenStreamComponents(tokenizer, new 
IndicNormalizationFilter(tokenizer));

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/It/TestItalianLightStemFilter.cs
----------------------------------------------------------------------
diff --git 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/It/TestItalianLightStemFilter.cs
 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/It/TestItalianLightStemFilter.cs
index d4f0fd7..1bd67dc 100644
--- 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/It/TestItalianLightStemFilter.cs
+++ 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/It/TestItalianLightStemFilter.cs
@@ -34,7 +34,7 @@ namespace Lucene.Net.Analysis.It
             {
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer source = new MockTokenizer(reader, 
MockTokenizer.WHITESPACE, false);
                 return new TokenStreamComponents(source, new 
ItalianLightStemFilter(source));
@@ -73,7 +73,7 @@ namespace Lucene.Net.Analysis.It
                 this.outerInstance = outerInstance;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new KeywordTokenizer(reader);
                 return new TokenStreamComponents(tokenizer, new 
ItalianLightStemFilter(tokenizer));

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Lv/TestLatvianStemmer.cs
----------------------------------------------------------------------
diff --git 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Lv/TestLatvianStemmer.cs 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Lv/TestLatvianStemmer.cs
index 52adc34..5192510 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Lv/TestLatvianStemmer.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Lv/TestLatvianStemmer.cs
@@ -34,7 +34,7 @@ namespace Lucene.Net.Analysis.Lv
             {
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new MockTokenizer(reader, 
MockTokenizer.WHITESPACE, false);
                 return new TokenStreamComponents(tokenizer, new 
LatvianStemFilter(tokenizer));
@@ -307,7 +307,7 @@ namespace Lucene.Net.Analysis.Lv
                 this.outerInstance = outerInstance;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new KeywordTokenizer(reader);
                 return new TokenStreamComponents(tokenizer, new 
LatvianStemFilter(tokenizer));

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestASCIIFoldingFilter.cs
----------------------------------------------------------------------
diff --git 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestASCIIFoldingFilter.cs
 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestASCIIFoldingFilter.cs
index 2b7cd61..d4d246f 100644
--- 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestASCIIFoldingFilter.cs
+++ 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestASCIIFoldingFilter.cs
@@ -242,7 +242,7 @@ namespace Lucene.Net.Analysis.Miscellaneous
                 this.outerInstance = outerInstance;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new MockTokenizer(reader, 
MockTokenizer.WHITESPACE, false);
                 return new TokenStreamComponents(tokenizer, new 
ASCIIFoldingFilter(tokenizer, Random().nextBoolean()));
@@ -265,7 +265,7 @@ namespace Lucene.Net.Analysis.Miscellaneous
                 this.outerInstance = outerInstance;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new KeywordTokenizer(reader);
                 return new TokenStreamComponents(tokenizer, new 
ASCIIFoldingFilter(tokenizer, Random().nextBoolean()));

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestCapitalizationFilter.cs
----------------------------------------------------------------------
diff --git 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestCapitalizationFilter.cs
 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestCapitalizationFilter.cs
index a95c131..781bdee 100644
--- 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestCapitalizationFilter.cs
+++ 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestCapitalizationFilter.cs
@@ -109,7 +109,7 @@ namespace Lucene.Net.Analysis.Miscellaneous
                 this.outerInstance = outerInstance;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new MockTokenizer(reader, 
MockTokenizer.WHITESPACE, false);
                 return new TokenStreamComponents(tokenizer, new 
CapitalizationFilter(tokenizer));
@@ -132,7 +132,7 @@ namespace Lucene.Net.Analysis.Miscellaneous
                 this.outerInstance = outerInstance;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new KeywordTokenizer(reader);
                 return new TokenStreamComponents(tokenizer, new 
CapitalizationFilter(tokenizer));

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestCodepointCountFilter.cs
----------------------------------------------------------------------
diff --git 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestCodepointCountFilter.cs
 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestCodepointCountFilter.cs
index e06097d..a53bab8 100644
--- 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestCodepointCountFilter.cs
+++ 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestCodepointCountFilter.cs
@@ -50,7 +50,7 @@ namespace Lucene.Net.Analysis.Miscellaneous
                 this.outerInstance = outerInstance;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new KeywordTokenizer(reader);
                 return new TokenStreamComponents(tokenizer, new 
CodepointCountFilter(TEST_VERSION_CURRENT, tokenizer, 0, 5));

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestHyphenatedWordsFilter.cs
----------------------------------------------------------------------
diff --git 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestHyphenatedWordsFilter.cs
 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestHyphenatedWordsFilter.cs
index a04f8ed..a2819c4 100644
--- 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestHyphenatedWordsFilter.cs
+++ 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestHyphenatedWordsFilter.cs
@@ -77,7 +77,7 @@ namespace Lucene.Net.Analysis.Miscellaneous
                 this.outerInstance = outerInstance;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new MockTokenizer(reader, 
MockTokenizer.WHITESPACE, false);
                 return new TokenStreamComponents(tokenizer, new 
HyphenatedWordsFilter(tokenizer));
@@ -100,7 +100,7 @@ namespace Lucene.Net.Analysis.Miscellaneous
                 this.outerInstance = outerInstance;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new KeywordTokenizer(reader);
                 return new TokenStreamComponents(tokenizer, new 
HyphenatedWordsFilter(tokenizer));

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestKeepWordFilter.cs
----------------------------------------------------------------------
diff --git 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestKeepWordFilter.cs
 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestKeepWordFilter.cs
index f92e74f..c8148e2 100644
--- 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestKeepWordFilter.cs
+++ 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestKeepWordFilter.cs
@@ -86,7 +86,7 @@ namespace Lucene.Net.Analysis.Miscellaneous
                 this.words = words;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new MockTokenizer(reader, 
MockTokenizer.WHITESPACE, false);
                 TokenStream stream = new KeepWordFilter(TEST_VERSION_CURRENT, 
tokenizer, new CharArraySet(TEST_VERSION_CURRENT, words, true));

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestLengthFilter.cs
----------------------------------------------------------------------
diff --git 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestLengthFilter.cs
 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestLengthFilter.cs
index 3c55dc8..7180aed 100644
--- 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestLengthFilter.cs
+++ 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestLengthFilter.cs
@@ -60,7 +60,7 @@ namespace Lucene.Net.Analysis.Miscellaneous
                 this.outerInstance = outerInstance;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new KeywordTokenizer(reader);
                 return new TokenStreamComponents(tokenizer, new 
LengthFilter(TEST_VERSION_CURRENT, tokenizer, 0, 5));

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestLimitTokenPositionFilter.cs
----------------------------------------------------------------------
diff --git 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestLimitTokenPositionFilter.cs
 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestLimitTokenPositionFilter.cs
index 3b4c014..c3afa22 100644
--- 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestLimitTokenPositionFilter.cs
+++ 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestLimitTokenPositionFilter.cs
@@ -55,7 +55,7 @@ namespace Lucene.Net.Analysis.Miscellaneous
                 this.consumeAll = consumeAll;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 MockTokenizer tokenizer = new MockTokenizer(reader, 
MockTokenizer.WHITESPACE, false);
                 // if we are consuming all tokens, we can use the checks, 
otherwise we can't

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestLucene47WordDelimiterFilter.cs
----------------------------------------------------------------------
diff --git 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestLucene47WordDelimiterFilter.cs
 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestLucene47WordDelimiterFilter.cs
index 3d97c0d..48d76d7 100644
--- 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestLucene47WordDelimiterFilter.cs
+++ 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestLucene47WordDelimiterFilter.cs
@@ -293,7 +293,7 @@ namespace Lucene.Net.Analysis.Miscellaneous
                 this.protWords = protWords;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new MockTokenizer(reader, 
MockTokenizer.WHITESPACE, false);
                 return new TokenStreamComponents(tokenizer, new 
Lucene47WordDelimiterFilter(tokenizer, flags, protWords));
@@ -314,7 +314,7 @@ namespace Lucene.Net.Analysis.Miscellaneous
                 this.protWords = protWords;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new MockTokenizer(reader, 
MockTokenizer.WHITESPACE, false);
                 return new TokenStreamComponents(tokenizer, new 
Lucene47WordDelimiterFilter(new LargePosIncTokenFilter(outerInstance, 
tokenizer), flags, protWords));
@@ -335,7 +335,7 @@ namespace Lucene.Net.Analysis.Miscellaneous
                 this.protWords = protWords;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new MockTokenizer(reader, 
MockTokenizer.WHITESPACE, false);
                 StopFilter filter = new StopFilter(TEST_VERSION_CURRENT, 
tokenizer, StandardAnalyzer.STOP_WORDS_SET);
@@ -381,7 +381,7 @@ namespace Lucene.Net.Analysis.Miscellaneous
                 this.protectedWords = protectedWords;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new MockTokenizer(reader, 
MockTokenizer.WHITESPACE, false);
                 return new TokenStreamComponents(tokenizer, new 
Lucene47WordDelimiterFilter(tokenizer, flags, protectedWords));
@@ -425,7 +425,7 @@ namespace Lucene.Net.Analysis.Miscellaneous
                 this.protectedWords = protectedWords;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new KeywordTokenizer(reader);
                 return new TokenStreamComponents(tokenizer, new 
Lucene47WordDelimiterFilter(tokenizer, flags, protectedWords));

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestPerFieldAnalyzerWrapper.cs
----------------------------------------------------------------------
diff --git 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestPerFieldAnalyzerWrapper.cs
 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestPerFieldAnalyzerWrapper.cs
index e3aa88e..7429429 100644
--- 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestPerFieldAnalyzerWrapper.cs
+++ 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestPerFieldAnalyzerWrapper.cs
@@ -90,12 +90,12 @@ namespace Lucene.Net.Analysis.Miscellaneous
                 this.outerInstance = outerInstance;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 return new TokenStreamComponents(new MockTokenizer(reader));
             }
 
-            protected internal override TextReader InitReader(string 
fieldName, TextReader reader)
+            protected override TextReader InitReader(string fieldName, 
TextReader reader)
             {
                 return new MockCharFilter(reader, 7);
             }

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestRemoveDuplicatesTokenFilter.cs
----------------------------------------------------------------------
diff --git 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestRemoveDuplicatesTokenFilter.cs
 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestRemoveDuplicatesTokenFilter.cs
index 5eb577b..7634702 100644
--- 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestRemoveDuplicatesTokenFilter.cs
+++ 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestRemoveDuplicatesTokenFilter.cs
@@ -171,7 +171,7 @@ namespace Lucene.Net.Analysis.Miscellaneous
                 this.ignoreCase = ignoreCase;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new MockTokenizer(reader, 
MockTokenizer.SIMPLE, true);
                 TokenStream stream = new SynonymFilter(tokenizer, map, 
ignoreCase);
@@ -195,7 +195,7 @@ namespace Lucene.Net.Analysis.Miscellaneous
                 this.outerInstance = outerInstance;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new KeywordTokenizer(reader);
                 return new TokenStreamComponents(tokenizer, new 
RemoveDuplicatesTokenFilter(tokenizer));

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestScandinavianFoldingFilter.cs
----------------------------------------------------------------------
diff --git 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestScandinavianFoldingFilter.cs
 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestScandinavianFoldingFilter.cs
index fcbbee9..9c1f248 100644
--- 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestScandinavianFoldingFilter.cs
+++ 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestScandinavianFoldingFilter.cs
@@ -31,7 +31,7 @@ namespace Lucene.Net.Analysis.Miscellaneous
             {
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new MockTokenizer(reader, 
MockTokenizer.WHITESPACE, false);
                 TokenStream stream = new ScandinavianFoldingFilter(tokenizer);
@@ -128,7 +128,7 @@ namespace Lucene.Net.Analysis.Miscellaneous
                 this.outerInstance = outerInstance;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new KeywordTokenizer(reader);
                 return new TokenStreamComponents(tokenizer, new 
ScandinavianFoldingFilter(tokenizer));

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestScandinavianNormalizationFilter.cs
----------------------------------------------------------------------
diff --git 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestScandinavianNormalizationFilter.cs
 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestScandinavianNormalizationFilter.cs
index 597d3f0..369d28d 100644
--- 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestScandinavianNormalizationFilter.cs
+++ 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestScandinavianNormalizationFilter.cs
@@ -33,7 +33,7 @@ namespace Lucene.Net.Analysis.Miscellaneous
             {
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new MockTokenizer(reader, 
MockTokenizer.WHITESPACE, false);
                 TokenStream stream = new 
ScandinavianNormalizationFilter(tokenizer);
@@ -128,7 +128,7 @@ namespace Lucene.Net.Analysis.Miscellaneous
                 this.outerInstance = outerInstance;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new KeywordTokenizer(reader);
                 return new TokenStreamComponents(tokenizer, new 
ScandinavianNormalizationFilter(tokenizer));

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestTrimFilter.cs
----------------------------------------------------------------------
diff --git 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestTrimFilter.cs 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestTrimFilter.cs
index c61e5b1..5b0d900 100644
--- 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestTrimFilter.cs
+++ 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestTrimFilter.cs
@@ -125,7 +125,7 @@ namespace Lucene.Net.Analysis.Miscellaneous
                 this.outerInstance = outerInstance;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new MockTokenizer(reader, 
MockTokenizer.KEYWORD, false);
                 return new TokenStreamComponents(tokenizer, new 
TrimFilter(LuceneVersion.LUCENE_43, tokenizer, true));
@@ -141,7 +141,7 @@ namespace Lucene.Net.Analysis.Miscellaneous
                 this.outerInstance = outerInstance;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new MockTokenizer(reader, 
MockTokenizer.KEYWORD, false);
                 return new TokenStreamComponents(tokenizer, new 
TrimFilter(TEST_VERSION_CURRENT, tokenizer, false));
@@ -164,7 +164,7 @@ namespace Lucene.Net.Analysis.Miscellaneous
                 this.outerInstance = outerInstance;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new KeywordTokenizer(reader);
                 bool updateOffsets = Random().nextBoolean();

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestWordDelimiterFilter.cs
----------------------------------------------------------------------
diff --git 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestWordDelimiterFilter.cs
 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestWordDelimiterFilter.cs
index 530682a..df9df2b 100644
--- 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestWordDelimiterFilter.cs
+++ 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Miscellaneous/TestWordDelimiterFilter.cs
@@ -287,7 +287,7 @@ namespace Lucene.Net.Analysis.Miscellaneous
                 this.protWords = protWords;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string field, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
field, TextReader reader)
             {
                 Tokenizer tokenizer = new MockTokenizer(reader, 
MockTokenizer.WHITESPACE, false);
                 return new TokenStreamComponents(tokenizer, new 
WordDelimiterFilter(TEST_VERSION_CURRENT, tokenizer, flags, protWords));
@@ -308,7 +308,7 @@ namespace Lucene.Net.Analysis.Miscellaneous
                 this.protWords = protWords;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string field, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
field, TextReader reader)
             {
                 Tokenizer tokenizer = new MockTokenizer(reader, 
MockTokenizer.WHITESPACE, false);
                 return new TokenStreamComponents(tokenizer, new 
WordDelimiterFilter(TEST_VERSION_CURRENT, new 
LargePosIncTokenFilter(outerInstance, tokenizer), flags, protWords));
@@ -329,7 +329,7 @@ namespace Lucene.Net.Analysis.Miscellaneous
                 this.protWords = protWords;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string field, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
field, TextReader reader)
             {
                 Tokenizer tokenizer = new MockTokenizer(reader, 
MockTokenizer.WHITESPACE, false);
                 StopFilter filter = new StopFilter(TEST_VERSION_CURRENT, 
tokenizer, StandardAnalyzer.STOP_WORDS_SET);
@@ -369,7 +369,7 @@ namespace Lucene.Net.Analysis.Miscellaneous
                 this.flags = flags;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string field, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
field, TextReader reader)
             {
                 Tokenizer tokenizer = new MockTokenizer(reader, 
MockTokenizer.WHITESPACE, false);
                 return new TokenStreamComponents(tokenizer, new 
WordDelimiterFilter(TEST_VERSION_CURRENT, tokenizer, flags, null));
@@ -409,7 +409,7 @@ namespace Lucene.Net.Analysis.Miscellaneous
                 this.flags = flags;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string field, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
field, TextReader reader)
             {
                 Tokenizer tokenizer = new MockTokenizer(reader, 
MockTokenizer.WHITESPACE, false);
                 return new TokenStreamComponents(tokenizer, new 
WordDelimiterFilter(TEST_VERSION_CURRENT, tokenizer, flags, null));
@@ -453,7 +453,7 @@ namespace Lucene.Net.Analysis.Miscellaneous
                 this.protectedWords = protectedWords;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new MockTokenizer(reader, 
MockTokenizer.WHITESPACE, false);
                 return new TokenStreamComponents(tokenizer, new 
WordDelimiterFilter(TEST_VERSION_CURRENT, tokenizer, flags, protectedWords));
@@ -499,7 +499,7 @@ namespace Lucene.Net.Analysis.Miscellaneous
             }
 
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new MockTokenizer(reader, 
MockTokenizer.WHITESPACE, false);
                 return new TokenStreamComponents(tokenizer, new 
WordDelimiterFilter(TEST_VERSION_CURRENT, tokenizer, flags, protectedWords));
@@ -543,7 +543,7 @@ namespace Lucene.Net.Analysis.Miscellaneous
                 this.protectedWords = protectedWords;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new KeywordTokenizer(reader);
                 return new TokenStreamComponents(tokenizer, new 
WordDelimiterFilter(TEST_VERSION_CURRENT, tokenizer, flags, protectedWords));

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/NGram/EdgeNGramTokenFilterTest.cs
----------------------------------------------------------------------
diff --git 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/NGram/EdgeNGramTokenFilterTest.cs
 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/NGram/EdgeNGramTokenFilterTest.cs
index ea6fbd7..bfdd72e 100644
--- 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/NGram/EdgeNGramTokenFilterTest.cs
+++ 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/NGram/EdgeNGramTokenFilterTest.cs
@@ -240,7 +240,7 @@ namespace Lucene.Net.Analysis.NGram
                 this.outerInstance = outerInstance;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new MockTokenizer(reader, 
MockTokenizer.WHITESPACE, false);
                 TokenFilter filters = new ASCIIFoldingFilter(tokenizer);
@@ -283,7 +283,7 @@ namespace Lucene.Net.Analysis.NGram
                 this.max = max;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new MockTokenizer(reader, 
MockTokenizer.WHITESPACE, false);
                 return new TokenStreamComponents(tokenizer, new 
EdgeNGramTokenFilter(TEST_VERSION_CURRENT, tokenizer, min, max));
@@ -299,7 +299,7 @@ namespace Lucene.Net.Analysis.NGram
                 this.outerInstance = outerInstance;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new MockTokenizer(reader, 
MockTokenizer.WHITESPACE, false);
 #pragma warning disable 612, 618
@@ -328,7 +328,7 @@ namespace Lucene.Net.Analysis.NGram
                 this.outerInstance = outerInstance;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new KeywordTokenizer(reader);
 #pragma warning disable 612, 618
@@ -346,7 +346,7 @@ namespace Lucene.Net.Analysis.NGram
                 this.outerInstance = outerInstance;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new KeywordTokenizer(reader);
 #pragma warning disable 612, 618

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/NGram/EdgeNGramTokenizerTest.cs
----------------------------------------------------------------------
diff --git 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/NGram/EdgeNGramTokenizerTest.cs 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/NGram/EdgeNGramTokenizerTest.cs
index 6cddafe..b9248fb 100644
--- 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/NGram/EdgeNGramTokenizerTest.cs
+++ 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/NGram/EdgeNGramTokenizerTest.cs
@@ -164,7 +164,7 @@ namespace Lucene.Net.Analysis.NGram
                 this.max = max;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, Reader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, Reader reader)
             {
                 Tokenizer tokenizer = new 
EdgeNGramTokenizer(TEST_VERSION_CURRENT, reader, min, max);
                 return new TokenStreamComponents(tokenizer, tokenizer);
@@ -180,7 +180,7 @@ namespace Lucene.Net.Analysis.NGram
                 this.outerInstance = outerInstance;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, Reader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, Reader reader)
             {
 #pragma warning disable 612, 618
                 Tokenizer tokenizer = new 
Lucene43EdgeNGramTokenizer(Version.LUCENE_43, reader, 
Lucene43EdgeNGramTokenizer.Side.BACK, 2, 4);

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/NGram/NGramTokenFilterTest.cs
----------------------------------------------------------------------
diff --git 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/NGram/NGramTokenFilterTest.cs 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/NGram/NGramTokenFilterTest.cs
index e485fc0..5d9fee4 100644
--- 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/NGram/NGramTokenFilterTest.cs
+++ 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/NGram/NGramTokenFilterTest.cs
@@ -142,7 +142,7 @@ namespace Lucene.Net.Analysis.NGram
                 this.outerInstance = outerInstance;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new MockTokenizer(reader, 
MockTokenizer.WHITESPACE, false);
                 TokenFilter filters = new ASCIIFoldingFilter(tokenizer);
@@ -179,7 +179,7 @@ namespace Lucene.Net.Analysis.NGram
                 this.max = max;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new MockTokenizer(reader, 
MockTokenizer.WHITESPACE, false);
                 return new TokenStreamComponents(tokenizer, new 
NGramTokenFilter(TEST_VERSION_CURRENT, tokenizer, min, max));
@@ -203,7 +203,7 @@ namespace Lucene.Net.Analysis.NGram
                 this.outerInstance = outerInstance;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new KeywordTokenizer(reader);
                 return new TokenStreamComponents(tokenizer, new 
NGramTokenFilter(TEST_VERSION_CURRENT, tokenizer, 2, 15));

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/NGram/NGramTokenizerTest.cs
----------------------------------------------------------------------
diff --git 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/NGram/NGramTokenizerTest.cs 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/NGram/NGramTokenizerTest.cs
index 2fc1356..b8db76e 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/NGram/NGramTokenizerTest.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/NGram/NGramTokenizerTest.cs
@@ -132,7 +132,7 @@ namespace Lucene.Net.Analysis.NGram
                 this.max = max;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new NGramTokenizer(TEST_VERSION_CURRENT, 
reader, min, max);
                 return new TokenStreamComponents(tokenizer, tokenizer);

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/No/TestNorwegianLightStemFilter.cs
----------------------------------------------------------------------
diff --git 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/No/TestNorwegianLightStemFilter.cs
 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/No/TestNorwegianLightStemFilter.cs
index cedb8c4..0b3e183 100644
--- 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/No/TestNorwegianLightStemFilter.cs
+++ 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/No/TestNorwegianLightStemFilter.cs
@@ -37,7 +37,7 @@ namespace Lucene.Net.Analysis.No
             {
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer source = new MockTokenizer(reader, 
MockTokenizer.WHITESPACE, false);
                 return new TokenStreamComponents(source, new 
NorwegianLightStemFilter(source, NorwegianStandard.BOKMAAL));
@@ -70,7 +70,7 @@ namespace Lucene.Net.Analysis.No
                 this.outerInstance = outerInstance;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer source = new MockTokenizer(reader, 
MockTokenizer.WHITESPACE, false);
                 return new TokenStreamComponents(source, new 
NorwegianLightStemFilter(source, NorwegianStandard.NYNORSK));
@@ -97,7 +97,7 @@ namespace Lucene.Net.Analysis.No
                 this.exclusionSet = exclusionSet;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer source = new MockTokenizer(reader, 
MockTokenizer.WHITESPACE, false);
                 TokenStream sink = new SetKeywordMarkerFilter(source, 
exclusionSet);
@@ -130,7 +130,7 @@ namespace Lucene.Net.Analysis.No
                 this.outerInstance = outerInstance;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new KeywordTokenizer(reader);
                 return new TokenStreamComponents(tokenizer, new 
NorwegianLightStemFilter(tokenizer));

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/No/TestNorwegianMinimalStemFilter.cs
----------------------------------------------------------------------
diff --git 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/No/TestNorwegianMinimalStemFilter.cs
 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/No/TestNorwegianMinimalStemFilter.cs
index e1c6df7..7aecc9f 100644
--- 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/No/TestNorwegianMinimalStemFilter.cs
+++ 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/No/TestNorwegianMinimalStemFilter.cs
@@ -37,7 +37,7 @@ namespace Lucene.Net.Analysis.No
             {
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer source = new MockTokenizer(reader, 
MockTokenizer.WHITESPACE, false);
                 return new TokenStreamComponents(source, new 
NorwegianMinimalStemFilter(source, NorwegianStandard.BOKMAAL));
@@ -70,7 +70,7 @@ namespace Lucene.Net.Analysis.No
                 this.outerInstance = outerInstance;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer source = new MockTokenizer(reader, 
MockTokenizer.WHITESPACE, false);
                 return new TokenStreamComponents(source, new 
NorwegianMinimalStemFilter(source, NorwegianStandard.NYNORSK));
@@ -97,7 +97,7 @@ namespace Lucene.Net.Analysis.No
                 this.exclusionSet = exclusionSet;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer source = new MockTokenizer(reader, 
MockTokenizer.WHITESPACE, false);
                 TokenStream sink = new SetKeywordMarkerFilter(source, 
exclusionSet);
@@ -130,7 +130,7 @@ namespace Lucene.Net.Analysis.No
                 this.outerInstance = outerInstance;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new KeywordTokenizer(reader);
                 return new TokenStreamComponents(tokenizer, new 
NorwegianMinimalStemFilter(tokenizer));

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Path/TestPathHierarchyTokenizer.cs
----------------------------------------------------------------------
diff --git 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Path/TestPathHierarchyTokenizer.cs
 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Path/TestPathHierarchyTokenizer.cs
index 6663527..d926e6c 100644
--- 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Path/TestPathHierarchyTokenizer.cs
+++ 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Path/TestPathHierarchyTokenizer.cs
@@ -168,7 +168,7 @@ namespace Lucene.Net.Analysis.Path
                 this.outerInstance = outerInstance;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new PathHierarchyTokenizer(reader);
                 return new TokenStreamComponents(tokenizer, tokenizer);
@@ -194,7 +194,7 @@ namespace Lucene.Net.Analysis.Path
                 this.outerInstance = outerInstance;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new PathHierarchyTokenizer(reader);
                 return new TokenStreamComponents(tokenizer, tokenizer);

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Path/TestReversePathHierarchyTokenizer.cs
----------------------------------------------------------------------
diff --git 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Path/TestReversePathHierarchyTokenizer.cs
 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Path/TestReversePathHierarchyTokenizer.cs
index 9c3ae0a..21ff553 100644
--- 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Path/TestReversePathHierarchyTokenizer.cs
+++ 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Path/TestReversePathHierarchyTokenizer.cs
@@ -137,7 +137,7 @@ namespace Lucene.Net.Analysis.Path
                 this.outerInstance = outerInstance;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new 
ReversePathHierarchyTokenizer(reader);
                 return new TokenStreamComponents(tokenizer, tokenizer);
@@ -163,7 +163,7 @@ namespace Lucene.Net.Analysis.Path
                 this.outerInstance = outerInstance;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new 
ReversePathHierarchyTokenizer(reader);
                 return new TokenStreamComponents(tokenizer, tokenizer);

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pattern/TestPatternCaptureGroupTokenFilter.cs
----------------------------------------------------------------------
diff --git 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pattern/TestPatternCaptureGroupTokenFilter.cs
 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pattern/TestPatternCaptureGroupTokenFilter.cs
index 6ca751e..c69cb84 100644
--- 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pattern/TestPatternCaptureGroupTokenFilter.cs
+++ 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pattern/TestPatternCaptureGroupTokenFilter.cs
@@ -192,7 +192,7 @@ namespace Lucene.Net.Analysis.Pattern
                 this.outerInstance = outerInstance;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new MockTokenizer(reader, 
MockTokenizer.WHITESPACE, false);
                 return new TokenStreamComponents(tokenizer, new 
PatternCaptureGroupTokenFilter(tokenizer, false, new Regex("((..)(..))", 
RegexOptions.Compiled)));

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pattern/TestPatternReplaceCharFilter.cs
----------------------------------------------------------------------
diff --git 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pattern/TestPatternReplaceCharFilter.cs
 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pattern/TestPatternReplaceCharFilter.cs
index 5e1b3ea..4976c7b 100644
--- 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pattern/TestPatternReplaceCharFilter.cs
+++ 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pattern/TestPatternReplaceCharFilter.cs
@@ -280,12 +280,12 @@ namespace Lucene.Net.Analysis.Pattern
                 this.replacement = replacement;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new MockTokenizer(reader, 
MockTokenizer.WHITESPACE, false);
                 return new TokenStreamComponents(tokenizer, tokenizer);
             }
-            protected internal override TextReader InitReader(string 
fieldName, TextReader reader)
+            protected override TextReader InitReader(string fieldName, 
TextReader reader)
             {
                 return new PatternReplaceCharFilter(p, replacement, reader);
             }

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pattern/TestPatternReplaceFilter.cs
----------------------------------------------------------------------
diff --git 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pattern/TestPatternReplaceFilter.cs
 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pattern/TestPatternReplaceFilter.cs
index bae6dbc..4ebe0f9 100644
--- 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pattern/TestPatternReplaceFilter.cs
+++ 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pattern/TestPatternReplaceFilter.cs
@@ -87,7 +87,7 @@ namespace Lucene.Net.Analysis.Pattern
                 this.outerInstance = outerInstance;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new MockTokenizer(reader, 
MockTokenizer.WHITESPACE, false);
                 TokenStream filter = new PatternReplaceFilter(tokenizer, new 
Regex("a", RegexOptions.Compiled), "b", false);
@@ -104,7 +104,7 @@ namespace Lucene.Net.Analysis.Pattern
                 this.outerInstance = outerInstance;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new MockTokenizer(reader, 
MockTokenizer.WHITESPACE, false);
                 TokenStream filter = new PatternReplaceFilter(tokenizer, new 
Regex("a", RegexOptions.Compiled), "b", true);
@@ -128,7 +128,7 @@ namespace Lucene.Net.Analysis.Pattern
                 this.outerInstance = outerInstance;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new KeywordTokenizer(reader);
                 return new TokenStreamComponents(tokenizer, new 
PatternReplaceFilter(tokenizer, new Regex("a", RegexOptions.Compiled), "b", 
true));

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pattern/TestPatternTokenizer.cs
----------------------------------------------------------------------
diff --git 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pattern/TestPatternTokenizer.cs 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pattern/TestPatternTokenizer.cs
index 98a4062..f3c0c47 100644
--- 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pattern/TestPatternTokenizer.cs
+++ 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pattern/TestPatternTokenizer.cs
@@ -135,7 +135,7 @@ namespace Lucene.Net.Analysis.Pattern
                 this.outerInstance = outerInstance;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new PatternTokenizer(reader, new 
Regex("a", RegexOptions.Compiled), -1);
                 return new TokenStreamComponents(tokenizer);
@@ -151,7 +151,7 @@ namespace Lucene.Net.Analysis.Pattern
                 this.outerInstance = outerInstance;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new PatternTokenizer(reader, new 
Regex("a", RegexOptions.Compiled), 0);
                 return new TokenStreamComponents(tokenizer);

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pt/TestPortugueseLightStemFilter.cs
----------------------------------------------------------------------
diff --git 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pt/TestPortugueseLightStemFilter.cs
 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pt/TestPortugueseLightStemFilter.cs
index 77641f9..14871c5 100644
--- 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pt/TestPortugueseLightStemFilter.cs
+++ 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pt/TestPortugueseLightStemFilter.cs
@@ -37,7 +37,7 @@ namespace Lucene.Net.Analysis.Pt
             {
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer source = new StandardTokenizer(TEST_VERSION_CURRENT, 
reader);
                 TokenStream result = new LowerCaseFilter(TEST_VERSION_CURRENT, 
source);
@@ -111,7 +111,7 @@ namespace Lucene.Net.Analysis.Pt
                 this.exclusionSet = exclusionSet;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer source = new MockTokenizer(reader, 
MockTokenizer.WHITESPACE, false);
                 TokenStream sink = new SetKeywordMarkerFilter(source, 
exclusionSet);
@@ -143,7 +143,7 @@ namespace Lucene.Net.Analysis.Pt
                 this.outerInstance = outerInstance;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new KeywordTokenizer(reader);
                 return new TokenStreamComponents(tokenizer, new 
PortugueseLightStemFilter(tokenizer));

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pt/TestPortugueseMinimalStemFilter.cs
----------------------------------------------------------------------
diff --git 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pt/TestPortugueseMinimalStemFilter.cs
 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pt/TestPortugueseMinimalStemFilter.cs
index 65bdc33..10bdf1c 100644
--- 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pt/TestPortugueseMinimalStemFilter.cs
+++ 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pt/TestPortugueseMinimalStemFilter.cs
@@ -37,7 +37,7 @@ namespace Lucene.Net.Analysis.Pt
             {
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer source = new StandardTokenizer(TEST_VERSION_CURRENT, 
reader);
                 TokenStream result = new LowerCaseFilter(TEST_VERSION_CURRENT, 
source);
@@ -83,7 +83,7 @@ namespace Lucene.Net.Analysis.Pt
                 this.exclusionSet = exclusionSet;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer source = new MockTokenizer(reader, 
MockTokenizer.WHITESPACE, false);
                 TokenStream sink = new SetKeywordMarkerFilter(source, 
exclusionSet);
@@ -115,7 +115,7 @@ namespace Lucene.Net.Analysis.Pt
                 this.outerInstance = outerInstance;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new KeywordTokenizer(reader);
                 return new TokenStreamComponents(tokenizer, new 
PortugueseMinimalStemFilter(tokenizer));

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pt/TestPortugueseStemFilter.cs
----------------------------------------------------------------------
diff --git 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pt/TestPortugueseStemFilter.cs 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pt/TestPortugueseStemFilter.cs
index 12acb52..7dca23e 100644
--- 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pt/TestPortugueseStemFilter.cs
+++ 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Pt/TestPortugueseStemFilter.cs
@@ -37,7 +37,7 @@ namespace Lucene.Net.Analysis.Pt
             {
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer source = new StandardTokenizer(TEST_VERSION_CURRENT, 
reader);
                 TokenStream result = new LowerCaseFilter(TEST_VERSION_CURRENT, 
source);
@@ -83,7 +83,7 @@ namespace Lucene.Net.Analysis.Pt
                 this.exclusionSet = exclusionSet;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer source = new MockTokenizer(reader, 
MockTokenizer.WHITESPACE, false);
                 TokenStream sink = new SetKeywordMarkerFilter(source, 
exclusionSet);
@@ -115,7 +115,7 @@ namespace Lucene.Net.Analysis.Pt
                 this.outerInstance = outerInstance;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new KeywordTokenizer(reader);
                 return new TokenStreamComponents(tokenizer, new 
PortugueseStemFilter(tokenizer));

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Reverse/TestReverseStringFilter.cs
----------------------------------------------------------------------
diff --git 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Reverse/TestReverseStringFilter.cs
 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Reverse/TestReverseStringFilter.cs
index cafde23..7dd43ed 100644
--- 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Reverse/TestReverseStringFilter.cs
+++ 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Reverse/TestReverseStringFilter.cs
@@ -125,7 +125,7 @@ namespace Lucene.Net.Analysis.Reverse
                 this.outerInstance = outerInstance;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new MockTokenizer(reader, 
MockTokenizer.WHITESPACE, false);
                 return new TokenStreamComponents(tokenizer, new 
ReverseStringFilter(TEST_VERSION_CURRENT, tokenizer));
@@ -148,7 +148,7 @@ namespace Lucene.Net.Analysis.Reverse
                 this.outerInstance = outerInstance;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new KeywordTokenizer(reader);
                 return new TokenStreamComponents(tokenizer, new 
ReverseStringFilter(TEST_VERSION_CURRENT, tokenizer));

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Ru/TestRussianLightStemFilter.cs
----------------------------------------------------------------------
diff --git 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Ru/TestRussianLightStemFilter.cs
 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Ru/TestRussianLightStemFilter.cs
index 2ed12ae..a1cae8a 100644
--- 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Ru/TestRussianLightStemFilter.cs
+++ 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Ru/TestRussianLightStemFilter.cs
@@ -36,7 +36,7 @@ namespace Lucene.Net.Analysis.Ru
             {
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer source = new MockTokenizer(reader, 
MockTokenizer.WHITESPACE, false);
                 return new TokenStreamComponents(source, new 
RussianLightStemFilter(source));
@@ -71,7 +71,7 @@ namespace Lucene.Net.Analysis.Ru
                 this.exclusionSet = exclusionSet;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer source = new MockTokenizer(reader, 
MockTokenizer.WHITESPACE, false);
                 TokenStream sink = new SetKeywordMarkerFilter(source, 
exclusionSet);
@@ -103,7 +103,7 @@ namespace Lucene.Net.Analysis.Ru
                 this.outerInstance = outerInstance;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new KeywordTokenizer(reader);
                 return new TokenStreamComponents(tokenizer, new 
RussianLightStemFilter(tokenizer));

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Shingle/ShingleAnalyzerWrapperTest.cs
----------------------------------------------------------------------
diff --git 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Shingle/ShingleAnalyzerWrapperTest.cs
 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Shingle/ShingleAnalyzerWrapperTest.cs
index e5caf56..d9e4f7d 100644
--- 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Shingle/ShingleAnalyzerWrapperTest.cs
+++ 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Shingle/ShingleAnalyzerWrapperTest.cs
@@ -233,7 +233,7 @@ namespace Lucene.Net.Analysis.Shingle
                 this.outerInstance = outerInstance;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 CharArraySet stopSet = 
StopFilter.MakeStopSet(TEST_VERSION_CURRENT, "into");
                 Tokenizer tokenizer = new MockTokenizer(reader, 
MockTokenizer.WHITESPACE, false);

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Shingle/ShingleFilterTest.cs
----------------------------------------------------------------------
diff --git 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Shingle/ShingleFilterTest.cs 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Shingle/ShingleFilterTest.cs
index 5e8a315..da2dc92 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Shingle/ShingleFilterTest.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Shingle/ShingleFilterTest.cs
@@ -523,7 +523,7 @@ namespace Lucene.Net.Analysis.Shingle
                 this.outerInstance = outerInstance;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new MockTokenizer(reader, 
MockTokenizer.WHITESPACE, false);
                 return new TokenStreamComponents(tokenizer, new 
ShingleFilter(tokenizer));
@@ -549,7 +549,7 @@ namespace Lucene.Net.Analysis.Shingle
                 this.outerInstance = outerInstance;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new MockTokenizer(reader, 
MockTokenizer.WHITESPACE, false);
                 return new TokenStreamComponents(tokenizer, new 
ShingleFilter(tokenizer));
@@ -572,7 +572,7 @@ namespace Lucene.Net.Analysis.Shingle
                 this.outerInstance = outerInstance;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new KeywordTokenizer(reader);
                 return new TokenStreamComponents(tokenizer, new 
ShingleFilter(tokenizer));

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Snowball/TestSnowball.cs
----------------------------------------------------------------------
diff --git 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Snowball/TestSnowball.cs 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Snowball/TestSnowball.cs
index e48f489..7ea9af1 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Snowball/TestSnowball.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Snowball/TestSnowball.cs
@@ -186,7 +186,7 @@ namespace Lucene.Net.Analysis.Snowball
                 this.lang = lang;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new KeywordTokenizer(reader);
                 return new TokenStreamComponents(tokenizer, new 
SnowballFilter(tokenizer, lang));
@@ -220,7 +220,7 @@ namespace Lucene.Net.Analysis.Snowball
                 this.snowballLanguage = snowballLanguage;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer t = new MockTokenizer(reader);
                 return new TokenStreamComponents(t, new SnowballFilter(t, 
snowballLanguage));

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Snowball/TestSnowballVocab.cs
----------------------------------------------------------------------
diff --git 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Snowball/TestSnowballVocab.cs 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Snowball/TestSnowballVocab.cs
index d86813a..a29e54b 100644
--- 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Snowball/TestSnowballVocab.cs
+++ 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Snowball/TestSnowballVocab.cs
@@ -87,7 +87,7 @@ namespace Lucene.Net.Analysis.Snowball
                 this.snowballLanguage = snowballLanguage;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer t = new KeywordTokenizer(reader);
                 return new TokenStreamComponents(t, new SnowballFilter(t, 
snowballLanguage));

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Sv/TestSwedishLightStemFilter.cs
----------------------------------------------------------------------
diff --git 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Sv/TestSwedishLightStemFilter.cs
 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Sv/TestSwedishLightStemFilter.cs
index 4b55f5f..4961365 100644
--- 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Sv/TestSwedishLightStemFilter.cs
+++ 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Sv/TestSwedishLightStemFilter.cs
@@ -36,7 +36,7 @@ namespace Lucene.Net.Analysis.Sv
             {
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer source = new MockTokenizer(reader, 
MockTokenizer.WHITESPACE, false);
                 return new TokenStreamComponents(source, new 
SwedishLightStemFilter(source));
@@ -71,7 +71,7 @@ namespace Lucene.Net.Analysis.Sv
                 this.exclusionSet = exclusionSet;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer source = new MockTokenizer(reader, 
MockTokenizer.WHITESPACE, false);
                 TokenStream sink = new SetKeywordMarkerFilter(source, 
exclusionSet);
@@ -103,7 +103,7 @@ namespace Lucene.Net.Analysis.Sv
                 this.outerInstance = outerInstance;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new KeywordTokenizer(reader);
                 return new TokenStreamComponents(tokenizer, new 
SwedishLightStemFilter(tokenizer));

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Synonym/TestSolrSynonymParser.cs
----------------------------------------------------------------------
diff --git 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Synonym/TestSolrSynonymParser.cs
 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Synonym/TestSolrSynonymParser.cs
index b9f2a74..da25523 100644
--- 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Synonym/TestSolrSynonymParser.cs
+++ 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Synonym/TestSolrSynonymParser.cs
@@ -63,7 +63,7 @@ namespace Lucene.Net.Analysis.Synonym
                 this.map = map;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new MockTokenizer(reader, 
MockTokenizer.WHITESPACE, true);
                 return new TokenStreamComponents(tokenizer, new 
SynonymFilter(tokenizer, map, true));
@@ -150,7 +150,7 @@ namespace Lucene.Net.Analysis.Synonym
                 this.map = map;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new MockTokenizer(reader, 
MockTokenizer.KEYWORD, false);
                 return new TokenStreamComponents(tokenizer, new 
SynonymFilter(tokenizer, map, false));

Reply via email to