Repository: lucenenet
Updated Branches:
  refs/heads/master b781e1178 -> 234c9b4c3


http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Synonym/TestSynonymMapFilter.cs
----------------------------------------------------------------------
diff --git 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Synonym/TestSynonymMapFilter.cs 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Synonym/TestSynonymMapFilter.cs
index f29882a..00df67c 100644
--- 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Synonym/TestSynonymMapFilter.cs
+++ 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Synonym/TestSynonymMapFilter.cs
@@ -184,7 +184,7 @@ namespace Lucene.Net.Analysis.Synonym
                 this.map = map;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new MockTokenizer(reader, 
MockTokenizer.SIMPLE, true);
                 return new TokenStreamComponents(tokenizer, new 
SynonymFilter(tokenizer, map, false));
@@ -224,7 +224,7 @@ namespace Lucene.Net.Analysis.Synonym
                 this.map = map;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new MockTokenizer(reader, 
MockTokenizer.SIMPLE, true);
                 return new TokenStreamComponents(tokenizer, new 
SynonymFilter(tokenizer, map, false));
@@ -604,7 +604,7 @@ namespace Lucene.Net.Analysis.Synonym
                 this.ignoreCase = ignoreCase;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new MockTokenizer(reader, 
MockTokenizer.SIMPLE, true);
                 return new TokenStreamComponents(tokenizer, new 
SynonymFilter(tokenizer, map, ignoreCase));
@@ -679,7 +679,7 @@ namespace Lucene.Net.Analysis.Synonym
                 this.ignoreCase = ignoreCase;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new MockTokenizer(reader, 
MockTokenizer.SIMPLE, true);
                 TokenStream syns = new SynonymFilter(tokenizer, map, 
ignoreCase);
@@ -724,7 +724,7 @@ namespace Lucene.Net.Analysis.Synonym
                 this.ignoreCase = ignoreCase;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new KeywordTokenizer(reader);
                 return new TokenStreamComponents(tokenizer, new 
SynonymFilter(tokenizer, map, ignoreCase));
@@ -774,7 +774,7 @@ namespace Lucene.Net.Analysis.Synonym
                 this.ignoreCase = ignoreCase;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new MockTokenizer(reader, 
MockTokenizer.SIMPLE, true);
                 return new TokenStreamComponents(tokenizer, new 
SynonymFilter(tokenizer, map, ignoreCase));
@@ -813,7 +813,7 @@ namespace Lucene.Net.Analysis.Synonym
                 this.map = map;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new MockTokenizer(reader, 
MockTokenizer.WHITESPACE, true);
                 return new TokenStreamComponents(tokenizer, new 
SynonymFilter(tokenizer, map, true));
@@ -895,7 +895,7 @@ namespace Lucene.Net.Analysis.Synonym
                 this.map = map;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new MockTokenizer(reader, 
MockTokenizer.WHITESPACE, false);
                 return new TokenStreamComponents(tokenizer, new 
SynonymFilter(tokenizer, map, true));
@@ -928,7 +928,7 @@ namespace Lucene.Net.Analysis.Synonym
                 this.map = map;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new MockTokenizer(reader, 
MockTokenizer.WHITESPACE, false);
                 return new TokenStreamComponents(tokenizer, new 
SynonymFilter(tokenizer, map, true));
@@ -961,7 +961,7 @@ namespace Lucene.Net.Analysis.Synonym
                 this.map = map;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new MockTokenizer(reader, 
MockTokenizer.WHITESPACE, false);
                 return new TokenStreamComponents(tokenizer, new 
SynonymFilter(tokenizer, map, true));
@@ -992,7 +992,7 @@ namespace Lucene.Net.Analysis.Synonym
                 this.map = map;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new MockTokenizer(reader, 
MockTokenizer.WHITESPACE, false);
                 return new TokenStreamComponents(tokenizer, new 
SynonymFilter(tokenizer, map, true));
@@ -1025,7 +1025,7 @@ namespace Lucene.Net.Analysis.Synonym
                 this.map = map;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new MockTokenizer(reader, 
MockTokenizer.WHITESPACE, false);
                 return new TokenStreamComponents(tokenizer, new 
SynonymFilter(tokenizer, map, true));
@@ -1094,7 +1094,7 @@ namespace Lucene.Net.Analysis.Synonym
                 this.map = map;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new MockTokenizer(reader, 
MockTokenizer.WHITESPACE, false);
                 return new TokenStreamComponents(tokenizer, new 
SynonymFilter(tokenizer, map, true));
@@ -1125,7 +1125,7 @@ namespace Lucene.Net.Analysis.Synonym
                 this.map = map;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new MockTokenizer(reader, 
MockTokenizer.WHITESPACE, false);
                 return new TokenStreamComponents(tokenizer, new 
SynonymFilter(tokenizer, map, true));
@@ -1157,7 +1157,7 @@ namespace Lucene.Net.Analysis.Synonym
                 this.map = map;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new MockTokenizer(reader, 
MockTokenizer.WHITESPACE, false);
                 return new TokenStreamComponents(tokenizer, new 
SynonymFilter(tokenizer, map, true));
@@ -1188,7 +1188,7 @@ namespace Lucene.Net.Analysis.Synonym
                 this.map = map;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new MockTokenizer(reader, 
MockTokenizer.WHITESPACE, false);
                 return new TokenStreamComponents(tokenizer, new 
SynonymFilter(tokenizer, map, true));

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Synonym/TestWordnetSynonymParser.cs
----------------------------------------------------------------------
diff --git 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Synonym/TestWordnetSynonymParser.cs
 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Synonym/TestWordnetSynonymParser.cs
index f545114..cc372cc 100644
--- 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Synonym/TestWordnetSynonymParser.cs
+++ 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Synonym/TestWordnetSynonymParser.cs
@@ -57,7 +57,7 @@ namespace Lucene.Net.Analysis.Synonym
                 this.map = map;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new MockTokenizer(reader, 
MockTokenizer.WHITESPACE, false);
                 return new TokenStreamComponents(tokenizer, new 
SynonymFilter(tokenizer, map, false));

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Tr/TestTurkishLowerCaseFilter.cs
----------------------------------------------------------------------
diff --git 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Tr/TestTurkishLowerCaseFilter.cs
 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Tr/TestTurkishLowerCaseFilter.cs
index e2cab3f..a49a557 100644
--- 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Tr/TestTurkishLowerCaseFilter.cs
+++ 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Tr/TestTurkishLowerCaseFilter.cs
@@ -86,7 +86,7 @@ namespace Lucene.Net.Analysis.Tr
                 this.outerInstance = outerInstance;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new KeywordTokenizer(reader);
                 return new TokenStreamComponents(tokenizer, new 
TurkishLowerCaseFilter(tokenizer));

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Util/TestCharTokenizers.cs
----------------------------------------------------------------------
diff --git 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Util/TestCharTokenizers.cs 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Util/TestCharTokenizers.cs
index 79249c8..87b8c22 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Util/TestCharTokenizers.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Util/TestCharTokenizers.cs
@@ -151,7 +151,7 @@ namespace Lucene.Net.Analysis.Util
 
         private sealed class AnalyzerAnonymousInnerClassHelper : Analyzer
         {
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new 
LetterTokenizerAnonymousInnerClassHelper(TEST_VERSION_CURRENT, reader);
                 return new TokenStreamComponents(tokenizer, tokenizer);
@@ -214,7 +214,7 @@ namespace Lucene.Net.Analysis.Util
 
         private sealed class AnalyzerAnonymousInnerClassHelper2 : Analyzer
         {
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new 
LetterTokenizerAnonymousInnerClassHelper2(TEST_VERSION_CURRENT, reader);
                 return new TokenStreamComponents(tokenizer, tokenizer);
@@ -260,7 +260,7 @@ namespace Lucene.Net.Analysis.Util
             public AnalyzerAnonymousInnerClassHelper3()
             { }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new 
NumberAndSurrogatePairTokenizer(TEST_VERSION_CURRENT, reader);
                 return new TokenStreamComponents(tokenizer, tokenizer);

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Util/TestElision.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Util/TestElision.cs 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Util/TestElision.cs
index 2bc4ab3..81c5b1d 100644
--- a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Util/TestElision.cs
+++ b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Util/TestElision.cs
@@ -71,7 +71,7 @@ namespace Lucene.Net.Analysis.Util
                 this.outerInstance = outerInstance;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new KeywordTokenizer(reader);
                 return new TokenStreamComponents(tokenizer, new 
ElisionFilter(tokenizer, FrenchAnalyzer.DEFAULT_ARTICLES));

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Analysis.Common/Analysis/Wikipedia/WikipediaTokenizerTest.cs
----------------------------------------------------------------------
diff --git 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Wikipedia/WikipediaTokenizerTest.cs
 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Wikipedia/WikipediaTokenizerTest.cs
index 7f1b0ea..c234e32 100644
--- 
a/src/Lucene.Net.Tests.Analysis.Common/Analysis/Wikipedia/WikipediaTokenizerTest.cs
+++ 
b/src/Lucene.Net.Tests.Analysis.Common/Analysis/Wikipedia/WikipediaTokenizerTest.cs
@@ -126,7 +126,7 @@ namespace Lucene.Net.Analysis.Wikipedia
                 this.outerInstance = outerInstance;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new WikipediaTokenizer(reader);
                 return new TokenStreamComponents(tokenizer, tokenizer);
@@ -152,7 +152,7 @@ namespace Lucene.Net.Analysis.Wikipedia
                 this.outerInstance = outerInstance;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 Tokenizer tokenizer = new WikipediaTokenizer(reader);
                 return new TokenStreamComponents(tokenizer, tokenizer);

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Highlighter/Highlight/HighlighterTest.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net.Tests.Highlighter/Highlight/HighlighterTest.cs 
b/src/Lucene.Net.Tests.Highlighter/Highlight/HighlighterTest.cs
index 03abf65..2d3c668 100644
--- a/src/Lucene.Net.Tests.Highlighter/Highlight/HighlighterTest.cs
+++ b/src/Lucene.Net.Tests.Highlighter/Highlight/HighlighterTest.cs
@@ -2060,7 +2060,7 @@ namespace Lucene.Net.Search.Highlight
          *      java.io.Reader)
          */
 
-        protected internal override TokenStreamComponents 
CreateComponents(String arg0, TextReader arg1)
+        protected override TokenStreamComponents CreateComponents(String arg0, 
TextReader arg1)
         {
             Tokenizer stream = new MockTokenizer(arg1, MockTokenizer.SIMPLE, 
true);
             stream.AddAttribute<ICharTermAttribute>();

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Highlighter/Highlight/OffsetLimitTokenFilterTest.cs
----------------------------------------------------------------------
diff --git 
a/src/Lucene.Net.Tests.Highlighter/Highlight/OffsetLimitTokenFilterTest.cs 
b/src/Lucene.Net.Tests.Highlighter/Highlight/OffsetLimitTokenFilterTest.cs
index c7a19cc..0e0d585 100644
--- a/src/Lucene.Net.Tests.Highlighter/Highlight/OffsetLimitTokenFilterTest.cs
+++ b/src/Lucene.Net.Tests.Highlighter/Highlight/OffsetLimitTokenFilterTest.cs
@@ -56,7 +56,7 @@ namespace Lucene.Net.Search.Highlight
 
         internal class AnalyzerAnonymousHelper : Analyzer
         {
-            protected internal override TokenStreamComponents 
CreateComponents(string fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(string 
fieldName, TextReader reader)
             {
                 MockTokenizer tokenizer = new MockTokenizer(reader, 
MockTokenizer.WHITESPACE, false);
                 tokenizer.EnableChecks = (false);

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Highlighter/VectorHighlight/AbstractTestCase.cs
----------------------------------------------------------------------
diff --git 
a/src/Lucene.Net.Tests.Highlighter/VectorHighlight/AbstractTestCase.cs 
b/src/Lucene.Net.Tests.Highlighter/VectorHighlight/AbstractTestCase.cs
index 21b4a98..4e8f696 100644
--- a/src/Lucene.Net.Tests.Highlighter/VectorHighlight/AbstractTestCase.cs
+++ b/src/Lucene.Net.Tests.Highlighter/VectorHighlight/AbstractTestCase.cs
@@ -209,7 +209,7 @@ namespace Lucene.Net.Search.VectorHighlight
 
         internal sealed class BigramAnalyzer : Analyzer
         {
-            protected internal override TokenStreamComponents 
CreateComponents(String fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(String 
fieldName, TextReader reader)
             {
                 return new TokenStreamComponents(new 
BasicNGramTokenizer(reader));
             }

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net.Tests.Highlighter/VectorHighlight/IndexTimeSynonymTest.cs
----------------------------------------------------------------------
diff --git 
a/src/Lucene.Net.Tests.Highlighter/VectorHighlight/IndexTimeSynonymTest.cs 
b/src/Lucene.Net.Tests.Highlighter/VectorHighlight/IndexTimeSynonymTest.cs
index 5dc704f..029fd3d 100644
--- a/src/Lucene.Net.Tests.Highlighter/VectorHighlight/IndexTimeSynonymTest.cs
+++ b/src/Lucene.Net.Tests.Highlighter/VectorHighlight/IndexTimeSynonymTest.cs
@@ -366,7 +366,7 @@ namespace Lucene.Net.Search.VectorHighlight
                 this.tokens = tokens;
             }
 
-            protected internal override TokenStreamComponents 
CreateComponents(String fieldName, TextReader reader)
+            protected override TokenStreamComponents CreateComponents(String 
fieldName, TextReader reader)
             {
                 Tokenizer ts = new 
TokenizerAnonymousHelper(Token.TOKEN_ATTRIBUTE_FACTORY, reader, tokens);
                 return new TokenStreamComponents(ts);

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net/Properties/AssemblyInfo.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net/Properties/AssemblyInfo.cs 
b/src/Lucene.Net/Properties/AssemblyInfo.cs
index f04b339..29b4581 100644
--- a/src/Lucene.Net/Properties/AssemblyInfo.cs
+++ b/src/Lucene.Net/Properties/AssemblyInfo.cs
@@ -42,12 +42,7 @@ using System.Runtime.CompilerServices;
 // must update overridden methods to match.
 [assembly: InternalsVisibleTo("Lucene.Net.Tests")]
 [assembly: InternalsVisibleTo("Lucene.Net.TestFramework")]
-[assembly: InternalsVisibleTo("Lucene.Net.Highlighter")] // For Automaton
-[assembly: InternalsVisibleTo("Lucene.Net.ICU")] // For Automaton
 [assembly: InternalsVisibleTo("Lucene.Net.Misc")]
-[assembly: InternalsVisibleTo("Lucene.Net.Suggest")] // For Automaton
-[assembly: InternalsVisibleTo("Lucene.Net.Tests.Analysis.Common")] // For 
Automaton
-[assembly: InternalsVisibleTo("Lucene.Net.Tests.Highlighter")] // For Automaton
 [assembly: InternalsVisibleTo("Lucene.Net.Tests.ICU")] // For 
Analysis.Util.TestSegmentingTokenizerBase
 [assembly: InternalsVisibleTo("Lucene.Net.Tests.Misc")]
 [assembly: InternalsVisibleTo("Lucene.Net.Tests.QueryParser")]

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net/Util/Automaton/BasicAutomata.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net/Util/Automaton/BasicAutomata.cs 
b/src/Lucene.Net/Util/Automaton/BasicAutomata.cs
index d7b50e0..fcebb46 100644
--- a/src/Lucene.Net/Util/Automaton/BasicAutomata.cs
+++ b/src/Lucene.Net/Util/Automaton/BasicAutomata.cs
@@ -39,7 +39,7 @@ namespace Lucene.Net.Util.Automaton
     /// <para/>
     /// @lucene.experimental
     /// </summary>
-    internal sealed class BasicAutomata
+    public sealed class BasicAutomata
     {
         private BasicAutomata()
         {

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net/Util/Automaton/BasicOperations.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net/Util/Automaton/BasicOperations.cs 
b/src/Lucene.Net/Util/Automaton/BasicOperations.cs
index bc4b488..d6efcb9 100644
--- a/src/Lucene.Net/Util/Automaton/BasicOperations.cs
+++ b/src/Lucene.Net/Util/Automaton/BasicOperations.cs
@@ -41,12 +41,8 @@ namespace Lucene.Net.Util.Automaton
     /// <para/>
     /// @lucene.experimental
     /// </summary>
-    internal sealed class BasicOperations
+    public static class BasicOperations // LUCENENET specific - made static 
since all members are static
     {
-        private BasicOperations()
-        {
-        }
-
         /// <summary>
         /// Returns an automaton that accepts the concatenation of the 
languages of the
         /// given automata.

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net/Util/Automaton/MinimizationOperations.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net/Util/Automaton/MinimizationOperations.cs 
b/src/Lucene.Net/Util/Automaton/MinimizationOperations.cs
index 2d4490b..b71c468 100644
--- a/src/Lucene.Net/Util/Automaton/MinimizationOperations.cs
+++ b/src/Lucene.Net/Util/Automaton/MinimizationOperations.cs
@@ -38,12 +38,8 @@ namespace Lucene.Net.Util.Automaton
     /// <para/>
     /// @lucene.experimental
     /// </summary>
-    internal sealed class MinimizationOperations
+    public static class MinimizationOperations // LUCENENET specific - made 
static since all members are static
     {
-        private MinimizationOperations()
-        {
-        }
-
         /// <summary>
         /// Minimizes (and determinizes if not already deterministic) the given
         /// automaton.

http://git-wip-us.apache.org/repos/asf/lucenenet/blob/e4f25624/src/Lucene.Net/Util/Automaton/SpecialOperations.cs
----------------------------------------------------------------------
diff --git a/src/Lucene.Net/Util/Automaton/SpecialOperations.cs 
b/src/Lucene.Net/Util/Automaton/SpecialOperations.cs
index d5f75db..b312fb9 100644
--- a/src/Lucene.Net/Util/Automaton/SpecialOperations.cs
+++ b/src/Lucene.Net/Util/Automaton/SpecialOperations.cs
@@ -1,5 +1,4 @@
 using Lucene.Net.Support;
-using System.Collections;
 using System.Collections.Generic;
 using System.Linq;
 using System.Text;
@@ -42,12 +41,8 @@ namespace Lucene.Net.Util.Automaton
     /// <para/>
     /// @lucene.experimental
     /// </summary>
-    internal sealed class SpecialOperations
+    public static class SpecialOperations // LUCENENET specific - made static 
since all members are static
     {
-        private SpecialOperations()
-        {
-        }
-
         /// <summary>
         /// Finds the largest entry whose value is less than or equal to 
<paramref name="c"/>, or 0 if
         /// there is no such entry.

Reply via email to