Repository: lucenenet Updated Branches: refs/heads/master cfeaf2841 -> f43d23261
http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net.Suggest/Suggest/Fst/FSTCompletionBuilder.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net.Suggest/Suggest/Fst/FSTCompletionBuilder.cs b/src/Lucene.Net.Suggest/Suggest/Fst/FSTCompletionBuilder.cs index c29c2f3..1146020 100644 --- a/src/Lucene.Net.Suggest/Suggest/Fst/FSTCompletionBuilder.cs +++ b/src/Lucene.Net.Suggest/Suggest/Fst/FSTCompletionBuilder.cs @@ -30,16 +30,16 @@ namespace Lucene.Net.Search.Suggest.Fst /// <para> /// The construction step in the object finalizer works as follows: /// <list type="bullet"> - /// <item>A set of input terms and their buckets is given.</item> - /// <item>All terms in the input are prefixed with a synthetic pseudo-character + /// <item><description>A set of input terms and their buckets is given.</description></item> + /// <item><description>All terms in the input are prefixed with a synthetic pseudo-character /// (code) of the weight bucket the term fell into. For example a term /// <c>abc</c> with a discretized weight equal '1' would become - /// <c>1abc</c>.</item> - /// <item>The terms are then sorted by their raw value of UTF-8 character values - /// (including the synthetic bucket code in front).</item> - /// <item>A finite state automaton (<see cref="FST"/>) is constructed from the input. The + /// <c>1abc</c>.</description></item> + /// <item><description>The terms are then sorted by their raw value of UTF-8 character values + /// (including the synthetic bucket code in front).</description></item> + /// <item><description>A finite state automaton (<see cref="FST"/>) is constructed from the input. The /// root node has arcs labeled with all possible weights. We cache all these - /// arcs, highest-weight first.</item> + /// arcs, highest-weight first.</description></item> /// </list> /// /// </para> @@ -47,21 +47,21 @@ namespace Lucene.Net.Search.Suggest.Fst /// At runtime, in <see cref="FSTCompletion.DoLookup(string, int)"/>, /// the automaton is utilized as follows: /// <list type="bullet"> - /// <item>For each possible term weight encoded in the automaton (cached arcs from + /// <item><description>For each possible term weight encoded in the automaton (cached arcs from /// the root above), starting with the highest one, we descend along the path of /// the input key. If the key is not a prefix of a sequence in the automaton - /// (path ends prematurely), we exit immediately -- no completions.</item> - /// <item>Otherwise, we have found an internal automaton node that ends the key. + /// (path ends prematurely), we exit immediately -- no completions.</description></item> + /// <item><description>Otherwise, we have found an internal automaton node that ends the key. /// <b>The entire subautomaton (all paths) starting from this node form the key's /// completions.</b> We start the traversal of this subautomaton. Every time we /// reach a final state (arc), we add a single suggestion to the list of results /// (the weight of this suggestion is constant and equal to the root path we /// started from). The tricky part is that because automaton edges are sorted and /// we scan depth-first, we can terminate the entire procedure as soon as we - /// collect enough suggestions the user requested.</item> - /// <item>In case the number of suggestions collected in the step above is still + /// collect enough suggestions the user requested.</description></item> + /// <item><description>In case the number of suggestions collected in the step above is still /// insufficient, we proceed to the next (smaller) weight leaving the root node - /// and repeat the same algorithm again.</item> + /// and repeat the same algorithm again.</description></item> /// </list> /// /// <h2>Runtime behavior and performance characteristic</h2> http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net/Analysis/Analyzer.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net/Analysis/Analyzer.cs b/src/Lucene.Net/Analysis/Analyzer.cs index ba27c90..80e5ffb 100644 --- a/src/Lucene.Net/Analysis/Analyzer.cs +++ b/src/Lucene.Net/Analysis/Analyzer.cs @@ -45,22 +45,22 @@ namespace Lucene.Net.Analysis /// <para/> /// For some concrete implementations bundled with Lucene, look in the analysis modules: /// <list type="bullet"> - /// <item>Common: - /// Analyzers for indexing content in different languages and domains.</item> - /// <item>ICU: - /// Exposes functionality from ICU to Apache Lucene.</item> - /// <item>Kuromoji: - /// Morphological analyzer for Japanese text.</item> - /// <item>Morfologik: - /// Dictionary-driven lemmatization for the Polish language.</item> - /// <item>Phonetic: - /// Analysis for indexing phonetic signatures (for sounds-alike search).</item> - /// <item>Smart Chinese: - /// Analyzer for Simplified Chinese, which indexes words.</item> - /// <item>Stempel: - /// Algorithmic Stemmer for the Polish Language.</item> - /// <item>UIMA: - /// Analysis integration with Apache UIMA.</item> + /// <item><description>Common: + /// Analyzers for indexing content in different languages and domains.</description></item> + /// <item><description>ICU: + /// Exposes functionality from ICU to Apache Lucene.</description></item> + /// <item><description>Kuromoji: + /// Morphological analyzer for Japanese text.</description></item> + /// <item><description>Morfologik: + /// Dictionary-driven lemmatization for the Polish language.</description></item> + /// <item><description>Phonetic: + /// Analysis for indexing phonetic signatures (for sounds-alike search).</description></item> + /// <item><description>Smart Chinese: + /// Analyzer for Simplified Chinese, which indexes words.</description></item> + /// <item><description>Stempel: + /// Algorithmic Stemmer for the Polish Language.</description></item> + /// <item><description>UIMA: + /// Analysis integration with Apache UIMA.</description></item> /// </list> /// </summary> public abstract class Analyzer : IDisposable http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net/Analysis/Token.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net/Analysis/Token.cs b/src/Lucene.Net/Analysis/Token.cs index 8e8cf07..be1938e 100644 --- a/src/Lucene.Net/Analysis/Token.cs +++ b/src/Lucene.Net/Analysis/Token.cs @@ -77,38 +77,38 @@ namespace Lucene.Net.Analysis /// for details.</para> /// <para>Typical Token reuse patterns: /// <list type="bullet"> - /// <item> Copying text from a string (type is reset to <see cref="TypeAttribute.DEFAULT_TYPE"/> if not specified): + /// <item><description> Copying text from a string (type is reset to <see cref="TypeAttribute.DEFAULT_TYPE"/> if not specified): /// <code> /// return reusableToken.Reinit(string, startOffset, endOffset[, type]); /// </code> - /// </item> - /// <item> Copying some text from a string (type is reset to <see cref="TypeAttribute.DEFAULT_TYPE"/> if not specified): + /// </description></item> + /// <item><description> Copying some text from a string (type is reset to <see cref="TypeAttribute.DEFAULT_TYPE"/> if not specified): /// <code> /// return reusableToken.Reinit(string, 0, string.Length, startOffset, endOffset[, type]); /// </code> - /// </item> - /// <item> Copying text from char[] buffer (type is reset to <see cref="TypeAttribute.DEFAULT_TYPE"/> if not specified): + /// </description></item> + /// <item><description> Copying text from char[] buffer (type is reset to <see cref="TypeAttribute.DEFAULT_TYPE"/> if not specified): /// <code> /// return reusableToken.Reinit(buffer, 0, buffer.Length, startOffset, endOffset[, type]); /// </code> - /// </item> - /// <item> Copying some text from a char[] buffer (type is reset to <see cref="TypeAttribute.DEFAULT_TYPE"/> if not specified): + /// </description></item> + /// <item><description> Copying some text from a char[] buffer (type is reset to <see cref="TypeAttribute.DEFAULT_TYPE"/> if not specified): /// <code> /// return reusableToken.Reinit(buffer, start, end - start, startOffset, endOffset[, type]); /// </code> - /// </item> - /// <item> Copying from one one <see cref="Token"/> to another (type is reset to <see cref="TypeAttribute.DEFAULT_TYPE"/> if not specified): + /// </description></item> + /// <item><description> Copying from one one <see cref="Token"/> to another (type is reset to <see cref="TypeAttribute.DEFAULT_TYPE"/> if not specified): /// <code> /// return reusableToken.Reinit(source.Buffer, 0, source.Length, source.StartOffset, source.EndOffset[, source.Type]); /// </code> - /// </item> + /// </description></item> /// </list> /// A few things to note: /// <list type="bullet"> - /// <item><see cref="Clear()"/> initializes all of the fields to default values. this was changed in contrast to Lucene 2.4, but should affect no one.</item> - /// <item>Because <see cref="TokenStream"/>s can be chained, one cannot assume that the <see cref="Token"/>'s current type is correct.</item> - /// <item>The startOffset and endOffset represent the start and offset in the source text, so be careful in adjusting them.</item> - /// <item>When caching a reusable token, clone it. When injecting a cached token into a stream that can be reset, clone it again.</item> + /// <item><description><see cref="Clear()"/> initializes all of the fields to default values. this was changed in contrast to Lucene 2.4, but should affect no one.</description></item> + /// <item><description>Because <see cref="TokenStream"/>s can be chained, one cannot assume that the <see cref="Token"/>'s current type is correct.</description></item> + /// <item><description>The startOffset and endOffset represent the start and offset in the source text, so be careful in adjusting them.</description></item> + /// <item><description>When caching a reusable token, clone it. When injecting a cached token into a stream that can be reset, clone it again.</description></item> /// </list> /// </para> /// <para> http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net/Analysis/TokenAttributes/IPositionIncrementAttribute.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net/Analysis/TokenAttributes/IPositionIncrementAttribute.cs b/src/Lucene.Net/Analysis/TokenAttributes/IPositionIncrementAttribute.cs index 3d47b7d..f3adee1 100644 --- a/src/Lucene.Net/Analysis/TokenAttributes/IPositionIncrementAttribute.cs +++ b/src/Lucene.Net/Analysis/TokenAttributes/IPositionIncrementAttribute.cs @@ -29,19 +29,19 @@ namespace Lucene.Net.Analysis.TokenAttributes /// <para/>Some common uses for this are: /// /// <list type="bullet"> - /// <item>Set it to zero to put multiple terms in the same position. this is + /// <item><description>Set it to zero to put multiple terms in the same position. this is /// useful if, e.g., a word has multiple stems. Searches for phrases /// including either stem will match. In this case, all but the first stem's /// increment should be set to zero: the increment of the first instance /// should be one. Repeating a token with an increment of zero can also be - /// used to boost the scores of matches on that token.</item> + /// used to boost the scores of matches on that token.</description></item> /// - /// <item>Set it to values greater than one to inhibit exact phrase matches. + /// <item><description>Set it to values greater than one to inhibit exact phrase matches. /// If, for example, one does not want phrases to match across removed stop /// words, then one could build a stop word filter that removes stop words and /// also sets the increment to the number of stop words removed before each /// non-stop word. Then exact phrase queries will only match when the terms - /// occur with no intervening stop words.</item> + /// occur with no intervening stop words.</description></item> /// </list> /// </summary> /// <seealso cref="Lucene.Net.Index.DocsAndPositionsEnum"/> http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net/Analysis/TokenStream.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net/Analysis/TokenStream.cs b/src/Lucene.Net/Analysis/TokenStream.cs index 7cec955..f9ec60f 100644 --- a/src/Lucene.Net/Analysis/TokenStream.cs +++ b/src/Lucene.Net/Analysis/TokenStream.cs @@ -31,9 +31,9 @@ namespace Lucene.Net.Analysis /// <para/> /// this is an abstract class; concrete subclasses are: /// <list type="bullet"> - /// <item><see cref="Tokenizer"/>, a <see cref="TokenStream"/> whose input is a <see cref="System.IO.TextReader"/>; and</item> - /// <item><see cref="TokenFilter"/>, a <see cref="TokenStream"/> whose input is another - /// <see cref="TokenStream"/>.</item> + /// <item><description><see cref="Tokenizer"/>, a <see cref="TokenStream"/> whose input is a <see cref="System.IO.TextReader"/>; and</description></item> + /// <item><description><see cref="TokenFilter"/>, a <see cref="TokenStream"/> whose input is another + /// <see cref="TokenStream"/>.</description></item> /// </list> /// A new <see cref="TokenStream"/> API has been introduced with Lucene 2.9. this API /// has moved from being <see cref="Token"/>-based to <see cref="Util.IAttribute"/>-based. While @@ -49,17 +49,17 @@ namespace Lucene.Net.Analysis /// <para/> /// <b>The workflow of the new <see cref="TokenStream"/> API is as follows:</b> /// <list type="number"> - /// <item>Instantiation of <see cref="TokenStream"/>/<see cref="TokenFilter"/>s which add/get - /// attributes to/from the <see cref="AttributeSource"/>.</item> - /// <item>The consumer calls <see cref="TokenStream.Reset()"/>.</item> - /// <item>The consumer retrieves attributes from the stream and stores local - /// references to all attributes it wants to access.</item> - /// <item>The consumer calls <see cref="IncrementToken()"/> until it returns false - /// consuming the attributes after each call.</item> - /// <item>The consumer calls <see cref="End()"/> so that any end-of-stream operations - /// can be performed.</item> - /// <item>The consumer calls <see cref="Dispose()"/> to release any resource when finished - /// using the <see cref="TokenStream"/>.</item> + /// <item><description>Instantiation of <see cref="TokenStream"/>/<see cref="TokenFilter"/>s which add/get + /// attributes to/from the <see cref="AttributeSource"/>.</description></item> + /// <item><description>The consumer calls <see cref="TokenStream.Reset()"/>.</description></item> + /// <item><description>The consumer retrieves attributes from the stream and stores local + /// references to all attributes it wants to access.</description></item> + /// <item><description>The consumer calls <see cref="IncrementToken()"/> until it returns false + /// consuming the attributes after each call.</description></item> + /// <item><description>The consumer calls <see cref="End()"/> so that any end-of-stream operations + /// can be performed.</description></item> + /// <item><description>The consumer calls <see cref="Dispose()"/> to release any resource when finished + /// using the <see cref="TokenStream"/>.</description></item> /// </list> /// To make sure that filters and consumers know which attributes are available, /// the attributes must be added during instantiation. Filters and consumers are http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net/Codecs/Codec.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net/Codecs/Codec.cs b/src/Lucene.Net/Codecs/Codec.cs index e70cd5e..a9f2448 100644 --- a/src/Lucene.Net/Codecs/Codec.cs +++ b/src/Lucene.Net/Codecs/Codec.cs @@ -31,13 +31,13 @@ namespace Lucene.Net.Codecs /// <para/> /// To implement your own codec: /// <list type="number"> - /// <item>Subclass this class.</item> - /// <item>Subclass <see cref="DefaultCodecFactory"/>, override the <see cref="DefaultCodecFactory.Initialize()"/> method, + /// <item><description>Subclass this class.</description></item> + /// <item><description>Subclass <see cref="DefaultCodecFactory"/>, override the <see cref="DefaultCodecFactory.Initialize()"/> method, /// and add the line <c>base.ScanForCodecs(typeof(YourCodec).GetTypeInfo().Assembly)</c>. /// If you have any codec classes in your assembly /// that are not meant for reading, you can add the <see cref="ExcludeCodecFromScanAttribute"/> - /// to them so they are ignored by the scan.</item> - /// <item>set the new <see cref="ICodecFactory"/> by calling <see cref="SetCodecFactory"/> at application startup.</item> + /// to them so they are ignored by the scan.</description></item> + /// <item><description>set the new <see cref="ICodecFactory"/> by calling <see cref="SetCodecFactory"/> at application startup.</description></item> /// </list> /// If your codec has dependencies, you may also override <see cref="DefaultCodecFactory.GetCodec(Type)"/> to inject /// them via pure DI or a DI container. See <a href="http://blog.ploeh.dk/2014/05/19/di-friendly-framework/">DI-Friendly Framework</a> http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net/Codecs/DocValuesFormat.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net/Codecs/DocValuesFormat.cs b/src/Lucene.Net/Codecs/DocValuesFormat.cs index 813d9a1..9ef0f4d 100644 --- a/src/Lucene.Net/Codecs/DocValuesFormat.cs +++ b/src/Lucene.Net/Codecs/DocValuesFormat.cs @@ -34,14 +34,14 @@ namespace Lucene.Net.Codecs /// <para/> /// To implement your own format: /// <list type="number"> - /// <item>Subclass this class.</item> - /// <item>Subclass <see cref="DefaultDocValuesFormatFactory"/>, override the <see cref="DefaultDocValuesFormatFactory.Initialize()"/> method, + /// <item><description>Subclass this class.</description></item> + /// <item><description>Subclass <see cref="DefaultDocValuesFormatFactory"/>, override the <see cref="DefaultDocValuesFormatFactory.Initialize()"/> method, /// and add the line <c>base.ScanForDocValuesFormats(typeof(YourDocValuesFormat).GetTypeInfo().Assembly)</c>. /// If you have any format classes in your assembly /// that are not meant for reading, you can add the <see cref="ExcludeDocValuesFormatFromScanAttribute"/> - /// to them so they are ignored by the scan.</item> - /// <item>Set the new <see cref="IDocValuesFormatFactory"/> by calling <see cref="SetDocValuesFormatFactory(IDocValuesFormatFactory)"/> - /// at application startup.</item> + /// to them so they are ignored by the scan.</description></item> + /// <item><description>Set the new <see cref="IDocValuesFormatFactory"/> by calling <see cref="SetDocValuesFormatFactory(IDocValuesFormatFactory)"/> + /// at application startup.</description></item> /// </list> /// If your format has dependencies, you may also override <see cref="DefaultDocValuesFormatFactory.GetDocValuesFormat(Type)"/> /// to inject them via pure DI or a DI container. See <a href="http://blog.ploeh.dk/2014/05/19/di-friendly-framework/">DI-Friendly Framework</a> http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net/Codecs/PostingsFormat.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net/Codecs/PostingsFormat.cs b/src/Lucene.Net/Codecs/PostingsFormat.cs index 2a74ea6..bc34e65 100644 --- a/src/Lucene.Net/Codecs/PostingsFormat.cs +++ b/src/Lucene.Net/Codecs/PostingsFormat.cs @@ -32,14 +32,14 @@ namespace Lucene.Net.Codecs /// <para/> /// If you implement your own format: /// <list type="number"> - /// <item>Subclass this class.</item> - /// <item>Subclass <see cref="DefaultPostingsFormatFactory"/>, override <see cref="DefaultPostingsFormatFactory.Initialize()"/>, + /// <item><description>Subclass this class.</description></item> + /// <item><description>Subclass <see cref="DefaultPostingsFormatFactory"/>, override <see cref="DefaultPostingsFormatFactory.Initialize()"/>, /// and add the line <c>base.ScanForPostingsFormats(typeof(YourPostingsFormat).GetTypeInfo().Assembly)</c>. /// If you have any format classes in your assembly /// that are not meant for reading, you can add the <see cref="ExcludePostingsFormatFromScanAttribute"/> - /// to them so they are ignored by the scan.</item> - /// <item>Set the new <see cref="IPostingsFormatFactory"/> by calling <see cref="SetPostingsFormatFactory(IPostingsFormatFactory)"/> - /// at application startup.</item> + /// to them so they are ignored by the scan.</description></item> + /// <item><description>Set the new <see cref="IPostingsFormatFactory"/> by calling <see cref="SetPostingsFormatFactory(IPostingsFormatFactory)"/> + /// at application startup.</description></item> /// </list> /// If your format has dependencies, you may also override <see cref="DefaultPostingsFormatFactory.GetPostingsFormat(Type)"/> to inject /// them via pure DI or a DI container. See <a href="http://blog.ploeh.dk/2014/05/19/di-friendly-framework/">DI-Friendly Framework</a> http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net/Document/Field.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net/Document/Field.cs b/src/Lucene.Net/Document/Field.cs index fbd7d13..54fe113 100644 --- a/src/Lucene.Net/Document/Field.cs +++ b/src/Lucene.Net/Document/Field.cs @@ -920,8 +920,8 @@ namespace Lucene.Net.Documents /// <exception cref="ArgumentNullException">if <paramref name="name"/> or <paramref name="value"/> is <c>null</c></exception> /// <exception cref="ArgumentException">in any of the following situations: /// <list type="bullet"> - /// <item>the field is neither stored nor indexed</item> - /// <item>the field is not indexed but termVector is <see cref="TermVector.YES"/></item> + /// <item><description>the field is neither stored nor indexed</description></item> + /// <item><description>the field is not indexed but termVector is <see cref="TermVector.YES"/></description></item> /// </list> /// </exception> [Obsolete("Use StringField, TextField instead.")] http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net/Index/AutomatonTermsEnum.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net/Index/AutomatonTermsEnum.cs b/src/Lucene.Net/Index/AutomatonTermsEnum.cs index e969235..e60a975 100644 --- a/src/Lucene.Net/Index/AutomatonTermsEnum.cs +++ b/src/Lucene.Net/Index/AutomatonTermsEnum.cs @@ -34,9 +34,9 @@ namespace Lucene.Net.Index /// <para/> /// The algorithm is such: /// <list type="number"> - /// <item>As long as matches are successful, keep reading sequentially.</item> - /// <item>When a match fails, skip to the next string in lexicographic order that - /// does not enter a reject state.</item> + /// <item><description>As long as matches are successful, keep reading sequentially.</description></item> + /// <item><description>When a match fails, skip to the next string in lexicographic order that + /// does not enter a reject state.</description></item> /// </list> /// <para> /// The algorithm does not attempt to actually skip to the next string that is http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net/Index/DocTermOrds.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net/Index/DocTermOrds.cs b/src/Lucene.Net/Index/DocTermOrds.cs index 1805499..237e3fc 100644 --- a/src/Lucene.Net/Index/DocTermOrds.cs +++ b/src/Lucene.Net/Index/DocTermOrds.cs @@ -74,36 +74,36 @@ namespace Lucene.Net.Index /// <remarks> /// Final form of the un-inverted field: /// <list type="bullet"> - /// <item>Each document points to a list of term numbers that are contained in that document.</item> - /// <item> + /// <item><description>Each document points to a list of term numbers that are contained in that document.</description></item> + /// <item><description> /// Term numbers are in sorted order, and are encoded as variable-length deltas from the /// previous term number. Real term numbers start at 2 since 0 and 1 are reserved. A /// term number of 0 signals the end of the termNumber list. - /// </item> - /// <item> + /// </description></item> + /// <item><description> /// There is a single int[maxDoc()] which either contains a pointer into a byte[] for /// the termNumber lists, or directly contains the termNumber list if it fits in the 4 /// bytes of an integer. If the first byte in the integer is 1, the next 3 bytes /// are a pointer into a byte[] where the termNumber list starts. - /// </item> - /// <item> + /// </description></item> + /// <item><description> /// There are actually 256 byte arrays, to compensate for the fact that the pointers /// into the byte arrays are only 3 bytes long. The correct byte array for a document /// is a function of it's id. - /// </item> - /// <item> + /// </description></item> + /// <item><description> /// To save space and speed up faceting, any term that matches enough documents will /// not be un-inverted... it will be skipped while building the un-inverted field structure, /// and will use a set intersection method during faceting. - /// </item> - /// <item> + /// </description></item> + /// <item><description> /// To further save memory, the terms (the actual string values) are not all stored in /// memory, but a TermIndex is used to convert term numbers to term values only /// for the terms needed after faceting has completed. Only every 128th term value /// is stored, along with it's corresponding term number, and this is used as an /// index to find the closest term and iterate until the desired number is hit (very /// much like Lucene's own internal term index). - /// </item> + /// </description></item> /// </list> /// </remarks> #if FEATURE_SERIALIZABLE http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net/Index/DocumentsWriterDeleteQueue.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net/Index/DocumentsWriterDeleteQueue.cs b/src/Lucene.Net/Index/DocumentsWriterDeleteQueue.cs index 36052ef..f035d65 100644 --- a/src/Lucene.Net/Index/DocumentsWriterDeleteQueue.cs +++ b/src/Lucene.Net/Index/DocumentsWriterDeleteQueue.cs @@ -52,13 +52,13 @@ namespace Lucene.Net.Index /// DWPT updates a document it: /// /// <list type="number"> - /// <item>consumes a document and finishes its processing</item> - /// <item>updates its private <see cref="DeleteSlice"/> either by calling + /// <item><description>consumes a document and finishes its processing</description></item> + /// <item><description>updates its private <see cref="DeleteSlice"/> either by calling /// <see cref="UpdateSlice(DeleteSlice)"/> or <see cref="Add(Term, DeleteSlice)"/> (if the - /// document has a delTerm)</item> - /// <item>applies all deletes in the slice to its private <see cref="BufferedUpdates"/> - /// and resets it</item> - /// <item>increments its internal document id</item> + /// document has a delTerm)</description></item> + /// <item><description>applies all deletes in the slice to its private <see cref="BufferedUpdates"/> + /// and resets it</description></item> + /// <item><description>increments its internal document id</description></item> /// </list> /// /// The DWPT also doesn't apply its current documents delete term until it has http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net/Index/FlushByRamOrCountsPolicy.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net/Index/FlushByRamOrCountsPolicy.cs b/src/Lucene.Net/Index/FlushByRamOrCountsPolicy.cs index 5ed2c7b..8c4da3c 100644 --- a/src/Lucene.Net/Index/FlushByRamOrCountsPolicy.cs +++ b/src/Lucene.Net/Index/FlushByRamOrCountsPolicy.cs @@ -28,13 +28,13 @@ namespace Lucene.Net.Index /// number of buffered delete terms. /// /// <list type="bullet"> - /// <item> + /// <item><description> /// <see cref="OnDelete(DocumentsWriterFlushControl, DocumentsWriterPerThreadPool.ThreadState)"/> /// - applies pending delete operations based on the global number of buffered /// delete terms iff <see cref="LiveIndexWriterConfig.MaxBufferedDeleteTerms"/> is /// enabled - /// </item> - /// <item> + /// </description></item> + /// <item><description> /// <see cref="OnInsert(DocumentsWriterFlushControl, DocumentsWriterPerThreadPool.ThreadState)"/> /// - flushes either on the number of documents per /// <see cref="DocumentsWriterPerThread"/> ( @@ -42,15 +42,15 @@ namespace Lucene.Net.Index /// memory consumption in the current indexing session iff /// <see cref="LiveIndexWriterConfig.MaxBufferedDocs"/> or /// <see cref="LiveIndexWriterConfig.RAMBufferSizeMB"/> is enabled respectively - /// </item> - /// <item> + /// </description></item> + /// <item><description> /// <see cref="FlushPolicy.OnUpdate(DocumentsWriterFlushControl, DocumentsWriterPerThreadPool.ThreadState)"/> /// - calls /// <see cref="OnInsert(DocumentsWriterFlushControl, DocumentsWriterPerThreadPool.ThreadState)"/> /// and /// <see cref="OnDelete(DocumentsWriterFlushControl, DocumentsWriterPerThreadPool.ThreadState)"/> /// in order - /// </item> + /// </description></item> /// </list> /// All <see cref="IndexWriterConfig"/> settings are used to mark /// <see cref="DocumentsWriterPerThread"/> as flush pending during indexing with http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net/Index/FlushPolicy.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net/Index/FlushPolicy.cs b/src/Lucene.Net/Index/FlushPolicy.cs index d342b0b..8eca198 100644 --- a/src/Lucene.Net/Index/FlushPolicy.cs +++ b/src/Lucene.Net/Index/FlushPolicy.cs @@ -30,10 +30,10 @@ namespace Lucene.Net.Index /// <para/> /// Segments are traditionally flushed by: /// <list type="bullet"> - /// <item>RAM consumption - configured via - /// <see cref="LiveIndexWriterConfig.RAMBufferSizeMB"/></item> - /// <item>Number of RAM resident documents - configured via - /// <see cref="LiveIndexWriterConfig.MaxBufferedDocs"/></item> + /// <item><description>RAM consumption - configured via + /// <see cref="LiveIndexWriterConfig.RAMBufferSizeMB"/></description></item> + /// <item><description>Number of RAM resident documents - configured via + /// <see cref="LiveIndexWriterConfig.MaxBufferedDocs"/></description></item> /// </list> /// The policy also applies pending delete operations (by term and/or query), /// given the threshold set in http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net/Index/IndexReader.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net/Index/IndexReader.cs b/src/Lucene.Net/Index/IndexReader.cs index c9ade02..3b72fc7 100644 --- a/src/Lucene.Net/Index/IndexReader.cs +++ b/src/Lucene.Net/Index/IndexReader.cs @@ -36,16 +36,16 @@ namespace Lucene.Net.Index /// /// <para/>There are two different types of <see cref="IndexReader"/>s: /// <list type="bullet"> - /// <item><see cref="AtomicReader"/>: These indexes do not consist of several sub-readers, + /// <item><description><see cref="AtomicReader"/>: These indexes do not consist of several sub-readers, /// they are atomic. They support retrieval of stored fields, doc values, terms, - /// and postings.</item> - /// <item><see cref="CompositeReader"/>: Instances (like <see cref="DirectoryReader"/>) + /// and postings.</description></item> + /// <item><description><see cref="CompositeReader"/>: Instances (like <see cref="DirectoryReader"/>) /// of this reader can only /// be used to get stored fields from the underlying <see cref="AtomicReader"/>s, /// but it is not possible to directly retrieve postings. To do that, get /// the sub-readers via <see cref="CompositeReader.GetSequentialSubReaders()"/>. /// Alternatively, you can mimic an <see cref="AtomicReader"/> (with a serious slowdown), - /// by wrapping composite readers with <see cref="SlowCompositeReaderWrapper"/>.</item> + /// by wrapping composite readers with <see cref="SlowCompositeReaderWrapper"/>.</description></item> /// </list> /// /// <para/><see cref="IndexReader"/> instances for indexes on disk are usually constructed http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net/Store/CompoundFileDirectory.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net/Store/CompoundFileDirectory.cs b/src/Lucene.Net/Store/CompoundFileDirectory.cs index 29ce68e..58cca90 100644 --- a/src/Lucene.Net/Store/CompoundFileDirectory.cs +++ b/src/Lucene.Net/Store/CompoundFileDirectory.cs @@ -42,29 +42,29 @@ namespace Lucene.Net.Store /// <para/> /// Files: /// <list type="bullet"> - /// <item><c>.cfs</c>: An optional "virtual" file consisting of all the other - /// index files for systems that frequently run out of file handles.</item> - /// <item><c>.cfe</c>: The "virtual" compound file's entry table holding all - /// entries in the corresponding .cfs file.</item> + /// <item><description><c>.cfs</c>: An optional "virtual" file consisting of all the other + /// index files for systems that frequently run out of file handles.</description></item> + /// <item><description><c>.cfe</c>: The "virtual" compound file's entry table holding all + /// entries in the corresponding .cfs file.</description></item> /// </list> /// <para>Description:</para> /// <list type="bullet"> - /// <item>Compound (.cfs) --> Header, FileData <sup>FileCount</sup></item> - /// <item>Compound Entry Table (.cfe) --> Header, FileCount, <FileName, - /// DataOffset, DataLength> <sup>FileCount</sup>, Footer</item> - /// <item>Header --> <see cref="CodecUtil.WriteHeader"/></item> - /// <item>FileCount --> <see cref="DataOutput.WriteVInt32"/></item> - /// <item>DataOffset,DataLength --> <see cref="DataOutput.WriteInt64"/></item> - /// <item>FileName --> <see cref="DataOutput.WriteString"/></item> - /// <item>FileData --> raw file data</item> - /// <item>Footer --> <see cref="CodecUtil.WriteFooter"/></item> + /// <item><description>Compound (.cfs) --> Header, FileData <sup>FileCount</sup></description></item> + /// <item><description>Compound Entry Table (.cfe) --> Header, FileCount, <FileName, + /// DataOffset, DataLength> <sup>FileCount</sup>, Footer</description></item> + /// <item><description>Header --> <see cref="CodecUtil.WriteHeader"/></description></item> + /// <item><description>FileCount --> <see cref="DataOutput.WriteVInt32"/></description></item> + /// <item><description>DataOffset,DataLength --> <see cref="DataOutput.WriteInt64"/></description></item> + /// <item><description>FileName --> <see cref="DataOutput.WriteString"/></description></item> + /// <item><description>FileData --> raw file data</description></item> + /// <item><description>Footer --> <see cref="CodecUtil.WriteFooter"/></description></item> /// </list> /// <para>Notes:</para> /// <list type="bullet"> - /// <item>FileCount indicates how many files are contained in this compound file. - /// The entry table that follows has that many entries.</item> - /// <item>Each directory entry contains a long pointer to the start of this file's data - /// section, the files length, and a <see cref="string"/> with that file's name.</item> + /// <item><description>FileCount indicates how many files are contained in this compound file. + /// The entry table that follows has that many entries.</description></item> + /// <item><description>Each directory entry contains a long pointer to the start of this file's data + /// section, the files length, and a <see cref="string"/> with that file's name.</description></item> /// </list> /// <para/> /// @lucene.experimental http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net/Store/Directory.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net/Store/Directory.cs b/src/Lucene.Net/Store/Directory.cs index 17cb98b..7565b8a 100644 --- a/src/Lucene.Net/Store/Directory.cs +++ b/src/Lucene.Net/Store/Directory.cs @@ -31,9 +31,9 @@ namespace Lucene.Net.Store /// .NET's i/o APIs not used directly, but rather all i/o is /// through this API. This permits things such as: /// <list type="bullet"> - /// <item> implementation of RAM-based indices;</item> - /// <item> implementation indices stored in a database;</item> - /// <item> implementation of an index as a single file;</item> + /// <item><description> implementation of RAM-based indices;</description></item> + /// <item><description> implementation indices stored in a database;</description></item> + /// <item><description> implementation of an index as a single file;</description></item> /// </list> /// <para/> /// Directory locking is implemented by an instance of @@ -67,9 +67,9 @@ namespace Lucene.Net.Store /// Returns the length of a file in the directory. this method follows the /// following contract: /// <list> - /// <item>Throws <see cref="System.IO.FileNotFoundException"/> - /// if the file does not exist.</item> - /// <item>Returns a value >=0 if the file exists, which specifies its length.</item> + /// <item><description>Throws <see cref="System.IO.FileNotFoundException"/> + /// if the file does not exist.</description></item> + /// <item><description>Returns a value >=0 if the file exists, which specifies its length.</description></item> /// </list> /// </summary> /// <param name="name"> the name of the file for which to return the length. </param> http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net/Store/FSDirectory.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net/Store/FSDirectory.cs b/src/Lucene.Net/Store/FSDirectory.cs index 42c5bfc..d1cbda1 100644 --- a/src/Lucene.Net/Store/FSDirectory.cs +++ b/src/Lucene.Net/Store/FSDirectory.cs @@ -36,14 +36,14 @@ namespace Lucene.Net.Store /// /// <list type="bullet"> /// - /// <item> <see cref="SimpleFSDirectory"/> is a straightforward + /// <item><description> <see cref="SimpleFSDirectory"/> is a straightforward /// implementation using <see cref="System.IO.FileStream"/>. /// However, it has poor concurrent performance /// (multiple threads will bottleneck) as it /// synchronizes when multiple threads read from the - /// same file.</item> + /// same file.</description></item> /// - /// <item> <see cref="NIOFSDirectory"/> uses java.nio's + /// <item><description> <see cref="NIOFSDirectory"/> uses java.nio's /// FileChannel's positional io when reading to avoid /// synchronization when reading from the same file. /// Unfortunately, due to a Windows-only <a @@ -53,9 +53,9 @@ namespace Lucene.Net.Store /// choice. Applications using <see cref="System.Threading.Thread.Interrupt()"/> or /// <see cref="System.Threading.Tasks.Task{TResult}"/> should use /// <see cref="SimpleFSDirectory"/> instead. See <see cref="NIOFSDirectory"/> java doc - /// for details.</item> + /// for details.</description></item> /// - /// <item> <see cref="MMapDirectory"/> uses memory-mapped IO when + /// <item><description> <see cref="MMapDirectory"/> uses memory-mapped IO when /// reading. This is a good choice if you have plenty /// of virtual memory relative to your index size, eg /// if you are running on a 64 bit runtime, or you are @@ -65,7 +65,7 @@ namespace Lucene.Net.Store /// Applications using <see cref="System.Threading.Thread.Interrupt()"/> or /// <see cref="System.Threading.Tasks.Task"/> should use /// <see cref="SimpleFSDirectory"/> instead. See <see cref="MMapDirectory"/> - /// doc for details.</item> + /// doc for details.</description></item> /// </list> /// /// Unfortunately, because of system peculiarities, there is http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net/Support/C5.Support.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net/Support/C5.Support.cs b/src/Lucene.Net/Support/C5.Support.cs index 50baa83..ab6600a 100644 --- a/src/Lucene.Net/Support/C5.Support.cs +++ b/src/Lucene.Net/Support/C5.Support.cs @@ -3961,13 +3961,13 @@ namespace Lucene.Net.Support.C5 /// <summary> /// A default generic equality comparer for type T. The procedure is as follows: /// <list> - /// <item>If the actual generic argument T implements the generic interface + /// <item><description>If the actual generic argument T implements the generic interface /// <see cref="T:C5.ISequenced`1"/> for some value W of its generic parameter T, - /// the equalityComparer will be <see cref="T:C5.SequencedCollectionEqualityComparer`2"/></item> - /// <item>If the actual generic argument T implements + /// the equalityComparer will be <see cref="T:C5.SequencedCollectionEqualityComparer`2"/></description></item> + /// <item><description>If the actual generic argument T implements /// <see cref="T:C5.ICollection`1"/> for some value W of its generic parameter T, - /// the equalityComparer will be <see cref="T:C5.UnsequencedCollectionEqualityComparer`2"/></item> - /// <item>Otherwise the SCG.EqualityComparer<T>.Default is returned</item> + /// the equalityComparer will be <see cref="T:C5.UnsequencedCollectionEqualityComparer`2"/></description></item> + /// <item><description>Otherwise the SCG.EqualityComparer<T>.Default is returned</description></item> /// </list> /// </summary> /// <value>The comparer</value> @@ -5311,9 +5311,9 @@ namespace Lucene.Net.Support.C5 /// whose only sign changes when going through items in increasing order /// can be /// <list> - /// <item>from positive to zero</item> - /// <item>from positive to negative</item> - /// <item>from zero to negative</item> + /// <item><description>from positive to zero</description></item> + /// <item><description>from positive to negative</description></item> + /// <item><description>from zero to negative</description></item> /// </list> /// The "cut" function is supplied as the <code>CompareTo</code> method /// of an object <code>c</code> implementing @@ -6030,10 +6030,10 @@ namespace Lucene.Net.Support.C5 /// /// <para>The methods are grouped according to /// <list> - /// <item>Extrema: report or report and delete an extremal item. This is reminiscent of simplified priority queues.</item> - /// <item>Nearest neighbor: report predecessor or successor in the collection of an item. Cut belongs to this group.</item> - /// <item>Range: report a view of a range of elements or remove all elements in a range.</item> - /// <item>AddSorted: add a collection of items known to be sorted in the same order (should be faster) (to be removed?)</item> + /// <item><description>Extrema: report or report and delete an extremal item. This is reminiscent of simplified priority queues.</description></item> + /// <item><description>Nearest neighbor: report predecessor or successor in the collection of an item. Cut belongs to this group.</description></item> + /// <item><description>Range: report a view of a range of elements or remove all elements in a range.</description></item> + /// <item><description>AddSorted: add a collection of items known to be sorted in the same order (should be faster) (to be removed?)</description></item> /// </list> /// </para> /// @@ -6175,9 +6175,9 @@ namespace Lucene.Net.Support.C5 /// whose only sign changes when going through items in increasing order /// can be /// <list> - /// <item>from positive to zero</item> - /// <item>from positive to negative</item> - /// <item>from zero to negative</item> + /// <item><description>from positive to zero</description></item> + /// <item><description>from positive to negative</description></item> + /// <item><description>from zero to negative</description></item> /// </list> /// The "cut" function is supplied as the <code>CompareTo</code> method /// of an object <code>c</code> implementing http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net/Support/Codecs/DefaultCodecFactory.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net/Support/Codecs/DefaultCodecFactory.cs b/src/Lucene.Net/Support/Codecs/DefaultCodecFactory.cs index 8c1ecb1..a08acc8 100644 --- a/src/Lucene.Net/Support/Codecs/DefaultCodecFactory.cs +++ b/src/Lucene.Net/Support/Codecs/DefaultCodecFactory.cs @@ -28,19 +28,19 @@ namespace Lucene.Net.Codecs /// <para/> /// The most common use cases are: /// <list type="bullet"> - /// <item>subclass <see cref="DefaultCodecFactory"/> and override + /// <item><description>subclass <see cref="DefaultCodecFactory"/> and override /// <see cref="DefaultCodecFactory.GetCodec(Type)"/> so an external dependency injection /// container can be used to supply the instances (lifetime should be singleton). Note that you could /// alternately use the "named type" feature that many DI containers have to supply the type based on name by - /// overriding <see cref="GetCodec(string)"/>.</item> - /// <item>subclass <see cref="DefaultCodecFactory"/> and override + /// overriding <see cref="GetCodec(string)"/>.</description></item> + /// <item><description>subclass <see cref="DefaultCodecFactory"/> and override /// <see cref="DefaultCodecFactory.GetCodecType(string)"/> so a type new type can be - /// supplied that is not in the <see cref="DefaultCodecFactory.codecNameToTypeMap"/>.</item> - /// <item>subclass <see cref="DefaultCodecFactory"/> to add new or override the default <see cref="Codec"/> - /// types by overriding <see cref="Initialize()"/> and calling <see cref="PutCodecType(Type)"/>.</item> - /// <item>subclass <see cref="DefaultCodecFactory"/> to scan additional assemblies for <see cref="Codec"/> + /// supplied that is not in the <see cref="DefaultCodecFactory.codecNameToTypeMap"/>.</description></item> + /// <item><description>subclass <see cref="DefaultCodecFactory"/> to add new or override the default <see cref="Codec"/> + /// types by overriding <see cref="Initialize()"/> and calling <see cref="PutCodecType(Type)"/>.</description></item> + /// <item><description>subclass <see cref="DefaultCodecFactory"/> to scan additional assemblies for <see cref="Codec"/> /// subclasses in by overriding <see cref="Initialize()"/> and calling <see cref="ScanForCodecs(Assembly)"/>. - /// For performance reasons, the default behavior only loads Lucene.Net codecs.</item> + /// For performance reasons, the default behavior only loads Lucene.Net codecs.</description></item> /// </list> /// <para/> /// To set the <see cref="ICodecFactory"/>, call <see cref="Codec.SetCodecFactory(ICodecFactory)"/>. http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net/Support/Codecs/DefaultDocValuesFormatFactory.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net/Support/Codecs/DefaultDocValuesFormatFactory.cs b/src/Lucene.Net/Support/Codecs/DefaultDocValuesFormatFactory.cs index a85d6af..f03c3ae 100644 --- a/src/Lucene.Net/Support/Codecs/DefaultDocValuesFormatFactory.cs +++ b/src/Lucene.Net/Support/Codecs/DefaultDocValuesFormatFactory.cs @@ -28,19 +28,19 @@ namespace Lucene.Net.Codecs /// <para/> /// The most common use cases are: /// <list type="bullet"> - /// <item>subclass <see cref="DefaultDocValuesFormatFactory"/> and override + /// <item><description>subclass <see cref="DefaultDocValuesFormatFactory"/> and override /// <see cref="DefaultDocValuesFormatFactory.GetDocValuesFormat(Type)"/> so an external dependency injection /// container can be used to supply the instances (lifetime should be singleton). Note that you could /// alternately use the "named type" feature that many DI containers have to supply the type based on name by - /// overriding <see cref="GetDocValuesFormat(string)"/>.</item> - /// <item>subclass <see cref="DefaultDocValuesFormatFactory"/> and override + /// overriding <see cref="GetDocValuesFormat(string)"/>.</description></item> + /// <item><description>subclass <see cref="DefaultDocValuesFormatFactory"/> and override /// <see cref="DefaultDocValuesFormatFactory.GetDocValuesFormatType(string)"/> so a type new type can be - /// supplied that is not in the <see cref="DefaultDocValuesFormatFactory.docValuesFormatNameToTypeMap"/>.</item> - /// <item>subclass <see cref="DefaultDocValuesFormatFactory"/> to add new or override the default <see cref="DocValuesFormat"/> - /// types by overriding <see cref="Initialize()"/> and calling <see cref="PutDocValuesFormatType(Type)"/>.</item> - /// <item>subclass <see cref="DefaultDocValuesFormatFactory"/> to scan additional assemblies for <see cref="DocValuesFormat"/> + /// supplied that is not in the <see cref="DefaultDocValuesFormatFactory.docValuesFormatNameToTypeMap"/>.</description></item> + /// <item><description>subclass <see cref="DefaultDocValuesFormatFactory"/> to add new or override the default <see cref="DocValuesFormat"/> + /// types by overriding <see cref="Initialize()"/> and calling <see cref="PutDocValuesFormatType(Type)"/>.</description></item> + /// <item><description>subclass <see cref="DefaultDocValuesFormatFactory"/> to scan additional assemblies for <see cref="DocValuesFormat"/> /// subclasses in by overriding <see cref="Initialize()"/> and calling <see cref="ScanForDocValuesFormats(Assembly)"/>. - /// For performance reasons, the default behavior only loads Lucene.Net codecs.</item> + /// For performance reasons, the default behavior only loads Lucene.Net codecs.</description></item> /// </list> /// <para/> /// To set the <see cref="IDocValuesFormatFactory"/>, call <see cref="DocValuesFormat.SetDocValuesFormatFactory(IDocValuesFormatFactory)"/>. http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net/Support/Codecs/DefaultPostingsFormatFactory.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net/Support/Codecs/DefaultPostingsFormatFactory.cs b/src/Lucene.Net/Support/Codecs/DefaultPostingsFormatFactory.cs index 0cbd907..08fb60e 100644 --- a/src/Lucene.Net/Support/Codecs/DefaultPostingsFormatFactory.cs +++ b/src/Lucene.Net/Support/Codecs/DefaultPostingsFormatFactory.cs @@ -28,19 +28,19 @@ namespace Lucene.Net.Codecs /// <para/> /// The most common use cases are: /// <list type="bullet"> - /// <item>subclass <see cref="DefaultPostingsFormatFactory"/> and override + /// <item><description>subclass <see cref="DefaultPostingsFormatFactory"/> and override /// <see cref="DefaultPostingsFormatFactory.GetPostingsFormat(Type)"/> so an external dependency injection /// container can be used to supply the instances (lifetime should be singleton). Note that you could /// alternately use the "named type" feature that many DI containers have to supply the type based on name by - /// overriding <see cref="GetPostingsFormat(string)"/>.</item> - /// <item>subclass <see cref="DefaultPostingsFormatFactory"/> and override + /// overriding <see cref="GetPostingsFormat(string)"/>.</description></item> + /// <item><description>subclass <see cref="DefaultPostingsFormatFactory"/> and override /// <see cref="DefaultPostingsFormatFactory.GetPostingsFormatType(string)"/> so a type new type can be - /// supplied that is not in the <see cref="DefaultPostingsFormatFactory.postingsFormatNameToTypeMap"/>.</item> - /// <item>subclass <see cref="DefaultPostingsFormatFactory"/> to add new or override the default <see cref="PostingsFormat"/> - /// types by overriding <see cref="Initialize()"/> and calling <see cref="PutPostingsFormatType(Type)"/>.</item> - /// <item>subclass <see cref="DefaultPostingsFormatFactory"/> to scan additional assemblies for <see cref="PostingsFormat"/> + /// supplied that is not in the <see cref="DefaultPostingsFormatFactory.postingsFormatNameToTypeMap"/>.</description></item> + /// <item><description>subclass <see cref="DefaultPostingsFormatFactory"/> to add new or override the default <see cref="PostingsFormat"/> + /// types by overriding <see cref="Initialize()"/> and calling <see cref="PutPostingsFormatType(Type)"/>.</description></item> + /// <item><description>subclass <see cref="DefaultPostingsFormatFactory"/> to scan additional assemblies for <see cref="PostingsFormat"/> /// subclasses in by overriding <see cref="Initialize()"/> and calling <see cref="ScanForPostingsFormats(Assembly)"/>. - /// For performance reasons, the default behavior only loads Lucene.Net codecs.</item> + /// For performance reasons, the default behavior only loads Lucene.Net codecs.</description></item> /// </list> /// <para/> /// To set the <see cref="IPostingsFormatFactory"/>, call <see cref="PostingsFormat.SetPostingsFormatFactory(IPostingsFormatFactory)"/>. http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net/Support/HashMap.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net/Support/HashMap.cs b/src/Lucene.Net/Support/HashMap.cs index aaf270b..6a293c3 100644 --- a/src/Lucene.Net/Support/HashMap.cs +++ b/src/Lucene.Net/Support/HashMap.cs @@ -57,15 +57,15 @@ namespace Lucene.Net.Support /// <remarks> /// <h2>Unordered Dictionaries</h2> /// <list type="bullet"> - /// <item><see cref="Dictionary{TKey, TValue}"/> - use when order is not important and all keys are non-null.</item> - /// <item><see cref="HashMap{TKey, TValue}"/> - use when order is not important and support for a null key is required.</item> + /// <item><description><see cref="Dictionary{TKey, TValue}"/> - use when order is not important and all keys are non-null.</description></item> + /// <item><description><see cref="HashMap{TKey, TValue}"/> - use when order is not important and support for a null key is required.</description></item> /// </list> /// <h2>Ordered Dictionaries</h2> /// <list type="bullet"> - /// <item><see cref="LinkedHashMap{TKey, TValue}"/> - use when you need to preserve entry insertion order. Keys are nullable.</item> - /// <item><see cref="SortedDictionary{TKey, TValue}"/> - use when you need natural sort order. Keys must be unique.</item> - /// <item><see cref="TreeDictionary{K, V}"/> - use when you need natural sort order. Keys may contain duplicates.</item> - /// <item><see cref="LurchTable{TKey, TValue}"/> - use when you need to sort by most recent access or most recent update. Works well for LRU caching.</item> + /// <item><description><see cref="LinkedHashMap{TKey, TValue}"/> - use when you need to preserve entry insertion order. Keys are nullable.</description></item> + /// <item><description><see cref="SortedDictionary{TKey, TValue}"/> - use when you need natural sort order. Keys must be unique.</description></item> + /// <item><description><see cref="TreeDictionary{K, V}"/> - use when you need natural sort order. Keys may contain duplicates.</description></item> + /// <item><description><see cref="LurchTable{TKey, TValue}"/> - use when you need to sort by most recent access or most recent update. Works well for LRU caching.</description></item> /// </list> /// </remarks> #if FEATURE_SERIALIZABLE http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net/Support/IO/Buffer.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net/Support/IO/Buffer.cs b/src/Lucene.Net/Support/IO/Buffer.cs index 3892365..79bd73e 100644 --- a/src/Lucene.Net/Support/IO/Buffer.cs +++ b/src/Lucene.Net/Support/IO/Buffer.cs @@ -27,39 +27,39 @@ namespace Lucene.Net.Support.IO /// <para/> /// A buffer can be described by the following properties: /// <list type="bullet"> - /// <item> + /// <item><description> /// Capacity: /// The number of elements a buffer can hold. Capacity may not be /// negative and never changes. - /// </item> - /// <item> + /// </description></item> + /// <item><description> /// Position: /// A cursor of this buffer. Elements are read or written at the /// position if you do not specify an index explicitly. Position may not be /// negative and not greater than the limit. - /// </item> - /// <item> + /// </description></item> + /// <item><description> /// Limit: /// Controls the scope of accessible elements. You can only read or /// write elements from index zero to <c>limit - 1</c>. Accessing /// elements out of the scope will cause an exception. Limit may not be negative /// and not greater than capacity. - /// </item> - /// <item> + /// </description></item> + /// <item><description> /// Mark: /// Used to remember the current position, so that you can reset the /// position later. Mark may not be negative and no greater than position. - /// </item> - /// <item> + /// </description></item> + /// <item><description> /// A buffer can be read-only or read-write. Trying to modify the elements /// of a read-only buffer will cause a <see cref="ReadOnlyBufferException"/>, /// while changing the position, limit and mark of a read-only buffer is OK. - /// </item> - /// <item> + /// </description></item> + /// <item><description> /// A buffer can be direct or indirect. A direct buffer will try its best to /// take advantage of native memory APIs and it may not stay in the heap, /// thus it is not affected by garbage collection. - /// </item> + /// </description></item> /// </list> /// <para/> /// Buffers are not thread-safe. If concurrent access to a buffer instance is http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net/Support/IO/ByteBuffer.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net/Support/IO/ByteBuffer.cs b/src/Lucene.Net/Support/IO/ByteBuffer.cs index 021ce02..709d1aa 100644 --- a/src/Lucene.Net/Support/IO/ByteBuffer.cs +++ b/src/Lucene.Net/Support/IO/ByteBuffer.cs @@ -30,11 +30,11 @@ namespace Lucene.Net.Support.IO /// <para/> /// A byte buffer can be created in either one of the following ways: /// <list type="bullet"> - /// <item><see cref="Allocate(int)"/> a new byte array and create a - /// buffer based on it</item> - /// <item><see cref="AllocateDirect(int)"/> a memory block and create a direct - /// buffer based on it</item> - /// <item><see cref="Wrap(byte[])"/> an existing byte array to create a new buffer</item> + /// <item><description><see cref="Allocate(int)"/> a new byte array and create a + /// buffer based on it</description></item> + /// <item><description><see cref="AllocateDirect(int)"/> a memory block and create a direct + /// buffer based on it</description></item> + /// <item><description><see cref="Wrap(byte[])"/> an existing byte array to create a new buffer</description></item> /// </list> /// </summary> #if FEATURE_SERIALIZABLE http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net/Support/IO/FileSupport.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net/Support/IO/FileSupport.cs b/src/Lucene.Net/Support/IO/FileSupport.cs index 82e8b1c..90f470a 100644 --- a/src/Lucene.Net/Support/IO/FileSupport.cs +++ b/src/Lucene.Net/Support/IO/FileSupport.cs @@ -134,8 +134,8 @@ namespace Lucene.Net.Support.IO /// Creates a new empty file in the specified directory, using the given prefix and suffix strings to generate its name. /// If this method returns successfully then it is guaranteed that: /// <list type="number"> - /// <item>The file denoted by the returned abstract pathname did not exist before this method was invoked, and</item> - /// <item>Neither this method nor any of its variants will return the same abstract pathname again in the current invocation of the virtual machine.</item> + /// <item><description>The file denoted by the returned abstract pathname did not exist before this method was invoked, and</description></item> + /// <item><description>Neither this method nor any of its variants will return the same abstract pathname again in the current invocation of the virtual machine.</description></item> /// </list> /// This method provides only part of a temporary-file facility.To arrange for a file created by this method to be deleted automatically, use the deleteOnExit() method. /// The prefix argument must be at least three characters long. It is recommended that the prefix be a short, meaningful string such as "hjb" or "mail". The suffix argument may be null, in which case the suffix ".tmp" will be used. http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net/Support/IO/LongBuffer.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net/Support/IO/LongBuffer.cs b/src/Lucene.Net/Support/IO/LongBuffer.cs index 82255c7..b278245 100644 --- a/src/Lucene.Net/Support/IO/LongBuffer.cs +++ b/src/Lucene.Net/Support/IO/LongBuffer.cs @@ -30,10 +30,10 @@ namespace Lucene.Net.Support.IO /// <para/> /// A long buffer can be created in either of the following ways: /// <list type="bullet"> - /// <item><see cref="Allocate(int)"/> a new long array and create a buffer - /// based on it</item> - /// <item><see cref="Wrap(long[])"/> an existing long array to create a new - /// buffer</item> + /// <item><description><see cref="Allocate(int)"/> a new long array and create a buffer + /// based on it</description></item> + /// <item><description><see cref="Wrap(long[])"/> an existing long array to create a new + /// buffer</description></item> /// </list> /// </summary> #if FEATURE_SERIALIZABLE http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net/Support/IO/LongToByteBufferAdapter.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net/Support/IO/LongToByteBufferAdapter.cs b/src/Lucene.Net/Support/IO/LongToByteBufferAdapter.cs index b83974d..ca5f4d7 100644 --- a/src/Lucene.Net/Support/IO/LongToByteBufferAdapter.cs +++ b/src/Lucene.Net/Support/IO/LongToByteBufferAdapter.cs @@ -28,10 +28,10 @@ namespace Lucene.Net.Support.IO /// <para/> /// Implementation notice: /// <list type="bullet"> - /// <item>After a byte buffer instance is wrapped, it becomes privately owned by - /// the adapter. It must NOT be accessed outside the adapter any more.</item> - /// <item>The byte buffer's position and limit are NOT linked with the adapter. - /// The adapter extends Buffer, thus has its own position and limit.</item> + /// <item><description>After a byte buffer instance is wrapped, it becomes privately owned by + /// the adapter. It must NOT be accessed outside the adapter any more.</description></item> + /// <item><description>The byte buffer's position and limit are NOT linked with the adapter. + /// The adapter extends Buffer, thus has its own position and limit.</description></item> /// </list> /// </summary> internal sealed class Int64ToByteBufferAdapter : Int64Buffer http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net/Support/LinkedHashMap.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net/Support/LinkedHashMap.cs b/src/Lucene.Net/Support/LinkedHashMap.cs index d241e68..e3d9c94 100644 --- a/src/Lucene.Net/Support/LinkedHashMap.cs +++ b/src/Lucene.Net/Support/LinkedHashMap.cs @@ -47,15 +47,15 @@ namespace Lucene.Net.Support /// <remarks> /// <h2>Unordered Dictionaries</h2> /// <list type="bullet"> - /// <item><see cref="Dictionary{TKey, TValue}"/> - use when order is not important and all keys are non-null.</item> - /// <item><see cref="HashMap{TKey, TValue}"/> - use when order is not important and support for a null key is required.</item> + /// <item><description><see cref="Dictionary{TKey, TValue}"/> - use when order is not important and all keys are non-null.</description></item> + /// <item><description><see cref="HashMap{TKey, TValue}"/> - use when order is not important and support for a null key is required.</description></item> /// </list> /// <h2>Ordered Dictionaries</h2> /// <list type="bullet"> - /// <item><see cref="LinkedHashMap{TKey, TValue}"/> - use when you need to preserve entry insertion order. Keys are nullable.</item> - /// <item><see cref="SortedDictionary{TKey, TValue}"/> - use when you need natural sort order. Keys must be unique.</item> - /// <item><see cref="TreeDictionary{K, V}"/> - use when you need natural sort order. Keys may contain duplicates.</item> - /// <item><see cref="LurchTable{TKey, TValue}"/> - use when you need to sort by most recent access or most recent update. Works well for LRU caching.</item> + /// <item><description><see cref="LinkedHashMap{TKey, TValue}"/> - use when you need to preserve entry insertion order. Keys are nullable.</description></item> + /// <item><description><see cref="SortedDictionary{TKey, TValue}"/> - use when you need natural sort order. Keys must be unique.</description></item> + /// <item><description><see cref="TreeDictionary{K, V}"/> - use when you need natural sort order. Keys may contain duplicates.</description></item> + /// <item><description><see cref="LurchTable{TKey, TValue}"/> - use when you need to sort by most recent access or most recent update. Works well for LRU caching.</description></item> /// </list> /// </remarks> #if FEATURE_SERIALIZABLE http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net/Support/StringExtensions.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net/Support/StringExtensions.cs b/src/Lucene.Net/Support/StringExtensions.cs index fe815d4..e8513f9 100644 --- a/src/Lucene.Net/Support/StringExtensions.cs +++ b/src/Lucene.Net/Support/StringExtensions.cs @@ -36,8 +36,8 @@ namespace Lucene.Net.Support /// <summary> /// This method mimics the Java String.compareTo(String) method in that it /// <list type="number"> - /// <item>Compares the strings using lexographic sorting rules</item> - /// <item>Performs a culture-insensitive comparison</item> + /// <item><description>Compares the strings using lexographic sorting rules</description></item> + /// <item><description>Performs a culture-insensitive comparison</description></item> /// </list> /// This method is a convenience to replace the .NET CompareTo method /// on all strings, provided the logic does not expect specific values http://git-wip-us.apache.org/repos/asf/lucenenet/blob/7099a846/src/Lucene.Net/Util/ArrayUtil.cs ---------------------------------------------------------------------- diff --git a/src/Lucene.Net/Util/ArrayUtil.cs b/src/Lucene.Net/Util/ArrayUtil.cs index afe75f8..2e45dfc 100644 --- a/src/Lucene.Net/Util/ArrayUtil.cs +++ b/src/Lucene.Net/Util/ArrayUtil.cs @@ -813,14 +813,14 @@ namespace Lucene.Net.Util /// <para/> /// The comparer returned depends on the <typeparam name="T"/> argument: /// <list type="number"> - /// <item>If the type is <see cref="string"/>, the comparer returned uses + /// <item><description>If the type is <see cref="string"/>, the comparer returned uses /// the <see cref="string.CompareOrdinal(string, string)"/> to make the comparison /// to ensure that the current culture doesn't affect the results. This is the - /// default string comparison used in Java, and what Lucene's design depends on.</item> - /// <item>If the type implements <see cref="IComparable{T}"/>, the comparer uses + /// default string comparison used in Java, and what Lucene's design depends on.</description></item> + /// <item><description>If the type implements <see cref="IComparable{T}"/>, the comparer uses /// <see cref="IComparable{T}.CompareTo(T)"/> for the comparison. This allows - /// the use of types with custom comparison schemes.</item> - /// <item>If neither of the above conditions are true, will default to <see cref="Comparer{T}.Default"/>.</item> + /// the use of types with custom comparison schemes.</description></item> + /// <item><description>If neither of the above conditions are true, will default to <see cref="Comparer{T}.Default"/>.</description></item> /// </list> /// <para/> /// NOTE: This was naturalComparer() in Lucene
