This is an automated email from the ASF dual-hosted git repository. nightowl888 pushed a commit to branch master in repository https://gitbox.apache.org/repos/asf/lucenenet.git
commit a0cec6174488826af4b970ff63d0bb70b5605bad Author: Shad Storhaug <[email protected]> AuthorDate: Tue Mar 30 19:39:40 2021 +0700 docs: Lucene.Net.Analysis (submodules): Fixed broken formatting and links (see #284, #300) --- src/Lucene.Net.Analysis.Kuromoji/overview.md | 8 ++++---- src/Lucene.Net.Analysis.Morfologik/overview.md | 8 ++++---- src/Lucene.Net.Analysis.Phonetic/overview.md | 8 ++++---- src/Lucene.Net.Analysis.SmartCn/overview.md | 6 +++--- src/Lucene.Net.Analysis.SmartCn/package.md | 8 +++----- 5 files changed, 18 insertions(+), 20 deletions(-) diff --git a/src/Lucene.Net.Analysis.Kuromoji/overview.md b/src/Lucene.Net.Analysis.Kuromoji/overview.md index 1ea768f..26b6544 100644 --- a/src/Lucene.Net.Analysis.Kuromoji/overview.md +++ b/src/Lucene.Net.Analysis.Kuromoji/overview.md @@ -1,4 +1,4 @@ ---- +--- uid: Lucene.Net.Analysis.Ja summary: *content --- @@ -20,8 +20,8 @@ summary: *content limitations under the License. --> - Kuromoji is a morphological analyzer for Japanese text. +Kuromoji is a morphological analyzer for Japanese text. - This module provides support for Japanese text analysis, including features such as part-of-speech tagging, lemmatization, and compound word analysis. +This module provides support for Japanese text analysis, including features such as part-of-speech tagging, lemmatization, and compound word analysis. - For an introduction to Lucene's analysis API, see the <xref:Lucene.Net.Analysis> package documentation. \ No newline at end of file +For an introduction to Lucene's analysis API, see the [Lucene.Net.Analysis](../core/Lucene.Net.Analysis.html) namespace documentation. \ No newline at end of file diff --git a/src/Lucene.Net.Analysis.Morfologik/overview.md b/src/Lucene.Net.Analysis.Morfologik/overview.md index a758f48..629e5df 100644 --- a/src/Lucene.Net.Analysis.Morfologik/overview.md +++ b/src/Lucene.Net.Analysis.Morfologik/overview.md @@ -1,4 +1,4 @@ ---- +--- uid: Lucene.Net.Analysis.Morfologik title: Lucene.Net.Analysis.Morfologik summary: *content @@ -21,8 +21,8 @@ summary: *content limitations under the License. --> - This package provides dictionary-driven lemmatization ("accurate stemming") filter and analyzer for the Polish Language, driven by the [Morfologik library](http://morfologik.blogspot.com/) developed by Dawid Weiss and Marcin Miłkowski. +This package provides dictionary-driven lemmatization ("accurate stemming") filter and analyzer for the Polish Language, driven by the [Morfologik library](http://morfologik.blogspot.com/) developed by Dawid Weiss and Marcin Miłkowski. - For an introduction to Lucene's analysis API, see the <xref:Lucene.Net.Analysis> package documentation. +For an introduction to Lucene's analysis API, see the [Lucene.Net.Analysis](../core/Lucene.Net.Analysis.html) namespace documentation. - The MorfologikFilter yields one or more terms for each token. Each of those terms is given the same position in the index. \ No newline at end of file +The MorfologikFilter yields one or more terms for each token. Each of those terms is given the same position in the index. \ No newline at end of file diff --git a/src/Lucene.Net.Analysis.Phonetic/overview.md b/src/Lucene.Net.Analysis.Phonetic/overview.md index caa5345..f1bd231 100644 --- a/src/Lucene.Net.Analysis.Phonetic/overview.md +++ b/src/Lucene.Net.Analysis.Phonetic/overview.md @@ -1,4 +1,4 @@ ---- +--- uid: Lucene.Net.Analysis.Phonetic summary: *content --- @@ -20,8 +20,8 @@ summary: *content limitations under the License. --> - Analysis for indexing phonetic signatures (for sounds-alike search) +Analysis for indexing phonetic signatures (for sounds-alike search) - For an introduction to Lucene's analysis API, see the <xref:Lucene.Net.Analysis> package documentation. +For an introduction to Lucene's analysis API, see the [Lucene.Net.Analysis](../core/Lucene.Net.Analysis.html) namespace documentation. - This module provides analysis components (using encoders from [Apache Commons Codec](http://commons.apache.org/codec/)) that index and search phonetic signatures. \ No newline at end of file +This module provides analysis components (using encoders ported to .NET from [Apache Commons Codec](http://commons.apache.org/codec/)) that index and search phonetic signatures. \ No newline at end of file diff --git a/src/Lucene.Net.Analysis.SmartCn/overview.md b/src/Lucene.Net.Analysis.SmartCn/overview.md index 0500484..061643b 100644 --- a/src/Lucene.Net.Analysis.SmartCn/overview.md +++ b/src/Lucene.Net.Analysis.SmartCn/overview.md @@ -1,4 +1,4 @@ ---- +--- uid: Lucene.Net.Analysis.Cn.Smart summary: *content --- @@ -20,6 +20,6 @@ summary: *content limitations under the License. --> - Analyzer for Simplified Chinese, which indexes words. +Analyzer for Simplified Chinese, which indexes words. - For an introduction to Lucene's analysis API, see the <xref:Lucene.Net.Analysis> package documentation. \ No newline at end of file +For an introduction to Lucene's analysis API, see the [Lucene.Net.Analysis](../core/Lucene.Net.Analysis.html) namespace documentation. \ No newline at end of file diff --git a/src/Lucene.Net.Analysis.SmartCn/package.md b/src/Lucene.Net.Analysis.SmartCn/package.md index 18dcfa9..5f52530 100644 --- a/src/Lucene.Net.Analysis.SmartCn/package.md +++ b/src/Lucene.Net.Analysis.SmartCn/package.md @@ -1,4 +1,4 @@ ---- +--- uid: Lucene.Net.Analysis.Cn.Smart summary: *content --- @@ -22,12 +22,12 @@ summary: *content Analyzer for Simplified Chinese, which indexes words. @lucene.experimental -<div> + Three analyzers are provided for Chinese, each of which treats Chinese text in a different way. * StandardAnalyzer: Index unigrams (individual Chinese characters) as a token. -* CJKAnalyzer (in the analyzers/cjk package): Index bigrams (overlapping groups of two adjacent Chinese characters) as tokens. +* CJKAnalyzer (in the <xref:Lucene.Net.Analysis.Cjk> namespace of <xref:Lucene.Net.Analysis.Common>): Index bigrams (overlapping groups of two adjacent Chinese characters) as tokens. * SmartChineseAnalyzer (in this package): Index words (attempt to segment Chinese text into words) as tokens. @@ -39,5 +39,3 @@ Example phrase: "我是中国人" 2. CJKAnalyzer: 我是-是中-中国-国人 3. SmartChineseAnalyzer: 我-是-中国-人 - -</div> \ No newline at end of file
