Hi - we are actually using some other filters for Chinese, although they are 
not specialized for Chinese:

        <tokenizer class="solr.StandardTokenizerFactory"/>
        <filter class="solr.CJKWidthFilterFactory"/>
        <filter class="solr.LowerCaseFilterFactory"/>
        <filter class="solr.CJKBigramFilterFactory"/>
 
 
-----Original message-----
> From:Zheng Lin Edwin Yeo <edwinye...@gmail.com>
> Sent: Thursday 25th June 2015 11:24
> To: solr-user@lucene.apache.org
> Subject: Re: Tokenizer and Filter Factory to index Chinese characters
> 
> Thank you.
> 
> I've tried that, but when I do a search, it's returning much more
> highlighted results that what it supposed to.
> 
> For example, if I enter the following query:
> http://localhost:8983/solr/chinese1/highlight?q=我国
> 
> I get the following results:
> 
> "highlighting":{
>     "chinese1":{
>       "id":["chinese1"],
>       
> "title":["<em>我国</em>1<em>月份</em>的制造业<em>产值</em><em>同比</em>仅<em>增长</em>0"],
>       
> "content":["<em>结束</em><em>连续</em>两个月的<em>萎缩</em>,但比经济师<em>普遍</em><em>预估</em>的<em>增长</em>3.3%<em>疲软</em>得多。这也意味着,<em>我国</em><em>今年</em><em>第一</em><em>季度</em>的<em>经济</em>很<em>可能</em>让人<em>失望</em>
> \n  "],
>       "author":["<em>Edwin</em>"]},
>     "chinese2":{
>       "id":["chinese2"],
>       
> "content":["<em>铜牌</em>,让<em>我国</em><em>暂时</em><em>高居</em><em>奖牌</em><em>荣誉</em>榜<em>榜首</em>。
> 你看好新加坡在本届的东运会中,会<em>夺得</em><em>多少</em>面<em>金牌</em>?
> 请在6月<em>12</em>日<em>中午</em>前,<em>投票</em>并<em>留言</em>为<em>我国</em><em>健将</em>寄上<em>祝语</em>吧
>  \n  "],
>       "author":["<em>Edwin</em>"]},
>     "chinese3":{
>       "id":["chinese3"],
>       
> "content":[")<em>组成</em>的<em>我国</em><em>女队</em>在<em>今天</em>的东运会保龄球<em>女子</em>三人赛中,
> 以六局3963总瓶分<em>夺冠</em>,为新加坡<em>赢得</em><em>本届</em><em>赛会</em>第三枚<em>金牌</em>。<em>队友</em>陈诗桦(Jazreel)、梁蕙芬和陈诗静以3707总瓶分<em>获得</em><em>亚军</em>,<em>季军</em>归菲律宾<em>女队</em>。(<em>联合</em><em>早报</em><em>记者</em>:郭嘉惠)
> \n  "],
>       "author":["<Edwin"]},
>     "chinese4":{
>       "id":["chinese4"],
>       
> "content":["<em>配套</em>的<em>读者</em>,则可<em>获得</em>一架<em>价值</em>309元的Philips
> Viva Collection <em>HD</em>9045面<em>包机</em>。 \n
> 欲订从速,<em>读者</em>可<em>登陆</em>www.wbsub.com.<em>sg</em>,或拨打客服<em>专线</em>6319
> 1800<em>订购</em>。 \n
> <em>此外</em>,一年一度的<em>晚报</em><em>保健</em><em>美容</em>展,将在<em>本月</em><em>23</em>日和<em>24</em>日,在新达新加坡<em>会展</em><em>中心</em>401、402<em>展厅</em><em>举行</em>。
> \n 
> <em>现场</em>将<em>开设</em>《<em>联合</em><em>晚报</em>》<em>订阅</em>展摊,<em>读者</em><em>当场</em><em>订阅</em><em>晚报</em>,<em>除了</em>可<em>获得</em><em>丰厚</em>的<em>赠品</em>,还有<em>机会</em><em>参与</em>“"],
>       "author":["<em>Edwin</em>"]}}}
> 
> 
> Is there any suitable filter factory to solve this issue?
> 
> I've tried WordDelimiterFilterFactory, PorterStemFilterFactory
> and StopFilterFactory, but there's no improvement in the search results.
> 
> 
> Regards,
> Edwin
> 
> 
> On 25 June 2015 at 17:17, Markus Jelsma <markus.jel...@openindex.io> wrote:
> 
> > Hello - you can use HMMChineseTokenizerFactory instead.
> >
> > http://lucene.apache.org/core/5_2_0/analyzers-smartcn/org/apache/lucene/analysis/cn/smart/HMMChineseTokenizerFactory.html
> >
> > -----Original message-----
> > > From:Zheng Lin Edwin Yeo <edwinye...@gmail.com>
> > > Sent: Thursday 25th June 2015 11:02
> > > To: solr-user@lucene.apache.org
> > > Subject: Tokenizer and Filter Factory to index Chinese characters
> > >
> > > Hi,
> > >
> > > Does anyone knows what is the correct replacement for these 2 tokenizer
> > and
> > > filter factory to index chinese into Solr?
> > > - SmartChineseSentenceTokenizerFactory
> > > - SmartChineseWordTokenFilterFactory
> > >
> > > I understand that these 2 tokenizer and filter factory are already
> > > deprecated in Solr 5.1, but I can't seem to find the correct replacement.
> > >
> > >
> > > <fieldType name="text_smartcn" class="solr.TextField"
> > > positionIncrementGap="0">
> > >           <analyzer type="index">
> > >             <tokenizer
> > >
> > class="org.apache.lucene.analysis.cn.smart.SmartChineseSentenceTokenizerFactory"/>
> > >             <filter
> > >
> > class="org.apache.lucene.analysis.cn.smart.SmartChineseWordTokenFilterFactory"/>
> > >           </analyzer>
> > >           <analyzer type="query">
> > >             <tokenizer
> > >
> > class="org.apache.lucene.analysis.cn.smart.SmartChineseSentenceTokenizerFactory"/>
> > >             <filter
> > >
> > class="org.apache.lucene.analysis.cn.smart.SmartChineseWordTokenFilterFactory"/>
> > >           </analyzer>
> > > </fieldType>
> > >
> > > Thank you.
> > >
> > >
> > > Regards,
> > > Edwin
> > >
> >
> 

Reply via email to