yep,lt is a bad idea to eliminate stopword during indexing,may be u
can eliminate stopword during querying,That is flexible
2012/11/28 Walter Underwood wun...@wunderwood.org
Eliminating stopwords is generally a bad idea. It means you cannot search
for vitamin a.
justt no stopwords are considered in that case
2012/11/28 Joe Zhang smartag...@gmail.com
t no stopwords are considered in
this case
MMSeg is better than ikanalyzer? is there anyone compare those open source
Chinese word analyzers? which one is better overall?
2012/10/9 Peter Keegan peterlkee...@gmail.com
We're using MMSeg with Lucene, but not Solr. Since each SolrCore is
independent, I'm not sure how you can avoid each
thank you . Hoss
Tokenization is not a problem in english,but in some other languages like
chinese, there are no space to seperate each term in article.Lt is a long
string like this “AABCDAEFSABS”,in which “AA” and BCD ...represent a
meaningful term ,so l want to boost some special and meaningful
if the query word is ABCD,then after being tokenized it is A BC D ,
l want to boost term BC ,so the query word is like this: A BC^10 D and
phrase query ABCD . all query words users typing in will be processed
like that automaticly.
l guess l can custom a filter to do it ,but l don't know if it is
if the query word is ABCD,then after being tokenized it is A BC D ,
l want to boost term BC ,so the query word is like this: A BC^10 D and
phrase query ABCD . all query words users typing in will be processed
like that automaticly.
l guess l can custom a filter to do it ,but l don't know if it is
u can use boost function query
2012/9/29 Erick Erickson erickerick...@gmail.com
How can Solr/Lucene distinguish between that which must be boosted
and that which must not? There must be some rule here, what is it?
Until one understands the rule, there's not much one can suggest...
Best