Cheolgoo Kang wrote:
>Thanks Bialecki,
>
>
Bialecki is my last name, my first name is Andrzej. No problem, it's
similarly confusing for Europeans to decide between the first and last
name in Asian names... :-) Is your first name Kang?
>I'm trying to test your program, thanks a lot!
>
>And also
Thanks Bialecki,
I'm trying to test your program, thanks a lot!
And also, can you give me the paper you've cited [1] and [2]? I've
googled(entire web and google scholar) about it but got nothing.
On 11/8/05, Andrzej Bialecki <[EMAIL PROTECTED]> wrote:
> KwonNam Son wrote:
>
> >First of all, I re
KwonNam Son wrote:
First of all, I really appreciate your work on Lucene for Korean words,
But If we cannot support stem analyzer for Korean words, I think one
token for one Korean character is better.
When we search a word, usually we use "검색" not "검색하다". ("하다" is like
"ed" of "searched").
If
t Appendix D. That's
> just simple enough to hold a list of Korean word endings, and check
> for each Token with matching word endings. :)
>
> >
> > I think there is some tradeoff here.
> >
> > Maybe need some good stop filter for korean etc...
> >
> &
- Original Message -
From: "Cheolgoo Kang" <[EMAIL PROTECTED]>
To: "Youngho Cho" <[EMAIL PROTECTED]>
Cc:
Sent: Tuesday, November 08, 2005 5:53 PM
Subject: Re: korean and lucene
> On 11/8/05, Youngho Cho <[EMAIL PROTECTED]> wrote:
> &g
ken with matching word endings. :)
>
> I think there is some tradeoff here.
>
> Maybe need some good stop filter for korean etc...
>
>
> Thanks
>
> Youngho
>
> - Original Message -
> From: "Youngho Cho" <[EMAIL PROTECTED]>
> To:
than good match.
I think there is some tradeoff here.
Maybe need some good stop filter for korean etc...
Thanks
Youngho
- Original Message -
From: "Youngho Cho" <[EMAIL PROTECTED]>
To:
Sent: Tuesday, November 08, 2005 4:44 PM
Subject: Re: korean and lu
Hello Cheolgoo,
I will test the patch.
Thanks,
Youngho
- Original Message -
From: "Cheolgoo Kang" <[EMAIL PROTECTED]>
To: ; "Youngho Cho" <[EMAIL PROTECTED]>
Sent: Tuesday, November 08, 2005 4:06 PM
Subject: Re: korean and lucene
> Hello,
>
ysis">
> >
> >
> > Japanese Test...
> >
> >
> >
> >
> >
> >
> >> description="Examples of Korean analysis">
> >
> >
> > Korean Test...
> >
> >
> >
> >
>
t; Japanese Test...
>
>
>
>
>
>
> description="Examples of Korean analysis">
>
>
> Korean Test...
>
>
>
>
>
>
>
> Thanks,
>
> Youngho
>
>
> ----- Original Messa
anese Test...
Korean Test...
Thanks,
Youngho
- Original Message -
From: "Youngho Cho" <[EMAIL PROTECTED]>
To: ; "Youngho Cho" <[EMAIL PROTECTED]>
Sent: Thursday, October 27, 2005 12:47 PM
Subject: Re: korean and lucene
> Hello all
thanks Koji and Cheolgoo
I will further test now...
Youngho
- Original Message -
From: "Youngho Cho" <[EMAIL PROTECTED]>
To:
Sent: Thursday, October 27, 2005 12:28 PM
Subject: Re: korean and lucene
> Hello Koji
>
> Here is test result.
> Japanese
iguchi" <[EMAIL PROTECTED]>
To: ; "Youngho Cho" <[EMAIL PROTECTED]>
Sent: Thursday, October 27, 2005 11:47 AM
Subject: RE: korean and lucene
> Hello Youngho,
>
> I don't understand why you couldn't get hits result in Japanese,
> though, you had better c
t line should be query = 경
to get hits result. Can you check why StandardAnalyzer
removes "경" during tokenizing?
Koji
> -Original Message-
> From: Youngho Cho [mailto:[EMAIL PROTECTED]
> Sent: Thursday, October 27, 2005 11:37 AM
> To: java-user@lucene.apache.org
&g
t;[EMAIL PROTECTED]>
To: ; "Youngho Cho" <[EMAIL PROTECTED]>
Sent: Thursday, October 27, 2005 9:48 AM
Subject: RE: korean and lucene
> Hi Youngho,
>
> With regard to Japanese, using StandardAnalyzer,
> I can search a word/phase.
>
> Did you use QueryParser?
ssage-
> From: Youngho Cho [mailto:[EMAIL PROTECTED]
> Sent: Thursday, October 27, 2005 8:18 AM
> To: java-user@lucene.apache.org; Cheolgoo Kang
> Subject: Re: korean and lucene
>
>
> Hello Cheolgoo,
>
> Now I updated my lucene version to 1.9 for using StandardAnalyz
ay, October 04, 2005 10:11 AM
Subject: Re: korean and lucene
> StandardAnalyzer's JavaCC based StandardTokenizer.jj cannot read
> Korean part of Unicode character blocks.
>
> You should 1) use CJKAnalyzer or 2) add Korean character
> block(0xAC00~0xD7AF) to the CJK token de
StandardAnalyzer's JavaCC based StandardTokenizer.jj cannot read
Korean part of Unicode character blocks.
You should 1) use CJKAnalyzer or 2) add Korean character
block(0xAC00~0xD7AF) to the CJK token definition on the
StandardTokenizer.jj file.
Hope it helps.
On 10/4/05, John Wang <[EMAIL PROT
-
From: "John Wang" <[EMAIL PROTECTED]>
To:
Sent: Tuesday, October 04, 2005 8:46 AM
Subject: korean and lucene
Hi:
We are running into problems with searching on korean documents. We are
using the StandardAnalyzer and everything works with Chinese and Japanese.
Are there kno
Hi:
We are running into problems with searching on korean documents. We are
using the StandardAnalyzer and everything works with Chinese and Japanese.
Are there known problems with Korean with Lucene?
Thanks
-John
20 matches
Mail list logo