Oh my god, forget it: it didn't index because of the line breaks... silly of me grmmm sorry :-\
On Apr 7, 5:29 pm, IvanHQ <[email protected]> wrote: > Thanx James! > > Well, the problem is that I need to "enable_star" on the model or it gives > "ERROR [...] enable_star=0" and indexes nothing. > If I enable star, then it gives no error, but does not index any string > containing accents :-\ > > Maybe I'm doing something wrong... I followed your tuto letter by letter :-\ > > Iván > > > > On Tue, Apr 7, 2009 at 2:51 PM, James Healy <[email protected]> wrote: > > > Hi Iván, > > > Iván Belmonte wrote: > > > Is it possible to make thinking_sphinx ignore accents on searches? I > > > mean... make "perdón" and "perdon" match the same? > > > > I've thought about indexing directly with no accents, and then parse > > > search strings eveytime to remove its accents before the match. The > > > question is: how can I make TS index without accents? > > > The trick is to use a sphinx feature called charset_table. I publised a > > blog article on how to use it last year: > > >http://yob.id.au/blog/2008/05/08/thinking_sphinx_and_unicode/ > > > -- James Healy <jimmy-at-deefa-dot-com> Tue, 07 Apr 2009 22:49:41 +1000 > > -- > > ____________________________________________________| > / > | Iván > | [email protected] > \________________________________ --~--~---------~--~----~------------~-------~--~----~ You received this message because you are subscribed to the Google Groups "Thinking Sphinx" group. To post to this group, send email to [email protected] To unsubscribe from this group, send email to [email protected] For more options, visit this group at http://groups.google.com/group/thinking-sphinx?hl=en -~----------~----~----~----~------~----~------~--~---
