Well, no - we don't care so much about the seconds, but hours & minutes are indeed crucial.

---
Alok K. Dhir
Symplicity Corporation
www.symplicity.com
(703) 351-0200 x 8080
[EMAIL PROTECTED]

On Oct 29, 2008, at 4:41 PM, Chris Harris wrote:

Do you need to search down to the minutes and seconds level? If searching by date provides sufficient granularity, for instance, you can normalize all the time-of-day portions of the timestamps to midnight while indexing. (So index any event happening on Oct 01, 2008 as 2008-10-01T00:00:00Z.) That
would give Solr many fewer unique timestamp values to go through.

On Wed, Oct 29, 2008 at 1:30 PM, Alok Dhir <[EMAIL PROTECTED]> wrote:

Hi -- using solr 1.3 -- roughly 11M docs on a 64 gig 8 core machine.

Fairly simple schema -- no large text fields, standard request handler. 4
small facet fields.

The index is an event log -- a primary search/retrieval requirement is date
range queries.

A simple query without a date range subquery is ridiculously fast - 2ms.
The same query with a date range takes up to 30s (30,000ms).

Concrete example, this query just look 18s:

instance:client\-csm.symplicity.com AND dt: [2008-10-01T04:00:00Z TO
2008-10-30T03:59:59Z] AND label_facet:"Added to Position"

The exact same query without the date range took 2ms.

I saw a thread from Apr 2008 which explains the problem being due to too much precision on the DateField type, and the range expansion leading to far too many elements being checked. Proposed solution appears to be a hack where you index date fields as strings and hacking together date functions
to generate proper queries/format results.

Does this remain the recommended solution to this issue?

Thanks

---
Alok K. Dhir
Symplicity Corporation
www.symplicity.com
(703) 351-0200 x 8080
[EMAIL PROTECTED]



Reply via email to