Re: Choosing what document to return.

2010-01-24 Thread Noam G.

Hi Mei

Thank you very much - looks good :-)

Noam.
-- 
View this message in context: 
http://old.nabble.com/Choosing-what-document-to-return.-tp27259310p27293083.html
Sent from the Solr - User mailing list archive at Nabble.com.



Re: Replication Handler Severe Error: Unable to move index file

2010-01-24 Thread Noble Paul നോബിള്‍ नोब्ळ्
On Fri, Jan 22, 2010 at 4:24 AM, Trey solrt...@gmail.com wrote:
 Unfortunately, when I went back to look at the logs this morning, the log
 file had been blown away... that puts a major damper on my debugging
 capabilities - so sorry about that.  As a double whammy, we optimize
 nightly, so the old index files have completely changed at this point.

 I do not remember seeing an exception / stack trace in the logs associated
 with the SEVERE *Unable to move file* entry, but we were grepping the
 logs, so if it was outputted onto another line it could have possibly been
 there.  I wouldn't really expect to see anything based upon the code in
 SnapPuller.java:

 /**
   * Copy a file by the File#renameTo() method. If it fails, it is
 considered a failure
   * p/
   * Todo may be we should try a simple copy if it fails
   */
  private boolean copyAFile(File tmpIdxDir, File indexDir, String fname,
 ListString copiedfiles) {
    File indexFileInTmpDir = new File(tmpIdxDir, fname);
    File indexFileInIndex = new File(indexDir, fname);
    boolean success = indexFileInTmpDir.renameTo(indexFileInIndex);
    if (!success) {
      LOG.error(Unable to move index file from:  + indexFileInTmpDir
              +  to:  + indexFileInIndex);
      for (String f : copiedfiles) {
        File indexFile = new File(indexDir, f);
        if (indexFile.exists())
          indexFile.delete();
      }
      delTree(tmpIdxDir);
      return false;
    }
    return true;
  }

 In terms of whether this is an off case: this is the first occurrence of
 this I have seen in the logs.  We tried to replicate the conditions under
 which the exception occurred, but were unable.  I'll send along some more
 useful info if this happens again.

 In terms of the behavior we saw: It appears that a replication occurred and
 the Unable to move file error occurred.  As a result, it looks like the
 ENTIRE index was subsequently replicated again into a temporary directory
 (several times, over and over).

 The end result was that we had multiple full copies of the index in
 temporary index folders on the slave, and the original still couldn't be
 updated (the move to ./index wouldn't work).  Does Solr ever hold files open
 in a manner that would prevent a file in the index directory from being
 overridden?

There is a TODO which says manual it try to copy if move (renameTo)
fails. We never did it because we never observed renameTo failing.


 2010/1/21 Noble Paul നോബിള്‍ नोब्ळ् noble.p...@corp.aol.com

 is it a one off case? do you observerve this frequently?

 On Thu, Jan 21, 2010 at 11:26 AM, Otis Gospodnetic
 otis_gospodne...@yahoo.com wrote:
  It's hard to tell without poking around, but one of the first things I'd
 do would be to look for /home/solr/cores/core8/index.20100119103919/_6qv.fnm
 - does this file/dir really exist?  Or, rather, did it exist when the error
 happened.
 
  I'm not looking at the source code now, but is that really the only error
 you got?  No exception stack trace?
 
   Otis
  --
  Sematext -- http://sematext.com/ -- Solr - Lucene - Nutch
 
 
 
  - Original Message 
  From: Trey solrt...@gmail.com
  To: solr-user@lucene.apache.org
  Sent: Wed, January 20, 2010 11:54:43 PM
  Subject: Replication Handler Severe Error: Unable to move index file
 
  Does anyone know what would cause the following error?:
 
  10:45:10 AM org.apache.solr.handler.SnapPuller copyAFile
 
       SEVERE: *Unable to move index file* from:
  /home/solr/cores/core8/index.20100119103919/_6qv.fnm to:
  /home/solr/cores/core8/index/_6qv.fnm
  This occurred a few days back and we noticed that several full copies of
 the
  index were subsequently pulled from the master to the slave, effectively
  evicting our live index from RAM (the linux os cache), and killing our
 query
  performance due to disk io contention.
 
  Has anyone experienced this behavior recently?  I found an old thread
 about
  this error from early 2009, but it looks like it was patched almost a
 year
  ago:
 
 http://old.nabble.com/%22Unable-to-move-index-file%22-error-during-replication-td21157722.html
 
 
  Additional Relevant information:
  -We are using the Solr 1.4 official release + a field collapsing patch
 from
  mid December (which I believe should only affect query side, not
 indexing /
  replication).
  -Our Replication PollInterval for slaves checking the master is very
 small
  (15 seconds)
  -We have a multi-box distributed search with each box possessing
 multiple
  cores
  -We issue a manual (rolling) optimize across the cores on the master
 once a
  day (occurred ~ 1-2 hours before the above timeline)
  -maxWarmingSearchers is set to 1.
 
 



 --
 -
 Noble Paul | Systems Architect| AOL | http://aol.com





-- 
-
Noble Paul | Systems Architect| AOL | http://aol.com


after migration to solr 1.4, queries always return 0 items

2010-01-24 Thread athir nuaimi
We have a working system based on Solr 1.3 that I tried to upgrade to 1.4 with 
no luck.  The new 1.4 Solr server is up and running but always returns 0 
results when you query it.  It seems to have the correct schema but the number 
of documents is 0 even though I used a backup of the production data.

I created the 1.4 solr server using a few machine and following the 
instructions on the solr wiki.   For the data, I used the 'backup' command to 
make a copy of the live index and moved to the data directory on the new 1.4 
server.   When I start the new server, the tomcat catinla logs don't have any 
errors (only info msgs).  The logs show the correct solr home and I can see its 
reading the schema.xml (acts_as_solr).   Finally, I see that the data directory 
where the index is has new subfolders for spellchecker  spellchecker2 so I 
assume it is looking the correct directory. 

I'd appreciate any tips on how to debug this.  Below are some more details on 
the setup..

thanks in advance.
Athir



Re: after migration to solr 1.4, queries always return 0 items

2010-01-24 Thread Erik Hatcher

You omitted the details.

Is there an index directory under your data directory?  That's the  
heart of the Lucene index there.  Look at Solr's admin, it'll tell you  
the data directory location it is using, be sure it is set where you  
think it should be.


Erik

On Jan 24, 2010, at 8:59 AM, athir nuaimi wrote:

We have a working system based on Solr 1.3 that I tried to upgrade  
to 1.4 with no luck.  The new 1.4 Solr server is up and running but  
always returns 0 results when you query it.  It seems to have the  
correct schema but the number of documents is 0 even though I used a  
backup of the production data.


I created the 1.4 solr server using a few machine and following the  
instructions on the solr wiki.   For the data, I used the 'backup'  
command to make a copy of the live index and moved to the data  
directory on the new 1.4 server.   When I start the new server, the  
tomcat catinla logs don't have any errors (only info msgs).  The  
logs show the correct solr home and I can see its reading the  
schema.xml (acts_as_solr).   Finally, I see that the data directory  
where the index is has new subfolders for spellchecker   
spellchecker2 so I assume it is looking the correct directory.


I'd appreciate any tips on how to debug this.  Below are some more  
details on the setup..


thanks in advance.
Athir





Re: after migration to solr 1.4, queries always return 0 items

2010-01-24 Thread athir nuaimi
Hi Eric.  Thanks for the quick reply.  I though the backup went in /data.  I 
have restored the backup to /data/index and am now getting results.


On 2010-01-24, at 9:46 AM, Erik Hatcher wrote:

 You omitted the details.
 
 Is there an index directory under your data directory?  That's the heart of 
 the Lucene index there.  Look at Solr's admin, it'll tell you the data 
 directory location it is using, be sure it is set where you think it should 
 be.
 
   Erik
 
 On Jan 24, 2010, at 8:59 AM, athir nuaimi wrote:
 
 We have a working system based on Solr 1.3 that I tried to upgrade to 1.4 
 with no luck.  The new 1.4 Solr server is up and running but always returns 
 0 results when you query it.  It seems to have the correct schema but the 
 number of documents is 0 even though I used a backup of the production data.
 
 I created the 1.4 solr server using a few machine and following the 
 instructions on the solr wiki.   For the data, I used the 'backup' command 
 to make a copy of the live index and moved to the data directory on the new 
 1.4 server.   When I start the new server, the tomcat catinla logs don't 
 have any errors (only info msgs).  The logs show the correct solr home and I 
 can see its reading the schema.xml (acts_as_solr).   Finally, I see that the 
 data directory where the index is has new subfolders for spellchecker  
 spellchecker2 so I assume it is looking the correct directory.
 
 I'd appreciate any tips on how to debug this.  Below are some more details 
 on the setup..
 
 thanks in advance.
 Athir
 
 



strategy for snapshottig Solr data directory on EC2

2010-01-24 Thread athir nuaimi
We are running two Solr servers (master/slave) on EC2 and have the solr home 
directories on EBS drives that we snapshot every 12 hours.  While that will 
mean that we will lose at most 12 hours of data, I wondered if there was a way 
I could reduce the window of data loss.   With our mysql servers, we snapshot 
every 12 hours but also copy the binary logs to S3 every 5 minutes.

We are doing commits every 10 minutes on the master and will be using the 
built-in java replication (today we are using snapshotting to replicate but are 
in the process of migrating from 1.3 to 1.4).

On a related note, are we doing the right thing in having our slave solr home 
directory on an EBS volume?  If the slave were to die and we had to create a 
fresh one, will it just resync the entire index from the master?  is the reason 
to have the slave on an EBS volume so that the slave has less data to resync on 
startup?

thanks in advance
Athir

Re: strategy for snapshottig Solr data directory on EC2

2010-01-24 Thread William Pierce

Our setup on ec2 is as follows:

a) mysql master on ebs volume.
b) solr master on its own ebs volume
c) solr slaves do not use ebs -- but rather use the ephemeral instance 
stores.  There is a small period of time where the solr slave has to re-sync 
the data from the solr master.


Cheers,

Bill

--
From: athir nuaimi at...@nuaim.com
Sent: Sunday, January 24, 2010 7:35 AM
To: solr-user@lucene.apache.org
Subject: strategy for snapshottig Solr data directory on EC2

We are running two Solr servers (master/slave) on EC2 and have the solr 
home directories on EBS drives that we snapshot every 12 hours.  While 
that will mean that we will lose at most 12 hours of data, I wondered if 
there was a way I could reduce the window of data loss.   With our mysql 
servers, we snapshot every 12 hours but also copy the binary logs to S3 
every 5 minutes.


We are doing commits every 10 minutes on the master and will be using the 
built-in java replication (today we are using snapshotting to replicate 
but are in the process of migrating from 1.3 to 1.4).


On a related note, are we doing the right thing in having our slave solr 
home directory on an EBS volume?  If the slave were to die and we had to 
create a fresh one, will it just resync the entire index from the master? 
is the reason to have the slave on an EBS volume so that the slave has 
less data to resync on startup?


thanks in advance
Athir 




Default value attribute in RSS DIH

2010-01-24 Thread David Stuart
Hey All,

Can anyone tell me what the attribute name is for defining a default value in 
the field tag of the RSS data import handler??

Basically I want to do something like
field column=type value=external_source commonField=true/   


Any Ideas?


Regards,


Dave

Retrieve Field Names:

2010-01-24 Thread Kumaravel Kandasami
How to retrieve field names from the dynamic schema in the client side
program ?
Any links or reference would be helpful.

(Similar to the schema browser on the SOLR admin screen.)



Kumar_/|\_
www.saisk.com
ku...@saisk.com
making a profound difference with knowledge and creativity...


using bq with standard request handler

2010-01-24 Thread gunjan_versata


Hi all,

I am using a query like:
http://localhost:8080/solr/select?q=product_category:Groceryfq=in_stock:truedebugQuery=true;
sort=display_priority+desc,prop_l_price+asc

To first sort on display_priority and then on price..

Is it possible to use display_priority/price fields in bq itself to acheive
the same result.. I tried forming some queries that but was unable to get
desired results...

Any kind of help would be highly appreciated...

Regards,
Gunjan
-- 
View this message in context: 
http://old.nabble.com/using-bq-with-standard-request-handler-tp27302951p27302951.html
Sent from the Solr - User mailing list archive at Nabble.com.