From another thread:
Spake Grant Ingersoll:
You can have multiple lat/lons per document, you just can't have multiple per
field.
Is this a temporary limitation, a quirk of the LatLon type, or is it an
architectural limitation in the compound type design?
Lance Norskog
).
What experience do others have using a large mergeFactor?
Tom
--
Lance Norskog
goks...@gmail.com
--
Lance Norskog
goks...@gmail.com
change?
[...]
You will need to reindex if the schema is changed.
Regards,
Gora
--
Lance Norskog
goks...@gmail.com
above also explains how these are handled.
Without going through the above could you not just set the timezone
for reg_date to UTC to get the result that you expect?
Regards,
Gora.
--
Lance Norskog
goks...@gmail.com
of a script transform vs.
the same transform done in java.
thanks,
Tim
--
Lance Norskog
goks...@gmail.com
.
peter.wola...@acquia.com : 978-296-5247
Get a free, hosted Drupal 7 site: http://www.drupalgardens.com;
--
Lance Norskog
goks...@gmail.com
-user@lucene.apache.org
hello,
Is there any api in SolrJ that calls the dataImportHandler to execute
commands like full-import and delta-import. Please help..
--
Lance Norskog
goks...@gmail.com
concerns would be really appreciated.
Thanks in advance.
--
Regards,
Tharindu
--
Regards,
Tharindu
--
Lance Norskog
goks...@gmail.com
. Thank you.
--
Lance Norskog
goks...@gmail.com
2007-06-19 09:08:48
Solr's input format is '2007-06-19T09:08:48Z'.
More to the point: you are creating a string and passing that in. The
date type will accept this, but the DIH has code to accept Java JDBC
datetime values directly. So, in your select you want to somehow cast
your field data
Correct. We used the Latin1 filter back then.
Chris Hostetter wrote:
: I am using solr 1.3. I get the below mentioned error when included the
: solr.ASCIIFoldingFilterFactory on 'text' field while index and query
: time:
I fairly certain ASCIIFoldingFilterFactory did not exist in Solr 1.3.
for your help.
--
Tanguy
--
Lance Norskog
goks...@gmail.com
in 3.1 compatible with both 3.1 and 1.4.1? If not, that's going to
make a graceful upgrade of my replicated distributed installation a little
harder.
Thanks,
Shawn
--
Lance Norskog
goks...@gmail.com
,
could that be enough to somehow increase memory usage enough to run into
OOM?
--
Lance Norskog
goks...@gmail.com
the memory overhead of adding cores (aside
from
needing to fit the new index in memory).
Thoughts?
-mike
--
Lance Norskog
goks...@gmail.com
--
Lance Norskog
goks...@gmail.com
index from a
text field which some people can see and other people can't.
On Tue, Oct 26, 2010 at 12:06 AM, Lance Norskog goks...@gmail.com wrote:
Filter queries are a set of bits which is ANDed against query results
at a very early stage of query processing. They are very useful. Note
from the Solr - User mailing list archive at Nabble.com.
--
Lance Norskog
goks...@gmail.com
in a couple more we'll have figured out ways to make indexing
engines like solr equally brainless, but not yet -- but it's still pretty
damn easy for what it is, the lucene/Solr folks have done a remarkable job.
--
Regards,
Tharindu
--
Lance Norskog
goks...@gmail.com
There is also a feature called a 'filter'. If you use certain words a
lot, you can make filter queries with just those words. Look for
'filter' and 'fq=' on the wiki.
But really you can have hundreds of words in a query and not have a
performance problem. Solr/Lucene is very fast. In
You may not sort on a tokenized field. You may not sort on a multiValued
field. You can only have one term in a field.
If there are more search terms than documents, A) sorting doesn't mean
anything and B) Lucene will throw an exception.
Erick Erickson wrote:
In general, the behavior when
It requires all of the jars that are packed into solr.war. It is a full
and complete implementation of indexing and searching.
Tharindu Mathew wrote:
Hi everyone,
Do we need all lucene jars in the class path for this? Seems that the
solr-solrj and solr-core jars are not enough
Did you restart all of these slave servers? That would help.
What garbage collection options do you use?
Which release of Solr?
How many Searchers are there in admin/stats.jsp?
Searchers hold open all kinds of memory. They are supposed to cycle out.
These are standard questions, but- what you
These directories are shown at the top of the admin/index.jsp page.
Check out all of the pages off of admin/index.jsp- there is a lot of
information there about what solr is doing.
Israel Ekpo wrote:
The Solr home is the -Dsolr.solr.home Java System property
Also make sure that
Please start new threads for new topics.
Xin Li wrote:
As we know we can use browser to check if Solr is running by going to
http://$hostName:$portNumber/$masterName/admin, say http://localhost:8080/solr1/admin. My questions
is: are there any ways to check it using command line? I used curl
CLOB is probably better for what you want.
Also, make sure the table is declared UTF-8 (or Unicode or whatever
mysql calls it.)
virtas wrote:
As it turns out issue was somewhere in mysql. Not sure exactly where, but
something to do to with BLOB.
Now, I changed text field from BLOB to varchar
The XPathEntityProcessor does not do full XPath. It is a very limited
set intended to be very fast.
You can add code in any scripting language, but that is not really
performant.
Is it possible to use the RegexTransformer to find your records with
regular expressions?
Ken Stanley wrote:
On
Yes, you can declare each field with the Spanish, French, etc. types.
The _t and other types are dynamic and don't have to be declared. This
feature is generally used when you have hundreds or thousands of fields.
It is more clear to declare your fields.
You're right- that error should not be
I'm not quite sure what Tika exceptions mean in this context.
You can give the 'fl=field1,field2' option to only return some fields in
a query.
You can get google-like results using highlighting and 'snippetizing'.
These are documented on the wiki.
satya swaroop wrote:
Hi ,
Can the
--
Lance Norskog
goks...@gmail.com
'arsenic' and
'arsenal' conflating
In my opinion, it would be best to re-index.
--
Lance Norskog
goks...@gmail.com
of highlighting.
After restarting tomcat all went fine again.
I'm trying to understand why I had to restart tomcat and solr and what
should I do to have it working 7/7 24/24.
Xavier
--
Lance Norskog
goks...@gmail.com
--
Lance Norskog
goks...@gmail.com
to create a new mail lists.
--
Lance Norskog
goks...@gmail.com
or
heap size to java or to tomcat where the solr is running
Regards,
satya
--
Lance Norskog
goks...@gmail.com
, QueryElevationComponent takes quite a lot of time. Any
suggestion how to improve this?
--
Chhorn Chamnap
http://chamnapchhorn.blogspot.com/
--
Chhorn Chamnap
http://chamnapchhorn.blogspot.com/
--
Chhorn Chamnap
http://chamnapchhorn.blogspot.com/
--
Lance Norskog
goks...@gmail.com
time which is
unacceptable. The size of elevation file is only 1 Mb. I wonder other
people using this component without problems (related to speed)? Am I
using it the wrong way or there is a limit when using this component?
On 10/29/10, Lance Norskog goks...@gmail.com wrote:
I do not know
when someone does actually want one.
You would be adding some overhead to the system in that Solr now has to
manage these files as stored fields. I guess I would do some benchmarking
to see.
--
Lance Norskog
goks...@gmail.com
when each document was indexed.
--
field name=timestamp type=date indexed=true stored=true
default=NOW multiValued=false/
--
Lance Norskog
goks...@gmail.com
location_details_s
dngythdb25fu (assuming tokenization would split on the
non-alphanumeric chars) in your text fields.
-Yonik
http://www.lucidimagination.com
--
Lance Norskog
goks...@gmail.com
How many items for each query?
On Sat, Oct 30, 2010 at 7:34 PM, Chamnap Chhorn chamnapchh...@gmail.com wrote:
Well, I use Solr 1.4.
There are 30698 lines in my elevation file. I need only 20 results response
back at a time.
On Sun, Oct 31, 2010 at 9:12 AM, Lance Norskog goks...@gmail.com
to replacate 6G index for solr in my env. But my
network can transfer file about 10-20M/s using scp. So solr's http
replcation is too slow, it's normal or I do something wrong?
--
Lance Norskog
goks...@gmail.com
over some other HTTP based
solution e.g. creating sending my own HTTP requests? Do I even
have to use HTTP?
I see the EmbeddedSolrServer exists. Any drawbacks using that?
Any hints are welcome, Thanks!
--
Lance Norskog
goks...@gmail.com
:
Is there an issue running Solr in /home/lib as opposed to running it
somewhere outside of the virtual hosts like /lib?
Eric
--
Lance Norskog
goks...@gmail.com
-multithreading-work-tp1776111p1776111.html
Sent from the Solr - User mailing list archive at Nabble.com.
--
Lance Norskog
goks...@gmail.com
. However, they
are fewer than the number of RUNNABLE threads as the
above.
Why is this happening? Is there anything we can do to
avoid getting in
these situations?
Thanks,
roxana
--
Lance Norskog
goks...@gmail.com
connection-method would suffice.
On 11/01/2010 03:23 AM, Lance Norskog wrote:
2.
The SolrJ library handling of content streams is pull, not push.
That is, you give it a reader and it pulls content when it feels like
it. If your software to feed the connection wants to write the data,
you have
and embeded solr in my app.I'm so confused.What I have
done wrong?
At 2010-11-01 10:12:38,Lance Norskog goks...@gmail.com wrote:
If you are copying from an indexer while you are indexing new content,
this would cause contention for the disk head. Does indexing slow down
during this period?
Lance
failures. Might Zoie help here? Or should we just query
against the Master?
Thanks,
Simon
--
Lance Norskog
goks...@gmail.com
solr again, and it cannot open the index because of the write lock.
Why is there a write lock file when I have not tried to index anything?
--
Lance Norskog
goks...@gmail.com
-3.
Where do I put the jar files?
How do I make Solr use the analyzer?
Thanks
--
Lance Norskog
goks...@gmail.com
rather than manually
apply patches and worry about consistency during upgrades...
Thanks in advance,
will
--
Lance Norskog
goks...@gmail.com
in context:
http://lucene.472066.n3.nabble.com/Disk-usage-per-field-tp934765p1827739.html
Sent from the Solr - User mailing list archive at Nabble.com.
--
Lance Norskog
goks...@gmail.com
,
Jakub.
--
Lance Norskog
goks...@gmail.com
and/or uses a
whole lot of CPU and disk io.
Lance Norskog wrote:
You should query against the indexer. I'm impressed that you got 5s
replication to work reliably.
On Mon, Nov 1, 2010 at 4:27 PM, Simon Wistow si...@thegestalt.org wrote:
We've been trying to get a setup in which a slave replicates
-field-tp1829480p1829480.html
Sent from the Solr - User mailing list archive at Nabble.com.
--
Lance Norskog
goks...@gmail.com
data? Also, are you getting any
errors
in your log?
- Ken
--
Lance Norskog
goks...@gmail.com
Lance found? Is my
understanding of how frequent commits can interact with time it takes to
warm a new collection correct? Appreciate any additional info.
Lance Norskog wrote:
Isn't that what this code does?
onDeckSearchers++;
if (onDeckSearchers 1) {
// should
AM org.apache.solr.common.SolrException log
SEVERE: java.lang.RuntimeException: [solrconfig.xml] requestHandler: missing
mandatory attribute 'class'
Anyone run into issues with Carrot2?
Eric
--
Lance Norskog
goks...@gmail.com
of the default xml's and if I find it
there, I will try and merge them. Does this sound I am on the right path now?
-Original Message-
From: Lance Norskog [mailto:goks...@gmail.com]
Sent: Sunday, November 07, 2010 12:41 PM
To: solr-user@lucene.apache.org
Subject: Re: Adding Carrot2
.
Can you explain my benefits with two solr instances via servlet, maybe more
performance?
Regards,
Nikola
--
Nikola Garafolic
SRCE, Sveucilisni racunski centar
tel: +385 1 6165 804
email: nikola.garafo...@srce.hr
--
Lance Norskog
goks...@gmail.com
McGraw-Hill employee e-mail addresses
without informing the sender or recipient of the message.
--
Lance Norskog
goks...@gmail.com
for
hte dest field.
-Hoss
--
Lance Norskog
goks...@gmail.com
/CoreAdmin
Cheers
--
Lance Norskog
goks...@gmail.com
can post the full
definition if helpful.
Any pointers as to how to debug this would be greatly appreciated!
sasank
--
Lance Norskog
goks...@gmail.com
them yourself.
from 'http://blogs.techrepublic.com.com/security/?p=4501tag=nl.e036'
EARTH has a Right To Life,
otherwise we all die.
- Original Message
From: Lance Norskog goks...@gmail.com
To: solr-user@lucene.apache.org
Sent: Tue, November 9, 2010 7:07:45 PM
Subject: Re: Using
please tell me
what is the preferred approach in such an environment? I am not a programmer
and would appreciate some hand-holding here :o)
Thanks in advance,
Sesh
--
Lance Norskog
goks...@gmail.com
I am close to the end of that subject.
Cheers,
Jakub Godawa.
2010/11/3 Lance Norskog goks...@gmail.com:
Here's the problem: Solr is a little dumb about these Filter classes,
and so you have to make a Factory object for the Stempel Filter.
There are a lot of other FilterFactory classes. You
racunski centar
tel: +385 1 6165 804
email: nikola.garafo...@srce.hr
--
Lance Norskog
goks...@gmail.com
There is a Random field type which returns random numbers. You might
try boosting with that.
Dave Searle wrote:
You could also try splitting the brand name from the product name into a
separate field and then boosting on the product name?
Sent from my iPhone
On 12 Nov 2010, at 20:32, Ahmet
in advance,
Sesh
--
Lance Norskog
goks...@gmail.com
. Is that the right approach to make it work?
Thanks for verbose explanation,
Jakub.
--
Lance Norskog
goks...@gmail.com
--
Lance Norskog
goks...@gmail.com
to index
something.
-Hoss
--
Lance Norskog
goks...@gmail.com
This feature would make the ReplicationHandler more robust in its own
practice of reserving previous commit points, by pushing that code out
into Solr proper.
Jason Rutherglen wrote:
The timed deletion policy is a bit too abstract, as is keeping a
numbered limit of commit points. How would
Here is a separate configuration: use separate Solr instances for
indexing and querying. Both point to the same data directory. A 'commit'
to the query Solr reloads the index. It works in read-only mode- for
production mode, I would make the indexer and queryer in different
permissions so that
in such an environment? I am not a
programmer
and would appreciate some hand-holding here :o)
Thanks in advance,
Sesh
--
Lance Norskog
goks...@gmail.com
It's not that EC2 instances have slow disks, it's that they have no
quota system to guarantee you X amount of throughput. I've benchmarked
1x to 3x on the same instance type at different times. That is, 300%
variation in disk speeds.
Filter queries are only slow once; after that they create
The XPathEntityProcessor has a very limited grammar of path expressions.
It has the ability to use an XSL script, which would then let you do
anything, but I have not used it.
Chantal Ackermann wrote:
Hi Jörg,
you could use the DataImportHandler's XPathEntityProcessor. There you
can specify
This is more complex than you need. The Solr update command can accept
streamed data, with the stream.url and stream.file options. You can just
use solr/update with stream.url=http://your.machine/your.php.script and
it will read as fast as it wants.
There is no parallel indexing support, but
.
So, if the parsing worked, the parsed document overwrote the first
document. If parsing failed, the metadata-only document went in.
Works quite well!
Ken Krugler wrote:
On Nov 14, 2010, at 3:02pm, Lance Norskog wrote:
Yes, the ExtractingRequestHandler uses Tika to parse many file formats
I think the deduplication signature field will work as a multiValued
field. So you can do copyField to it from all of the source fields.
Dan Lynn wrote:
Hi,
I just finished reading on the wiki about deduplication and the
solr.UUIDField type. What I'd like to do is generate an ID for a
Nobody has ever reported seeing a collision 'in the wild' with MD5. It
is broken, but that takes an algorithm.
As to cosmic rays: it's a real problem. A recent Google paper reported
that some ram chips will have 1 bit error per gigabit per century, while
others have that much per hour. I've
This very definitely should not happen.
Can you remove the index and reindex everything? And then do these
iterative tests?
Which version of Solr are you running?
If these changes do not cure the problem, can you post your schema.xml
and solrconfig.xml?
Pramod Goyal wrote:
Hi,
I am
Upload the files independently of Solr. Solr is not a content management
system.
One problem is getting the links put together so that the link that
comes out with the document can be turned into a link the user can open.
Chad Salamon wrote:
I would like to save files sent to the
(multiply the probability curves) and you will get
really variable response times. I don't know how to guide you.
lu.rongbin wrote:
thanks,Lance Norskog-2. I've tested the EBS, but it's not better. so ,maybe I
have to optimize my solr config for ec2 m2.4xlarge.this kind computer config
is :
cpu units
slower
- I'd check with 4.0.x if someone can point me at a tool that can
migrate indexes. I seem to be unable to find one and Lucene 3.0 informs
me that it's incompatible with 2.9.x
--
Lance Norskog
goks...@gmail.com
TextBlock ID=P9_TB1 HPOS=1045 VPOS=196 WIDTH=173 HEIGHT=28
STYLEREFS=TXT_0 PAR_CENTER
TextLine ID=P9_TL1 HPOS=1045 VPOS=197 WIDTH=173 HEIGHT=27
String ID=P9_ST1 HPOS=1045 VPOS=197 WIDTH=173 HEIGHT=27
CONTENT=Preface WC=0.98 CC=000/
/TextLine
--
Lance Norskog
goks...@gmail.com
the document. I tried
setting
null values for the copy field when i am updating the document but it
didnt
solve the problem.
--
Lance Norskog
goks...@gmail.com
is the CPU of the master so high?
--
Lance Norskog
goks...@gmail.com
. will it consider
the word test also ?
Please help
--
Thanks,
Pawan Darira
--
Lance Norskog
goks...@gmail.com
is that it failed to replicate due to the OOM and since then
tries
to do a full replication again and again?
but why is the CPU of the master so high?
--
Lance Norskog
goks...@gmail.com
--
Lance Norskog
goks...@gmail.com
was everything.Any help would be appreciated.-Viswa
--
Lance Norskog
goks...@gmail.com
the slave doesn't do any
indexing, i will call commit and it will do nothing to the index itself, but
will reload it?
thanks
On Sun, Nov 21, 2010 at 8:26 AM, Lance Norskog goks...@gmail.com wrote:
Ah! If the program doing the indexing has manual commits, the program
could send a commit
really need a commit? or can i use the
*readercycle*http://wiki.apache.org/solr/SolrOperationsToolsscript? since
i don't need to comit anything, just reopen the reader.
thanks
On Sun, Nov 21, 2010 at 12:17 PM, Lance Norskog goks...@gmail.com wrote:
Yes, the Solr commit operations always
/advice.
Jason.
If you wish to view the St. James's Place email disclaimer, please use the
link below
http://www.sjp.co.uk/portal/internet/SJPemaildisclaimer
--
Lance Norskog
goks...@gmail.com
- Should I move the per-user permissions into a separate index?
What
technique to implement?
--
Lance Norskog
goks...@gmail.com
? That shouldn't be possible (if you mean the index is truly
corrupt - i.e. you can't open it).
-Yonik
http://www.lucidimagination.com
--
Lance Norskog
goks...@gmail.com
only'?
--
Lance Norskog
goks...@gmail.com
Cariens richcari...@gmail.com wrote:
Anyone ever use Solr to present a view of a document with hit-terms
highlighted within? Kind of like Google's cached
http://bit.ly/hgudWqcopies?
--
Lance Norskog
goks...@gmail.com
901 - 1000 of 1360 matches
Mail list logo