.**
tgels.org http://rankingalgorithm.tgels.org
On 8/14/2011 10:37 AM, Naveen Gupta wrote:
Hi Mark/Erick/Nagendra,
I was not very confident about NRT at that point of time, when we
started
project almost 1 year ago, definitely i would try NRT and see the
performance.
The current requirement
or the old searchers closed or new searchers opened, and
warmed (error that you are facing).
Regards
- Nagendra Nagarajayya
http://solr-ra.tgels.org
http://rankingalgorithm.tgels.**orghttp://rankingalgorithm.tgels.org
On 8/14/2011 10:37 AM, Naveen Gupta wrote:
Hi Mark/Erick/Nagendra
Hi Mark/Erick/Nagendra,
I was not very confident about NRT at that point of time, when we started
project almost 1 year ago, definitely i would try NRT and see the
performance.
The current requirement was working fine till we were using commitWithin 10
millisecs in the XMLDocument which we were
Hi,
Most of the settings are default.
We have single node (Memory 1 GB, Index Size 4GB)
We have a requirement where we are doing very fast commit. This is kind of
real time requirement where we are polling many threads from third party and
indexes into our system.
We want these results to be
the indexing operations.
On Thu, Aug 11, 2011 at 1:02 PM, Naveen Gupta nkgiit...@gmail.com wrote:
Yes this was happening because of JVM heap size
But the real issue is that if our index size is growing (very high)
then indexing time is taking very long (using streaming)
earlier for indexing
this exception with no other error or explananation in
the logs, this is almost always because the JVM has run out of memory.
Have you checked/profiled your mem usage/GC during the stream operation?
On Thu, Aug 11, 2011 at 3:18 AM, Naveen Gupta nkgiit...@gmail.com wrote:
Hi,
We are doing
Hi,
We are doing streaming update to solr for multiple user,
We are getting
Aug 10, 2011 11:56:55 AM org.apache.solr.common.SolrException log
SEVERE: org.apache.lucene.store.LockObtainFailedException: Lock obtain timed
out: NativeFSLock@/var/lib/solr/data/index/write.lock
at
is to simply have multiple clients
indexing simultaneously. But you haven't indicated how the indexing is
being
done. Are you using DIH? SolrJ? Streaming documents to Solr? You have to
provide those kinds of details to get meaningful help.
Best
Erick
On Aug 2, 2011 8:06 AM, Naveen Gupta nkgiit
to get meaningful help.
Best
Erick
On Aug 2, 2011 8:06 AM, Naveen Gupta nkgiit...@gmail.com wrote:
Hi
We have a requirement where we are indexing all the messages of a a
thread,
a thread may have attachment too . We are adding to the solr for indexing
and searching for applying few
Hi,
We are having a requirement where we are having almost 100,000 documents to
be indexed (atleast 20 fields). These fields are not having length greater
than 10 KB.
Also we are running parallel search for the same index.
We found that it is taking almost 3 min to index the entire documents.
Sorry for 15k Docs, it is taking 3 mins.
On Thu, Aug 4, 2011 at 10:07 PM, Naveen Gupta nkgiit...@gmail.com wrote:
Hi,
We are having a requirement where we are having almost 100,000 documents to
be indexed (atleast 20 fields). These fields are not having length greater
than 10 KB.
Also we
Hi
We have a requirement where we are indexing all the messages of a a thread,
a thread may have attachment too . We are adding to the solr for indexing
and searching for applying few business rule.
For a user, we have almost many threads (100k) in number and each thread may
be having 10-20
Can somebody answer this?
What should be the best strategy for optimize (when million of messages we
are indexing for a new registered user)
Thanks
Naveen
On Tue, Aug 2, 2011 at 5:36 PM, Naveen Gupta nkgiit...@gmail.com wrote:
Hi
We have a requirement where we are indexing all the messages
Hi,
I am trying to achieve this use case with following expectation
three fields
1. field1
2. field2
3. field3
field1 should have the max relevance
field2 should have the next
field3 is the last
the term will be entered by end user (say* rock roll*)
i want to show the results which will
.
On 09/06/2011 03:00, Naveen Gupta wrote:
Hi Gary
It started working .. though i did not test for Zip files, but for rar
files, it is working fine ..
only thing what i wanted to do is to index the metadata (text mapped to
content) not store the data Also in search result, i want to filter
Hi
This is my document
in php
$xmldoc = 'adddocfield name=idF_146/fieldfield
name=userid74/fieldfield name=groupuseidgmail.com/fieldfield
name=attachment_size121/fieldfield
name=attachment_namesample.pptx/field/doc/add';
$ch = curl_init(http://localhost:8080/solr/update;);
Hi,
curl http://localhost:8983/solr/update?commit=true -H Content-Type:
text/xml --data-binary 'adddocfield
name=idtestdoc/field/doc/add'
Regards
Naveen
On Fri, Jun 10, 2011 at 10:18 AM, Naveen Gupta nkgiit...@gmail.com wrote:
Hi
This is my document
in php
$xmldoc = 'adddocfield name
-and-Solr-1-4-1-td2327933.html
I got it to work with the 3.1 source and a couple of patches.
Hope this helps.
Regards,
Gary.
On 08/06/2011 04:12, Naveen Gupta wrote:
Hi Can somebody answer this ...
3. can somebody tell me an idea how to do indexing for a zip file ?
1. while sending docx
Hi
We are using requestextractinghandler and we are getting following error. we
are giving microsoft docx file for indexing.
I think that this is something to do with field date definition .. but now
very sure ...what field type should we use?
2. we are trying to index jpg (when we search over
)
at
org.apache.coyote.http11.Http11Protocol$Http11ConnectionHandler.process(Http11Protocol.java:583)
at
org.apache.tomcat.util.net.JIoEndpoint$Worker.run(JIoEndpoint.java:447)
at java.lang.Thread.run(Thread.java:619)
Thanks
Naveen
On Tue, Jun 7, 2011 at 3:33 PM, Naveen Gupta nkgiit
?
Tomás
On Mon, Jun 6, 2011 at 2:54 AM, Naveen Gupta nkgiit...@gmail.com wrote:
Hi
Since it is php, we are using solphp for calling curl based call,
what my concern here is that for each user, we might be having 20-40
attachments needed to be indexed each day, and there are various users
Hi
Since it is php, we are using solphp for calling curl based call,
what my concern here is that for each user, we might be having 20-40
attachments needed to be indexed each day, and there are various users
..daily we are targeting around 500-1000 users ..
right now if you see, we
?php
$ch =
Hi
I want to implement different index strategy where we want to keep indexes
with respect to each tennant and we want to maintain indexes separately ...
first level of category -- company name
second level of category - company name + fields to be indexed
then further categories - group of
Hi Romi
As per me, you need to understand how ajax with jquery works .. then go for
json and then jsonp (if you are fetching from different)
query here is dynamic query which you will be trying to hit solr .. (it
could be simple text, or more advanced query string)
Hi
We want to post to solr server with some of the files (rtf,doc,etc) using
php .. one way is to post using curl
is there any client like java client (solrcell)
urls will also help
Thanks
Naveen
Hi Pravesh
We don't have that setup right now .. we are thinking of doing that
for writes we are going to have one instance and for read, we are going to
have another...
do you have other design in mind .. kindly share
Thanks
Naveen
On Fri, Jun 3, 2011 at 2:50 PM, pravesh
Yes,
that one i used and it is working fine .thanks to nabble ..
Thanks
Naveen
On Fri, Jun 3, 2011 at 4:02 PM, Gora Mohanty g...@mimirtech.com wrote:
On Fri, Jun 3, 2011 at 3:55 PM, Naveen Gupta nkgiit...@gmail.com wrote:
Hi
We want to post to solr server with some of the files (rtf,doc
Hi
I am trying to integrate solr 3.1 and tika (which comes default with the
version)
and using curl command trying to index few of the documents, i am getting
this error. the error is attr_meta field is unknown. i checked the
solrconfig, it looks perfect to me.
can you please tell me what i am
Hi
I am trying to integrate solr 3.1 and tika (which comes default with the
version)
and using curl command trying to index few of the documents, i am getting
this error. the error is attr_meta field is unknown. i checked the
solrconfig, it looks perfect to me.
can you please tell me what i am
if there is a dynamic field named attr_* in the schema. The
uprefix=attr_ parameter means that if Solr can't find an extracted field
in the schema, it'll add the prefix attr_ and try again.
*Juan*
On Thu, Jun 2, 2011 at 4:21 AM, Naveen Gupta nkgiit...@gmail.com wrote:
Hi
I am trying
Hi
We are having an application where every 10 mins, we are doing indexing of
users docs repository, and eventually, if some thread is being added in that
particular discussion, we need to index the thread again (please note we are
not doing blind indexing each time, we have various rules to
31 matches
Mail list logo