Check out https://issues.apache.org/jira/browse/SOLR-4722, which will
return matching terms (and their offsets). Patch can be applied cleanly to
Solr 4; doesn't appear to have been tried with Solr 5
-Simon
On Tue, Aug 18, 2015 at 11:30 AM, Jack Krupansky
wrote:
> Maybe a spe
we've been using with
some success for this task.
best
-Simon
On Mon, Aug 24, 2015 at 2:13 PM, afrooz wrote:
> Thanks Erick,
> I will explain the detail scenario so you might give me a solution:
> I want to annotate a medical document base on only medical dictionary. I
> don&
it is ingested into our main
Solr collection.
How many documents/product leaflets do you have ? The tagger is very fast
at the Solr level but I'm seeing quite a bit of HTTP overhead.
best
-Simon
On Fri, Sep 11, 2015 at 1:39 PM, Sujit Pal wrote:
> Hi Francisco,
>
> >>
https://github.com/OpenSextant/SolrTextTagger/
We're using it for country tagging successfully.
On Wed, Nov 4, 2015 at 3:10 PM, Doug Turnbull <
dturnb...@opensourceconnections.com> wrote:
> David Smiley had a place name and general tagging engine that for the life
> of me I can't find.
>
> It di
Please do push your script to github - I (re)-compile custom code
infrequently and never remember how to setup the environment.
On Thu, Nov 12, 2015 at 5:14 AM, Upayavira wrote:
> Okay, makes sense. As to your question - making a new ValueSourceParser
> that handles 'equals' sounds pretty straig
There's a JIRA ( https://issues.apache.org/jira/browse/SOLR-4722 )
describing a highlighter which returns term positions rather than
snippets, which could then be mapped to the matching words in the indexed
document (assuming that it's stored or that you have a copy elsewhere).
-Sim
every
place where a date format conversion is needed is proving painful indeed ;=(
My thought is to write a custom function of the form
datereformatter(, ) but I thought I'd
check if it's already been done or if someone can suggest a better approach.
regards
-Simon
thal/defsolr/server/logs --module=http
solrconfig.xml: basically the default with some minor tweaks in the
indexConfig section
5.0
200
1
20
60
20
... everything else is default
Insights as to why this is happening would be welcome.
-Simon
Do you already have a set of terms for which you would want to find out
their co-occurence, or are you trying to do data mining, looking in a
collection for terms which occur together more often than by chance ?
On Sun, Oct 16, 2016 at 3:45 AM, Yangrui Guo wrote:
> Hello
>
> I'm curious to know
o it
looks like a bug.
-Simon
On Fri, Jun 9, 2017 at 5:14 AM, Andreas Hubold wrote:
> Hi,
>
> I just tried to update from Solr 6.5.1 to Solr 6.6.0 and observed a
> changed behaviour with regard to unloading cores in Solr standalone mode.
>
> After unloading a core using the Core
with multiple
tokens. Then construct a query which searches both field1 for an exact
match, and field2 using ComplexQueryParser (use the localparams syntax) to
combine them. Boost the field1 (exact match).
HTH
-Simon
On Thu, Jun 15, 2017 at 1:20 PM, Max Bridgewater
wrote:
> Thanks Susheel. The c
deleted in current versions of Solr - so you'll
have to find a way (outside Solr) to copy it or re-create it.
What is the use case here ?
best
-Simon
On Tue, Sep 12, 2017 at 1:27 PM, Shashank Pedamallu
wrote:
> Hi,
>
> I wanted to know how does Solr pick up cores on startup. Bas
ctory might work for this.
best
-Simon
On Thu, Sep 14, 2017 at 1:46 PM, Arnold Bronley
wrote:
> I know I can apply PatternReplaceFilterFactory to remove control characters
> from indexed value. However, is it possible to do similar thing for stored
> value? Because of some cont
@Arnold: are these non UTF-8 control characters (which is what the Nutch
issue was about) or otherwise legal UTF-8 characters which Solr for some
reason is choking on ?
If you could provide a full stack trace it would be really helpful.
On Thu, Sep 14, 2017 at 2:55 PM, Markus Jelsma
wrote:
>
, and you
could live with dropping the offending document(s) then you might want to
investigate the TolerantUpdateProcessorFactory Solr 6.1 or later)
-Simon
On Thu, Sep 14, 2017 at 3:56 PM, arnoldbronley
wrote:
> Thanks for information. Here is the full stack trace. I thought to handle
>
though see SOLR-11078 , which is reporting significant query slowdowns
after converting *Trie to *Point fields in 7.1, compared with 6.4.2
On Wed, Nov 1, 2017 at 9:06 PM, Yonik Seeley wrote:
> On Wed, Nov 1, 2017 at 2:36 PM, Erick Erickson
> wrote:
> > I _always_ prefer to reindex if possible.
y - I
can't say.
best
-Simon
On Tue, Nov 7, 2017 at 1:44 AM, Amin Raeiszadeh
wrote:
> Hi
> i want to use more than one ssd in each server of solr cluster but i don't
> know how to set multiple hdd in solr.xml configurations.
> i set on hdd path in solr.xml by:
> /media
You might want to take a look at
https://issues.apache.org/jira/browse/SOLR-4722
( 'highlighter which generates a list of query term positions'). We used it
a while back and doesn't appear to have been used in any Solr > 4.10)
-Simon
On Tue, Nov 29, 2016 at 11:43 AM, John
ystem
logs and didn't see any evidence of hardware errors
I'm puzzled as to why this would start happening out of the blue and I
can't find any partiuclarly relevant posts to this forum or Stackexchange.
Anyone have an idea what's going on ?
-Simon
sitions with no need for actual highlighting. The
patch is pretty old - I applied it to Solr 4.10 I think, so will probably
need some work for later releases.
HTH
-Simon
On Tue, Mar 28, 2017 at 4:59 AM, forest_soup wrote:
> Thanks Eric.
>
> Actually solr highlighting function does not
will return a boolean if the term is in a specific field.
I've used this for simple cases where it worked well, though I wouldn't
like to speculate on how well this scales if you have an edismax query
where you might need to generate multiple term/field combinations.
HTH
-Simon
On Thu, Ap
reposting this as the problem described is happening again and there were
no responses to the original email. Anyone ?
I'm seeing an odd error during indexing for which I can't find any reason.
The relevant solr log entry:
2017-03-24 19:09:35.363 ERROR (commitSchedule
W
> if you look now and have free space it still may have been all used up
> but had some space reclaimed.
>
> Best,
> Erick
>
> On Wed, Apr 26, 2017 at 12:02 PM, simon wrote:
> > reposting this as the problem described is happening again and there were
> > no
I ran into the exact same situation recently. I unloaded from the browser
GUI which does not delete the data or instance dirs, but does delete
core.properties. I couldn't find any API either so I eventually manually
recreated core.properties and restarted Solr.
Would be nice if the core.propert
-Simon
On Tue, May 2, 2017 at 4:04 PM, Erick Erickson
wrote:
> IIRC, the core.properties file _is_ renamed to
> core.properties.unloaded or something like that.
>
> Yeah, this is something of a pain. The inverse of "unload" is "create"
> but you have to know e
dexer scripts running concurrently, but the duration goes up
proportionately.
-Simon
On Thu, Apr 27, 2017 at 9:26 AM, simon wrote:
> Nope ... huge file system (600gb) only 50% full, and a complete index
> would be 80gb max.
>
> On Wed, Apr 26, 2017 at 4:04 PM, Erick Erickson
> wr
Your updateRequestProcessorChain config snippet specifies the "id" field
to generate a signature, but the sample data doesn't contain an "id" field
... check that out first.
-Simon
On Wed, May 31, 2017 at 12:06 PM, Lebin Sebastian
wrote:
> Hello,
>
> I am
Also seeing this on Mac OS X.
java version = Java(TM) SE Runtime Environment (build 1.7.0_51-b13)
On Wed, Apr 2, 2014 at 11:01 AM, Joshua P wrote:
> Hi there!
>
> I'm recieving the following errors when trying to run luke-with-deps.jar
>
> SLF4J: Failed to load class "org.slf4j.impl.StaticLogg
adding that worked - thanks.
On Thu, Apr 3, 2014 at 4:18 AM, Dmitry Kan wrote:
> Hi Joshua, Simon,
>
> do you pass the -XX:MaxPermSize=512m to your jvm?
>
> java -XX:MaxPermSize=512m -jar luke-with-deps.jar
>
> My java runtime environment is of the same version as Simon
derstanding solr uniqueKey is like a
database primary key. I am wondering how could I end up with two documents
with same uniqueKey in the index.
Thanks,
Simon
--
View this message in context:
http://lucene.472066.n3.nabble.com/Duplicate-Unique-Key-tp4129651.html
Sent from the Solr - User mailing list archive at Nabble.com.
Erick,
It's indeed quite odd. And after I trigger re-indexing all documents (via
the normal process of existing program). The duplication is gone. It can
not be reproduced easily. But it did occur occasionally and that makes it a
frustrating task to troubleshoot.
Thanks,
Simon
--
MergingIndex is not the case here as I am not doing that. Even the issue is
gone for now, it is not a relief for me as I am not sure how to explain this
to others (peer, boss and user). I am thinking of implement a watch dog to
check whenever the total Solr documents exceeds the number of items i
problems (and DBI takes care of writing to a database).
I'm probably going to rewrite in Python since the final destination of many
of our extracts is Tableau, which has a Python API for creating TDEs
(Tableau data extracts)
regards
-Simon
On Fri, May 2, 2014 at 7:43 AM, Siegfried Goeschl
e to share your solutions?
Thanks,
Simon
--
View this message in context:
http://lucene.472066.n3.nabble.com/Solr-block-join-tp4098128.html
Sent from the Solr - User mailing list archive at Nabble.com.
Tika code as I am not using
it).
You should replace StreamingUpdateSolrServer by ConcurrentUpdateSolrServer
and experiment to find the optimal number of threads to configure.
-Simon
On Sun, Jan 26, 2014 at 11:28 AM, Erick Erickson wrote:
> 1> That's what I'd do. For incremen
n user configure a field to be auto
completion.
Thanks,
Simon
--
View this message in context:
http://lucene.472066.n3.nabble.com/Suggester-on-Dynamic-fields-tp4165270p4165329.html
Sent from the Solr - User mailing list archive at Nabble.com.
That's basically it.
remove all /update URLs from the slave config
On Tue, Aug 30, 2011 at 8:34 AM, Miguel Valencia <
miguel.valen...@juntadeandalucia.es> wrote:
> Hi
>
>I've read that it's possible add documents to slave machine:
>
> http://wiki.apache.org/solr/**SolrReplication#What_if_I_a
what issues exactly ?
are you using 32 bit Java ? That will restrict the JVM heap size to 2GB max.
-Simon
On Tue, Aug 30, 2011 at 11:26 AM, Tirthankar Chatterjee <
tchatter...@commvault.com> wrote:
> Hi,
>
> I have a machine (win 2008R2) with 16GB RAM, I am having issue indexin
So if I understand you, you are using Tika /SolrJ together in a Solr client
process which talks to your Solr server ? What is the heap size ? Can you
give us a stack trace from the OOM exception ?
-Simon
On Wed, Aug 31, 2011 at 10:58 AM, Tirthankar Chatterjee <
tchatter...@commvault.com>
That error has nothing to do with Solr - it looks as though you are trying
to start the JVM with a heap size that is too big for the available physical
memory.
-Simon
On Fri, Sep 2, 2011 at 2:15 AM, Rohit wrote:
> Hi All,
>
>
>
> I am using Solr 3.0 and have 4 cores build i
Not sure about the exact reason for the error. However, there's a related
email thread today with a code fragment that you might find useful -- see
http://www.lucidimagination.com/search/document/a553f89beb41e39a/how_to_use_solrj_self_signed_cert_ssl_basic_auth#a553f89beb41e39a
-Simon
O
You need to give us more information. The code which throws this exception
will be most helpful.
-Simon
On Fri, Sep 2, 2011 at 5:43 AM, Kissue Kissue wrote:
> Hi,
>
> I am using apache solr 3.3.0 with SolrJ on a linux box.
>
> I am getting the error below when indexing kicks in:
27;stale file handle'
nastiness).
That said, an API which allows you to open an IndexSearcher to a previous
commit, or indeed to a snapshot saved in another subdirectory would be a
really useful improvement. Maybe an extension to the CoreAdmin API ?
-Simon
On Tue, Sep 6, 2011 at 5:16 PM, Emm
ambiguity, but that would be very slow indeed
-Simon
On Tue, Sep 6, 2011 at 12:58 PM, Leonardo Souza wrote:
> Hi Mark,
>
> The implementation is logging anyway, we have subclassed
> StreamingUpdateSolrServer and used handleError to log, but inspecting the
> stack trace in in
XML schema to
what Solr expects.
SolrJ would normally be used to construct Solr Documents from the parsed XML
input, whatever its format, and use the SolrJ API to ingest these documents
for indexing.
-Simon
On Fri, Sep 30, 2011 at 9:03 AM, kiran.bodigam wrote:
> We can post the documents from
ance and memory use.
-Simon
On Fri, Sep 30, 2011 at 8:27 AM, Pranav Prakash wrote:
> How will the number of fields increase the amount of RAM usage in Solr 3.4?
> I have about 37 different fields because i've made every field for every
> language. I shall be doing dismax search a
There's an open issue -
https://issues.apache.org/jira/browse/SOLR-2731which addresses adding
this kind of metadata to csv output. There's a patch
there which may be useful, and could probably be adapted if needed
-Simon
On Fri, Oct 14, 2011 at 4:37 PM, Fred Zimmerman wrote:
> Hi,
ersions of Lucene
I'm not sure how easily this would all would backport to Solr 3.1, but you
could try....
best
-Simon
On Tue, Nov 22, 2011 at 1:05 AM, Rahul Mehta wrote:
> Hello,
>
> I want to Run surround query .
>
>
> 1. Downloading from
>
oops, didn't see all of the thread before I hit send. Good work, Erik
On Fri, Dec 2, 2011 at 5:21 PM, simon wrote:
> Take a look at https://issues.apache.org/jira/browse/SOLR-2703, which
> integrates the surround parser into Solr trunk. There's a dependency on a
> Lucene pat
wo cores over as soon as the indexing
on "standby" is complete.
It seems to work well on my development box, but I have not seen this
approach discussed elsewhere so I was wondering if I was missing
something here.
Feedback gratefully received!
Simon
The extended dismax parser (see SOLR-1553) may do what you are looking for
From its feature list..
'Supports the "boost" parameter.. like the dismax bf param, but multiplies
the function query instead of adding it in'
On Sun, Aug 29, 2010 at 12:27 AM, Andy wrote:
> But isn't it the case that
take a look at SOLR-2703, which was committed for 4.0. It provides a Solr
wrapper for the surround query parser, which supports span queries.
On Fri, Nov 23, 2012 at 3:38 PM, Anirudha Jadhav wrote:
> What is the best way to use span queries in solr ?
>
> I see https://issues.apache.org/jira/brow
That won't work. External file fields are currently only usable within
function queries, according to the Javadocs
On Wed, May 4, 2011 at 12:16 PM, Rih wrote:
> This could work. Are there search/index performance drawbacks when using
> it?
>
>
> On Mon, May 2, 2011 at 6:22 PM, Ahmet Arslan wrot
this
time. If your application is written in java, take a look at the
Quartz scheduler - http://www.quartz-scheduler.org/
-Simon
How long are the documents ? indexing a large document can be slow
(although 2 seconds is very slow indeed).
2011/6/22 Rode González (libnova) :
> Hi !
>
>
>
> We are using Zend Search based on Lucene. Our indexing pdf consultations
> take longer than 2 seconds.
>
> We want to change to solr to tr
have you checked out the deduplication process that's available at
indexing time ? This includes a fuzzy hash algorithm .
http://wiki.apache.org/solr/Deduplication
-Simon
On Thu, Jun 23, 2011 at 5:55 AM, Pranav Prakash wrote:
> This approach would definitely work is the two docum
You could configure a PostCommit event listener on the master which
would send a HTTP fetchindex request to the slave you want to carry
out replication - see
http://wiki.apache.org/solr/SolrReplication#HTTP_API
But why do you want the master to push to the slave ?
-Simon
On Mon, Aug 8, 2011 at
Only one should be returned, but it's non-deterministic. See
http://wiki.apache.org/solr/DistributedSearch#Distributed_Searching_Limitations
-Simon
On Sat, Aug 6, 2011 at 6:27 AM, Pooja Verlani wrote:
> Hi,
>
> We have a multicore solr with 6 cores. We merge the results using shar
eisey wrote:
> On 8/8/2011 4:07 PM, simon wrote:
>>
>> Only one should be returned, but it's non-deterministic. See
>>
>> http://wiki.apache.org/solr/DistributedSearch#Distributed_Searching_Limitations
>
> I had heard it was based on which one responded fi
Could you tell us what you're trying to achieve with the range query ?
It's not clear.
-Simon
On Wed, Aug 10, 2011 at 5:57 AM, Amit Sawhney wrote:
> Hi All,
>
> I am trying to sort the results on a unix timestamp using this query.
>
> http://url.com:8983/solr/db/select
I meant the frange query, of course
On Wed, Aug 10, 2011 at 10:21 AM, simon wrote:
> Could you tell us what you're trying to achieve with the range query ?
> It's not clear.
>
> -Simon
>
> On Wed, Aug 10, 2011 at 5:57 AM, Amit Sawhney wrote:
>> Hi All,
>>
Worth remembering there are some performance penalties with deep
paging, if you use the page-by-page approach. may not be too much of a
problem if you really are only looking to retrieve 10K docs.
-Simon
On Wed, Aug 10, 2011 at 10:32 AM, Erick Erickson
wrote:
> Well, if you really want to
Th attachment isn't showing up (in gmail, at least). Can you inline
the relevant bits of code ?
On Wed, Aug 10, 2011 at 11:05 AM, Tom Mortimer wrote:
> Hi,
> Apologies if this is really basic. I'm trying to learn how to create a
> custom request handler, so I wrote the minimal class (attached), c
It's working for me. Compiled, inserted in solr/lib, added the config
line to solrconfig.
when I send a /flaxtest request i get
0
16
Hello!
I was doing this within a core defined in solr.xml
-Simon
On Wed, Aug 10, 2011 at 11:46 AM, Tom Mortimer wrote:
> Sure -
&g
This is in trunk (up to date). Compiler is 1.6.0_26
classpath was
dist/apache-solr-solrj-4.0-SNAPSHOT.jar:dist/apache-solr-core-4.0-SNAPSHOT.jar
built from trunk just prior by 'ant dist'
I'd try again with a clean trunk .
-Simon
On Wed, Aug 10, 2011 at 1:20 PM, Tom
me at
the browser ?
You can monitor cache statistics from the admin console 'statistics' page
Are you seeing anything untoward in the solr logs ?
-Simon
On Wed, Aug 10, 2011 at 1:11 PM, Charles-Andre Martin
wrote:
> Hi,
>
>
>
> I've noticed poor performance for m
an hl.fragsize of 1000 is problematical, as Solr parses that
parameter as a 32 bit int... that's several bits more.
-Simon
On Wed, Aug 10, 2011 at 4:59 PM, Sang Yum wrote:
> Hi,
>
> I have been trying to increase the size of the highlight snippets using
> "
alues.
It would help if you described exactly what problem you're trying to solve.
-Simon
On Mon, Aug 15, 2011 at 1:02 PM, Donald J. Organ IV
wrote:
> Is there a way to set a minimum score requirement so that matches below a
> given score are not return/included in facet counts.
An UpdateRequestProcessor would do the trick. Look at the (rather minimal)
documentation and code example in
http://wiki.apache.org/solr/UpdateRequestProcessor
-Simon
On Thu, Aug 18, 2011 at 4:15 PM, bhawna singh wrote:
> Hi All,
> I have a requirement to update a certain field
your code ?
and - what problem are you trying to solve with the
CachingKeyworkMarkerFilter ?
FWIW, I've been looking at a more generalized way of tracking changes in
protwords/stopwords/ etc and it's turning out to be quite complex.
-Simon
>
> cheers
> Charlie
>
> --
> Vie
http://wiki.apache.org/solr/SolrQuerySyntax has answers for you.
-Simon
On Thu, Aug 25, 2011 at 1:04 AM, Ranveer Kumar wrote:
> any help...
>
> On Wed, Aug 24, 2011 at 12:58 PM, Ranveer Kumar >wrote:
>
> > Hi,
> >
> > is it right way to do :
>
That's not a question we can answer in this group - you need to take it up
with your hosting provider - they may already have it available.
On Thu, Aug 25, 2011 at 2:59 PM, Devora wrote:
> Thank you!
>
> Since it's shared hosting, how do I install java?
>
> -Original Message-
> From: Jae
It sounds as though you are optimizing the index after the delta import. If
you don't do that, then only new segments will be replicated and syncing
will be much faster.
On Fri, Aug 26, 2011 at 12:08 PM, Mark wrote:
> We are currently delta-importing using DIH after which all of our servers
> h
The multicore API (see http://wiki.apache.org/solr/CoreAdmin ) allows you to
swap, unload, reload cores. That should allow you to do what you want,
-Simon
On Fri, Aug 26, 2011 at 11:13 AM, Mike Austin wrote:
> I would like to have the ability to keep requests from being slowed from
&g
The built-in DIH scheduling was never implemented as far as I know - the
Wiki section is just a design proposal and explicitly says "Hasn't been
committed to SVN (published only here) "
On Windows, you can use the Task Scheduler to do the kinds of things that
cron does on Unix/Lin
some time back I used dreamhost for a Solr based project. Looks as though
all their offerings, including shared hosting have Java support - see
http://wiki.dreamhost.com/What_We_Support. I was very happy with their
service and support.
-Simon
On Tue, Oct 9, 2012 at 10:44 AM, Michael Della Bitta
to it...
-Simon
On Fri, Oct 12, 2012 at 12:27 PM, Phil Hoy wrote:
> Hi,
>
> We have a multi-core set up with a fairly large synonym file, all cores
> share the same schema.xml and synonym file but when solr loads the cores,
> it loads multiple instances of the synonym map, this is a
first, check what port 8983 is bound to - should be listening on all
interfaces
netstat -an |grep 8983
You should see
tcp0 0 0.0.0.0:8983 0.0.0.0:* LISTEN
-Simon
On Wed, Nov 25, 2009 at 3:55 PM, Joel Nylund wrote:
> Hi, if I try to use any other hostn
uess that the
second line is the key to the problem -looks as though TCP over IPv4 is onl
y listening on the localhost interface, which is a network configuration
issue.
what does the Solr log say after it's started - should be a line
INFO: Started SelectChannelConnector @ 0.0.0.0:8983
unlikely mixes of alpha/numeric/punctuation, and also
eliminated longer words which consisted of runs of not-ocurring-in-English
bigrams.
Hope this helps
-Simon
>
> --
>
problem Eventually I cloned our
environment to a new AWS instance, which proved to be the solution. Why, I
have no idea...
-Simon
On Mon, Sep 24, 2018 at 1:13 PM, Susheel Kumar
wrote:
> Got it. I'll have first hardware folks check and if they don't see/find
> anything suspicious then
Tim:
How up to date is the Solr-5410 patch/zip in JIRA ?. Looking to use the
Span Query parser in 6.5.1, migrating to 7.x sometime soon.
Would love to see these committed !
-Simon
On Mon, Feb 12, 2018 at 10:41 AM, Allison, Timothy B.
wrote:
> That requires a SpanNotQuery. AFAIK, there
r configuration so that it's not tied to one particular
external API - defining a macro, if you will, so that you could supply
'fl='a,b,c,%numcites%,...' in the request and have Solr do the expansion.
Is there some way to do this that I've overlooked ? if not, I think it
would be a useful new feature.
-Simon
hat it's not tied to one particular
> external API - defining a macro, if you will, so that you could supply
> 'fl='a,b,c,%numcites%,...' in the request and have Solr do the expansion.
>
> Is there some way to do this that I've overlooked ? if not, I think it
> would be a useful new feature.
>
>
> -Simon
>
>
>
Thanks Mikhail:
I considered that, but not all queries would request that field, and there
are in fact a couple more similar DocTransformer-generated aliased fields
which we can optionally request, so it's not a general enough solution.
-Simon
On Wed, Feb 28, 2018 at 1:18 AM, Mikhail Khl
Could it be that the header should be 'Content-Type' (which is what I see
in the relevant RFC) rather than 'Content-type' as shown in your email ? I
don't know if headers are case-sensitive, but it's worth checking.
-Simon
On Tue, May 29, 2018 at 11:02 AM, Roee
Looking carefully at the documentation for JSON facets, it looks as though
the offset parameter is not supported for range facets, only for term
facets. You'd have to do pagination in your application.
-Simon
On Tue, Jul 10, 2018 at 11:45 AM, Anil wrote:
> HI Eric,
>
> i mean
ard Club of New York City
Nationalist dictatorships versus open society / by George Soros
15891
Soros, George
The new paradigm for financial markets : the credit crisis of 2008 and what
it means / George Soros
Thank you for the help in advance,
Simon.
Hi Alex,
It's simply defined like this in the schema.xml :
and it is cloned to the other multi-valued field o_title :
Should I simply change the type to be "string" instead?
Thanks again,
Simon.
On Wed, Feb 18, 2015 at 12:00 PM, Alexandre Rafalovitch
wrote:
>
ess releases and articles on policy changes affecting the Singapore
property market] / compiled by the Information Resource Centre, Monetary
Authority of Singapore
dataq
Simon is testing Solr - This one is in English. Color of the Wind. 我是中国人 ,
БOΛbШ OЙ PYCCKO-KИTAЙCKИЙ CΛOBAPb , Français-Chinois
e Analysis
> screen.
>
> Regards,
> Alex.
>
>
> Sign up for my Solr resources newsletter at http://www.solr-start.com/
>
> On 17 February 2015 at 22:36, Simon Cheng wrote:
> > Hi Alex,
> >
> > It's okay after I added in a new field "s_tit
hdfs.security.kerberos.principal">solr/@CLUSTER.HADOOP
and on Hadoop' core-site.xml, my hadoop.security.authentication
parameter is set to Kerberos.
Am I missing something ?
Thank you very much for your input, have a great day.
Simon M.
compound of
lindor and schlitten
but i get
lindor dorsch schlitten
so the filter is extracting dorsch but the word before (lin) and after
(litten) are not valid word parts.
Is there any better compound word filter for German?
Thanks, Simon
Hi,
Have a look at the generated terms to see how they look.
Simon
On Thu, Apr 2, 2015 at 9:43 AM, Palagiri, Jayasankar <
jayashankar.palag...@honeywell.com> wrote:
> Hello Team,
>
> Below is my field type
>
> positionIncrementGap="100"
Good morning,
I used Solr 4.7 to post 186,745 XML files and 186,622 files have been
indexed. That means there are 123 XML files with errors. How can I trace
what these files are?
Thank you in advance,
Simon Cheng.
Hi,
I am using Solr 6.0 on Ubuntu 14.04.
I am ending up with loads of junk in the text body. It starts like,
The JSON entry output of a search result shows the indexed text starting
with...
body_txt_en: " stream_size 36499 X-Parsed-By
org.apache.tika.parser.DefaultParser X-Parsed-By"
An
uot;extractOnly" mode resulting in a XML output.
The difference between selecting "text" or "xml" format is that the
escaped document in the tag is either the original HTML (xml
mode) or stripped HTML (text mode). It seems some Javascript creeps into
the text version.
ng a bug report.
Regards,
Simon
On 27/05/16 20:22, Alexandre Rafalovitch wrote:
I think Solr's layer above Tika was merging in metadata and text all
together without a way (that I could see) to separate them.
That's all I remember of my examination of this issue when I run into
something sim
Thanks Timothy,
Will give the DIH a try. I have submitted a bug report.
Regards,
Simon
On 31/05/16 13:22, Allison, Timothy B. wrote:
From the same page, extractFormat=text only applies when extractOnly
is true, which just shows the output from tika without indexing the document.
Y, sorry
xt mode). It seems some Javascript
creeps into the text version. (See below)
Regards,
Simon
HTML mode sample:
051<?xml
version="1.0" encoding="UTF-8"?>
<html xmlns="http://www.w3.org/1999/xhtml">;
<head>
<link
rel="styleshee
1 - 100 of 282 matches
Mail list logo