Hi,
The heap part was resolved. Thanks for adding it to the wiki. Any ideas as to
why the JMX doesn't show the info when attached?
~Mohit
- Original Message
From: Mattias Persson matt...@neotechnology.com
To: Neo4j user discussions user@lists.neo4j.org
Sent: Tue, September 14, 2010
Mohit,
are you connecting via JConsole to the running process to see the JMX data?
Cheers,
/peter neubauer
VP Product Development, Neo Technology
GTalk: neubauer.peter
Skype peter.neubauer
Phone +46 704 106975
LinkedIn http://www.linkedin.com/in/neubauer
Twitter
Hi,
My suggestion is to look at the BestFirstSelectorFactory abstract
classhttp://components.neo4j.org/neo4j-graph-algo/apidocs/org/neo4j/graphalgo/util/BestFirstSelectorFactory.html.
Extend that and fill in the methods.
It worked wonders for me trying to traverse a weighted graph and worked
Hi, all.
We're trying to use Lucene for fulltext indexing of some textual
content that is stored in Neo, and we've hit a bit of a roadblock. In
some cases, that content will be updated/edited and/or nodes will be
removed, but the process by which index information is removed
Doh! Seems like we just overlooked the method signature with
removeIndex(Node,key), which will do exactly what we want.
Have to lay off the Duff for a while...
Original Message
Subject: [Neo4j] Possible_functional_gap_in_Lucene_indexing?
From:
I just added a way to do this (not as a persistent config, since they
control write behaviour), but instead as an addition to QueryContext. So you
can do like this:
myNodeIndex.query( new QueryContext( name:Mattias occupation:developer
).defaultOperator( Operator.AND ) );
I know it's a bit
2010/9/15 rick.bullo...@burningskysoftware.com
Doh! Seems like we just overlooked the method signature with
removeIndex(Node,key), which will do exactly what we want.
Excellent!
Have to lay off the Duff for a while...
Original Message
Subject: [Neo4j]
I've noticed that when indexing full text, the last term/word is always
ignored. This is a major issue, but I'm not sure if it is in the index
utils or in Lucene itself.
Any thoughts?
Thanks,
Rick
___
Neo4j mailing list
That sounds weird. Look at TestLuceneFulltextIndexService#testSimpleFulltext
method, it queries for the last word and it seems to work.
Could you provide more info on this?
2010/9/15 rick.bullo...@burningskysoftware.com
I've noticed that when indexing full text, the last term/word is always
Using neo4j-index-1.1 and lucene-core-2.9.2, by the way.
Original Message
Subject: Re: [Neo4j] Bug: LuceneFullTextQueryIndex service ignoring
last word/term
From: Mattias Persson [1]matt...@neotechnology.com
Date: Wed, September 15, 2010 10:37 am
To:
Hi Alex, sorry this response took me so long, see responses inline!
2010/9/10 Alexandru Popescu ☀ the.mindstorm.mailingl...@gmail.com
On Tuesday, September 7, 2010, Jacob Hansson ja...@voltvoodoo.com wrote:
2010/9/6 Alexandru Popescu ☀ the.mindstorm.mailingl...@gmail.com
On Monday,
Couldn't it be that sentences ends with a dot... so Cheese is good. will
index the words: [Cheese, is, good.] ? Observe the last word isn't
good, it's good. with a dot. I know that has messed up some searches for
me at least. You could perhaps override the implementation and instantiate
an
Actually, it seems like a deeper bug/design flaw in Lucene's
analyzer/tokenizer. The actual text is HTML text, with p and /p
wrappers. Lucene somewhat randomly seems to treat the last two words
as a single token, and in other cases ignore it altogether. The dot
character screws
Hi
I might be overly simplistic here, but why not lowercase the text, remove
html markup, then remove all non-word-or-space-characters, store this as the
stripped version of the text on the node (for de-indexing) and index this?
/Barklund
On Wed, Sep 15, 2010 at 18:07,
Removing HTML markup is not a trivial task, but luckily, the Apache
Solr team has already created additional analyzers for Lucene that do
what I need (the analysis package in solr has a lot of really good
stuff in it);
I will still need some help from the Neo team to understand
Seems like neo4j-index-1.1 is using a lowercase whitespace tokenizer.
Lucene's StandardTokenizer splits on punctuation (for specifics see
http://grepcode.com/file/repo1.maven.org/maven2/org.apache.lucene/lucene-core/2.9.2/org/apache/lucene/analysis/standard/StandardTokenizer.java?av=f).
I think
On Wed, Sep 15, 2010 at 5:37 PM, Jacob Hansson ja...@voltvoodoo.com wrote:
Hi Alex, sorry this response took me so long, see responses inline!
2010/9/10 Alexandru Popescu ☀ the.mindstorm.mailingl...@gmail.com
On Tuesday, September 7, 2010, Jacob Hansson ja...@voltvoodoo.com wrote:
2010/9/6
Actually, I ended up coming with a workaround that involved using
HTMLStripReader/HTMLStripCharFilter for pre-parsing the text before
passing it into the neo .index(node,key,value) method. Works great, though
there's a little extra string allocation involved. It won't be invoked
often, so it
Well, I have it implemented, and it is cleaning up the content, but the
standard Lucene analyzer still isn't working correctly. Random words are
completely ignored with no special markup in the content, sometimes words
are combined, punctuation is never removed, etc.. Something is really
wrong,
I tried something similar but went block when I couldn't find a way to
retrieve indexes stored for a node so I'm wondering if Lucene can do that
with a decent performance... Don't know if it can retrieve a relationship
like indexed fields - node.
Anyone knows if is that possible?
On Wed, Sep 15,
Solr's LowerCaseTokenizer drops all non-letters (contrast with Lucene's
LowerCaseFilter which just lowercases letters and doesn't drop anything).
See
http://wiki.apache.org/solr/AnalyzersTokenizersTokenFilters#solr.LowerCaseTokenizerFactory
Might be worth trying to tokenize with that after
21 matches
Mail list logo