to make sure
the
articles being returned are what you expect?
-Original Message-
From: [EMAIL PROTECTED]
[mailto:[EMAIL PROTECTED] On Behalf Of Alan
Darnell
Sent: Tuesday, April 03, 2007 6:45 PM
To: General Mark Logic Developer Discussion
Subject: [MarkLogic Dev General] MarkLogic XQuery
I have a set of documents that have volume and issue numbers
represented as strings (e.g. 1, 10, 2, 23). If I use the following
collation, I get a nice numeric sort without having to pad or
otherwise normalize the data, but performance suffers. Is there an
index I can add that would
James,
You're right -- the data is mixed. For example, you might get s2 or
pt2 for an issue or a roman numeral for a volume. The cast fails on
these but the collation based sort seems to treat them as text and
keeps going.
Alan
On 4-Jun-07, at 6:52 PM, James A. Robinson wrote:
I
of any good strategies for converting math in
TeX format to MathML?
Thanks,
Alan
Alan Darnell
University of Toronto
?xml version=1.0 encoding=iso-8859-1??xml-stylesheet
type=text/xsl href=file://batchgate1\StyleS\bpg4
0.xsl?
!DOCTYPE content PUBLIC -//BLACKWELL PUBLISHING GROUP//DTD 4.0
I've got a few records in my database in languages / scripts that
read from right to left.
If I cut and paste some of this text into a search box, I get no
results. I'm just
wondering how MarkLogic stores these kinds of alphabets and is there
something I
need to do in my XQuery to let the
Great -- thanks Mary. Will give this a try.
On 3-Aug-07, at 7:36 PM, Mary Holstege wrote:
On Fri, 03 Aug 2007 14:00:25 -0700, Alan Darnell
[EMAIL PROTECTED] wrote:
I've got a few records in my database in languages / scripts that
read from right to left.
If I cut and paste some
I love Mark Logic. I love my Mac. I'd like to run Mark Logic on my
Mac. Is there any way (supported or not) that I can do this?
Alan
Alan Darnell
University of Toronto
___
General mailing list
General@developer.marklogic.com
http://xqzone.com
. a list of uris of documents a user has selected
during a session), is it better to store this in the database and
delete it later (e.g. after a user logs out) rather than try to
maintain it in a session variable?
Alan
Alan Darnell
University of Toronto
with it? Do we need to warm the indexes periodically by running
searches against them?
Alan Darnell
University of Toronto
___
General mailing list
General@developer.marklogic.com
http://xqzone.com/mailman/listinfo/general
://kerneltrap.org/node/3000 for some discussion).
One could determine which process is causing the paging, and disable
it (cron jobs are likely candidates). One could also tune down the
VM swappiness, per the kerneltrap link.
-- Mike
Alan Darnell wrote:
We have recently moved from a single host
/hate namespaces dept, it doesn't appear in
your
example that there are any namespaces on your facet-def elements. Try
search:facet-defs, etc.
Hope this helps!
Chris
-Original Message-
From: [EMAIL PROTECTED]
[mailto:[EMAIL PROTECTED] On Behalf Of Alan
Darnell
Sent: Monday, June 16
When I combine a text search with a value search and include an option
to get facet information, the facet information seems to be based only
on the text search and not the combination of the text and value
search. For example, if I run:
import module namespace
out the logical test following the line $search-criteria/
search:values, the facet counts adjust as new values are added to the
search criteria.
Alan
On Jun 20, 2008, at 12:15 AM, Alan Darnell wrote:
When I combine a text search with a value search and include an
option to get facet
Ashish,
Here's the facet definition section in the lib-search-custom file that
we use and that works for us. Of course, we have range indexes
defined for each of the element or attributes we want to facet on.
Alan
facets
default-do-counttrue/default-do-count
I can't find examples in the documentation of how to set up mappings
for lib-parser for element-attribute searching.
Here's what I have now in lib-search-custom.xqy
config
xmlns=http://www.marklogic.com/ps/lib/lib-search/config;
$a
else
VMARGS=$VMARGS $a
fi
done
if [ -d $JAVA_HOME ]; then
JAVA=$JAVA_HOME/bin/java
else
JAVA=java
fi
$JAVA -cp $CP $VMARGS com.marklogic.ps.RecordLoader $FILES
# end recordloader.sh
Alan Darnell wrote:
I'm trying to load some documents that come to me with the
following error
We've been running into some problems using record loader with the
property
CONTENT_FACTORY_CLASSNAME =
com.marklogic.recordloader.xcc.XccModuleContentFactory
and
CONTENT_MODULE_URI = /path-to-our/load.xqy
Load.xqy invokes another module which runs an XQuery transformation on
the input
I'm having difficulty understanding the lib-search documentation on
how to incorporate collection limits into searches.
It looks like lib-search is looking for the parameter coll for
collection limits, but I can't figure out what the value should to
pair with coll.
For instance, say I
-
From: [EMAIL PROTECTED]
[mailto:[EMAIL PROTECTED] On Behalf Of Alan
Darnell
Sent: Thursday, November 20, 2008 5:58 PM
To: General Mark Logic Developer Discussion
Subject: [MarkLogic Dev General] lib-search and collections
I'm having difficulty understanding the lib-search documentation on
how
valueorca/value
/collections
/search-criteria
And you're very welcome.
Cheers!
Chris
-Original Message-
From: [EMAIL PROTECTED]
[mailto:[EMAIL PROTECTED] On Behalf Of Alan
Darnell
Sent: Thursday, November 20, 2008 8:32 PM
To: General Mark Logic Developer Discussion
Subject: Re: [MarkLogic Dev
URI, but that might not match the URI actually inserted by your
XQuery module or Java class.
-- Mike
Alan Darnell wrote:
Sometimes when we are bulk loading records, usually with record
loader, we experience every slow rates of ingest and see a flurry
of messages about DEADLOCKs
/text
/term
collections
set-idrep/set-id
valueorca/value
/collections
/search-criteria
And you're very welcome.
Cheers!
Chris
-Original Message-
From: [EMAIL PROTECTED]
[mailto:[EMAIL PROTECTED] On Behalf Of Alan
Darnell
Sent: Thursday, November 20, 2008 8:32 PM
To: General Mark Logic
We've just upgrade to 4.0 and want to try out some of the sample apps,
including the heat map demo in the /opt/MarkLogic/Samples
directory. We don't currently have a database with geo information in
it -- I wonder if someone could point me in the direction of a public
domain source for a
I'm wondering if there is a way to capture app server errors (as
reported to the error log) and have them sent via the new admin api to
an email address. I'd like to send these on to JIRA to automatically
create project issues as we run our apps through various tests. Would
something
Of Alan
Darnell
Sent: Sunday, January 25, 2009 11:39 AM
To: General Mark Logic Developer Discussion
Subject: [MarkLogic Dev General] capturing app server errors
I'm wondering if there is a way to capture app server errors (as
reported to the error log) and have them sent via the new admin api
I'm wondering if anyone has experience they could share on loading PDFs into ML
and indexing these for text retrieval whie leaving the PDF in the database for
users to download.
Do you use the CPF to extract text from the PDF and store that as a new text
document in ML?
If so, how do you link
...@developer.marklogic.com
[mailto:general-boun...@developer.marklogic.com] On Behalf Of Mary Holstege
Sent: Friday, May 29, 2009 9:16 AM
To: General Mark Logic Developer Discussion
Subject: Re: [MarkLogic Dev General] Experience loading PDFs into ML
On Thu, 28 May 2009 17:18:04 -0700, Alan
Thanks Jason. Very helpful.
Alan
On 6/2/09 8:37 PM, Jason Hunter jhun...@marklogic.com wrote:
On May 28, 2009, at 5:18 PM, Alan Darnell wrote:
I'm wondering if anyone has experience they could share on loading
PDFs into ML and indexing these for text retrieval whie leaving the
PDF
We have a cluster of 6 servers - two e-nodes and 4 d-nodes. All six machines
are configured with 32GB of RAM. The 2 e-nodes behave reasonably, eating up
about 16GB of RAM. All four d-nodes eat up 32GB of RAM quickly after MarkLogic
is started. Two of the machines start eating into swap after
Wondering if anyone has used MarkLogic yet as a platform for
creating/distributing ebooks in the epub format? Adobe has added DRM support
for epubs in its content platform Adobe Content Server. That may not be
available in ML server, but since epub is an XML format it seems that ML would
be
We have been working on migrating a 2TB MarkLogic database from a v 4 cluster
to a v 5 cluster. We transferred the data and reindexed it for ML5. We are
seeing slower response times from the ML5 cluster (which is using more powerful
hardware) when we use frequency-order as an option to
I have some XSLT code that I'm trying to use as a transformation step in
information studio. The XSLT code came from the same vendor as the content.
It uses a function that MarkLogic doesn't recognize
xsl:value-of select=saxon:line-number() /
Read a bit on the web and determined that this is
I'd like to use the Search API and offer users the option to specify boosting
options in their search queries. Something like this:
(TI:water)^ OR water
Where TI is a constraint on the document title element and ^ instructs the
Search API to boost the ranking of documents with the word water
Is there a way to indicate a weight with a constraint in the Search API? I was
thinking a term-option element within the constraint definition with
something like weight=XX might work, but it doesn't seem to. The idea is that
the weight would be passed on to the cts-query constructed from the
13, 2012, at 9:59 PM, Alan Darnell alan.darn...@utoronto.ca wrote:
Is there a way to indicate a weight with a constraint in the Search API? I
was thinking a term-option element within the constraint definition with
something like weight=XX might work, but it doesn't seem to. The idea
I suspect this is an old question answered elsewhere but I can't seem to find a
complete description of best practices for handling empty HTML elements. Say
I'm outputting some text content from an XML document to the browser.
Typically, I'll wrap that content in a div or some other HTML
I have this section in my search options node:
grammar
implicit
cts:near-query strength=20
xmlns:cts=http://marklogic.com/cts/
/implicit
/grammar
This turns the default search from a cts;and-query to cts:near-query,
I'm wondering if anyone has tried to create n-grams from a large body of XML
documents stored in MarkLogic? Single word n-grams can be derived from word
lexicons. But what about 2, 3, 4, or 5 word n-grams? Are there efficient ways
to do this, maybe using Hadoop perhaps, and then storing the
that offloading the extra compute time is a net win.
Would stemming be desirable? If so that makes it harder to benefit from
hadoop - but maybe not impossible.
-- Mike
On 15 Jan 2013, at 16:44 , Alan Darnell alan.darn...@utoronto.ca
wrote:
I'm wondering if anyone has tried to create n-grams from
Putting in a supporting plug for N-gram support in ML. This would be a great
feature for text-mining applications.
Alan
On Feb 4, 2016, at 4:28 PM, Geert Josten
> wrote:
Hi Danny,
Word lexicons don’t expose frequency counts, and
Having a little trouble setting up a field constraint in the Search API and
can’t find an answer in the documentation.
We’ve created a field called “anywhere” composed of a number of elements that
appear in our standard XML document.
We’ve also created a constraint called ANY defined as:
ogic.com<http://www.marklogic.com/>
From:
general-boun...@developer.marklogic.com<mailto:general-boun...@developer.marklogic.com>
[general-boun...@developer.marklogic.com<mailto:general-boun...@developer.marklogic.com>]
on behalf of Alan Darnell
[alan.darn...
Apologies for a very general, newbie question. I understand that ML implements
a triple-store index, can ingest RDF and other formats, and supports SPARQL.
But does ML implement the WC3 specification for Linked Data Platforms?
https://www.w3.org/TR/ldp/
Can I interact with the ML triple
ww.w3.org/TR/sparql11-protocol/
https://www.w3.org/TR/sparql11-http-rdf-update/
John
On 23/08/2017 10:37, Alan Darnell wrote:
Apologies for a very general, newbie question. I understand that ML implements
a triple-store index, can ingest RDF and other formats, and supports SPARQL.
But does
.
>>
>> Alternatively, you might find that cts:parse() provides a faster solution
>> than search:parse() for parsing query text.
>>
>>
>> Hoping that helps,
>>
>>
>> Erik Hennum
>>
>>
>>
>> __
We’ve noticed that when using a Field constraint with the Search API, that
search strings such as:
KW:(war peace)
are parsed in such a way that the two words are treated as a phrase. The
resulting query looks like this:
(cts:search(/book, cts:field-word-query("keywords", "war peace",
46 matches
Mail list logo