Inline:
On 04/04/2020 12:23, Dave Reynolds wrote:
Hi,
On 03/04/2020 15:38, Benjamin Geer wrote:
I’ve been reading the documentation and list archives about Fuseki
assembler configurations with TDB and reasoners, and I’m trying to
figure out whether the setup I’d like to use is possible. I have three
questions:
1. I’d like to use a forward-chaining reasoner to improve query
performance with a large TDB dataset by inferring some frequently
queried relations. To avoid having to recompute all the inferred
triples every time Fuseki is started (which could take a long time),
I’d like to persist the inferred triples in TDB as well. Is that
possible? I looked for this scenario in the Jena documentation but
didn’t find it.
Basically this isn't supported, sorry.
Benjamin,
What complexity of reasoning are you doing?
There is a tradeoff of complexity/performance at scale/effort needed.
And "effort needed" for complex+scale can be huge.
But RDFS+ level of complixity could take a different approach to the
current rules. Essentially, be backwards chaining (sees updates) with
materialized transitive properties.
It's might be possible to go a fit further that that. Rules that
generate a single triple from a BGP+FILTER(+BIND), together with
transitive properties (not written as rules) might be possible.
The forward chaining engine keeps a *lot* of state in memory in the
RETE-like network. Which means unless you have very selective patterns
in your rules you can end up with large parts of the data in memory. In
worst cases you can have multiple copies.
This has several implications:
First, it means that it's not scalable. If you have a very large TDB
dataset then the reasoner is likely to run out memory. Plus the internal
format is really not optimised for large scale data and inference speed
will take a hit.
Second, it means that there's no point persisting the inference results
on their own, unless they are static. If, as in your case, you want to
continue to add new data and get incremental inferencing then you would
need some way to preserve and restore the intermediate state in the
engine, which is not supported.
So given this there's little point in supporting having the deductions
graph in TDB because that doesn't solve the problems of scaling and
restart.
2. For queries, I’d like a default graph containing the union of all
named graphs plus the inferred statements. Can this be done along with
(1)?
The first part can be done manually but not along with (1).
It's possible to use some offline process to generate a static set of
inferences (whether using the rule engine or e.g. SPARQL construct
queries) to one named graph, put the data in another graph and then have
the default graph be the union.
However, your data isn't static so this doesn't help.
3. The named graphs in the base model need to be continually updated
(always using SPARQL quad patterns), and I’d like the reasoner to
update its inferences when that happens. After reading some old
messages on this list, I think this might not be possible, because if
I understand correctly, the only way to update the base model would be
via a separate Fuseki service that updates the underlying TDB dataset
directly, and in that case, the reasoner won’t see those updates until
Fuseki is restarted. Did I understand that correctly, and if so, is it
still true?
I thought you could configure fuseki to have a reasoner as the source
model and so have updates do to the reasoner rather than a base graph.
However, given none of the rest of what you need to do is supported this
point is moot.
Yes,but not for the union default graph. That actually only exists for
query (and WHERE in SPARQL Update). It isn't updatable. If you update
the named graphs it sees the change but that's bypassed the reasoner on
the graph.
Sorry to not be able to support your use case.
Dave