Hi Andreas,

On 11/03/2019 14:37, Walker, Andreas wrote:
Hi Andy,


the database was created from the web interface, and I've only used the web 
interface to add data to it, so no other version has ever touched it.

OK - so you have only run v3.10.0.

If I understand you correctly, the problem is with the database as a whole, so 
dropping and reloading the graph might not solve the problem. I have now 
switched to a fresh database and am currently reloading the data, so I can see 
whether the problem persists beyond the original database.

If it does then we have a reproducible test case. That said, I can think of a way of a single load or a single load and a sequence of updates not in a parallel can break the node table.

The email of Osma's is a compaction - have you compacted the database? (Fuseki must not be running at the time - this is supposed to be caught using OS file locks but ... I'm told VMs can get this wrong (but I don't know which and when).

The only backup I have done is by making snapshots of the entire virtual server 
this is running on, so I don't think that is related in any way.

Probably not related but is it an instantaneous backup of the filesystem? If not, then it isn't a reliable backup (in the same way that copying all the files isn't a safe backup procedure).

The problem is that if the copy is done while a write transaction is running, some files may be copied before the commit point and some after, which risks chaos.

    Andy



Thanks again for your help,

Andreas

________________________________
Von: Andy Seaborne <[email protected]>
Gesendet: Freitag, 8. März 2019 10:50:28
An: [email protected]
Betreff: Re: AW: AW: Error 500: No conversion to a Node: <RDF_Term >

Hi Andreas,

Is this a database that has only ever been used with 3.10.0 or was the
data loaded with a previous version at some time in the past?

The problem occurs silently during loading. There is no sign of the
problem at the time and the system works just fine while the RDF term,
or terms, are also in the node table cache.

Then the system is restarted.

Then the RDF term is needed for a query and the errors are reported.

But the problem originated back when the data was loaded or updated, may
be several restarts ago.

Of course, it may be a different issue in which case, but the error
message is consistent with the known bug.

Have you been backing up the server on a regular basis? A backup is
NQuads so it is pulling every RDF term from disk (subject to already
being cached).

      Andy

On 07/03/2019 20:47, Walker, Andreas wrote:
Hi Andy,


I am running Version 3.10.0. The problem with reloading the database is the 
regular (multiple times a day) recurrence of the problem, so if there are any 
strategies to avoid it, I'd appreciate any advice.


Best,

Andreas


________________________________
Von: Andy Seaborne <[email protected]>
Gesendet: Donnerstag, 7. März 2019 21:12
An: [email protected]
Betreff: Re: AW: Error 500: No conversion to a Node: <RDF_Term >

Hi Andreas - which version are you running?

It does not look like the  corruption problem, which is now fixed.

The best thing to do is reload the database again. Whatever terms were
messed up are permanently damaged I'm afraid.

       Andy

On 07/03/2019 10:49, Walker, Andreas wrote:
Dear all,


as a quick follow-up which might be helpful in identifying the error; I can 
currently run a SPARQL query (just listing any triples) with LIMIT 80, but no 
higher, before I run into the error, so it seems like there might indeed be a 
particular part of the database that is corrupted.


Best,

Andreas

________________________________
Von: Walker, Andreas <[email protected]>
Gesendet: Mittwoch, 6. März 2019 10:42:32
An: [email protected]
Betreff: Error 500: No conversion to a Node: <RDF_Term >

Dear all,


from time to time, my Fuseki server starts throwing the following error message 
on any SPARQL query I pose to one of my graphs:


"Error 500: No conversion to a Node: <RDF_Term >"


Unfortunately, I couldn't find any explanation of this error message, beyond a 
discussion of a corrupted TDB2 database.


(https://users.jena.apache.narkive.com/LF4XE801/corrupted-tdb2-database)<https://users.jena.apache.narkive.com/LF4XE801/corrupted-tdb2-database>


Once this happens, the only thing I could do so far is to drop the entire 
afflicted graph and rebuild it, but of course that isn't going to be a viable 
solution in the long term.


The only way I interact with Fuseki is by starting and stopping it, querying it 
via the SPARQL endpoint (and sometimes through the web interface, e.g. when 
troubleshooting my application), and uploading new triples (as turtle files) 
via the web interface. I haven't been able to find a pattern in when the error 
appears so far.


Any insights into why this error appears, and what to do in order to avoid it? 
I'd appreciate any help.


Best,

Andreas



Reply via email to