Dear all, I have been tweaking the tdb.node2nodeid_cache_size and tdb.nodeid2node_cache_size parameters as Andy suggested. They indeed reduce the RAM used by Fuseki, but not to a point where it becomes usable. In attachment you can find a chart plotting memory use increase against dataset size. There is no visible correlation, but on average each additional triplet requires upwards of 30 MB of RAM.
The actual datasets I work with count triplets in the millions (from relational databases with tens of thousands of records). Even if I ever convince a data centre to provide the required amounts of RAM to a single container, the costs will be prohibitive. Can anyone provide their experiences with Fuseki in production? Particularly in micro-services/containerised platforms? Thank you. -- Luís
