I remember listening to a Keith presentation about the testing against the multi-level RFile index which was introduced in 1.4.0.

You also want to think about caching at the operating system level. I'm not entirely positive what Keith did to try to mitigate this, but I imagine writing a bunch of garbage from /dev/urandom out to disk should work. That, or you could actually reboot the nodes.

On 8/3/2012 8:55 PM, William Slacum wrote:
Steve, I'm a little confused. The Rfile block cache is tied to a
TServer, so if you kill a node, its cache should go away. Are you
querying for the same data after you kill the node that hosted the
tablet which contained the data? Also, between runs, you could stop and
restart everything, thereby eliminating the cache.

On Fri, Aug 3, 2012 at 5:50 PM, Steven Troxell <[email protected]
<mailto:[email protected]>> wrote:

    Hi  all,

    I am running a benchmarking project on accumulo looking at RDF
    queries for clusters with different node sizes.   While I intend to
    look at caching for each optimizing each individual run, I do NOT
    want caching to interfere for example between runs involving the use
    of 10 and 8 tablet servers.

    Up to now I'd just been killing nodes via the bin/stop-here.sh
    script but I realize that may have allowed caching from previous
    runs with different node sizes to influence my results.   It seemed
    weird to me for exmaple when I realized dropping nodes actually
    increased performance (as measured by query return times) in some
    cases (though I acknowledge the code I'm working with has some
    serious issues with how ineffectively it is actually utilizing
    accumulo, but that's an issue I intend to address later).

    I suppose one way would be between a change of node sizes,  stop and
    restart ALL nodes ( as opposed to what I'd been doing in just
    killing 2 nodes for example in transitioning from a 10 to 8 node
    test).  Will this be sure to clear the influence of caching across
    runs, and is there any cleaner way to do this?

    thanks,
    Steve


Reply via email to