Hi all, We are running a number of Map/Reduce jobs on top of HBase. We are not using HBase for any of its realtime capabilities, only for batch-processing. So we aren't doing lookups, just scans.
Each one of our jobs has *scan.setCaching(false)* to turn off block-caching, since each block will only be accessed once. We recently started using Cloudera Manager, and I’m seeing something that doesn’t add up. See image below. It’s clear from the graphs that Block Cache is being used currently, and blocks are being cached and evicted. We do have *hfile.block.cache.size* set to 0.4 (default), but my understanding is that the jobs setting scan.setCaching(false) should override this. Since it’s set in every job, there should be no blocks being cached. Can anyone help me understand what we’re seeing? Thanks, -Matt [image: Inline image 1]
