Yup, that property can be changed on the fly, and future compactions will use the new codec (all of the queued compactions will see it). It will not actively seek out and re-write files which are not compressed with the current codec.

And, yes, depending on the distro of Hadoop you're using, there might be some extra bits to install. I believe that your normal Hadoop-2.x.y release from Apache will have the native libs available for you and might just to install libsnappy via your OS's package manager.

On 5/6/14, 11:52 AM, David Medinets wrote:
Can that property be changed on the fly? And Snappy needs to be
installed throughout the cluster before making the configuration change?


On Tue, May 6, 2014 at 12:43 AM, Josh Elser <[email protected]
<mailto:[email protected]>> wrote:

    Depending on the CPU/IO ratio for your system, switching to a
    different compression codec might help. Snappy tends to be a bit
    quicker writing out data as opposed to gzip at the cost of being
    larger on disk. The increase in final size on disk might be
    prohibitive depending on your requirements though.

    I forget the table property off hand, but, if you haven't changed
    this already, it will be the property with a defauly value of 'gz' :)

    On May 5, 2014 6:43 PM, "Dickson, Matt MR"
    <[email protected] <mailto:[email protected]>>
    wrote:

        __

        *UNOFFICIAL*

        I'm trying to compact a table and have had a queue of 60,000
        compactions with only 800 running at a time.  This has now been
        running for 4 days and only decreased the queue by 8,000.
        I've increased tserver.compaction.major.concurrent.max=12 and
        stopped all ingest but not seen a change in progress.  Are there
        other accumulo settings I can alter to improve this?  I also saw
        tserver.compaction.major.thread.files.open.max=10 should this be
        increased?
        Thanks in advance,
        Matt


Reply via email to