Any update from anyone on this? We are invoking major compaction manually
and we see too many skipping of major compaction since we have MIN_VERSION
to 1.

Thanks

On Tue, Aug 11, 2015 at 5:09 PM, mukund murrali <mukundmurra...@gmail.com>
wrote:

> Yes. We have min version to be 1, hence I suppose there wont be any sort
> of time based compaction. After triggering major compaction we had only the
> following logs
>
> compactions.RatioBasedCompactionPolicy: Selecting compaction from 3 store
> files, 0 compacting, 3 eligible, 50 blocking
>
> Thanks
>
> On Tue, Aug 11, 2015 at 3:25 PM, Ted Yu <yuzhih...@gmail.com> wrote:
>
>> bq. oldestTime -9223370597787064221ms
>>
>> This is due to minTimestamp missing from store file:
>>
>>         Long minTimestamp = sf.getMinimumTimestamp();
>>
>>         long oldest = (minTimestamp == null)
>>
>>             ? Long.MIN_VALUE
>>
>>             : now - minTimestamp.longValue();
>>
>> Can you pastebin log from RatioBasedCompactionPolicy in region server log
>> after the manual compaction ?
>>
>> Thanks
>>
>> On Tue, Aug 11, 2015 at 2:06 AM, mukund murrali <mukundmurra...@gmail.com
>> >
>> wrote:
>>
>> > We are using hbase-1.0.0. The following logs appears for all regions in
>> the
>> > regionserver
>> >
>> > 2015-08-08 14:01:51,586 DEBUG [regionserver//R1:16020.compactionChecker]
>> > compactions.RatioBasedCompactionPolicy: Skipping major compaction of
>> >
>> >
>> org.apache.hadoop.hbase.regionserver.compactions.ExploringCompactionPolicy@7bc4e8d8
>> > because one (major) compacted file only, oldestTime
>> -9223370597787064221ms
>> > is < ttl=9223372036854775807 and blockLocalityIndex is 1.0 (min 0.0)
>> >
>> >
>> > Yes after manual triggering the deletes purged. But we don't want to
>> have
>> > it manual. Any other config to avoid such scenario?
>> >
>> > Thanks
>> >
>> >
>> >
>> > On Mon, Aug 10, 2015 at 8:01 PM, Ted Yu <yuzhih...@gmail.com> wrote:
>> >
>> > > What release of hbase are you using ?
>> > >
>> > > Can you pastebin region server log with DEBUG logging ?
>> > >
>> > > I guess you have tried issuing manual command. Did it work ?
>> > >
>> > > Thanks
>> > >
>> > > On Mon, Aug 10, 2015 at 7:02 AM, mukund murrali <
>> > mukundmurra...@gmail.com>
>> > > wrote:
>> > >
>> > > > Any one help us in this :(  Are we missing somewhere in the use
>> case?
>> > > None
>> > > > of the deleted cells are undergoing major compaction.
>> > > >
>> > > > Thanks
>> > > >
>> > > > On Wed, Aug 5, 2015 at 12:04 PM, mukund murrali <
>> > > mukundmurra...@gmail.com>
>> > > > wrote:
>> > > >
>> > > > > Hi,
>> > > > >
>> > > > > We wanted to have deleted data for a week. So we configured
>> > > > >
>> > > > > MIN_VERSIONS => 1
>> > > > > KEEP_DELETED_CELLS => TTL
>> > > > > TTL => 1 week.
>> > > > >
>> > > > > As per our understanding, after 1 week the deleted data becomes
>> > > available
>> > > > > for major compaction and should be purged (correct if wrong).
>> Since
>> > we
>> > > > have
>> > > > > time series data, we don't have any write operations in those
>> regions
>> > > > after
>> > > > > a week . But major compaction never took place for any regions and
>> > our
>> > > > > overall size grew drastically though we have deletes happening.
>> After
>> > > > > analyzing, we found that major compaction takes place if any one
>> of
>> > > the 2
>> > > > > condition is satisfied.
>> > > > >
>> > > > > 1. If the time interval between major compaction is greater than a
>> > week
>> > > > > (default config).
>> > > > > 2. if the block locality index falls below a threshold.
>> > > > >
>> > > > > In our case, since we have min_versions to be 1, the first case
>> > > condition
>> > > > > fails. Time to verify is set to Long.Max value, if min versions is
>> > not
>> > > 0.
>> > > > >
>> > > > > Second is block locality.  To check the block locality index we
>> > enabled
>> > > > > fine logs. And we found the  block locality is always 1, and we
>> got
>> > > logs
>> > > > > stating "Skipping major compaction......".
>> > > > >
>> > > > > So, in this case is manually triggering major compaction the only
>> > > choice?
>> > > > >
>> > > > > Thanks
>> > > > >
>> > > >
>> > >
>> >
>>
>
>

Reply via email to