[jira] [Commented] (HBASE-4578) NPE when altering a table that has moving regions
[ https://issues.apache.org/jira/browse/HBASE-4578?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13133885#comment-13133885 ] gaojinchao commented on HBASE-4578: --- The real cluster test passed. I found that shell show regions is inconsistent with real online regions looks the logs: hbase(main):049:0 alter 't1', NAME = 'f1', VERSIONS = 3 Updating all regions with the new schema... // It said that the cluster has 10 regions. but the cluster has 8 regions and 2 offlined regions. // This needs modify or not? 10/10 regions updated. Done. 0 row(s) in 3.3680 seconds 2011-10-24 04:01:06,510 INFO org.apache.hadoop.hbase.master.handler.TableEventHandler: Skip REGION = {NAME = 't1,20,1319077909713.0b5fa089e2de0430e4b3087492fc239a.', TableName = 't1', STARTKEY = '20', ENDKEY = '30', ENCODED = 0b5fa089e2de0430e4b3087492fc239a, OFFLINE = true, SPLIT = true,} 2011-10-24 04:01:06,510 INFO org.apache.hadoop.hbase.master.handler.TableEventHandler: Skip REGION = {NAME = 't1,40,1319077909713.e2db32ef132229401ecb5b57305b8c8e.', TableName = 't1', STARTKEY = '40', ENDKEY = '', ENCODED = e2db32ef132229401ecb5b57305b8c8e, OFFLINE = true, SPLIT = true,} 2011-10-24 04:01:06,510 INFO org.apache.hadoop.hbase.master.handler.TableEventHandler: Reopening 8 regions on 3 region servers. 2011-10-24 04:01:06,510 DEBUG org.apache.hadoop.hbase.master.BulkReOpen: Creating threads for each region server NPE when altering a table that has moving regions - Key: HBASE-4578 URL: https://issues.apache.org/jira/browse/HBASE-4578 Project: HBase Issue Type: Bug Affects Versions: 0.92.0 Reporter: Jean-Daniel Cryans Assignee: gaojinchao Priority: Blocker Fix For: 0.92.0 Attachments: HBASE-4578_trial_Trunk.patch I'm still not a 100% sure on the source of this error, but here's what I was able to get twice while altering a table that was doing a bunch of splits: {quote} 2011-10-11 23:48:59,344 INFO org.apache.hadoop.hbase.master.handler.SplitRegionHandler: Handled SPLIT report); parent=TestTable,0002608338,1318376880454.a75d6815fdfc513fb1c8aabe086c6763. daughter a=TestTable,0002608338,1318376938764.ef170ff6cd8695dc8aec92e542dc9ac1.daughter b=TestTable,0003301408,1318376938764.36eb2530341bd46888ede312c5559b5d. 2011-10-11 23:49:09,579 DEBUG org.apache.hadoop.hbase.master.handler.TableEventHandler: Ignoring table not disabled exception for supporting online schema changes. 2011-10-11 23:49:09,580 INFO org.apache.hadoop.hbase.master.handler.TableEventHandler: Handling table operation C_M_MODIFY_TABLE on table TestTable 2011-10-11 23:49:09,612 INFO org.apache.hadoop.hbase.util.FSUtils: TableInfoPath = hdfs://sv4r11s38:9100/hbase/TestTable/.tableinfo tmpPath = hdfs://sv4r11s38:9100/hbase/TestTable/.tmp/.tableinfo.1318376949612 2011-10-11 23:49:09,692 INFO org.apache.hadoop.hbase.util.FSUtils: TableDescriptor stored. TableInfoPath = hdfs://sv4r11s38:9100/hbase/TestTable/.tableinfo 2011-10-11 23:49:09,693 INFO org.apache.hadoop.hbase.util.FSUtils: Updated tableinfo=hdfs://sv4r11s38:9100/hbase/TestTable/.tableinfo to blah 2011-10-11 23:49:09,695 INFO org.apache.hadoop.hbase.master.handler.TableEventHandler: Bucketing regions by region server... 2011-10-11 23:49:09,695 DEBUG org.apache.hadoop.hbase.client.MetaScanner: Scanning .META. starting at row=TestTable,,00 for max=2147483647 rows 2011-10-11 23:49:09,709 DEBUG org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation: The connection to hconnection-0x132f043bbde02e9 has been closed. 2011-10-11 23:49:09,709 ERROR org.apache.hadoop.hbase.executor.EventHandler: Caught throwable while processing event C_M_MODIFY_TABLE java.lang.NullPointerException at java.util.TreeMap.getEntry(TreeMap.java:324) at java.util.TreeMap.containsKey(TreeMap.java:209) at org.apache.hadoop.hbase.master.handler.TableEventHandler.reOpenAllRegions(TableEventHandler.java:114) at org.apache.hadoop.hbase.master.handler.TableEventHandler.process(TableEventHandler.java:90) at org.apache.hadoop.hbase.executor.EventHandler.run(EventHandler.java:168) at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) at java.lang.Thread.run(Thread.java:662) {quote} The first time the shell reported that all the regions were updated correctly, the second time it got stuck for a while: {quote} 6/14 regions updated. 0/14 regions updated. ... 0/14 regions updated. 2/16 regions updated. ... 2/16 regions updated. 8/9 regions updated. ... 8/9 regions updated. {quote}
[jira] [Updated] (HBASE-4578) NPE when altering a table that has moving regions
[ https://issues.apache.org/jira/browse/HBASE-4578?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] gaojinchao updated HBASE-4578: -- Attachment: HBASE-4578_Trunk_V1.patch NPE when altering a table that has moving regions - Key: HBASE-4578 URL: https://issues.apache.org/jira/browse/HBASE-4578 Project: HBase Issue Type: Bug Affects Versions: 0.92.0 Reporter: Jean-Daniel Cryans Assignee: gaojinchao Priority: Blocker Fix For: 0.92.0 Attachments: HBASE-4578_Trunk_V1.patch, HBASE-4578_trial_Trunk.patch I'm still not a 100% sure on the source of this error, but here's what I was able to get twice while altering a table that was doing a bunch of splits: {quote} 2011-10-11 23:48:59,344 INFO org.apache.hadoop.hbase.master.handler.SplitRegionHandler: Handled SPLIT report); parent=TestTable,0002608338,1318376880454.a75d6815fdfc513fb1c8aabe086c6763. daughter a=TestTable,0002608338,1318376938764.ef170ff6cd8695dc8aec92e542dc9ac1.daughter b=TestTable,0003301408,1318376938764.36eb2530341bd46888ede312c5559b5d. 2011-10-11 23:49:09,579 DEBUG org.apache.hadoop.hbase.master.handler.TableEventHandler: Ignoring table not disabled exception for supporting online schema changes. 2011-10-11 23:49:09,580 INFO org.apache.hadoop.hbase.master.handler.TableEventHandler: Handling table operation C_M_MODIFY_TABLE on table TestTable 2011-10-11 23:49:09,612 INFO org.apache.hadoop.hbase.util.FSUtils: TableInfoPath = hdfs://sv4r11s38:9100/hbase/TestTable/.tableinfo tmpPath = hdfs://sv4r11s38:9100/hbase/TestTable/.tmp/.tableinfo.1318376949612 2011-10-11 23:49:09,692 INFO org.apache.hadoop.hbase.util.FSUtils: TableDescriptor stored. TableInfoPath = hdfs://sv4r11s38:9100/hbase/TestTable/.tableinfo 2011-10-11 23:49:09,693 INFO org.apache.hadoop.hbase.util.FSUtils: Updated tableinfo=hdfs://sv4r11s38:9100/hbase/TestTable/.tableinfo to blah 2011-10-11 23:49:09,695 INFO org.apache.hadoop.hbase.master.handler.TableEventHandler: Bucketing regions by region server... 2011-10-11 23:49:09,695 DEBUG org.apache.hadoop.hbase.client.MetaScanner: Scanning .META. starting at row=TestTable,,00 for max=2147483647 rows 2011-10-11 23:49:09,709 DEBUG org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation: The connection to hconnection-0x132f043bbde02e9 has been closed. 2011-10-11 23:49:09,709 ERROR org.apache.hadoop.hbase.executor.EventHandler: Caught throwable while processing event C_M_MODIFY_TABLE java.lang.NullPointerException at java.util.TreeMap.getEntry(TreeMap.java:324) at java.util.TreeMap.containsKey(TreeMap.java:209) at org.apache.hadoop.hbase.master.handler.TableEventHandler.reOpenAllRegions(TableEventHandler.java:114) at org.apache.hadoop.hbase.master.handler.TableEventHandler.process(TableEventHandler.java:90) at org.apache.hadoop.hbase.executor.EventHandler.run(EventHandler.java:168) at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) at java.lang.Thread.run(Thread.java:662) {quote} The first time the shell reported that all the regions were updated correctly, the second time it got stuck for a while: {quote} 6/14 regions updated. 0/14 regions updated. ... 0/14 regions updated. 2/16 regions updated. ... 2/16 regions updated. 8/9 regions updated. ... 8/9 regions updated. {quote} After which I killed it, redid the alter and it worked. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HBASE-4224) Need a flush by regionserver rather than by table option
[ https://issues.apache.org/jira/browse/HBASE-4224?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Akash Ashok updated HBASE-4224: --- Attachment: HBase-4224.patch This is not tested completly. I'm yet to do a complete round of testing and write test cases. I've attached the patch so that If there are any review comments I could incorporate them. Cheers Need a flush by regionserver rather than by table option Key: HBASE-4224 URL: https://issues.apache.org/jira/browse/HBASE-4224 Project: HBase Issue Type: Bug Components: shell Reporter: stack Assignee: Akash Ashok Attachments: HBase-4224.patch This evening needed to clean out logs on the cluster. logs are by regionserver. to let go of logs, we need to have all edits emptied from memory. only flush is by table or region. We need to be able to flush the regionserver. Need to add this. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HBASE-4577) Region server reports storefileSizeMB bigger than storefileUncompressedSizeMB
[ https://issues.apache.org/jira/browse/HBASE-4577?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13133922#comment-13133922 ] gaojinchao commented on HBASE-4577: --- I am studying the HFile version 2. Can I handle this issue? Region server reports storefileSizeMB bigger than storefileUncompressedSizeMB - Key: HBASE-4577 URL: https://issues.apache.org/jira/browse/HBASE-4577 Project: HBase Issue Type: Bug Affects Versions: 0.92.0 Reporter: Jean-Daniel Cryans Assignee: gaojinchao Priority: Minor Fix For: 0.92.0 Minor issue while looking at the RS metrics: bq. numberOfStorefiles=8, storefileUncompressedSizeMB=2418, storefileSizeMB=2420, compressionRatio=1.0008 I guess there's a truncation somewhere when it's adding the numbers up. FWIW there's no compression on that table. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Assigned] (HBASE-4577) Region server reports storefileSizeMB bigger than storefileUncompressedSizeMB
[ https://issues.apache.org/jira/browse/HBASE-4577?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] gaojinchao reassigned HBASE-4577: - Assignee: gaojinchao Region server reports storefileSizeMB bigger than storefileUncompressedSizeMB - Key: HBASE-4577 URL: https://issues.apache.org/jira/browse/HBASE-4577 Project: HBase Issue Type: Bug Affects Versions: 0.92.0 Reporter: Jean-Daniel Cryans Assignee: gaojinchao Priority: Minor Fix For: 0.92.0 Minor issue while looking at the RS metrics: bq. numberOfStorefiles=8, storefileUncompressedSizeMB=2418, storefileSizeMB=2420, compressionRatio=1.0008 I guess there's a truncation somewhere when it's adding the numbers up. FWIW there's no compression on that table. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HBASE-4577) Region server reports storefileSizeMB bigger than storefileUncompressedSizeMB
[ https://issues.apache.org/jira/browse/HBASE-4577?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13134013#comment-13134013 ] gaojinchao commented on HBASE-4577: --- Hi J-D It seems that: storefileUncompressedSizeMB only includes datablock size. but storefileSizeMB is the size of total Hfile. So they is different. Do we need add other size? Region server reports storefileSizeMB bigger than storefileUncompressedSizeMB - Key: HBASE-4577 URL: https://issues.apache.org/jira/browse/HBASE-4577 Project: HBase Issue Type: Bug Affects Versions: 0.92.0 Reporter: Jean-Daniel Cryans Assignee: gaojinchao Priority: Minor Fix For: 0.92.0 Minor issue while looking at the RS metrics: bq. numberOfStorefiles=8, storefileUncompressedSizeMB=2418, storefileSizeMB=2420, compressionRatio=1.0008 I guess there's a truncation somewhere when it's adding the numbers up. FWIW there's no compression on that table. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Assigned] (HBASE-2675) Quick smoke tests testsuite
[ https://issues.apache.org/jira/browse/HBASE-2675?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] stack reassigned HBASE-2675: Assignee: nkeywal Assigning this to you Nicolas since it looks like you are fixing this at mo. Quick smoke tests testsuite - Key: HBASE-2675 URL: https://issues.apache.org/jira/browse/HBASE-2675 Project: HBase Issue Type: Test Reporter: Benoit Sigoure Assignee: nkeywal Priority: Minor It would be nice if there was a known subset of the tests that run fast (e.g. not more than a few seconds) and quickly help us check whether the code isn't horribly broken. This way one could run those tests at a frequent interval when iterating and only run the entire testsuite at the end, when they think they're done, since doing so is very time consuming. Someone would need to identify which tests really focus on the core functionality and add a target in the build system to just run those tests. As a bonus, it would be awesome++ if the core tests ran, say, 10x faster than they currently do. There's a lot of sleep-based synchronization in the tests and it would be nice to remove some of that where possible to make the tests run as fast as the machine can handle them. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HBASE-4550) When master passed regionserver different address , because regionserver didn't create new zookeeper znode, as a result stop-hbase.sh is hang
[ https://issues.apache.org/jira/browse/HBASE-4550?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] wanbin updated HBASE-4550: -- Attachment: patch.txt When master passed regionserver different address , because regionserver didn't create new zookeeper znode, as a result stop-hbase.sh is hang --- Key: HBASE-4550 URL: https://issues.apache.org/jira/browse/HBASE-4550 Project: HBase Issue Type: Bug Components: regionserver Affects Versions: 0.90.3 Reporter: wanbin Assignee: wanbin Fix For: 0.90.5 Attachments: hbase-0.90.3.patch, patch, patch.txt Original Estimate: 2h Remaining Estimate: 2h when master passed regionserver different address, regionserver didn't create new zookeeper znode, master store new address in ServerManager, when call stop-hbase.sh , RegionServerTracker.nodeDeleted received path is old address, serverManager.expireServer is not be called. so stop-hbase.sh is hang. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HBASE-4652) Mechanism for atomic recovery from partial failure when atomically bulk-loading multi-CF rows.
[ https://issues.apache.org/jira/browse/HBASE-4652?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jonathan Hsieh updated HBASE-4652: -- Description: The suggestions from HBASE-4552 and initial implementations could violate row atomicity if there is a failure while bulk-loading multiple CF rows. A mechanism to atomically rollback or commit only upon completion (mvcc at the store level) is needed if a partial failures encountered when attempting to bulk load multiple families. For example, let's say I want to bulk load a region with cfs A, B, C. I issue a call to an RS region to atomically bulkload the HFiles. The RS loads A and B successfully but fails on C (hdfs failure, or rs goes down, etc). We could rollback A and B - if we don't we would have A and B loaded but not C and have an atomicity violation. was: The suggestions from HBASE-4552 and initial implementations could violate row atomicity if there are multiple CFs. A mechanism to atomically rollback or commit only upon completion (mvcc at the store level) is needed if a partial failures encountered when attempting to bulk load multiple families. For example, let's say I want to bulk load a region with cfs A, B, C. I issue a call to an RS region to atomically bulkload the HFiles. The RS loads A and B successfully but fails on C (hdfs failure, or rs goes down, etc). We could rollback A and B - if we don't we would have A and B loaded but not C and have an atomicity violation. Mechanism for atomic recovery from partial failure when atomically bulk-loading multi-CF rows. -- Key: HBASE-4652 URL: https://issues.apache.org/jira/browse/HBASE-4652 Project: HBase Issue Type: Bug Reporter: Jonathan Hsieh The suggestions from HBASE-4552 and initial implementations could violate row atomicity if there is a failure while bulk-loading multiple CF rows. A mechanism to atomically rollback or commit only upon completion (mvcc at the store level) is needed if a partial failures encountered when attempting to bulk load multiple families. For example, let's say I want to bulk load a region with cfs A, B, C. I issue a call to an RS region to atomically bulkload the HFiles. The RS loads A and B successfully but fails on C (hdfs failure, or rs goes down, etc). We could rollback A and B - if we don't we would have A and B loaded but not C and have an atomicity violation. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Created] (HBASE-4652) Mechanism for atomic recovery from partial failure when atomically bulk-loading multi-CF rows.
Mechanism for atomic recovery from partial failure when atomically bulk-loading multi-CF rows. -- Key: HBASE-4652 URL: https://issues.apache.org/jira/browse/HBASE-4652 Project: HBase Issue Type: Bug Reporter: Jonathan Hsieh The suggestions from HBASE-4552 and initial implementations could violate row atomicity if there are multiple CFs. A mechanism to atomically rollback or commit only upon completion (mvcc at the store level) is needed if a partial failures encountered when attempting to bulk load multiple families. For example, let's say I want to bulk load a region with cfs A, B, C. I issue a call to an RS region to atomically bulkload the HFiles. The RS loads A and B successfully but fails on C (hdfs failure, or rs goes down, etc). We could rollback A and B - if we don't we would have A and B loaded but not C and have an atomicity violation. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HBASE-4552) multi-CF bulk load is not atomic across column families
[ https://issues.apache.org/jira/browse/HBASE-4552?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13134138#comment-13134138 ] Jonathan Hsieh commented on HBASE-4552: --- Created recovery mechanism jira at HBASE-4652 multi-CF bulk load is not atomic across column families --- Key: HBASE-4552 URL: https://issues.apache.org/jira/browse/HBASE-4552 Project: HBase Issue Type: Bug Components: regionserver Affects Versions: 0.92.0 Reporter: Todd Lipcon Assignee: Jonathan Hsieh Fix For: 0.92.0 Currently the bulk load API simply imports one HFile at a time. With multi-column-family support, this is inappropriate, since different CFs show up separately. Instead, the IPC endpoint should take a of CF - HFiles, so we can online them all under a single region-wide lock. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HBASE-4550) When master passed regionserver different address , because regionserver didn't create new zookeeper znode, as a result stop-hbase.sh is hang
[ https://issues.apache.org/jira/browse/HBASE-4550?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13134137#comment-13134137 ] wanbin commented on HBASE-4550: --- ok. When master passed regionserver different address , because regionserver didn't create new zookeeper znode, as a result stop-hbase.sh is hang --- Key: HBASE-4550 URL: https://issues.apache.org/jira/browse/HBASE-4550 Project: HBase Issue Type: Bug Components: regionserver Affects Versions: 0.90.3 Reporter: wanbin Assignee: wanbin Fix For: 0.90.5 Attachments: hbase-0.90.3.patch, patch, patch.txt Original Estimate: 2h Remaining Estimate: 2h when master passed regionserver different address, regionserver didn't create new zookeeper znode, master store new address in ServerManager, when call stop-hbase.sh , RegionServerTracker.nodeDeleted received path is old address, serverManager.expireServer is not be called. so stop-hbase.sh is hang. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HBASE-4649) Add atomic bulk load function to region server
[ https://issues.apache.org/jira/browse/HBASE-4649?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13134153#comment-13134153 ] Jonathan Hsieh commented on HBASE-4649: --- review here: https://reviews.apache.org/r/2545/ Add atomic bulk load function to region server -- Key: HBASE-4649 URL: https://issues.apache.org/jira/browse/HBASE-4649 Project: HBase Issue Type: Sub-task Components: regionserver Reporter: Jonathan Hsieh Assignee: Jonathan Hsieh Fix For: 0.92.0 Attachments: 0001-HBASE-4649-Add-atomic-bulk-load-function-to-region-s.patch Add a method that atomically bulk load multiple hfiles. Row atomicity guarantees for multi-column family rows require this functionality. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HBASE-4649) Add atomic bulk load function to region server
[ https://issues.apache.org/jira/browse/HBASE-4649?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jonathan Hsieh updated HBASE-4649: -- Attachment: 0001-HBASE-4649-Add-atomic-bulk-load-function-to-region-s.patch Add atomic bulk load function to region server -- Key: HBASE-4649 URL: https://issues.apache.org/jira/browse/HBASE-4649 Project: HBase Issue Type: Sub-task Components: regionserver Reporter: Jonathan Hsieh Assignee: Jonathan Hsieh Fix For: 0.92.0 Attachments: 0001-HBASE-4649-Add-atomic-bulk-load-function-to-region-s.patch Add a method that atomically bulk load multiple hfiles. Row atomicity guarantees for multi-column family rows require this functionality. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HBASE-4649) Add atomic bulk load function to region server
[ https://issues.apache.org/jira/browse/HBASE-4649?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jonathan Hsieh updated HBASE-4649: -- Affects Version/s: 0.92.0 0.90.4 Status: Patch Available (was: Open) Add atomic bulk load function to region server -- Key: HBASE-4649 URL: https://issues.apache.org/jira/browse/HBASE-4649 Project: HBase Issue Type: Sub-task Components: regionserver Affects Versions: 0.90.4, 0.92.0 Reporter: Jonathan Hsieh Assignee: Jonathan Hsieh Fix For: 0.92.0 Attachments: 0001-HBASE-4649-Add-atomic-bulk-load-function-to-region-s.patch Add a method that atomically bulk load multiple hfiles. Row atomicity guarantees for multi-column family rows require this functionality. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HBASE-4649) Add atomic bulk load function to region server
[ https://issues.apache.org/jira/browse/HBASE-4649?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jonathan Hsieh updated HBASE-4649: -- Fix Version/s: (was: 0.92.0) Add atomic bulk load function to region server -- Key: HBASE-4649 URL: https://issues.apache.org/jira/browse/HBASE-4649 Project: HBase Issue Type: Sub-task Components: regionserver Affects Versions: 0.90.4, 0.92.0 Reporter: Jonathan Hsieh Assignee: Jonathan Hsieh Attachments: 0001-HBASE-4649-Add-atomic-bulk-load-function-to-region-s.patch Add a method that atomically bulk load multiple hfiles. Row atomicity guarantees for multi-column family rows require this functionality. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HBASE-4634) test.build.data property overused leading to write data at the wrong place
[ https://issues.apache.org/jira/browse/HBASE-4634?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13134171#comment-13134171 ] stack commented on HBASE-4634: -- @nkeywal Mind refreshing your TRUNK? Patch did not apply cleanly and I got this when trying to build it: {code} [INFO] [ERROR] BUILD FAILURE [INFO] [INFO] Compilation failure /Users/stack/checkout/clean-trunk/src/test/java/org/apache/hadoop/hbase/master/TestDistributedLogSplitting.java:[265,8] cannot find symbol symbol: variable ShowTime ShowTime.start(); /Users/stack/checkout/clean-trunk/src/test/java/org/apache/hadoop/hbase/master/TestDistributedLogSplitting.java:[270,8] cannot find symbol symbol: variable ShowTime ShowTime.pub(public void run() ); /Users/stack/checkout/clean-trunk/src/test/java/org/apache/hadoop/hbase/master/TestDistributedLogSplitting.java:[401,4] cannot find symbol symbol : variable ShowTime location: class org.apache.hadoop.hbase.master.TestDistributedLogSplitting /Users/stack/checkout/clean-trunk/src/test/java/org/apache/hadoop/hbase/master/TestDistributedLogSplitting.java:[403,4] cannot find symbol symbol : variable ShowTime location: class org.apache.hadoop.hbase.master.TestDistributedLogSplitting /Users/stack/checkout/clean-trunk/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java:[228,39] cannot find symbol symbol : method getTestDir() location: class org.apache.hadoop.hbase.HBaseTestingUtility {code} Thanks boss. test.build.data property overused leading to write data at the wrong place Key: HBASE-4634 URL: https://issues.apache.org/jira/browse/HBASE-4634 Project: HBase Issue Type: Bug Components: test Affects Versions: 0.92.0 Environment: all Reporter: nkeywal Assignee: nkeywal Attachments: 2011020_4634_all.patch, 20111020_4639_TestStoreFile.patch, 20111021_4634_all.v2.patch, 20111022_4634_all.v3.patch, 20111023_4634_all.v4.patch test.build.data is overloaded in HBase.At the beginning, it's the Default parent directory for test output., but then it's rewritten to be the directory itself in functions like HBaseTestingUtility#startMiniDFSCluster It seems that this value is already used by MiniDFS (i.e. outside of HBase): Name is as it is because mini dfs has hard-codings to put test data here. As it is today, there is at least a bug in HBaseTestingUtility: {noformat} public void initTestDir() { if (System.getProperty(TEST_DIRECTORY_KEY) == null) { clusterTestBuildDir = setupClusterTestBuildDir(); System.setProperty(TEST_DIRECTORY_KEY, clusterTestBuildDir.getPath()); } } {noformat} if you set a value for test.build.data, the test dir will be the parent directory and not a temp subdir, leading to issues as multiple tests will end-ups in the same (bad) directory. This function is barely used today, hence it's not visible, but I would like to use it in some new code. A possible fix is to remove the check for null and continue with the overloading, but I don't think it would be a big issue to create a new key(like test.build.data.rootdirectory) specific to the root directory and to use test.build.data only to communicate with MiniDFS. Feedback welcome. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HBASE-4634) test.build.data property overused leading to write data at the wrong place
[ https://issues.apache.org/jira/browse/HBASE-4634?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13134181#comment-13134181 ] nkeywal commented on HBASE-4634: that's an error I missed in my patch. Ok to refresh. On Mon, Oct 24, 2011 at 5:51 PM, stack (Commented) (JIRA) test.build.data property overused leading to write data at the wrong place Key: HBASE-4634 URL: https://issues.apache.org/jira/browse/HBASE-4634 Project: HBase Issue Type: Bug Components: test Affects Versions: 0.92.0 Environment: all Reporter: nkeywal Assignee: nkeywal Attachments: 2011020_4634_all.patch, 20111020_4639_TestStoreFile.patch, 20111021_4634_all.v2.patch, 20111022_4634_all.v3.patch, 20111023_4634_all.v4.patch test.build.data is overloaded in HBase.At the beginning, it's the Default parent directory for test output., but then it's rewritten to be the directory itself in functions like HBaseTestingUtility#startMiniDFSCluster It seems that this value is already used by MiniDFS (i.e. outside of HBase): Name is as it is because mini dfs has hard-codings to put test data here. As it is today, there is at least a bug in HBaseTestingUtility: {noformat} public void initTestDir() { if (System.getProperty(TEST_DIRECTORY_KEY) == null) { clusterTestBuildDir = setupClusterTestBuildDir(); System.setProperty(TEST_DIRECTORY_KEY, clusterTestBuildDir.getPath()); } } {noformat} if you set a value for test.build.data, the test dir will be the parent directory and not a temp subdir, leading to issues as multiple tests will end-ups in the same (bad) directory. This function is barely used today, hence it's not visible, but I would like to use it in some new code. A possible fix is to remove the check for null and continue with the overloading, but I don't think it would be a big issue to create a new key(like test.build.data.rootdirectory) specific to the root directory and to use test.build.data only to communicate with MiniDFS. Feedback welcome. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HBASE-4648) Bytes.toBigDecimal() doesn't use offset
[ https://issues.apache.org/jira/browse/HBASE-4648?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13134179#comment-13134179 ] stack commented on HBASE-4648: -- You have a patch please Bryan? Bytes.toBigDecimal() doesn't use offset --- Key: HBASE-4648 URL: https://issues.apache.org/jira/browse/HBASE-4648 Project: HBase Issue Type: Bug Components: util Affects Versions: 0.90.4 Environment: Java 1.6.0_26, Mac OS X 10.7 and CentOS 6 Reporter: Bryan Keller The Bytes.toBigDecimal(byte[], offset, len) method does not use the offset, thus you will get an incorrect result for the BigDecimal unless the BigDecimal's bytes are at the beginning of the byte array. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HBASE-4578) NPE when altering a table that has moving regions
[ https://issues.apache.org/jira/browse/HBASE-4578?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13134227#comment-13134227 ] stack commented on HBASE-4578: -- bq. // It said that the cluster has 10 regions. but the cluster has 8 regions and 2 offlined regions. I think the fact that it says 10 regions updated seems fine. When you say offlined, what do you mean? They are offline parents of splits? If so, they should probably not be counted but no harm done I'd say. You thinking that we should not be altering split, offlined regions? That sounds good but in another issue? NPE when altering a table that has moving regions - Key: HBASE-4578 URL: https://issues.apache.org/jira/browse/HBASE-4578 Project: HBase Issue Type: Bug Affects Versions: 0.92.0 Reporter: Jean-Daniel Cryans Assignee: gaojinchao Priority: Blocker Fix For: 0.92.0 Attachments: HBASE-4578_Trunk_V1.patch, HBASE-4578_trial_Trunk.patch I'm still not a 100% sure on the source of this error, but here's what I was able to get twice while altering a table that was doing a bunch of splits: {quote} 2011-10-11 23:48:59,344 INFO org.apache.hadoop.hbase.master.handler.SplitRegionHandler: Handled SPLIT report); parent=TestTable,0002608338,1318376880454.a75d6815fdfc513fb1c8aabe086c6763. daughter a=TestTable,0002608338,1318376938764.ef170ff6cd8695dc8aec92e542dc9ac1.daughter b=TestTable,0003301408,1318376938764.36eb2530341bd46888ede312c5559b5d. 2011-10-11 23:49:09,579 DEBUG org.apache.hadoop.hbase.master.handler.TableEventHandler: Ignoring table not disabled exception for supporting online schema changes. 2011-10-11 23:49:09,580 INFO org.apache.hadoop.hbase.master.handler.TableEventHandler: Handling table operation C_M_MODIFY_TABLE on table TestTable 2011-10-11 23:49:09,612 INFO org.apache.hadoop.hbase.util.FSUtils: TableInfoPath = hdfs://sv4r11s38:9100/hbase/TestTable/.tableinfo tmpPath = hdfs://sv4r11s38:9100/hbase/TestTable/.tmp/.tableinfo.1318376949612 2011-10-11 23:49:09,692 INFO org.apache.hadoop.hbase.util.FSUtils: TableDescriptor stored. TableInfoPath = hdfs://sv4r11s38:9100/hbase/TestTable/.tableinfo 2011-10-11 23:49:09,693 INFO org.apache.hadoop.hbase.util.FSUtils: Updated tableinfo=hdfs://sv4r11s38:9100/hbase/TestTable/.tableinfo to blah 2011-10-11 23:49:09,695 INFO org.apache.hadoop.hbase.master.handler.TableEventHandler: Bucketing regions by region server... 2011-10-11 23:49:09,695 DEBUG org.apache.hadoop.hbase.client.MetaScanner: Scanning .META. starting at row=TestTable,,00 for max=2147483647 rows 2011-10-11 23:49:09,709 DEBUG org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation: The connection to hconnection-0x132f043bbde02e9 has been closed. 2011-10-11 23:49:09,709 ERROR org.apache.hadoop.hbase.executor.EventHandler: Caught throwable while processing event C_M_MODIFY_TABLE java.lang.NullPointerException at java.util.TreeMap.getEntry(TreeMap.java:324) at java.util.TreeMap.containsKey(TreeMap.java:209) at org.apache.hadoop.hbase.master.handler.TableEventHandler.reOpenAllRegions(TableEventHandler.java:114) at org.apache.hadoop.hbase.master.handler.TableEventHandler.process(TableEventHandler.java:90) at org.apache.hadoop.hbase.executor.EventHandler.run(EventHandler.java:168) at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) at java.lang.Thread.run(Thread.java:662) {quote} The first time the shell reported that all the regions were updated correctly, the second time it got stuck for a while: {quote} 6/14 regions updated. 0/14 regions updated. ... 0/14 regions updated. 2/16 regions updated. ... 2/16 regions updated. 8/9 regions updated. ... 8/9 regions updated. {quote} After which I killed it, redid the alter and it worked. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HBASE-4578) NPE when altering a table that has moving regions
[ https://issues.apache.org/jira/browse/HBASE-4578?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13134229#comment-13134229 ] stack commented on HBASE-4578: -- +1 on patch. Will apply in a little while unless objection. NPE when altering a table that has moving regions - Key: HBASE-4578 URL: https://issues.apache.org/jira/browse/HBASE-4578 Project: HBase Issue Type: Bug Affects Versions: 0.92.0 Reporter: Jean-Daniel Cryans Assignee: gaojinchao Priority: Blocker Fix For: 0.92.0 Attachments: HBASE-4578_Trunk_V1.patch, HBASE-4578_trial_Trunk.patch I'm still not a 100% sure on the source of this error, but here's what I was able to get twice while altering a table that was doing a bunch of splits: {quote} 2011-10-11 23:48:59,344 INFO org.apache.hadoop.hbase.master.handler.SplitRegionHandler: Handled SPLIT report); parent=TestTable,0002608338,1318376880454.a75d6815fdfc513fb1c8aabe086c6763. daughter a=TestTable,0002608338,1318376938764.ef170ff6cd8695dc8aec92e542dc9ac1.daughter b=TestTable,0003301408,1318376938764.36eb2530341bd46888ede312c5559b5d. 2011-10-11 23:49:09,579 DEBUG org.apache.hadoop.hbase.master.handler.TableEventHandler: Ignoring table not disabled exception for supporting online schema changes. 2011-10-11 23:49:09,580 INFO org.apache.hadoop.hbase.master.handler.TableEventHandler: Handling table operation C_M_MODIFY_TABLE on table TestTable 2011-10-11 23:49:09,612 INFO org.apache.hadoop.hbase.util.FSUtils: TableInfoPath = hdfs://sv4r11s38:9100/hbase/TestTable/.tableinfo tmpPath = hdfs://sv4r11s38:9100/hbase/TestTable/.tmp/.tableinfo.1318376949612 2011-10-11 23:49:09,692 INFO org.apache.hadoop.hbase.util.FSUtils: TableDescriptor stored. TableInfoPath = hdfs://sv4r11s38:9100/hbase/TestTable/.tableinfo 2011-10-11 23:49:09,693 INFO org.apache.hadoop.hbase.util.FSUtils: Updated tableinfo=hdfs://sv4r11s38:9100/hbase/TestTable/.tableinfo to blah 2011-10-11 23:49:09,695 INFO org.apache.hadoop.hbase.master.handler.TableEventHandler: Bucketing regions by region server... 2011-10-11 23:49:09,695 DEBUG org.apache.hadoop.hbase.client.MetaScanner: Scanning .META. starting at row=TestTable,,00 for max=2147483647 rows 2011-10-11 23:49:09,709 DEBUG org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation: The connection to hconnection-0x132f043bbde02e9 has been closed. 2011-10-11 23:49:09,709 ERROR org.apache.hadoop.hbase.executor.EventHandler: Caught throwable while processing event C_M_MODIFY_TABLE java.lang.NullPointerException at java.util.TreeMap.getEntry(TreeMap.java:324) at java.util.TreeMap.containsKey(TreeMap.java:209) at org.apache.hadoop.hbase.master.handler.TableEventHandler.reOpenAllRegions(TableEventHandler.java:114) at org.apache.hadoop.hbase.master.handler.TableEventHandler.process(TableEventHandler.java:90) at org.apache.hadoop.hbase.executor.EventHandler.run(EventHandler.java:168) at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) at java.lang.Thread.run(Thread.java:662) {quote} The first time the shell reported that all the regions were updated correctly, the second time it got stuck for a while: {quote} 6/14 regions updated. 0/14 regions updated. ... 0/14 regions updated. 2/16 regions updated. ... 2/16 regions updated. 8/9 regions updated. ... 8/9 regions updated. {quote} After which I killed it, redid the alter and it worked. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HBASE-4649) Add atomic bulk load function to region server
[ https://issues.apache.org/jira/browse/HBASE-4649?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13134233#comment-13134233 ] Ted Yu commented on HBASE-4649: --- There are a few blank lines in TestHRegionServerBulkLoad. There seems to be some overlap between the test and TestLoadIncrementalHFiles.java Since the patch is part of fix for HBASE-4552, I wonder what the plan is to modify LoadIncrementalHFiles where the list of family, path pairs would be built by multiple threads. In other words, what data structure would be used to accommodate concurrency, assuming the current form of LoadIncrementalHFiles is kept ? The above is related to my comment on HBASE-4552 @ 12/Oct/11 19:00. Add atomic bulk load function to region server -- Key: HBASE-4649 URL: https://issues.apache.org/jira/browse/HBASE-4649 Project: HBase Issue Type: Sub-task Components: regionserver Affects Versions: 0.90.4, 0.92.0 Reporter: Jonathan Hsieh Assignee: Jonathan Hsieh Attachments: 0001-HBASE-4649-Add-atomic-bulk-load-function-to-region-s.patch Add a method that atomically bulk load multiple hfiles. Row atomicity guarantees for multi-column family rows require this functionality. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HBASE-3992) Evaluate Lock Free Skip Tree
[ https://issues.apache.org/jira/browse/HBASE-3992?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13134234#comment-13134234 ] stack commented on HBASE-3992: -- There is also this that Ted turned up: https://github.com/nbronson/snaptree/ Evaluate Lock Free Skip Tree Key: HBASE-3992 URL: https://issues.apache.org/jira/browse/HBASE-3992 Project: HBase Issue Type: Improvement Reporter: Jason Rutherglen Priority: Minor We can test out this variant of the ConcurrentSkipListMap. Drop-in replacement for java.util.concurrent.ConcurrentSkipList[Map|Set] https://github.com/mspiegel/lockfreeskiptree -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HBASE-4649) Add atomic bulk load function to region server
[ https://issues.apache.org/jira/browse/HBASE-4649?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13134243#comment-13134243 ] Jonathan Hsieh commented on HBASE-4649: --- @Ted The existing LoadIncrementalHFiles emphasizes behavior when bulk imported HFiles have the wrong region boundaries and require split. The new tests in this patch focuses on the the atomicity properties. createHFile was borrowed and modified so that the expected values would easy to demonstrate atomicity failures. More code was actually borrowed from TestAcidGuarentees. I'll comment about LoadIncrementalHFiles in HBASE-4649 -- I'm in the process of cleaning up a first cut patch. Add atomic bulk load function to region server -- Key: HBASE-4649 URL: https://issues.apache.org/jira/browse/HBASE-4649 Project: HBase Issue Type: Sub-task Components: regionserver Affects Versions: 0.90.4, 0.92.0 Reporter: Jonathan Hsieh Assignee: Jonathan Hsieh Attachments: 0001-HBASE-4649-Add-atomic-bulk-load-function-to-region-s.patch Add a method that atomically bulk load multiple hfiles. Row atomicity guarantees for multi-column family rows require this functionality. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HBASE-4649) Add atomic bulk load function to region server
[ https://issues.apache.org/jira/browse/HBASE-4649?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13134244#comment-13134244 ] Jonathan Hsieh commented on HBASE-4649: --- I meant comment about LoadIncrementalHFiles in HBASE-4650 Add atomic bulk load function to region server -- Key: HBASE-4649 URL: https://issues.apache.org/jira/browse/HBASE-4649 Project: HBase Issue Type: Sub-task Components: regionserver Affects Versions: 0.90.4, 0.92.0 Reporter: Jonathan Hsieh Assignee: Jonathan Hsieh Attachments: 0001-HBASE-4649-Add-atomic-bulk-load-function-to-region-s.patch Add a method that atomically bulk load multiple hfiles. Row atomicity guarantees for multi-column family rows require this functionality. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HBASE-4577) Region server reports storefileSizeMB bigger than storefileUncompressedSizeMB
[ https://issues.apache.org/jira/browse/HBASE-4577?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13134252#comment-13134252 ] Jean-Daniel Cryans commented on HBASE-4577: --- I'm not sure what the best solution is, but I think the current one is confusing since both metrics share the same name with only Uncompressed being different. Looking at the code it seems to be also confused. The javadoc for Store.getStoreSizeUncompressed: bq. The size of the store files, in bytes, uncompressed. And for Store.getStorefilesSize: bq. The size of the store files, in bytes. Even the name of the variable is misleading because it's total data size, not total size: bq. totalUncompressedBytes They read the same, but the first one is really the size of the uncompressed data only. I think names should be changed to reflect what they really mean, no need for new metrics. Region server reports storefileSizeMB bigger than storefileUncompressedSizeMB - Key: HBASE-4577 URL: https://issues.apache.org/jira/browse/HBASE-4577 Project: HBase Issue Type: Bug Affects Versions: 0.92.0 Reporter: Jean-Daniel Cryans Assignee: gaojinchao Priority: Minor Fix For: 0.92.0 Minor issue while looking at the RS metrics: bq. numberOfStorefiles=8, storefileUncompressedSizeMB=2418, storefileSizeMB=2420, compressionRatio=1.0008 I guess there's a truncation somewhere when it's adding the numbers up. FWIW there's no compression on that table. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HBASE-4650) Update LoadIncrementalHFiles to use atomic bulk load RS mechanism
[ https://issues.apache.org/jira/browse/HBASE-4650?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13134253#comment-13134253 ] Jonathan Hsieh commented on HBASE-4650: --- I'm in the process of cleaning up the modifications to LoadIncrementalHFiles and adding more tests before submitting HBASE-4650 for a for review. This cut passes the two tests that use LoadIncrementalHFiles (TestLoadIncrementalHFiles and TestHFileOutputFormat). I'll post a preliminary version for those interested. In the code, there are significant logic changes due to grouping so I've chosen to take out the concurrency on the first cut because gathering and splitting HFiles into proper groups introduces a synchronization point that prevents some of the concurrency as before. This is because groups need to be fully gathered before bulk loads in a region is attempted. I'll include comments where concurrency is ok. Before I spend effort to parallelize this implementation more, I want to add another test to verify that this works while splits are going on. Update LoadIncrementalHFiles to use atomic bulk load RS mechanism - Key: HBASE-4650 URL: https://issues.apache.org/jira/browse/HBASE-4650 Project: HBase Issue Type: Sub-task Components: regionserver Reporter: Jonathan Hsieh Fix For: 0.92.0 MR jobs and command line bulk load execution runs use the LoadIncrementalHFile.doBulkLoad. This needs to be updated to group HFiles by row/region so that rows can be atomically loaded multiple column families. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HBASE-4649) Add atomic bulk load function to region server
[ https://issues.apache.org/jira/browse/HBASE-4649?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13134266#comment-13134266 ] Jonathan Hsieh commented on HBASE-4649: --- With respect to using a Queue vs List, I think he argument is moot -- because of the gather step we have a different instance of the region map, and different list of things to bulk load per iteration. Add atomic bulk load function to region server -- Key: HBASE-4649 URL: https://issues.apache.org/jira/browse/HBASE-4649 Project: HBase Issue Type: Sub-task Components: regionserver Affects Versions: 0.90.4, 0.92.0 Reporter: Jonathan Hsieh Assignee: Jonathan Hsieh Attachments: 0001-HBASE-4649-Add-atomic-bulk-load-function-to-region-s.patch Add a method that atomically bulk load multiple hfiles. Row atomicity guarantees for multi-column family rows require this functionality. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HBASE-4532) Avoid top row seek by dedicated bloom filter for delete family bloom filter
[ https://issues.apache.org/jira/browse/HBASE-4532?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13134277#comment-13134277 ] Nicolas Spiegelberg commented on HBASE-4532: +1 on commit. TestHCM is an issue unrelated to this JIRA and shouldn't hold it up. Should use 'git bisect' to figure out where it was introduced and comment on that JIRA. Avoid top row seek by dedicated bloom filter for delete family bloom filter --- Key: HBASE-4532 URL: https://issues.apache.org/jira/browse/HBASE-4532 Project: HBase Issue Type: Improvement Reporter: Liyin Tang Assignee: Liyin Tang Attachments: D27.1.patch, D27.1.patch, HBASE-4532-apache-trunk.patch, hbase-4532-89-fb.patch The previous jira, HBASE-4469, is to avoid the top row seek operation if row-col bloom filter is enabled. This jira tries to avoid top row seek for all the cases by creating a dedicated bloom filter only for delete family The only subtle use case is when we are interested in the top row with empty column. For example, we are interested in row1/cf1:/1/put. So we seek to the top row: row1/cf1:/MAX_TS/MAXIMUM. And the delete family bloom filter will say there is NO delete family. Then it will avoid the top row seek and return a fake kv, which is the last kv for this row (createLastOnRowCol). In this way, we have already missed the real kv we are interested in. The solution for the above problem is to disable this optimization if we are trying to GET/SCAN a row with empty column. Evaluation from TestSeekOptimization: Previously: For bloom=NONE, compr=NONE total seeks without optimization: 2506, with optimization: 1714 (68.40%), savings: 31.60% For bloom=ROW, compr=NONE total seeks without optimization: 2506, with optimization: 1714 (68.40%), savings: 31.60% For bloom=ROWCOL, compr=NONE total seeks without optimization: 2506, with optimization: 1458 (58.18%), savings: 41.82% For bloom=NONE, compr=GZ total seeks without optimization: 2506, with optimization: 1714 (68.40%), savings: 31.60% For bloom=ROW, compr=GZ total seeks without optimization: 2506, with optimization: 1714 (68.40%), savings: 31.60% For bloom=ROWCOL, compr=GZ total seeks without optimization: 2506, with optimization: 1458 (58.18%), savings: 41.82% So we can get about 10% more seek savings ONLY if the ROWCOL bloom filter is enabled.[HBASE-4469] After this change: For bloom=NONE, compr=NONE total seeks without optimization: 2506, with optimization: 1458 (58.18%), savings: 41.82% For bloom=ROW, compr=NONE total seeks without optimization: 2506, with optimization: 1458 (58.18%), savings: 41.82% For bloom=ROWCOL, compr=NONE total seeks without optimization: 2506, with optimization: 1458 (58.18%), savings: 41.82% For bloom=NONE, compr=GZ total seeks without optimization: 2506, with optimization: 1458 (58.18%), savings: 41.82% For bloom=ROW, compr=GZ total seeks without optimization: 2506, with optimization: 1458 (58.18%), savings: 41.82% For bloom=ROWCOL, compr=GZ total seeks without optimization: 2506, with optimization: 1458 (58.18%), savings: 41.82% So we can get about 10% more seek savings for ALL kinds of bloom filter. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HBASE-4627) Ability to specify a custom start/end to RegionSplitter
[ https://issues.apache.org/jira/browse/HBASE-4627?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Nicolas Spiegelberg updated HBASE-4627: --- Comment: was deleted (was: nspiegelberg requested code review of [jira] [HBASE-4627] Ability to specify a custom start/end to RegionSplitter. Reviewers: DUMMY_REVIEWER [HBASE-4627] added a custom start/end row to RegionSplitter. Also solved an off-by-one error because the end row is prefix-inclusive and not exclusive. a href=https://issues.apache.org/jira/browse/HBASE-4489; title=Better key splitting in RegionSplitterdelHBASE-4489/del/a changed the default endKey on HexStringSplit from 7FFF... to ... While this is correct, existing users of 0.90 RegionSplitter have 7FFF as the end key in their schema and the last region will not split properly under this new code. We need to let the user specify a custom start/end key range for when situations like this arise. Optimally, we should also write the start/end key in META so we could figure this out implicitly instead of requiring the user to explicitly specify it. TEST PLAN - mvn test -Dtest=TestRegionSplitter REVISION DETAIL http://reviews.facebook.net/D39 AFFECTED FILES src/main/java/org/apache/hadoop/hbase/util/Bytes.java src/main/java/org/apache/hadoop/hbase/util/RegionSplitter.java src/test/java/org/apache/hadoop/hbase/util/TestRegionSplitter.java MANAGE HERALD DIFFERENTIAL RULES http://reviews.facebook.net/herald/view/differential/ WHY DID I GET THIS EMAIL? http://reviews.facebook.net/herald/transcript/75/ Tip: use the X-Herald-Rules header to filter Herald messages in your client. ) Ability to specify a custom start/end to RegionSplitter --- Key: HBASE-4627 URL: https://issues.apache.org/jira/browse/HBASE-4627 Project: HBase Issue Type: Improvement Affects Versions: 0.94.0 Reporter: Nicolas Spiegelberg Assignee: Nicolas Spiegelberg Attachments: D39.1.patch, D39.1.patch HBASE-4489 changed the default endKey on HexStringSplit from 7FFF... to ... While this is correct, existing users of 0.90 RegionSplitter have 7FFF as the end key in their schema and the last region will not split properly under this new code. We need to let the user specify a custom start/end key range for when situations like this arise. Optimally, we should also write the start/end key in META so we could figure this out implicitly instead of requiring the user to explicitly specify it. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HBASE-4549) Add thrift API to read version and build date of HBase
[ https://issues.apache.org/jira/browse/HBASE-4549?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13134285#comment-13134285 ] Song Liu commented on HBASE-4549: - ping Jon Add thrift API to read version and build date of HBase --- Key: HBASE-4549 URL: https://issues.apache.org/jira/browse/HBASE-4549 Project: HBase Issue Type: Improvement Components: thrift Reporter: Song Liu Priority: Minor Attachments: patch-hbase-4549.txt Original Estimate: 2h Remaining Estimate: 2h Adding API to get the hbase server version and build date will be helpful for the client to communicate with different versions of the server accordingly. class VersionInfo can be reused to provide required information. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HBASE-4650) Update LoadIncrementalHFiles to use atomic bulk load RS mechanism
[ https://issues.apache.org/jira/browse/HBASE-4650?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jonathan Hsieh updated HBASE-4650: -- Attachment: 0001-HBASE-4650-Update-LoadIncrementalHFiles-to-use-atomi.prelim.patch Here's my preliminary non parallelized implementation. Update LoadIncrementalHFiles to use atomic bulk load RS mechanism - Key: HBASE-4650 URL: https://issues.apache.org/jira/browse/HBASE-4650 Project: HBase Issue Type: Sub-task Components: regionserver Reporter: Jonathan Hsieh Fix For: 0.92.0 Attachments: 0001-HBASE-4650-Update-LoadIncrementalHFiles-to-use-atomi.prelim.patch MR jobs and command line bulk load execution runs use the LoadIncrementalHFile.doBulkLoad. This needs to be updated to group HFiles by row/region so that rows can be atomically loaded multiple column families. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HBASE-4651) ConcurrentModificationException might be thrown in TestHCM.testConnectionUniqueness
[ https://issues.apache.org/jira/browse/HBASE-4651?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Ted Yu updated HBASE-4651: -- Issue Type: Test (was: Bug) Integrated to 0.92 and TRUNK. Thanks for the patch Jinchao. ConcurrentModificationException might be thrown in TestHCM.testConnectionUniqueness --- Key: HBASE-4651 URL: https://issues.apache.org/jira/browse/HBASE-4651 Project: HBase Issue Type: Test Affects Versions: 0.92.0 Reporter: gaojinchao Assignee: gaojinchao Priority: Minor Fix For: 0.92.0 Attachments: HBASE-4651_Trunk.patch See https://builds.apache.org/view/G-L/view/HBase/job/HBase-TRUNK/2357/testReport/junit/org.apache.hadoop.hbase.client/TestHCM/testConnectionUniqueness 363 EntryK,V nextEntry() { 364 if (modCount != expectedModCount) 365 throw new ConcurrentModificationExceptionhttp://kickjava.com/src/java/util/ConcurrentModificationException.java.htm [image: JavaDoc] http://kickjava.com/2487.htm(); Read more: http://kickjava.com/src/java/util/LinkedHashMap.java.htm#ixzz1bbCC0gaT HCM uses proper synchronization when accessing HBASE_INSTANCES. Looking at TestHCM.getValidKeyCount(), it puts values of HBASE_INSTANCES in a Set and returns the size of the Set. However, post HBASE-3777, the values (HConnectionImplementation's) in HBASE_INSTANCES would be unique. TestHCM.getValidKeyCount() can be removed from the test. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Resolved] (HBASE-4651) ConcurrentModificationException might be thrown in TestHCM.testConnectionUniqueness
[ https://issues.apache.org/jira/browse/HBASE-4651?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Ted Yu resolved HBASE-4651. --- Resolution: Fixed Hadoop Flags: Reviewed ConcurrentModificationException might be thrown in TestHCM.testConnectionUniqueness --- Key: HBASE-4651 URL: https://issues.apache.org/jira/browse/HBASE-4651 Project: HBase Issue Type: Test Affects Versions: 0.92.0 Reporter: gaojinchao Assignee: gaojinchao Priority: Minor Fix For: 0.92.0 Attachments: HBASE-4651_Trunk.patch See https://builds.apache.org/view/G-L/view/HBase/job/HBase-TRUNK/2357/testReport/junit/org.apache.hadoop.hbase.client/TestHCM/testConnectionUniqueness 363 EntryK,V nextEntry() { 364 if (modCount != expectedModCount) 365 throw new ConcurrentModificationExceptionhttp://kickjava.com/src/java/util/ConcurrentModificationException.java.htm [image: JavaDoc] http://kickjava.com/2487.htm(); Read more: http://kickjava.com/src/java/util/LinkedHashMap.java.htm#ixzz1bbCC0gaT HCM uses proper synchronization when accessing HBASE_INSTANCES. Looking at TestHCM.getValidKeyCount(), it puts values of HBASE_INSTANCES in a Set and returns the size of the Set. However, post HBASE-3777, the values (HConnectionImplementation's) in HBASE_INSTANCES would be unique. TestHCM.getValidKeyCount() can be removed from the test. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HBASE-4191) Utilize getTopBlockLocations in load balancer
[ https://issues.apache.org/jira/browse/HBASE-4191?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13134308#comment-13134308 ] Karthik Ranganathan commented on HBASE-4191: I think we should also go one step further and store the preferred assignments in a column-family in META. That way, upon restarts the locality is automatically preserved. Utilize getTopBlockLocations in load balancer - Key: HBASE-4191 URL: https://issues.apache.org/jira/browse/HBASE-4191 Project: HBase Issue Type: Improvement Reporter: Ted Yu Assignee: Liyin Tang HBASE-4114 implemented getTopBlockLocations(). Load balancer should utilize this method and assign the region to be moved to the region server with the highest block affinity. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Created] (HBASE-4653) Master can't easily get rid of LZO compressed tables when the codec isn't available
Master can't easily get rid of LZO compressed tables when the codec isn't available --- Key: HBASE-4653 URL: https://issues.apache.org/jira/browse/HBASE-4653 Project: HBase Issue Type: Bug Components: master Affects Versions: 0.90.4 Reporter: Benoit Sigoure Priority: Minor I mistakenly created a table with {{COMPRESSION = LZO}}, and I don't have LZO installed. I'm running a vanilla 0.90.4 release. The master is unable to deploy the region of that table because the codec is missing. I can't get rid of it. I can't drop the table from the shell, although it seems I could disable it. Thankfully I found a workaround for this bug (see further below). {code} hbase(main):003:0 disable 'mytable' 0 row(s) in 1.1010 seconds hbase(main):004:0 drop 'mytable' [hung forever] {code} in the logs: {code} 2011-10-22 03:05:42,153 DEBUG org.apache.hadoop.hbase.regionserver.HRegion: Instantiated mytable,,1319278131519.6eb6891a8b072402b5064f4cc68c210d. 2011-10-22 03:05:42,154 ERROR org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler: Failed open of region=mytable,,1319278131519.6eb6891a8b072402b5064f4cc68c210d. java.io.IOException: java.lang.RuntimeException: java.lang.ClassNotFoundException: com.hadoop.compression.lzo.LzoCodec at org.apache.hadoop.hbase.util.CompressionTest.testCompression(CompressionTest.java:89) at org.apache.hadoop.hbase.regionserver.HRegion.checkCompressionCodecs(HRegion.java:2573) at org.apache.hadoop.hbase.regionserver.HRegion.openHRegion(HRegion.java:2562) at org.apache.hadoop.hbase.regionserver.HRegion.openHRegion(HRegion.java:2550) at org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler.openRegion(OpenRegionHandler.java:272) at org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler.process(OpenRegionHandler.java:99) at org.apache.hadoop.hbase.executor.EventHandler.run(EventHandler.java:156) at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) at java.lang.Thread.run(Thread.java:680) Caused by: java.lang.RuntimeException: java.lang.ClassNotFoundException: com.hadoop.compression.lzo.LzoCodec at org.apache.hadoop.hbase.io.hfile.Compression$Algorithm$1.getCodec(Compression.java:92) at org.apache.hadoop.hbase.io.hfile.Compression$Algorithm.getCompressor(Compression.java:197) at org.apache.hadoop.hbase.util.CompressionTest.testCompression(CompressionTest.java:84) ... 9 more Caused by: java.lang.ClassNotFoundException: com.hadoop.compression.lzo.LzoCodec at java.net.URLClassLoader$1.run(URLClassLoader.java:202) at java.security.AccessController.doPrivileged(Native Method) at java.net.URLClassLoader.findClass(URLClassLoader.java:190) at java.lang.ClassLoader.loadClass(ClassLoader.java:306) at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:301) at java.lang.ClassLoader.loadClass(ClassLoader.java:247) at org.apache.hadoop.hbase.io.hfile.Compression$Algorithm$1.getCodec(Compression.java:87) ... 11 more [...] 2011-10-22 03:15:30,620 DEBUG org.apache.hadoop.hbase.master.handler.DeleteTableHandler: Waiting on region to clear regions in transition; mytable,,1319278131519.6eb6891a8b072402b5064f4cc68c210d. state=OPENING, ts=1319278483001 2011-10-22 03:15:31,621 DEBUG org.apache.hadoop.hbase.master.handler.DeleteTableHandler: Waiting on region to clear regions in transition; mytable,,1319278131519.6eb6891a8b072402b5064f4cc68c210d. state=OPENING, ts=1319278483001 [repeat message above indefinitely every 1s] {code} I tried restarting HBase, no luck. How do I get rid of this table so I can recreate it without {{COMPRESSION = LZO}}? h2. Workaround Change the schema for each family, restart HBase, drop the table. {code} hbase(main):004:0 alter 'mytable', {NAME = 'fam1', COMPRESSION = 'NONE'} 0 row(s) in 0.1160 seconds hbase(main):005:0 alter 'mytable', {NAME = 'fam2', COMPRESSION = 'NONE'} 0 row(s) in 0.0480 seconds hbase(main):007:0 drop 'mytable' ^C [hung forever] {code} [restart HBase] :( {code} hbase(main):001:0 disable 'mytable' 0 row(s) in 2.5010 seconds hbase(main):002:0 drop 'mytable' 0 row(s) in 1.1240 seconds {code} -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HBASE-4653) Master can't easily get rid of LZO compressed tables when the codec isn't available
[ https://issues.apache.org/jira/browse/HBASE-4653?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13134334#comment-13134334 ] Lars Hofhansl commented on HBASE-4653: -- Yep... I've found that too. Didn't think to file a case, though. :) Maybe the answer is to disallow creation of CFs with unsupported compression. Master can't easily get rid of LZO compressed tables when the codec isn't available --- Key: HBASE-4653 URL: https://issues.apache.org/jira/browse/HBASE-4653 Project: HBase Issue Type: Bug Components: master Affects Versions: 0.90.4 Reporter: Benoit Sigoure Priority: Minor I mistakenly created a table with {{COMPRESSION = LZO}}, and I don't have LZO installed. I'm running a vanilla 0.90.4 release. The master is unable to deploy the region of that table because the codec is missing. I can't get rid of it. I can't drop the table from the shell, although it seems I could disable it. Thankfully I found a workaround for this bug (see further below). {code} hbase(main):003:0 disable 'mytable' 0 row(s) in 1.1010 seconds hbase(main):004:0 drop 'mytable' [hung forever] {code} in the logs: {code} 2011-10-22 03:05:42,153 DEBUG org.apache.hadoop.hbase.regionserver.HRegion: Instantiated mytable,,1319278131519.6eb6891a8b072402b5064f4cc68c210d. 2011-10-22 03:05:42,154 ERROR org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler: Failed open of region=mytable,,1319278131519.6eb6891a8b072402b5064f4cc68c210d. java.io.IOException: java.lang.RuntimeException: java.lang.ClassNotFoundException: com.hadoop.compression.lzo.LzoCodec at org.apache.hadoop.hbase.util.CompressionTest.testCompression(CompressionTest.java:89) at org.apache.hadoop.hbase.regionserver.HRegion.checkCompressionCodecs(HRegion.java:2573) at org.apache.hadoop.hbase.regionserver.HRegion.openHRegion(HRegion.java:2562) at org.apache.hadoop.hbase.regionserver.HRegion.openHRegion(HRegion.java:2550) at org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler.openRegion(OpenRegionHandler.java:272) at org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler.process(OpenRegionHandler.java:99) at org.apache.hadoop.hbase.executor.EventHandler.run(EventHandler.java:156) at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) at java.lang.Thread.run(Thread.java:680) Caused by: java.lang.RuntimeException: java.lang.ClassNotFoundException: com.hadoop.compression.lzo.LzoCodec at org.apache.hadoop.hbase.io.hfile.Compression$Algorithm$1.getCodec(Compression.java:92) at org.apache.hadoop.hbase.io.hfile.Compression$Algorithm.getCompressor(Compression.java:197) at org.apache.hadoop.hbase.util.CompressionTest.testCompression(CompressionTest.java:84) ... 9 more Caused by: java.lang.ClassNotFoundException: com.hadoop.compression.lzo.LzoCodec at java.net.URLClassLoader$1.run(URLClassLoader.java:202) at java.security.AccessController.doPrivileged(Native Method) at java.net.URLClassLoader.findClass(URLClassLoader.java:190) at java.lang.ClassLoader.loadClass(ClassLoader.java:306) at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:301) at java.lang.ClassLoader.loadClass(ClassLoader.java:247) at org.apache.hadoop.hbase.io.hfile.Compression$Algorithm$1.getCodec(Compression.java:87) ... 11 more [...] 2011-10-22 03:15:30,620 DEBUG org.apache.hadoop.hbase.master.handler.DeleteTableHandler: Waiting on region to clear regions in transition; mytable,,1319278131519.6eb6891a8b072402b5064f4cc68c210d. state=OPENING, ts=1319278483001 2011-10-22 03:15:31,621 DEBUG org.apache.hadoop.hbase.master.handler.DeleteTableHandler: Waiting on region to clear regions in transition; mytable,,1319278131519.6eb6891a8b072402b5064f4cc68c210d. state=OPENING, ts=1319278483001 [repeat message above indefinitely every 1s] {code} I tried restarting HBase, no luck. How do I get rid of this table so I can recreate it without {{COMPRESSION = LZO}}? h2. Workaround Change the schema for each family, restart HBase, drop the table. {code} hbase(main):004:0 alter 'mytable', {NAME = 'fam1', COMPRESSION = 'NONE'} 0 row(s) in 0.1160 seconds hbase(main):005:0 alter 'mytable', {NAME = 'fam2', COMPRESSION = 'NONE'} 0 row(s) in 0.0480 seconds hbase(main):007:0 drop 'mytable' ^C [hung forever] {code} [restart HBase] :( {code} hbase(main):001:0 disable 'mytable' 0 row(s) in 2.5010 seconds hbase(main):002:0 drop 'mytable' 0 row(s) in
[jira] [Commented] (HBASE-4648) Bytes.toBigDecimal() doesn't use offset
[ https://issues.apache.org/jira/browse/HBASE-4648?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13134336#comment-13134336 ] Lars Hofhansl commented on HBASE-4648: -- Whoa... Good find Bryan. Bytes.toBigDecimal() doesn't use offset --- Key: HBASE-4648 URL: https://issues.apache.org/jira/browse/HBASE-4648 Project: HBase Issue Type: Bug Components: util Affects Versions: 0.90.4 Environment: Java 1.6.0_26, Mac OS X 10.7 and CentOS 6 Reporter: Bryan Keller The Bytes.toBigDecimal(byte[], offset, len) method does not use the offset, thus you will get an incorrect result for the BigDecimal unless the BigDecimal's bytes are at the beginning of the byte array. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HBASE-4224) Need a flush by regionserver rather than by table option
[ https://issues.apache.org/jira/browse/HBASE-4224?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13134355#comment-13134355 ] Ted Yu commented on HBASE-4224: --- Overall, patch looks good. A better name for getRegionServerPairs() would be getRegionServerPairsForTable(). The LOG.info() in isValidRegionServerPair() should be LOG.debug(). Also, offline region isn't covered in the log message. There are some formatting improvements, such as: {code} if(flush(serverName, pair.getFirst())) { {code} There should be a space between if and (. Please also add support for this feature to hbase shell. Need a flush by regionserver rather than by table option Key: HBASE-4224 URL: https://issues.apache.org/jira/browse/HBASE-4224 Project: HBase Issue Type: Bug Components: shell Reporter: stack Assignee: Akash Ashok Attachments: HBase-4224.patch This evening needed to clean out logs on the cluster. logs are by regionserver. to let go of logs, we need to have all edits emptied from memory. only flush is by table or region. We need to be able to flush the regionserver. Need to add this. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HBASE-4629) enable automated patch testing for hbase
[ https://issues.apache.org/jira/browse/HBASE-4629?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13134363#comment-13134363 ] stack commented on HBASE-4629: -- @Giri Are we supposed to commit this to hbase? It creates a dir named dev-support at top level in hbase. Is that what you want? enable automated patch testing for hbase Key: HBASE-4629 URL: https://issues.apache.org/jira/browse/HBASE-4629 Project: HBase Issue Type: New Feature Reporter: Giridharan Kesavan Assignee: Giridharan Kesavan Attachments: HBASE-4629.patch enable jenkins automated patch testing for hbase project -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HBASE-4648) Bytes.toBigDecimal() doesn't use offset
[ https://issues.apache.org/jira/browse/HBASE-4648?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13134386#comment-13134386 ] Bryan Keller commented on HBASE-4648: - I'll upload the patch a little later today, sorry for the delay. Bytes.toBigDecimal() doesn't use offset --- Key: HBASE-4648 URL: https://issues.apache.org/jira/browse/HBASE-4648 Project: HBase Issue Type: Bug Components: util Affects Versions: 0.90.4 Environment: Java 1.6.0_26, Mac OS X 10.7 and CentOS 6 Reporter: Bryan Keller The Bytes.toBigDecimal(byte[], offset, len) method does not use the offset, thus you will get an incorrect result for the BigDecimal unless the BigDecimal's bytes are at the beginning of the byte array. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HBASE-4634) test.build.data property overused leading to write data at the wrong place
[ https://issues.apache.org/jira/browse/HBASE-4634?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] nkeywal updated HBASE-4634: --- Attachment: 20111024_4634_all.v5.patch v5, all tests ok, after a refresh. test.build.data property overused leading to write data at the wrong place Key: HBASE-4634 URL: https://issues.apache.org/jira/browse/HBASE-4634 Project: HBase Issue Type: Bug Components: test Affects Versions: 0.92.0 Environment: all Reporter: nkeywal Assignee: nkeywal Attachments: 2011020_4634_all.patch, 20111020_4639_TestStoreFile.patch, 20111021_4634_all.v2.patch, 20111022_4634_all.v3.patch, 20111023_4634_all.v4.patch, 20111024_4634_all.v5.patch test.build.data is overloaded in HBase.At the beginning, it's the Default parent directory for test output., but then it's rewritten to be the directory itself in functions like HBaseTestingUtility#startMiniDFSCluster It seems that this value is already used by MiniDFS (i.e. outside of HBase): Name is as it is because mini dfs has hard-codings to put test data here. As it is today, there is at least a bug in HBaseTestingUtility: {noformat} public void initTestDir() { if (System.getProperty(TEST_DIRECTORY_KEY) == null) { clusterTestBuildDir = setupClusterTestBuildDir(); System.setProperty(TEST_DIRECTORY_KEY, clusterTestBuildDir.getPath()); } } {noformat} if you set a value for test.build.data, the test dir will be the parent directory and not a temp subdir, leading to issues as multiple tests will end-ups in the same (bad) directory. This function is barely used today, hence it's not visible, but I would like to use it in some new code. A possible fix is to remove the check for null and continue with the overloading, but I don't think it would be a big issue to create a new key(like test.build.data.rootdirectory) specific to the root directory and to use test.build.data only to communicate with MiniDFS. Feedback welcome. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HBASE-4577) Region server reports storefileSizeMB bigger than storefileUncompressedSizeMB
[ https://issues.apache.org/jira/browse/HBASE-4577?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13134408#comment-13134408 ] Matt Corgan commented on HBASE-4577: I think the intention of adding it was to see how big the data would be in memory as opposed to on disk, which is a valuable metric. However, we're already jumping ahead to doing delta encoding and prefix compression, so there will soon be a need for a third metric to track encoded size. Maybe these 3 names would be better: storefileSize: size as reported by the filesystem (lzo/gzip compressed) encodedDataSize: size in the block cache (with delta encoding or prefix compression, but no gzip) rawDataSize (instead of uncompressedBytes): size when stored in the current concatenated KeyValue format (the biggest of the 3) The last 2 would only count datablocks of KeyValues. I'm not sure where bloomfilters and indexblocks should be counted into these. Possibly separate metrics? Region server reports storefileSizeMB bigger than storefileUncompressedSizeMB - Key: HBASE-4577 URL: https://issues.apache.org/jira/browse/HBASE-4577 Project: HBase Issue Type: Bug Affects Versions: 0.92.0 Reporter: Jean-Daniel Cryans Assignee: gaojinchao Priority: Minor Fix For: 0.92.0 Minor issue while looking at the RS metrics: bq. numberOfStorefiles=8, storefileUncompressedSizeMB=2418, storefileSizeMB=2420, compressionRatio=1.0008 I guess there's a truncation somewhere when it's adding the numbers up. FWIW there's no compression on that table. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HBASE-4629) enable automated patch testing for hbase
[ https://issues.apache.org/jira/browse/HBASE-4629?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13134418#comment-13134418 ] Giridharan Kesavan commented on HBASE-4629: --- @Stack. Please commit this patch to hbase. This patch would create a dir - dev-support with two files test-patch.sh and test-patch.properties. Instead of having test-patch script as part of the source code it's better to have it at the top level in a diff directory. enable automated patch testing for hbase Key: HBASE-4629 URL: https://issues.apache.org/jira/browse/HBASE-4629 Project: HBase Issue Type: New Feature Reporter: Giridharan Kesavan Assignee: Giridharan Kesavan Attachments: HBASE-4629.patch enable jenkins automated patch testing for hbase project -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HBASE-4532) Avoid top row seek by dedicated bloom filter for delete family bloom filter
[ https://issues.apache.org/jira/browse/HBASE-4532?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13134421#comment-13134421 ] Hudson commented on HBASE-4532: --- Integrated in HBase-TRUNK #2363 (See [https://builds.apache.org/job/HBase-TRUNK/2363/]) HBASE-4532 Avoid top row seek by dedicated bloom filter for delete family bloom filter nspiegelberg : Files : * /hbase/trunk/src/main/java/org/apache/hadoop/hbase/KeyValue.java * /hbase/trunk/src/main/java/org/apache/hadoop/hbase/io/hfile/BlockType.java * /hbase/trunk/src/main/java/org/apache/hadoop/hbase/io/hfile/HFile.java * /hbase/trunk/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileBlock.java * /hbase/trunk/src/main/java/org/apache/hadoop/hbase/io/hfile/HFilePrettyPrinter.java * /hbase/trunk/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileReaderV1.java * /hbase/trunk/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileReaderV2.java * /hbase/trunk/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileWriterV1.java * /hbase/trunk/src/main/java/org/apache/hadoop/hbase/io/hfile/HFileWriterV2.java * /hbase/trunk/src/main/java/org/apache/hadoop/hbase/regionserver/ScanQueryMatcher.java * /hbase/trunk/src/main/java/org/apache/hadoop/hbase/regionserver/StoreFile.java * /hbase/trunk/src/main/java/org/apache/hadoop/hbase/regionserver/StoreFileScanner.java * /hbase/trunk/src/main/java/org/apache/hadoop/hbase/regionserver/StoreScanner.java * /hbase/trunk/src/main/java/org/apache/hadoop/hbase/util/BloomFilterFactory.java * /hbase/trunk/src/test/java/org/apache/hadoop/hbase/regionserver/TestBlocksRead.java * /hbase/trunk/src/test/java/org/apache/hadoop/hbase/regionserver/TestCompoundBloomFilter.java * /hbase/trunk/src/test/java/org/apache/hadoop/hbase/regionserver/TestStoreFile.java Avoid top row seek by dedicated bloom filter for delete family bloom filter --- Key: HBASE-4532 URL: https://issues.apache.org/jira/browse/HBASE-4532 Project: HBase Issue Type: Improvement Reporter: Liyin Tang Assignee: Liyin Tang Attachments: D27.1.patch, D27.1.patch, HBASE-4532-apache-trunk.patch, hbase-4532-89-fb.patch The previous jira, HBASE-4469, is to avoid the top row seek operation if row-col bloom filter is enabled. This jira tries to avoid top row seek for all the cases by creating a dedicated bloom filter only for delete family The only subtle use case is when we are interested in the top row with empty column. For example, we are interested in row1/cf1:/1/put. So we seek to the top row: row1/cf1:/MAX_TS/MAXIMUM. And the delete family bloom filter will say there is NO delete family. Then it will avoid the top row seek and return a fake kv, which is the last kv for this row (createLastOnRowCol). In this way, we have already missed the real kv we are interested in. The solution for the above problem is to disable this optimization if we are trying to GET/SCAN a row with empty column. Evaluation from TestSeekOptimization: Previously: For bloom=NONE, compr=NONE total seeks without optimization: 2506, with optimization: 1714 (68.40%), savings: 31.60% For bloom=ROW, compr=NONE total seeks without optimization: 2506, with optimization: 1714 (68.40%), savings: 31.60% For bloom=ROWCOL, compr=NONE total seeks without optimization: 2506, with optimization: 1458 (58.18%), savings: 41.82% For bloom=NONE, compr=GZ total seeks without optimization: 2506, with optimization: 1714 (68.40%), savings: 31.60% For bloom=ROW, compr=GZ total seeks without optimization: 2506, with optimization: 1714 (68.40%), savings: 31.60% For bloom=ROWCOL, compr=GZ total seeks without optimization: 2506, with optimization: 1458 (58.18%), savings: 41.82% So we can get about 10% more seek savings ONLY if the ROWCOL bloom filter is enabled.[HBASE-4469] After this change: For bloom=NONE, compr=NONE total seeks without optimization: 2506, with optimization: 1458 (58.18%), savings: 41.82% For bloom=ROW, compr=NONE total seeks without optimization: 2506, with optimization: 1458 (58.18%), savings: 41.82% For bloom=ROWCOL, compr=NONE total seeks without optimization: 2506, with optimization: 1458 (58.18%), savings: 41.82% For bloom=NONE, compr=GZ total seeks without optimization: 2506, with optimization: 1458 (58.18%), savings: 41.82% For bloom=ROW, compr=GZ total seeks without optimization: 2506, with optimization: 1458 (58.18%), savings: 41.82% For bloom=ROWCOL, compr=GZ total seeks without optimization: 2506, with optimization: 1458 (58.18%), savings: 41.82% So we can get about 10% more seek savings for ALL kinds of bloom filter. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please
[jira] [Commented] (HBASE-4651) ConcurrentModificationException might be thrown in TestHCM.testConnectionUniqueness
[ https://issues.apache.org/jira/browse/HBASE-4651?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13134420#comment-13134420 ] Hudson commented on HBASE-4651: --- Integrated in HBase-TRUNK #2363 (See [https://builds.apache.org/job/HBase-TRUNK/2363/]) HBASE-4651 ConcurrentModificationException might be thrown in TestHCM.testConnectionUniqueness (Jinchao) tedyu : Files : * /hbase/trunk/CHANGES.txt * /hbase/trunk/src/test/java/org/apache/hadoop/hbase/client/TestHCM.java ConcurrentModificationException might be thrown in TestHCM.testConnectionUniqueness --- Key: HBASE-4651 URL: https://issues.apache.org/jira/browse/HBASE-4651 Project: HBase Issue Type: Test Affects Versions: 0.92.0 Reporter: gaojinchao Assignee: gaojinchao Priority: Minor Fix For: 0.92.0 Attachments: HBASE-4651_Trunk.patch See https://builds.apache.org/view/G-L/view/HBase/job/HBase-TRUNK/2357/testReport/junit/org.apache.hadoop.hbase.client/TestHCM/testConnectionUniqueness 363 EntryK,V nextEntry() { 364 if (modCount != expectedModCount) 365 throw new ConcurrentModificationExceptionhttp://kickjava.com/src/java/util/ConcurrentModificationException.java.htm [image: JavaDoc] http://kickjava.com/2487.htm(); Read more: http://kickjava.com/src/java/util/LinkedHashMap.java.htm#ixzz1bbCC0gaT HCM uses proper synchronization when accessing HBASE_INSTANCES. Looking at TestHCM.getValidKeyCount(), it puts values of HBASE_INSTANCES in a Set and returns the size of the Set. However, post HBASE-3777, the values (HConnectionImplementation's) in HBASE_INSTANCES would be unique. TestHCM.getValidKeyCount() can be removed from the test. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Resolved] (HBASE-4629) enable automated patch testing for hbase
[ https://issues.apache.org/jira/browse/HBASE-4629?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] stack resolved HBASE-4629. -- Resolution: Fixed Fix Version/s: 0.94.0 Hadoop Flags: Reviewed Committed to TRUNK. Thanks for the patch Giri. enable automated patch testing for hbase Key: HBASE-4629 URL: https://issues.apache.org/jira/browse/HBASE-4629 Project: HBase Issue Type: New Feature Reporter: Giridharan Kesavan Assignee: Giridharan Kesavan Fix For: 0.94.0 Attachments: HBASE-4629.patch enable jenkins automated patch testing for hbase project -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Created] (HBASE-4654) [replication] Add a check to make sure we don't replicate to ourselves
[replication] Add a check to make sure we don't replicate to ourselves -- Key: HBASE-4654 URL: https://issues.apache.org/jira/browse/HBASE-4654 Project: HBase Issue Type: Improvement Affects Versions: 0.90.4 Reporter: Jean-Daniel Cryans Fix For: 0.90.5 It's currently possible to add a peer for replication and point it to the local cluster, which I believe could very well happen for those like us that use only one ZK ensemble per DC so that only the root znode changes when you want to set up replication intra-DC. I don't think comparing just the cluster ID would be enough because you would normally use a different one for another cluster and nothing will block you from pointing elsewhere. Comparing the ZK ensemble address doesn't work either when you have multiple DNS entries that point at the same place. I think this could be resolved by looking up the master address in the relevant znode as it should be exactly the same thing in the case where you have the same cluster. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HBASE-4191) Utilize getTopBlockLocations in load balancer
[ https://issues.apache.org/jira/browse/HBASE-4191?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13134457#comment-13134457 ] Ted Yu commented on HBASE-4191: --- Currently assignAllUserRegions() reads hbase.master.startup.retainassign and if the value is true, balancer.retainAssignment() is called. Looks like new config parameter should be introduce, e.g. hbase.master.startup.policy, to accommodate the above and future enhancements. Utilize getTopBlockLocations in load balancer - Key: HBASE-4191 URL: https://issues.apache.org/jira/browse/HBASE-4191 Project: HBase Issue Type: Improvement Reporter: Ted Yu Assignee: Liyin Tang HBASE-4114 implemented getTopBlockLocations(). Load balancer should utilize this method and assign the region to be moved to the region server with the highest block affinity. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HBASE-4591) TTL for old HLogs should be calculated from last modification time.
[ https://issues.apache.org/jira/browse/HBASE-4591?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Nicolas Spiegelberg updated HBASE-4591: --- Attachment: 0007-HBASE-4591-Increase-TTL-for-old-HLogs-and-change-TTL.patch patch generated from our internal branch, but applies cleanly to all open source branches with 'patch -p1' TTL for old HLogs should be calculated from last modification time. --- Key: HBASE-4591 URL: https://issues.apache.org/jira/browse/HBASE-4591 Project: HBase Issue Type: Improvement Components: master Affects Versions: 0.89.20100621 Reporter: Madhuwanti Vaidya Assignee: Madhuwanti Vaidya Priority: Minor Attachments: 0007-HBASE-4591-Increase-TTL-for-old-HLogs-and-change-TTL.patch -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HBASE-4591) TTL for old HLogs should be calculated from last modification time.
[ https://issues.apache.org/jira/browse/HBASE-4591?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13134478#comment-13134478 ] Ted Yu commented on HBASE-4591: --- {code} +} catch (IOException e) { + LOG.error(Unable to get modification time of file + filePath.getName() + {code} Only the final component of this path is logged. Should the full path be logged ? TTL for old HLogs should be calculated from last modification time. --- Key: HBASE-4591 URL: https://issues.apache.org/jira/browse/HBASE-4591 Project: HBase Issue Type: Improvement Components: master Affects Versions: 0.89.20100621 Reporter: Madhuwanti Vaidya Assignee: Madhuwanti Vaidya Priority: Minor Attachments: 0007-HBASE-4591-Increase-TTL-for-old-HLogs-and-change-TTL.patch -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HBASE-4591) TTL for old HLogs should be calculated from last modification time.
[ https://issues.apache.org/jira/browse/HBASE-4591?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13134495#comment-13134495 ] Nicolas Spiegelberg commented on HBASE-4591: can change that to toString() to print the full pathname. thanks Ted TTL for old HLogs should be calculated from last modification time. --- Key: HBASE-4591 URL: https://issues.apache.org/jira/browse/HBASE-4591 Project: HBase Issue Type: Improvement Components: master Affects Versions: 0.89.20100621 Reporter: Madhuwanti Vaidya Assignee: Madhuwanti Vaidya Priority: Minor Attachments: 0007-HBASE-4591-Increase-TTL-for-old-HLogs-and-change-TTL.patch -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HBASE-4191) hbase load balancer needs locality awareness
[ https://issues.apache.org/jira/browse/HBASE-4191?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Liyin Tang updated HBASE-4191: -- Description: Previously, HBASE-4114 implements the metrics for HFile HDFS block locality, which provides the HFile level locality information. But in order to work with load balancer and region assignment, we need the region level locality information. Let's define the region locality information first, which is almost the same as HFile locality index. HRegion locality index (HRegion A, RegionServer B) = (Total number of HDFS blocks that can be retrieved locally by the RegionServer B for the HRegion A) / ( Total number of the HDFS blocks for the Region A) So the HRegion locality index tells us that how much locality we can get if the HMaster assign the HRegion A to the RegionServer B. So there will be 2 steps involved to assign regions based on the locality. 1) During the cluster start up time, the master will scan the hdfs to calculate the HRegion locality index for each pair of HRegion and Region Server. It is pretty expensive to scan the dfs. So we only needs to do this once during the start up time. 2) During the cluster run time, each region server will update the HRegion locality index as metrics periodically as HBASE-4114 did. The Region Server can expose them to the Master through ZK, meta table, or just RPC messages. Based on the HRegion locality index, the assignment manager in the master would have a global knowledge about the region locality distribution. Imaging the HRegion locality index as the capacity between the region server set and region set, the assignment manager could the run the MAXIMUM FLOW solver to reach the global optimization. In addition, the HBASE-4491 (Locality Checker) is the tool, which is based on the same metrics, to proactively to scan dfs to calculate the global locality information in the cluster. It will help us to verify data locality information during the run time. was: HBASE-4114 implemented getTopBlockLocations(). Load balancer should utilize this method and assign the region to be moved to the region server with the highest block affinity. Issue Type: New Feature (was: Improvement) Summary: hbase load balancer needs locality awareness (was: Utilize getTopBlockLocations in load balancer) hbase load balancer needs locality awareness Key: HBASE-4191 URL: https://issues.apache.org/jira/browse/HBASE-4191 Project: HBase Issue Type: New Feature Reporter: Ted Yu Assignee: Liyin Tang Previously, HBASE-4114 implements the metrics for HFile HDFS block locality, which provides the HFile level locality information. But in order to work with load balancer and region assignment, we need the region level locality information. Let's define the region locality information first, which is almost the same as HFile locality index. HRegion locality index (HRegion A, RegionServer B) = (Total number of HDFS blocks that can be retrieved locally by the RegionServer B for the HRegion A) / ( Total number of the HDFS blocks for the Region A) So the HRegion locality index tells us that how much locality we can get if the HMaster assign the HRegion A to the RegionServer B. So there will be 2 steps involved to assign regions based on the locality. 1) During the cluster start up time, the master will scan the hdfs to calculate the HRegion locality index for each pair of HRegion and Region Server. It is pretty expensive to scan the dfs. So we only needs to do this once during the start up time. 2) During the cluster run time, each region server will update the HRegion locality index as metrics periodically as HBASE-4114 did. The Region Server can expose them to the Master through ZK, meta table, or just RPC messages. Based on the HRegion locality index, the assignment manager in the master would have a global knowledge about the region locality distribution. Imaging the HRegion locality index as the capacity between the region server set and region set, the assignment manager could the run the MAXIMUM FLOW solver to reach the global optimization. In addition, the HBASE-4491 (Locality Checker) is the tool, which is based on the same metrics, to proactively to scan dfs to calculate the global locality information in the cluster. It will help us to verify data locality information during the run time. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HBASE-4591) TTL for old HLogs should be calculated from last modification time.
[ https://issues.apache.org/jira/browse/HBASE-4591?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13134496#comment-13134496 ] Nicolas Spiegelberg commented on HBASE-4591: received +1s internally from karthik.ranga kannanm. Will commit TTL for old HLogs should be calculated from last modification time. --- Key: HBASE-4591 URL: https://issues.apache.org/jira/browse/HBASE-4591 Project: HBase Issue Type: Improvement Components: master Affects Versions: 0.89.20100621 Reporter: Madhuwanti Vaidya Assignee: Madhuwanti Vaidya Priority: Minor Attachments: 0007-HBASE-4591-Increase-TTL-for-old-HLogs-and-change-TTL.patch -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Resolved] (HBASE-1294) Add automated patch testing to issue tracker.
[ https://issues.apache.org/jira/browse/HBASE-1294?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] stack resolved HBASE-1294. -- Resolution: Duplicate Marking as duplicate of HBASE-1294 Add automated patch testing to issue tracker. - Key: HBASE-1294 URL: https://issues.apache.org/jira/browse/HBASE-1294 Project: HBase Issue Type: Improvement Components: build Reporter: Nitay Joffe Hadoop has a system which automatically runs every submitted patch through a slew of checks (e.g. ant test, checkstyle, findbugs, javadoc) which each return +1/-1 on the patch. We should add a similar automated system to our issue tracker. The ZooKeeper folks recently did this. Here's some info on their progress: https://issues.apache.org/jira/browse/ZOOKEEPER-349 http://hudson.zones.apache.org/hudson/view/ZooKeeper/ -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Resolved] (HBASE-4591) TTL for old HLogs should be calculated from last modification time.
[ https://issues.apache.org/jira/browse/HBASE-4591?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Nicolas Spiegelberg resolved HBASE-4591. Resolution: Fixed Fix Version/s: 0.94.0 0.92.0 TTL for old HLogs should be calculated from last modification time. --- Key: HBASE-4591 URL: https://issues.apache.org/jira/browse/HBASE-4591 Project: HBase Issue Type: Improvement Components: master Affects Versions: 0.89.20100621 Reporter: Madhuwanti Vaidya Assignee: Madhuwanti Vaidya Priority: Minor Fix For: 0.92.0, 0.94.0 Attachments: 0007-HBASE-4591-Increase-TTL-for-old-HLogs-and-change-TTL.patch, HBASE-4591-2.patch -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Created] (HBASE-4655) Document architecture of backups
Document architecture of backups Key: HBASE-4655 URL: https://issues.apache.org/jira/browse/HBASE-4655 Project: HBase Issue Type: Sub-task Reporter: Karthik Ranganathan Assignee: Karthik Ranganathan Basic idea behind the backup architecture for HBase -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HBASE-4591) TTL for old HLogs should be calculated from last modification time.
[ https://issues.apache.org/jira/browse/HBASE-4591?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13134522#comment-13134522 ] Nicolas Spiegelberg commented on HBASE-4591: Breaks TestLogsCleaner. Will come out with a test fix patch shortly. TTL for old HLogs should be calculated from last modification time. --- Key: HBASE-4591 URL: https://issues.apache.org/jira/browse/HBASE-4591 Project: HBase Issue Type: Improvement Components: master Affects Versions: 0.89.20100621 Reporter: Madhuwanti Vaidya Assignee: Madhuwanti Vaidya Priority: Minor Fix For: 0.92.0, 0.94.0 Attachments: 0007-HBASE-4591-Increase-TTL-for-old-HLogs-and-change-TTL.patch, HBASE-4591-2.patch -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HBASE-4437) Update hadoop in 0.92 (0.20.205?)
[ https://issues.apache.org/jira/browse/HBASE-4437?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] stack updated HBASE-4437: - Attachment: 4437.txt Move to 0.20.205.0 hadoop. Also includes edits to our jsp and jamon templates commenting out DOCTYPE to get around bug where css is served as text/html. See tail of HBASE-2110 for discussion. Update hadoop in 0.92 (0.20.205?) - Key: HBASE-4437 URL: https://issues.apache.org/jira/browse/HBASE-4437 Project: HBase Issue Type: Task Reporter: stack Attachments: 4437.txt We ship with branch-0.20-append a few versions back from the tip. If 205 comes out and hbase works on it, we should ship 0.92 with it (while also ensuring it work w/ 0.22 and 0.23 branches). -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Created] (HBASE-4656) Note how dfs.support.append has to be enabled in 0.20.205.0 clusters
Note how dfs.support.append has to be enabled in 0.20.205.0 clusters Key: HBASE-4656 URL: https://issues.apache.org/jira/browse/HBASE-4656 Project: HBase Issue Type: Task Reporter: stack -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Resolved] (HBASE-4437) Update hadoop in 0.92 (0.20.205?)
[ https://issues.apache.org/jira/browse/HBASE-4437?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] stack resolved HBASE-4437. -- Resolution: Fixed Fix Version/s: 0.92.0 Assignee: stack Release Note: We ship with hadoop 0.20.205.0 now. Make sure you have dfs.support.append enabled everywhere on your cluster. Hadoop Flags: Reviewed Update hadoop in 0.92 (0.20.205?) - Key: HBASE-4437 URL: https://issues.apache.org/jira/browse/HBASE-4437 Project: HBase Issue Type: Task Reporter: stack Assignee: stack Fix For: 0.92.0 Attachments: 4437.txt We ship with branch-0.20-append a few versions back from the tip. If 205 comes out and hbase works on it, we should ship 0.92 with it (while also ensuring it work w/ 0.22 and 0.23 branches). -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HBASE-4591) TTL for old HLogs should be calculated from last modification time.
[ https://issues.apache.org/jira/browse/HBASE-4591?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13134545#comment-13134545 ] Ted Yu commented on HBASE-4591: --- In HBASE-4591-2.patch, I still see filePath.getName() Do you mind changing that in the addendum ? TTL for old HLogs should be calculated from last modification time. --- Key: HBASE-4591 URL: https://issues.apache.org/jira/browse/HBASE-4591 Project: HBase Issue Type: Improvement Components: master Affects Versions: 0.89.20100621 Reporter: Madhuwanti Vaidya Assignee: Madhuwanti Vaidya Priority: Minor Fix For: 0.92.0, 0.94.0 Attachments: 0007-HBASE-4591-Increase-TTL-for-old-HLogs-and-change-TTL.patch, HBASE-4591-2.patch -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HBASE-2611) Handle RS that fails while processing the failure of another one
[ https://issues.apache.org/jira/browse/HBASE-2611?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13134549#comment-13134549 ] Chris Trezzo commented on HBASE-2611: - @J-D If you don't mind, I was thinking about taking a crack at this using your 4 types of znode strategy. I'll start working on a sketch patch. At a first glance, it seems as though most of the code changes are going to be in ReplicationSourceManager.NodeFailoverWorker.run(). Handle RS that fails while processing the failure of another one Key: HBASE-2611 URL: https://issues.apache.org/jira/browse/HBASE-2611 Project: HBase Issue Type: Sub-task Components: replication Reporter: Jean-Daniel Cryans Assignee: Jean-Daniel Cryans HBASE-2223 doesn't manage region servers that fail while doing the transfer of HLogs queues from other region servers that failed. Devise a reliable way to do it. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HBASE-4650) Update LoadIncrementalHFiles to use atomic bulk load RS mechanism
[ https://issues.apache.org/jira/browse/HBASE-4650?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13134553#comment-13134553 ] Jonathan Hsieh commented on HBASE-4650: --- initial parallel implementation here: https://reviews.apache.org/r/2557/ Update LoadIncrementalHFiles to use atomic bulk load RS mechanism - Key: HBASE-4650 URL: https://issues.apache.org/jira/browse/HBASE-4650 Project: HBase Issue Type: Sub-task Components: regionserver Reporter: Jonathan Hsieh Fix For: 0.92.0 Attachments: 0001-HBASE-4650-Update-LoadIncrementalHFiles-to-use-atomi.patch, 0001-HBASE-4650-Update-LoadIncrementalHFiles-to-use-atomi.prelim.patch MR jobs and command line bulk load execution runs use the LoadIncrementalHFile.doBulkLoad. This needs to be updated to group HFiles by row/region so that rows can be atomically loaded multiple column families. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HBASE-4650) Update LoadIncrementalHFiles to use atomic bulk load RS mechanism
[ https://issues.apache.org/jira/browse/HBASE-4650?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jonathan Hsieh updated HBASE-4650: -- Attachment: 0001-HBASE-4650-Update-LoadIncrementalHFiles-to-use-atomi.patch Update LoadIncrementalHFiles to use atomic bulk load RS mechanism - Key: HBASE-4650 URL: https://issues.apache.org/jira/browse/HBASE-4650 Project: HBase Issue Type: Sub-task Components: regionserver Reporter: Jonathan Hsieh Fix For: 0.92.0 Attachments: 0001-HBASE-4650-Update-LoadIncrementalHFiles-to-use-atomi.patch, 0001-HBASE-4650-Update-LoadIncrementalHFiles-to-use-atomi.prelim.patch MR jobs and command line bulk load execution runs use the LoadIncrementalHFile.doBulkLoad. This needs to be updated to group HFiles by row/region so that rows can be atomically loaded multiple column families. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HBASE-4650) Update LoadIncrementalHFiles to use atomic bulk load RS mechanism
[ https://issues.apache.org/jira/browse/HBASE-4650?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jonathan Hsieh updated HBASE-4650: -- Status: Patch Available (was: Open) Update LoadIncrementalHFiles to use atomic bulk load RS mechanism - Key: HBASE-4650 URL: https://issues.apache.org/jira/browse/HBASE-4650 Project: HBase Issue Type: Sub-task Components: regionserver Reporter: Jonathan Hsieh Fix For: 0.92.0 Attachments: 0001-HBASE-4650-Update-LoadIncrementalHFiles-to-use-atomi.patch, 0001-HBASE-4650-Update-LoadIncrementalHFiles-to-use-atomi.prelim.patch MR jobs and command line bulk load execution runs use the LoadIncrementalHFile.doBulkLoad. This needs to be updated to group HFiles by row/region so that rows can be atomically loaded multiple column families. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Assigned] (HBASE-4650) Update LoadIncrementalHFiles to use atomic bulk load RS mechanism
[ https://issues.apache.org/jira/browse/HBASE-4650?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jonathan Hsieh reassigned HBASE-4650: - Assignee: Jonathan Hsieh Update LoadIncrementalHFiles to use atomic bulk load RS mechanism - Key: HBASE-4650 URL: https://issues.apache.org/jira/browse/HBASE-4650 Project: HBase Issue Type: Sub-task Components: regionserver Reporter: Jonathan Hsieh Assignee: Jonathan Hsieh Fix For: 0.92.0 Attachments: 0001-HBASE-4650-Update-LoadIncrementalHFiles-to-use-atomi.patch, 0001-HBASE-4650-Update-LoadIncrementalHFiles-to-use-atomi.prelim.patch MR jobs and command line bulk load execution runs use the LoadIncrementalHFile.doBulkLoad. This needs to be updated to group HFiles by row/region so that rows can be atomically loaded multiple column families. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HBASE-4629) enable automated patch testing for hbase
[ https://issues.apache.org/jira/browse/HBASE-4629?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13134555#comment-13134555 ] Hudson commented on HBASE-4629: --- Integrated in HBase-TRUNK #2364 (See [https://builds.apache.org/job/HBase-TRUNK/2364/]) HBASE-4629 enable automated patch testing for hbase stack : Files : * /hbase/trunk/CHANGES.txt * /hbase/trunk/dev-support * /hbase/trunk/dev-support/test-patch.properties * /hbase/trunk/dev-support/test-patch.sh enable automated patch testing for hbase Key: HBASE-4629 URL: https://issues.apache.org/jira/browse/HBASE-4629 Project: HBase Issue Type: New Feature Reporter: Giridharan Kesavan Assignee: Giridharan Kesavan Fix For: 0.94.0 Attachments: HBASE-4629.patch enable jenkins automated patch testing for hbase project -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HBASE-2611) Handle RS that fails while processing the failure of another one
[ https://issues.apache.org/jira/browse/HBASE-2611?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13134568#comment-13134568 ] Chris Trezzo commented on HBASE-2611: - ...and of course ReplicationZookeeper. Handle RS that fails while processing the failure of another one Key: HBASE-2611 URL: https://issues.apache.org/jira/browse/HBASE-2611 Project: HBase Issue Type: Sub-task Components: replication Reporter: Jean-Daniel Cryans Assignee: Jean-Daniel Cryans HBASE-2223 doesn't manage region servers that fail while doing the transfer of HLogs queues from other region servers that failed. Devise a reliable way to do it. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HBASE-2110) Move to ivy broke our finding stylesheet (Tables don't have blue lines around them any more).
[ https://issues.apache.org/jira/browse/HBASE-2110?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13134601#comment-13134601 ] stack commented on HBASE-2110: -- I commented out DOCTYPE for now in 0.92 branch and in TRUNK. Move to ivy broke our finding stylesheet (Tables don't have blue lines around them any more). - Key: HBASE-2110 URL: https://issues.apache.org/jira/browse/HBASE-2110 Project: HBase Issue Type: Bug Reporter: stack Labels: noob Move to ivy broke pulling in hbase.css from the static webapp; investigate. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HBASE-4656) Note how dfs.support.append has to be enabled in 0.20.205.0 clusters
[ https://issues.apache.org/jira/browse/HBASE-4656?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] stack updated HBASE-4656: - Attachment: 4656.txt Added note around our hadoop section to book and set dfs.support.append to true in hbase-default.xml Note how dfs.support.append has to be enabled in 0.20.205.0 clusters Key: HBASE-4656 URL: https://issues.apache.org/jira/browse/HBASE-4656 Project: HBase Issue Type: Task Reporter: stack Fix For: 0.92.0 Attachments: 4656.txt -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Resolved] (HBASE-4656) Note how dfs.support.append has to be enabled in 0.20.205.0 clusters
[ https://issues.apache.org/jira/browse/HBASE-4656?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] stack resolved HBASE-4656. -- Resolution: Fixed Fix Version/s: 0.92.0 Assignee: stack Hadoop Flags: Reviewed Note how dfs.support.append has to be enabled in 0.20.205.0 clusters Key: HBASE-4656 URL: https://issues.apache.org/jira/browse/HBASE-4656 Project: HBase Issue Type: Task Reporter: stack Assignee: stack Fix For: 0.92.0 Attachments: 4656.txt -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HBASE-4388) Second start after migration from 90 to trunk crashes
[ https://issues.apache.org/jira/browse/HBASE-4388?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13134610#comment-13134610 ] stack commented on HBASE-4388: -- All unit tests passed for me. Second start after migration from 90 to trunk crashes - Key: HBASE-4388 URL: https://issues.apache.org/jira/browse/HBASE-4388 Project: HBase Issue Type: Bug Components: master, migration Affects Versions: 0.92.0 Reporter: Todd Lipcon Assignee: stack Priority: Blocker Fix For: 0.92.0 Attachments: 4388-v2.txt, 4388-v3.txt, 4388-v4.txt, 4388.txt, hbase-4388-root.dir.tgz, hbase-master-nase.log, meta.tgz I started a trunk cluster to upgrade from 90, inserted a ton of data, then did a clean shutdown. When I started again, I got the following exception: 11/09/13 12:29:09 INFO master.HMaster: Meta has HRI with HTDs. Updating meta now. 11/09/13 12:29:09 FATAL master.HMaster: Unhandled exception. Starting shutdown. java.lang.NegativeArraySizeException: -102 at org.apache.hadoop.hbase.util.Bytes.readByteArray(Bytes.java:147) at org.apache.hadoop.hbase.HTableDescriptor.readFields(HTableDescriptor.java:606) at org.apache.hadoop.hbase.migration.HRegionInfo090x.readFields(HRegionInfo090x.java:641) at org.apache.hadoop.hbase.util.Writables.getWritable(Writables.java:133) at org.apache.hadoop.hbase.util.Writables.getWritable(Writables.java:103) at org.apache.hadoop.hbase.util.Writables.getHRegionInfoForMigration(Writables.java:228) at org.apache.hadoop.hbase.catalog.MetaEditor.getHRegionInfoForMigration(MetaEditor.java:350) at org.apache.hadoop.hbase.catalog.MetaEditor$1.visit(MetaEditor.java:273) at org.apache.hadoop.hbase.catalog.MetaReader.fullScan(MetaReader.java:633) at org.apache.hadoop.hbase.catalog.MetaReader.fullScan(MetaReader.java:255) at org.apache.hadoop.hbase.catalog.MetaReader.fullScan(MetaReader.java:235) at org.apache.hadoop.hbase.catalog.MetaEditor.updateMetaWithNewRegionInfo(MetaEditor.java:284) at org.apache.hadoop.hbase.catalog.MetaEditor.migrateRootAndMeta(MetaEditor.java:298) at org.apache.hadoop.hbase.master.HMaster.updateMetaWithNewHRI(HMaster.java:529) at org.apache.hadoop.hbase.master.HMaster.finishInitialization(HMaster.java:472) at org.apache.hadoop.hbase.master.HMaster.run(HMaster.java:309) -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Resolved] (HBASE-4578) NPE when altering a table that has moving regions
[ https://issues.apache.org/jira/browse/HBASE-4578?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] stack resolved HBASE-4578. -- Resolution: Fixed Hadoop Flags: Reviewed Applied to trunk and branch. Thanks for the patch gaojinchao. NPE when altering a table that has moving regions - Key: HBASE-4578 URL: https://issues.apache.org/jira/browse/HBASE-4578 Project: HBase Issue Type: Bug Affects Versions: 0.92.0 Reporter: Jean-Daniel Cryans Assignee: gaojinchao Priority: Blocker Fix For: 0.92.0 Attachments: HBASE-4578_Trunk_V1.patch, HBASE-4578_trial_Trunk.patch I'm still not a 100% sure on the source of this error, but here's what I was able to get twice while altering a table that was doing a bunch of splits: {quote} 2011-10-11 23:48:59,344 INFO org.apache.hadoop.hbase.master.handler.SplitRegionHandler: Handled SPLIT report); parent=TestTable,0002608338,1318376880454.a75d6815fdfc513fb1c8aabe086c6763. daughter a=TestTable,0002608338,1318376938764.ef170ff6cd8695dc8aec92e542dc9ac1.daughter b=TestTable,0003301408,1318376938764.36eb2530341bd46888ede312c5559b5d. 2011-10-11 23:49:09,579 DEBUG org.apache.hadoop.hbase.master.handler.TableEventHandler: Ignoring table not disabled exception for supporting online schema changes. 2011-10-11 23:49:09,580 INFO org.apache.hadoop.hbase.master.handler.TableEventHandler: Handling table operation C_M_MODIFY_TABLE on table TestTable 2011-10-11 23:49:09,612 INFO org.apache.hadoop.hbase.util.FSUtils: TableInfoPath = hdfs://sv4r11s38:9100/hbase/TestTable/.tableinfo tmpPath = hdfs://sv4r11s38:9100/hbase/TestTable/.tmp/.tableinfo.1318376949612 2011-10-11 23:49:09,692 INFO org.apache.hadoop.hbase.util.FSUtils: TableDescriptor stored. TableInfoPath = hdfs://sv4r11s38:9100/hbase/TestTable/.tableinfo 2011-10-11 23:49:09,693 INFO org.apache.hadoop.hbase.util.FSUtils: Updated tableinfo=hdfs://sv4r11s38:9100/hbase/TestTable/.tableinfo to blah 2011-10-11 23:49:09,695 INFO org.apache.hadoop.hbase.master.handler.TableEventHandler: Bucketing regions by region server... 2011-10-11 23:49:09,695 DEBUG org.apache.hadoop.hbase.client.MetaScanner: Scanning .META. starting at row=TestTable,,00 for max=2147483647 rows 2011-10-11 23:49:09,709 DEBUG org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation: The connection to hconnection-0x132f043bbde02e9 has been closed. 2011-10-11 23:49:09,709 ERROR org.apache.hadoop.hbase.executor.EventHandler: Caught throwable while processing event C_M_MODIFY_TABLE java.lang.NullPointerException at java.util.TreeMap.getEntry(TreeMap.java:324) at java.util.TreeMap.containsKey(TreeMap.java:209) at org.apache.hadoop.hbase.master.handler.TableEventHandler.reOpenAllRegions(TableEventHandler.java:114) at org.apache.hadoop.hbase.master.handler.TableEventHandler.process(TableEventHandler.java:90) at org.apache.hadoop.hbase.executor.EventHandler.run(EventHandler.java:168) at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) at java.lang.Thread.run(Thread.java:662) {quote} The first time the shell reported that all the regions were updated correctly, the second time it got stuck for a while: {quote} 6/14 regions updated. 0/14 regions updated. ... 0/14 regions updated. 2/16 regions updated. ... 2/16 regions updated. 8/9 regions updated. ... 8/9 regions updated. {quote} After which I killed it, redid the alter and it worked. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HBASE-2312) Possible data loss when RS goes into GC pause while rolling HLog
[ https://issues.apache.org/jira/browse/HBASE-2312?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] stack updated HBASE-2312: - Fix Version/s: (was: 0.92.0) 0.94.0 We don't have the hdfs support for this (0.20.205.0 or CDH). Moving out to 0.94.0. Possible data loss when RS goes into GC pause while rolling HLog Key: HBASE-2312 URL: https://issues.apache.org/jira/browse/HBASE-2312 Project: HBase Issue Type: Bug Components: master, regionserver Affects Versions: 0.90.0 Reporter: Karthik Ranganathan Assignee: Nicolas Spiegelberg Priority: Critical Fix For: 0.94.0 There is a very corner case when bad things could happen(ie data loss): 1)RS #1 is going to roll its HLog - not yet created the new one, old one will get no more writes 2)RS #1 enters GC Pause of Death 3)Master lists HLog files of RS#1 that is has to split as RS#1 is dead, starts splitting 4)RS #1 wakes up, created the new HLog (previous one was rolled) and appends an edit - which is lost The following seems like a possible solution: 1)Master detects RS#1 is dead 2)The master renames the /hbase/.logs/regionserver name directory to something else (say /hbase/.logs/regionserver name-dead) 3)Add mkdir support (as opposed to mkdirs) to HDFS - so that a file create fails if the directory doesn't exist. Dhruba tells me this is very doable. 4)RS#1 comes back up and is not able create the new hlog. It restarts itself. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HBASE-4070) [Coprocessors] Improve region server metrics to report loaded coprocessors to master
[ https://issues.apache.org/jira/browse/HBASE-4070?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Andrew Purtell updated HBASE-4070: -- Resolution: Fixed Fix Version/s: 0.94.0 0.92.0 Hadoop Flags: Reviewed Status: Resolved (was: Patch Available) Committed to trunk and 0.92 [Coprocessors] Improve region server metrics to report loaded coprocessors to master Key: HBASE-4070 URL: https://issues.apache.org/jira/browse/HBASE-4070 Project: HBase Issue Type: Improvement Affects Versions: 0.90.3 Reporter: Mingjie Lai Assignee: Eugene Koontz Fix For: 0.92.0, 0.94.0 Attachments: HBASE-4070.patch, HBASE-4070.patch, HBASE-4070.patch, HBASE-4070.patch, master-web-ui.jpg, rs-status-web-ui.jpg HBASE-3512 is about listing loaded cp classes at shell. To make it more generic, we need a way to report this piece of information from region to master (or just at region server level). So later on, we can display the loaded class names at shell as well as web console. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HBASE-4374) Up default regions size from 256M to 1G
[ https://issues.apache.org/jira/browse/HBASE-4374?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13134622#comment-13134622 ] Jean-Daniel Cryans commented on HBASE-4374: --- +1 Up default regions size from 256M to 1G --- Key: HBASE-4374 URL: https://issues.apache.org/jira/browse/HBASE-4374 Project: HBase Issue Type: Task Reporter: stack Priority: Blocker Fix For: 0.92.0 HBASE-4365 has some discussion of why we default for a table should tend to fewer bigger regions. It doesn't look like this issue will be done for 0.92. For 0.92, lets up default region size from 256M to 1G and talk up pre-split on table creation in manual. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HBASE-4648) Bytes.toBigDecimal() doesn't use offset
[ https://issues.apache.org/jira/browse/HBASE-4648?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13134625#comment-13134625 ] Lars Hofhansl commented on HBASE-4648: -- This: {code} public static BigDecimal toBigDecimal(byte[] bytes, int offset) { return toBigDecimal(bytes, offset, bytes.length); } {code} Is also wrong. Should be: {code} return toBigDecimal(bytes, offset, bytes.length-offset); {code} Bytes.toBigDecimal() doesn't use offset --- Key: HBASE-4648 URL: https://issues.apache.org/jira/browse/HBASE-4648 Project: HBase Issue Type: Bug Components: util Affects Versions: 0.90.4 Environment: Java 1.6.0_26, Mac OS X 10.7 and CentOS 6 Reporter: Bryan Keller The Bytes.toBigDecimal(byte[], offset, len) method does not use the offset, thus you will get an incorrect result for the BigDecimal unless the BigDecimal's bytes are at the beginning of the byte array. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HBASE-4254) Get tests passing on Hadoop 23
[ https://issues.apache.org/jira/browse/HBASE-4254?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] stack updated HBASE-4254: - Priority: Major (was: Critical) Knocking this down to major from critical. Get tests passing on Hadoop 23 -- Key: HBASE-4254 URL: https://issues.apache.org/jira/browse/HBASE-4254 Project: HBase Issue Type: Bug Affects Versions: 0.92.0 Reporter: Todd Lipcon Assignee: Todd Lipcon Fix For: 0.92.0 Currently some 30 or so tests are failing on the HBase-trunk-on-hadoop-23 build. It looks like most are reflection-based issues. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HBASE-3512) Coprocessors: Shell support for listing currently loaded coprocessor set
[ https://issues.apache.org/jira/browse/HBASE-3512?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Andrew Purtell updated HBASE-3512: -- Resolution: Fixed Fix Version/s: 0.94.0 Hadoop Flags: Reviewed Status: Resolved (was: Patch Available) Committed to trunk and 0.92 branch. Coprocessors: Shell support for listing currently loaded coprocessor set Key: HBASE-3512 URL: https://issues.apache.org/jira/browse/HBASE-3512 Project: HBase Issue Type: Improvement Components: coprocessors Reporter: Andrew Purtell Assignee: Eugene Koontz Fix For: 0.92.0, 0.94.0 Attachments: HBASE-3512-only.patch, HBASE-3512-only.patch, HBASE-3512.patch, HBASE-3512.patch, HBASE-3512.patch, hbase-shell-session.txt Add support to the shell for listing the coprocessors loaded globally on the regionserver and those loaded on a per-table basis. Perhaps by extending the 'status' command. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HBASE-2312) Possible data loss when RS goes into GC pause while rolling HLog
[ https://issues.apache.org/jira/browse/HBASE-2312?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13134629#comment-13134629 ] Nicolas Spiegelberg commented on HBASE-2312: @stack: since we have proper reflection, can we not apply this jira mark the test as broken? that way, customers can have this fix if they upgrade to the proper version of HDFS instead of requiring an HBase upgrade as well Possible data loss when RS goes into GC pause while rolling HLog Key: HBASE-2312 URL: https://issues.apache.org/jira/browse/HBASE-2312 Project: HBase Issue Type: Bug Components: master, regionserver Affects Versions: 0.90.0 Reporter: Karthik Ranganathan Assignee: Nicolas Spiegelberg Priority: Critical Fix For: 0.94.0 There is a very corner case when bad things could happen(ie data loss): 1)RS #1 is going to roll its HLog - not yet created the new one, old one will get no more writes 2)RS #1 enters GC Pause of Death 3)Master lists HLog files of RS#1 that is has to split as RS#1 is dead, starts splitting 4)RS #1 wakes up, created the new HLog (previous one was rolled) and appends an edit - which is lost The following seems like a possible solution: 1)Master detects RS#1 is dead 2)The master renames the /hbase/.logs/regionserver name directory to something else (say /hbase/.logs/regionserver name-dead) 3)Add mkdir support (as opposed to mkdirs) to HDFS - so that a file create fails if the directory doesn't exist. Dhruba tells me this is very doable. 4)RS#1 comes back up and is not able create the new hlog. It restarts itself. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HBASE-4648) Bytes.toBigDecimal() doesn't use offset
[ https://issues.apache.org/jira/browse/HBASE-4648?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13134628#comment-13134628 ] Bryan Keller commented on HBASE-4648: - I was actually thinking that method isn't useful, it seems it should be removed, as the BigDecimal serialization is variable-length. It would only be used if the BigDecimal comes at the end of the byte array. Bytes.toBigDecimal() doesn't use offset --- Key: HBASE-4648 URL: https://issues.apache.org/jira/browse/HBASE-4648 Project: HBase Issue Type: Bug Components: util Affects Versions: 0.90.4 Environment: Java 1.6.0_26, Mac OS X 10.7 and CentOS 6 Reporter: Bryan Keller The Bytes.toBigDecimal(byte[], offset, len) method does not use the offset, thus you will get an incorrect result for the BigDecimal unless the BigDecimal's bytes are at the beginning of the byte array. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HBASE-4648) Bytes.toBigDecimal() doesn't use offset
[ https://issues.apache.org/jira/browse/HBASE-4648?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Bryan Keller updated HBASE-4648: Attachment: bigdec.patch Here is the patch. I also added some tests for testing serialization with offsets for other data types. Bytes.toBigDecimal() doesn't use offset --- Key: HBASE-4648 URL: https://issues.apache.org/jira/browse/HBASE-4648 Project: HBase Issue Type: Bug Components: util Affects Versions: 0.90.4 Environment: Java 1.6.0_26, Mac OS X 10.7 and CentOS 6 Reporter: Bryan Keller Attachments: bigdec.patch The Bytes.toBigDecimal(byte[], offset, len) method does not use the offset, thus you will get an incorrect result for the BigDecimal unless the BigDecimal's bytes are at the beginning of the byte array. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HBASE-4648) Bytes.toBigDecimal() doesn't use offset
[ https://issues.apache.org/jira/browse/HBASE-4648?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Bryan Keller updated HBASE-4648: Attachment: bigdec2.patch Sorry bad patch, here's a new one. Bytes.toBigDecimal() doesn't use offset --- Key: HBASE-4648 URL: https://issues.apache.org/jira/browse/HBASE-4648 Project: HBase Issue Type: Bug Components: util Affects Versions: 0.90.4 Environment: Java 1.6.0_26, Mac OS X 10.7 and CentOS 6 Reporter: Bryan Keller Attachments: bigdec.patch, bigdec2.patch The Bytes.toBigDecimal(byte[], offset, len) method does not use the offset, thus you will get an incorrect result for the BigDecimal unless the BigDecimal's bytes are at the beginning of the byte array. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HBASE-4645) Edits Log recovery losing data across column families
[ https://issues.apache.org/jira/browse/HBASE-4645?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Amitanand Aiyer updated HBASE-4645: --- Attachment: 4645-v9.diff v9 of the revision from https://reviews.apache.org/r/2524/diff Edits Log recovery losing data across column families - Key: HBASE-4645 URL: https://issues.apache.org/jira/browse/HBASE-4645 Project: HBase Issue Type: Bug Affects Versions: 0.89.20100924, 0.92.0 Reporter: Amitanand Aiyer Assignee: Amitanand Aiyer Attachments: 4645-v9.diff There is a data loss happening (for some of the column families) when we do the replay logs. The bug seems to be from the fact that during replay-logs we only choose to replay the logs from the maximumSequenceID across *ALL* the stores. This is wrong. If a column family is ahead of others (because the crash happened before all the column families were flushed), then we lose data for the column families that have not yet caught up. The correct logic for replay should begin the replay from the minimum across the maximum in each store. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HBASE-4645) Edits Log recovery losing data across column families
[ https://issues.apache.org/jira/browse/HBASE-4645?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Amitanand Aiyer updated HBASE-4645: --- Status: Patch Available (was: Open) Edits Log recovery losing data across column families - Key: HBASE-4645 URL: https://issues.apache.org/jira/browse/HBASE-4645 Project: HBase Issue Type: Bug Affects Versions: 0.89.20100924, 0.92.0 Reporter: Amitanand Aiyer Assignee: Amitanand Aiyer Attachments: 4645-v9.diff There is a data loss happening (for some of the column families) when we do the replay logs. The bug seems to be from the fact that during replay-logs we only choose to replay the logs from the maximumSequenceID across *ALL* the stores. This is wrong. If a column family is ahead of others (because the crash happened before all the column families were flushed), then we lose data for the column families that have not yet caught up. The correct logic for replay should begin the replay from the minimum across the maximum in each store. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Created] (HBASE-4657) Improve the efficiency of our MR jobs with a few configurations
Improve the efficiency of our MR jobs with a few configurations --- Key: HBASE-4657 URL: https://issues.apache.org/jira/browse/HBASE-4657 Project: HBase Issue Type: Improvement Affects Versions: 0.90.4 Reporter: Jean-Daniel Cryans Fix For: 0.92.0 This is a low hanging fruit, some of our MR jobs like RowCounter and CopyTable don't even setCacheBlocks on the scan object which out of the box completely screws up a running system. Another thing would be to disable speculative execution. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HBASE-4591) TTL for old HLogs should be calculated from last modification time.
[ https://issues.apache.org/jira/browse/HBASE-4591?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13134653#comment-13134653 ] Hudson commented on HBASE-4591: --- Integrated in HBase-TRUNK #2365 (See [https://builds.apache.org/job/HBase-TRUNK/2365/]) HBASE-4591 (2) Fix broken unit test HBASE-4591 TTL for old HLogs should be calculated from last modification time. nspiegelberg : Files : * /hbase/trunk/src/test/java/org/apache/hadoop/hbase/master/TestLogsCleaner.java nspiegelberg : Files : * /hbase/trunk/CHANGES.txt * /hbase/trunk/src/main/java/org/apache/hadoop/hbase/master/TimeToLiveLogCleaner.java TTL for old HLogs should be calculated from last modification time. --- Key: HBASE-4591 URL: https://issues.apache.org/jira/browse/HBASE-4591 Project: HBase Issue Type: Improvement Components: master Affects Versions: 0.89.20100621 Reporter: Madhuwanti Vaidya Assignee: Madhuwanti Vaidya Priority: Minor Fix For: 0.92.0, 0.94.0 Attachments: 0007-HBASE-4591-Increase-TTL-for-old-HLogs-and-change-TTL.patch, HBASE-4591-2.patch -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HBASE-4437) Update hadoop in 0.92 (0.20.205?)
[ https://issues.apache.org/jira/browse/HBASE-4437?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13134654#comment-13134654 ] Hudson commented on HBASE-4437: --- Integrated in HBase-TRUNK #2365 (See [https://builds.apache.org/job/HBase-TRUNK/2365/]) HBASE-4437 Update hadoop in 0.92 (0.20.205?) stack : Files : * /hbase/trunk/CHANGES.txt * /hbase/trunk/pom.xml * /hbase/trunk/src/main/jamon/org/apache/hbase/tmpl/master/MasterStatusTmpl.jamon * /hbase/trunk/src/main/jamon/org/apache/hbase/tmpl/regionserver/RSStatusTmpl.jamon * /hbase/trunk/src/main/resources/hbase-webapps/master/table.jsp * /hbase/trunk/src/main/resources/hbase-webapps/master/tablesDetailed.jsp * /hbase/trunk/src/main/resources/hbase-webapps/master/zk.jsp Update hadoop in 0.92 (0.20.205?) - Key: HBASE-4437 URL: https://issues.apache.org/jira/browse/HBASE-4437 Project: HBase Issue Type: Task Reporter: stack Assignee: stack Fix For: 0.92.0 Attachments: 4437.txt We ship with branch-0.20-append a few versions back from the tip. If 205 comes out and hbase works on it, we should ship 0.92 with it (while also ensuring it work w/ 0.22 and 0.23 branches). -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Created] (HBASE-4658) Put attributes are not exposed via the ThriftServer
Put attributes are not exposed via the ThriftServer --- Key: HBASE-4658 URL: https://issues.apache.org/jira/browse/HBASE-4658 Project: HBase Issue Type: Bug Components: thrift Reporter: dhruba borthakur Assignee: dhruba borthakur The Put api also takes in a bunch of arbitrary attributes that an application can use to associate metadata with each put operation. This is not exposed via Thrift. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HBASE-4374) Up default regions size from 256M to 1G
[ https://issues.apache.org/jira/browse/HBASE-4374?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] stack updated HBASE-4374: - Attachment: 4374.txt Flush size of 128M and split size of 1G. Up default regions size from 256M to 1G --- Key: HBASE-4374 URL: https://issues.apache.org/jira/browse/HBASE-4374 Project: HBase Issue Type: Task Reporter: stack Priority: Blocker Fix For: 0.92.0 Attachments: 4374.txt HBASE-4365 has some discussion of why we default for a table should tend to fewer bigger regions. It doesn't look like this issue will be done for 0.92. For 0.92, lets up default region size from 256M to 1G and talk up pre-split on table creation in manual. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Resolved] (HBASE-4374) Up default regions size from 256M to 1G
[ https://issues.apache.org/jira/browse/HBASE-4374?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] stack resolved HBASE-4374. -- Resolution: Fixed Release Note: Upped default split size to 1G from 256M and the flush size from 64M to 128M. Up default regions size from 256M to 1G --- Key: HBASE-4374 URL: https://issues.apache.org/jira/browse/HBASE-4374 Project: HBase Issue Type: Task Reporter: stack Priority: Blocker Fix For: 0.92.0 Attachments: 4374.txt HBASE-4365 has some discussion of why we default for a table should tend to fewer bigger regions. It doesn't look like this issue will be done for 0.92. For 0.92, lets up default region size from 256M to 1G and talk up pre-split on table creation in manual. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HBASE-4374) Up default regions size from 256M to 1G
[ https://issues.apache.org/jira/browse/HBASE-4374?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13134729#comment-13134729 ] stack commented on HBASE-4374: -- Committed to branch and trunk. Up default regions size from 256M to 1G --- Key: HBASE-4374 URL: https://issues.apache.org/jira/browse/HBASE-4374 Project: HBase Issue Type: Task Reporter: stack Priority: Blocker Fix For: 0.92.0 Attachments: 4374.txt HBASE-4365 has some discussion of why we default for a table should tend to fewer bigger regions. It doesn't look like this issue will be done for 0.92. For 0.92, lets up default region size from 256M to 1G and talk up pre-split on table creation in manual. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Created] (HBASE-4659) SingleColumnValueFilter is not working correctly
SingleColumnValueFilter is not working correctly Key: HBASE-4659 URL: https://issues.apache.org/jira/browse/HBASE-4659 Project: HBase Issue Type: Bug Components: filters Affects Versions: 0.90.4 Reporter: guozhw scan.addColumn(Bytes.toBytes(Info), Bytes.toBytes(value)); scan.setMaxVersions(); SingleColumnValueFilter scvf = new SingleColumnValueFilter(Bytes.toBytes(Info), Bytes.toBytes(value), CompareOp.EQUAL, Bytes.toBytes(test)); scvf.setLatestVersionOnly(false); scan.setFilter(scvf); it can not filter previous versions -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Created] (HBASE-4660) Place to publish RegionServer information such as webuiport and coprocessors loaded
Place to publish RegionServer information such as webuiport and coprocessors loaded --- Key: HBASE-4660 URL: https://issues.apache.org/jira/browse/HBASE-4660 Project: HBase Issue Type: Bug Reporter: stack HBASE-4070 added loaded CoProcessors to HServerLoad which seems like wrong place to carry this info. We need a locus for static info of this type such as loaded CoProcessors and webuiport as well as stuff like how many cpus, RAM, etc: e.g. in regionserver znode or available on invocation of an HRegionServer method (master can ask HRegionServer when it needs it). -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HBASE-4070) [Coprocessors] Improve region server metrics to report loaded coprocessors to master
[ https://issues.apache.org/jira/browse/HBASE-4070?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13134734#comment-13134734 ] stack commented on HBASE-4070: -- Here was my last comment on the patch over on RB in response to Eugene's Hi Michael, As Ted mentioned, HServerInfo is deprecated (per https://issues.apache.org/jira/browse/HBASE-1502): {code} Michael Stack 5 days, 4 hours ago (October 19th, 2011, 11:34 p.m.) Yes but that is not reason to shoe horn it in elsewhere (smile). There is another server attribute that is currently without a home (webuiport) and I think there are other a bunch of server attrbutes -- as opposed to server load -- that servers might export and have aggregated over in the master and/or used calculating cluster balance and the like. We could let this patch go in as is or we could build the server attributes mechanism -- ServerInfoV2 or json in regionserver znode, etc. -- and put stuff like this in there. {code} I think it a little perverse passing loaded coprocessors as part of HServerLoad. I opened HBASE-4660 to address it. [Coprocessors] Improve region server metrics to report loaded coprocessors to master Key: HBASE-4070 URL: https://issues.apache.org/jira/browse/HBASE-4070 Project: HBase Issue Type: Improvement Affects Versions: 0.90.3 Reporter: Mingjie Lai Assignee: Eugene Koontz Fix For: 0.92.0, 0.94.0 Attachments: HBASE-4070.patch, HBASE-4070.patch, HBASE-4070.patch, HBASE-4070.patch, master-web-ui.jpg, rs-status-web-ui.jpg HBASE-3512 is about listing loaded cp classes at shell. To make it more generic, we need a way to report this piece of information from region to master (or just at region server level). So later on, we can display the loaded class names at shell as well as web console. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Updated] (HBASE-2312) Possible data loss when RS goes into GC pause while rolling HLog
[ https://issues.apache.org/jira/browse/HBASE-2312?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] stack updated HBASE-2312: - Fix Version/s: (was: 0.94.0) 0.92.0 Yes Nicolas. Pulling it back in. Let me try and apply latest version of the patch. Possible data loss when RS goes into GC pause while rolling HLog Key: HBASE-2312 URL: https://issues.apache.org/jira/browse/HBASE-2312 Project: HBase Issue Type: Bug Components: master, regionserver Affects Versions: 0.90.0 Reporter: Karthik Ranganathan Assignee: Nicolas Spiegelberg Priority: Critical Fix For: 0.92.0 There is a very corner case when bad things could happen(ie data loss): 1)RS #1 is going to roll its HLog - not yet created the new one, old one will get no more writes 2)RS #1 enters GC Pause of Death 3)Master lists HLog files of RS#1 that is has to split as RS#1 is dead, starts splitting 4)RS #1 wakes up, created the new HLog (previous one was rolled) and appends an edit - which is lost The following seems like a possible solution: 1)Master detects RS#1 is dead 2)The master renames the /hbase/.logs/regionserver name directory to something else (say /hbase/.logs/regionserver name-dead) 3)Add mkdir support (as opposed to mkdirs) to HDFS - so that a file create fails if the directory doesn't exist. Dhruba tells me this is very doable. 4)RS#1 comes back up and is not able create the new hlog. It restarts itself. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Commented] (HBASE-2312) Possible data loss when RS goes into GC pause while rolling HLog
[ https://issues.apache.org/jira/browse/HBASE-2312?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13134740#comment-13134740 ] Nicolas Spiegelberg commented on HBASE-2312: tried to apply this onto our internal 92 branch and noticed that the Distributed Log Splitter broke this patch. Will work and resubmit. I think this is critical to get into 92 because of data loss issues, however. Possible data loss when RS goes into GC pause while rolling HLog Key: HBASE-2312 URL: https://issues.apache.org/jira/browse/HBASE-2312 Project: HBase Issue Type: Bug Components: master, regionserver Affects Versions: 0.90.0 Reporter: Karthik Ranganathan Assignee: Nicolas Spiegelberg Priority: Critical Fix For: 0.92.0 There is a very corner case when bad things could happen(ie data loss): 1)RS #1 is going to roll its HLog - not yet created the new one, old one will get no more writes 2)RS #1 enters GC Pause of Death 3)Master lists HLog files of RS#1 that is has to split as RS#1 is dead, starts splitting 4)RS #1 wakes up, created the new HLog (previous one was rolled) and appends an edit - which is lost The following seems like a possible solution: 1)Master detects RS#1 is dead 2)The master renames the /hbase/.logs/regionserver name directory to something else (say /hbase/.logs/regionserver name-dead) 3)Add mkdir support (as opposed to mkdirs) to HDFS - so that a file create fails if the directory doesn't exist. Dhruba tells me this is very doable. 4)RS#1 comes back up and is not able create the new hlog. It restarts itself. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Created] (HBASE-4661) Ability to export the list of files for a some or all column families for a given region
Ability to export the list of files for a some or all column families for a given region Key: HBASE-4661 URL: https://issues.apache.org/jira/browse/HBASE-4661 Project: HBase Issue Type: Sub-task Reporter: Karthik Ranganathan Assignee: Karthik Ranganathan Should be able to query the regionservers to figure out the list of files in one/some/all column families for a given regions to determine which files to copy for a backup. -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira
[jira] [Created] (HBASE-4662) Replay the required hlog edits to make the backup preserve row atomicity.
Replay the required hlog edits to make the backup preserve row atomicity. - Key: HBASE-4662 URL: https://issues.apache.org/jira/browse/HBASE-4662 Project: HBase Issue Type: Sub-task Reporter: Karthik Ranganathan Assignee: Karthik Ranganathan The algorithm is as follows: A. For HFiles: 1. Need to track t1,t2 for each backup (start and end times of the backup) 2. For point in time restore to time t, pick a HFile snapshot which has t2 t 3. Copy HFile snapshot to a temp location - HTABLE_RESTORE_t B. For HLogs: for each regionserver do for .logs and .oldlogs do 1. log file is hlog.TIME 2. if (t TIME and hlog.TIME is open for write) fail restore for t 3. Pick the latest HLog whose create time is t1 4. Pick all HLogs whose create time is t1 and = t2 5. Copy hlogs to the right structures inside HTABLE_RESTORE_t C. Split logs 1. Enhance HLog.splitLog to take timestamp t 2. Enhance distributed log split tool to pass HTABLE_RESTORE_t, so that log split is picked up and put in the right location 3. Enhance distributed log split tool to pass t so that all edits till t are included and others ignored D. Import the directory into the running HBase with META entries, etc (this already exists) -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira