[ https://issues.apache.org/jira/browse/HBASE-3581?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13120689#comment-13120689 ]
stack commented on HBASE-3581: ------------------------------ Tried it on 0.92 branch head and got these failures. Look into it: {code} Results : Failed tests: testLogRollOnDatanodeDeath(org.apache.hadoop.hbase.regionserver.wal.TestLogRolling): New log file should have the default replication Tests in error: testMultiSlaveReplication(org.apache.hadoop.hbase.replication.TestMultiSlaveReplication): test timed out after 300000 milliseconds testCyclicReplication(org.apache.hadoop.hbase.replication.TestMasterReplication): test timed out after 300000 milliseconds testSimplePutDelete(org.apache.hadoop.hbase.replication.TestMasterReplication): Cluster already running at /home/stack/0.92/target/test-data/18b3c603-4c68-434f-8701-7cfcf558129b testExceptionFromCoprocessorWhenCreatingTable(org.apache.hadoop.hbase.coprocessor.TestMasterCoprocessorExceptionWithRemove): test timed out after 30000 milliseconds testClusterRestart(org.apache.hadoop.hbase.master.TestRestartCluster): test timed out after 300000 milliseconds testBasicRollingRestart(org.apache.hadoop.hbase.master.TestRollingRestart): test timed out after 300000 milliseconds testUsingMetaAndBinary(org.apache.hadoop.hbase.regionserver.TestGetClosestAtOrBefore): Cannot lock storage /home/stack/0.92/build/hbase/test/dfs/name1. The directory is already locked. testGetClosestRowBefore3(org.apache.hadoop.hbase.regionserver.TestGetClosestAtOrBefore): Cannot lock storage /home/stack/0.92/build/hbase/test/dfs/name1. The directory is already locked. testGetClosestRowBefore2(org.apache.hadoop.hbase.regionserver.TestGetClosestAtOrBefore): Cannot lock storage /home/stack/0.92/build/hbase/test/dfs/name1. The directory is already locked. testWideScanBatching(org.apache.hadoop.hbase.regionserver.TestWideScanner): Cannot lock storage /home/stack/0.92/build/hbase/test/dfs/name1. The directory is already locked. testBasicHalfMapFile(org.apache.hadoop.hbase.regionserver.TestStoreFile): Call to localhost/127.0.0.1:0 failed on connection exception: java.net.ConnectException: Connection refused testReference(org.apache.hadoop.hbase.regionserver.TestStoreFile): Call to localhost/127.0.0.1:0 failed on connection exception: java.net.ConnectException: Connection refused testBloomFilter(org.apache.hadoop.hbase.regionserver.TestStoreFile): Call to localhost/127.0.0.1:0 failed on connection exception: java.net.ConnectException: Connection refused testBloomTypes(org.apache.hadoop.hbase.regionserver.TestStoreFile): Call to localhost/127.0.0.1:0 failed on connection exception: java.net.ConnectException: Connection refused testBloomEdgeCases(org.apache.hadoop.hbase.regionserver.TestStoreFile): Call to localhost/127.0.0.1:0 failed on connection exception: java.net.ConnectException: Connection refused testFlushTimeComparator(org.apache.hadoop.hbase.regionserver.TestStoreFile): Call to localhost/127.0.0.1:0 failed on connection exception: java.net.ConnectException: Connection refused testMROnTableWithCustomMapper(org.apache.hadoop.hbase.mapreduce.TestImportTsv): java.io.IOException: File /tmp/hadoop-stack/mapred/system/job_local_0002/libjars/zookeeper-3.3.3.jar could only be replicated to 0 nodes, instead of 1 testSimpleLoad(org.apache.hadoop.hbase.mapreduce.TestLoadIncrementalHFiles): No such file or directory testReconstruction(org.apache.hadoop.hbase.TestFullLogReconstruction): test timed out after 300000 milliseconds testFlushCommitsWithAbort(org.apache.hadoop.hbase.client.TestMultiParallel): test timed out after 300000 milliseconds testFlushCommitsNoAbort(org.apache.hadoop.hbase.client.TestMultiParallel): test timed out after 300000 milliseconds test2481(org.apache.hadoop.hbase.client.TestScannerTimeout): test timed out after 300000 milliseconds test2772(org.apache.hadoop.hbase.client.TestScannerTimeout): test timed out after 300000 milliseconds Tests run: 897, Failures: 1, Errors: 20, Skipped: 18 {code} > hbase rpc should send size of response > -------------------------------------- > > Key: HBASE-3581 > URL: https://issues.apache.org/jira/browse/HBASE-3581 > Project: HBase > Issue Type: Improvement > Reporter: ryan rawson > Assignee: stack > Priority: Critical > Fix For: 0.92.0 > > Attachments: 3581-v2.txt, 3581-v3.txt, 3581-v4.txt, > HBASE-rpc-response.txt > > > The RPC reply from Server->Client does not include the size of the payload, > it is framed like so: > <i32> callId > <byte> errorFlag > <byte[]> data > The data segment would contain enough info about how big the response is so > that it could be decoded by a writable reader. > This makes it difficult to write buffering clients, who might read the entire > 'data' then pass it to a decoder. While less memory efficient, if you want to > easily write block read clients (eg: nio) it would be necessary to send the > size along so that the client could snarf into a local buf. > The new proposal is: > <i32> callId > <i32> size > <byte> errorFlag > <byte[]> data > the size being sizeof(data) + sizeof(errorFlag). -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa For more information on JIRA, see: http://www.atlassian.com/software/jira