[ 
https://issues.apache.org/jira/browse/HBASE-19289?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16261449#comment-16261449
 ] 

Steve Loughran commented on HBASE-19289:
----------------------------------------

If people really want hbase -> file://

* they'd need a distributed file:// or some shared NFS server
* it'd presumably need its own RAID > 0 to do checksumming; so checksum fs is 
moot

I'd look at seeing whether checksumfs could actually bypass its checksum, say 
if we set the property "bytes per checksum == 0" as the secret no, turn me off" 
switch. But people would probably then use it for performance and then be upset 
when all their data got corrupted without anything noticing. It's too critical 
a layer under HDFS really.

I was thinking about what if we added a raw:// URL which bonded directly to raw 
local fs, but RawLocalFileSystem has an expectation that file:// is its schema 
and returns it in getURI(), so forcing you back to CheckummedFS

I believe the way to do this is
* subclass RawLocalFileSystem
* give it a new schema, like say "raw"
* have it remember its URI in initialize() and return it in getURI()
* register it (statically, dynamically)


> CommonFSUtils$StreamLacksCapabilityException: hflush when running test 
> against hadoop3 beta1
> --------------------------------------------------------------------------------------------
>
>                 Key: HBASE-19289
>                 URL: https://issues.apache.org/jira/browse/HBASE-19289
>             Project: HBase
>          Issue Type: Test
>            Reporter: Ted Yu
>         Attachments: 19289.v1.txt, 19289.v2.txt
>
>
> As of commit d8fb10c8329b19223c91d3cda6ef149382ad4ea0 , I encountered the 
> following exception when running unit test against hadoop3 beta1:
> {code}
> testRefreshStoreFiles(org.apache.hadoop.hbase.regionserver.TestHStore)  Time 
> elapsed: 0.061 sec  <<< ERROR!
> java.io.IOException: cannot get log writer
>       at 
> org.apache.hadoop.hbase.regionserver.TestHStore.initHRegion(TestHStore.java:215)
>       at 
> org.apache.hadoop.hbase.regionserver.TestHStore.init(TestHStore.java:220)
>       at 
> org.apache.hadoop.hbase.regionserver.TestHStore.init(TestHStore.java:195)
>       at 
> org.apache.hadoop.hbase.regionserver.TestHStore.init(TestHStore.java:190)
>       at 
> org.apache.hadoop.hbase.regionserver.TestHStore.init(TestHStore.java:185)
>       at 
> org.apache.hadoop.hbase.regionserver.TestHStore.init(TestHStore.java:179)
>       at 
> org.apache.hadoop.hbase.regionserver.TestHStore.init(TestHStore.java:173)
>       at 
> org.apache.hadoop.hbase.regionserver.TestHStore.testRefreshStoreFiles(TestHStore.java:962)
> Caused by: 
> org.apache.hadoop.hbase.util.CommonFSUtils$StreamLacksCapabilityException: 
> hflush
>       at 
> org.apache.hadoop.hbase.regionserver.TestHStore.initHRegion(TestHStore.java:215)
>       at 
> org.apache.hadoop.hbase.regionserver.TestHStore.init(TestHStore.java:220)
>       at 
> org.apache.hadoop.hbase.regionserver.TestHStore.init(TestHStore.java:195)
>       at 
> org.apache.hadoop.hbase.regionserver.TestHStore.init(TestHStore.java:190)
>       at 
> org.apache.hadoop.hbase.regionserver.TestHStore.init(TestHStore.java:185)
>       at 
> org.apache.hadoop.hbase.regionserver.TestHStore.init(TestHStore.java:179)
>       at 
> org.apache.hadoop.hbase.regionserver.TestHStore.init(TestHStore.java:173)
>       at 
> org.apache.hadoop.hbase.regionserver.TestHStore.testRefreshStoreFiles(TestHStore.java:962)
> {code}



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

Reply via email to