[ 
https://issues.apache.org/jira/browse/HADOOP-3348?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12594973#action_12594973
 ] 

Raghu Angadi commented on HADOOP-3348:
--------------------------------------

Two options :
1. FSInputChecker.read() should not do readFully(). This is harder to change 
since multiple users implicitly depend on current behaviour. But in long run it 
should change.
2. ChecksumFileSystem should use the base filesystem (RawLocalFileSystem in 
this case) directly when there is no .crc file. 

I think #2 is simpler to do and will reduce strange surprises. It is required 
anyway since even after #1, it will still try to read 512 bytes. 

> TestUrlStreamHandler hangs on LINUX
> -----------------------------------
>
>                 Key: HADOOP-3348
>                 URL: https://issues.apache.org/jira/browse/HADOOP-3348
>             Project: Hadoop Core
>          Issue Type: Bug
>          Components: fs
>    Affects Versions: 0.18.0
>         Environment: LINUX 2.6.9
>            Reporter: lohit vijayarenu
>         Attachments: Datanode_jstack.txt
>
>
> TestUrlStreamHandler sets setURLStreamHandlerFactory as
> {noformat}
> FsUrlStreamHandlerFactory factory =
>         new org.apache.hadoop.fs.FsUrlStreamHandlerFactory();
>     java.net.URL.setURLStreamHandlerFactory(factory);
> {noformat}
> After this, MiniDFSCluster seems to hang while Datanodes tries to register in 
> setNewStorageID, specifically at
> {noformat}
> rand = SecureRandom.getInstance("SHA1PRNG").nextInt(Integer.MAX_VALUE);
> {noformat}
> jstack output shows that the main thread is stuck in 
> RawLocalFileSystem$LocalFSFileInputStream.read
> (Attaching the jstack)

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.

Reply via email to