[ 
https://issues.apache.org/jira/browse/RANGER-1501?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15961462#comment-15961462
 ] 

Ramesh Mani commented on RANGER-1501:
-------------------------------------

[~ yzhou2001] This patch might need a good amount of testing especially KAFKA 
and HDFS ranger plugin where  there would be a continues logging happening and 
this doesn't get blocked in synchronized call to hflush() and when number of 
data nodes are more. 
Can we  come up with testing scenarios which can be done for this?
To address the loss of audit logs we already did an enhancement via 
https://issues.apache.org/jira/browse/RANGER-1310

> Audit Flush to HDFS does not actually cause the audit logs to be flushed to 
> HDFS 
> ---------------------------------------------------------------------------------
>
>                 Key: RANGER-1501
>                 URL: https://issues.apache.org/jira/browse/RANGER-1501
>             Project: Ranger
>          Issue Type: Bug
>          Components: audit
>    Affects Versions: 0.7.0
>            Reporter: Yan
>            Assignee: Yan
>             Fix For: master
>
>         Attachments: 
> 0001-RANGER-1501-Audit-Flush-to-HDFS-does-not-actually-ca.patch
>
>
> The reason is that HDFS file stream's flush() call does not really flush the 
> data all the way to disk, nor even makes the data visible to HDFS users. See 
> the HDFS semantics of the flush/sync at 
> https://issues.apache.org/jira/browse/HADOOP-6313.
> Consequently the audit logs on HDFS won't be visible/durable from HDFS client 
> until the log file is closed. This will, among other issues, boost chances of 
> losing audit logs in case of system failure.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

Reply via email to