You can also refer : 
http://eagle.apache.org/docs/import-hdfs-auditLog.html 
<http://eagle.apache.org/docs/import-hdfs-auditLog.html>

Try having the log stash conf filter and output similar to this:


filter {
 grok {
   match => ["message", 
"ugi=(?<user>([\w\d\-]+))@|ugi=(?<user>([\w\d\-]+))/[\w\d\-.]+@|ugi=(?<user>([\w\d.\-_]+))[\s(]+"]
  }

}

output {
   stdout {
     codec => rubydebug
   }
   kafka {
     codec => plain {
                 format => "%{message}"
             }
   bootstrap_servers => "127.0.0.1:6667"
   topic_id => "TestTopic"
   retry_backoff_ms => 100
   send_buffer_bytes => 102400



 }
 }


-Sudha Jenslin



> On Mar 21, 2017, at 7:53 PM, Jean Rossier <[email protected]> wrote:
> 
> Hello,
> 
> I'm trying to setup the HDFS data activity monitoring stream (as explained 
> here http://eagle.apache.org/docs/hdfs-data-activity-monitoring.html 
> <http://eagle.apache.org/docs/hdfs-data-activity-monitoring.html>), but I 
> don't see any logs flowing into my Kafka topic.
> 
> Is there anything to do that is not on the quick start page (mentioned above) 
> ?
> 
> One thing that surprised me, is that I don't have any log4jkafka folder in 
> /usr/hdp/current/eagle/lib/ (I'm using eagle 0.5.0-SNAPSHOT, from current 
> master branch). I'm supposed to add the libs from this folder to the 
> HADOOP_CLASSPATH env variable. I moved them by hand in the proper folder, 
> but, still, I don't see any logs flowing into my Kafka topic.
> 
> I don't see any error regarding kafka in the HDFS audit logs. There are, 
> though, some solr exceptions but I guess it shouldn't impact the stream I'm 
> interested in.
> 
> Thanks
> Jean
> 

Reply via email to