i still see backtype.storm packages in ur stacktrace. u need to clean and package or install whole of eagle, also you ll have to stop, delete the application and re-deploy it.
- Jayesh On Sat, Mar 18, 2017 at 2:57 PM, <[email protected]> wrote: > Hi Jayesh, > > > > I’ve done pull and rebuild Eagle 0.5. > > Config problem has gone, but I’ve still failed to start any application. > > > > 1. Kafka Topic for Auditlog Event Sink is filled with > hdfs_audit_event_${site} > – last part is not resolving to my site_id. > 2. I’ve created two kafka topics: hdfs_audit_log_{SITE_ID}, > hdfs_audit_log_enriched_{SITE_ID} and not hdfs_audit_event_{site_id} – > is this ok? > 3. Launching HDFS Audit logMonitor in cluster mode has no success and > no errors, Here is some log: > > > > INFO [2017-03-18 21:29:17,042] > org.apache.eagle.app.environment.impl.StormExecutionRuntime: > Starting HDFS_AUDIT_LOG_MONITOR_APP_DEMO(org.apache.eagle. > security.auditlog.HdfsAuditLogApplication), mode: CL > > USTER > > INFO [2017-03-18 21:29:17,050] > org.apache.eagle.app.environment.impl.StormExecutionRuntime: > Overriding application.storm.nimbusHost = demo5 > > INFO [2017-03-18 21:29:17,050] > org.apache.eagle.app.environment.impl.StormExecutionRuntime: > Overriding application.storm.nimbusThriftPort = 6627 > > INFO [2017-03-18 21:29:17,050] > org.apache.eagle.app.environment.impl.StormExecutionRuntime: > Submitting as cluster mode ... > > INFO [2017-03-18 21:29:21,415] > org.apache.eagle.storage.jdbc.entity.impl.JdbcEntityReaderImpl: > Read 0 records in 3 ms (sql: SELECT alert_detail_alert_detail.* FROM > alert_detail_alert_detail WHERE (alert_det > > ail_alert_detail.timestamp>=? AND alert_detail_alert_detail.timestamp<?) > LIMIT 10000 Replacements: [1489829363000,1489872563000]) > > INFO [2017-03-18 21:29:28,231] org.apache.eagle.app.service.impl. > ApplicationStatusUpdateServiceImpl: Updating application status > > > > 1. Nothing in Storm UI and logs. > 2. Launched in Local mode: > > > > INFO [2017-03-18 21:31:35,906] backtype.storm.daemon.executor: Prepared > bolt kafkaSink:(7) > > INFO [2017-03-18 21:31:35,912] storm.kafka.PartitionManager: Read > partition information from: > /consumers/hdfs_audit_log_demo/eagleConsumer/partition_0 > --> null > > INFO [2017-03-18 21:31:35,917] backtype.storm.daemon.executor: Prepared > bolt kafkaSink:(6) > > ERROR [2017-03-18 21:31:35,983] backtype.storm.util: Async loop died! > > ! java.lang.NullPointerException: null > > ! at org.apache.kafka.common.utils.Utils.formatAddress(Utils.java:312) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at > kafka.consumer.SimpleConsumer$$anonfun$disconnect$1.apply(SimpleConsumer.scala:49) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at > kafka.consumer.SimpleConsumer$$anonfun$disconnect$1.apply(SimpleConsumer.scala:49) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at kafka.utils.Logging$class.debug(Logging.scala:52) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at kafka.consumer.SimpleConsumer.debug(SimpleConsumer.scala:30) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at kafka.consumer.SimpleConsumer.disconnect(SimpleConsumer.scala:49) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at kafka.consumer.SimpleConsumer.liftedTree1$1(SimpleConsumer.scala:82) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at kafka.consumer.SimpleConsumer.kafka$consumer$SimpleConsumer$ > $sendRequest(SimpleConsumer.scala:68) ~[eagle-topology-0.5.0- > SNAPSHOT-assembly.jar:na] > > ! at kafka.consumer.SimpleConsumer.getOffsetsBefore(SimpleConsumer.scala:127) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at > kafka.javaapi.consumer.SimpleConsumer.getOffsetsBefore(SimpleConsumer.scala:79) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at storm.kafka.KafkaUtils.getOffset(KafkaUtils.java:74) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at storm.kafka.KafkaUtils.getOffset(KafkaUtils.java:64) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at storm.kafka.PartitionManager.<init>(PartitionManager.java:89) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at storm.kafka.ZkCoordinator.refresh(ZkCoordinator.java:98) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! ... 6 common frames omitted > > ! Causing: java.lang.RuntimeException: java.lang.NullPointerException > > ! at storm.kafka.ZkCoordinator.refresh(ZkCoordinator.java:103) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at storm.kafka.ZkCoordinator.getMyManagedPartitions(ZkCoordinator.java:69) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at storm.kafka.KafkaSpout.nextTuple(KafkaSpout.java:135) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at > backtype.storm.daemon.executor$fn__3373$fn__3388$fn__3417.invoke(executor.clj:565) > ~[storm-core-0.9.3.jar:0.9.3] > > ! at backtype.storm.util$async_loop$fn__464.invoke(util.clj:463) > ~[storm-core-0.9.3.jar:0.9.3] > > ! at clojure.lang.AFn.run(AFn.java:24) [eagle-topology-0.5.0- > SNAPSHOT-assembly.jar:na] > > ! at java.lang.Thread.run(Thread.java:745) [na:1.8.0_101] > > ERROR [2017-03-18 21:31:35,984] backtype.storm.daemon.executor: > > ! java.lang.NullPointerException: null > > ! at org.apache.kafka.common.utils.Utils.formatAddress(Utils.java:312) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at > kafka.consumer.SimpleConsumer$$anonfun$disconnect$1.apply(SimpleConsumer.scala:49) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at > kafka.consumer.SimpleConsumer$$anonfun$disconnect$1.apply(SimpleConsumer.scala:49) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at kafka.utils.Logging$class.debug(Logging.scala:52) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at kafka.consumer.SimpleConsumer.debug(SimpleConsumer.scala:30) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at kafka.consumer.SimpleConsumer.disconnect(SimpleConsumer.scala:49) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at kafka.consumer.SimpleConsumer.liftedTree1$1(SimpleConsumer.scala:82) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at kafka.consumer.SimpleConsumer.kafka$consumer$SimpleConsumer$ > $sendRequest(SimpleConsumer.scala:68) ~[eagle-topology-0.5.0- > SNAPSHOT-assembly.jar:na] > > ! at kafka.consumer.SimpleConsumer.getOffsetsBefore(SimpleConsumer.scala:127) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at > kafka.javaapi.consumer.SimpleConsumer.getOffsetsBefore(SimpleConsumer.scala:79) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at storm.kafka.KafkaUtils.getOffset(KafkaUtils.java:74) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at storm.kafka.KafkaUtils.getOffset(KafkaUtils.java:64) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at storm.kafka.PartitionManager.<init>(PartitionManager.java:89) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at storm.kafka.ZkCoordinator.refresh(ZkCoordinator.java:98) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! ... 6 common frames omitted > > ! Causing: java.lang.RuntimeException: java.lang.NullPointerException > > ! at storm.kafka.ZkCoordinator.refresh(ZkCoordinator.java:103) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > > > > > Also, first launched with Hbase as deep storage and got this: > > > > > > INFO [2017-03-18 21:16:31,744] org.apache.zookeeper.ClientCnxn: Session > establishment complete on server localhost/127.0.0.1:2181, sessionid = > 0x35ad9214d4700c0, negotiated timeout = 40000 > > WARN [2017-03-18 21:16:31,942] > org.apache.hadoop.hbase.util.DynamicClassLoader: > Failed to identify the fs of dir /tmp/hbase-root/hbase/lib, ignored > > ! java.io.IOException: No FileSystem for scheme: file > > ! at org.apache.hadoop.fs.FileSystem.getFileSystemClass(FileSystem.java:2607) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2614) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:91) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2653) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2635) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:370) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:169) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:354) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at org.apache.hadoop.fs.Path.getFileSystem(Path.java:296) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at > org.apache.hadoop.hbase.util.DynamicClassLoader.<init>(DynamicClassLoader.java:104) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at > org.apache.hadoop.hbase.protobuf.ProtobufUtil.<clinit>(ProtobufUtil.java:223) > [eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at org.apache.hadoop.hbase.ClusterId.parseFrom(ClusterId.java:64) > [eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at > org.apache.hadoop.hbase.zookeeper.ZKClusterId.readClusterIdZNode(ZKClusterId.java:75) > [eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at > org.apache.hadoop.hbase.client.ZooKeeperRegistry.getClusterId(ZooKeeperRegistry.java:106) > [eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at org.apache.hadoop.hbase.client.ConnectionManager$ > HConnectionImplementation.retrieveClusterId(ConnectionManager.java:890) > [eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at org.apache.hadoop.hbase.client.ConnectionManager$ > HConnectionImplementation.<init>(ConnectionManager.java:667) > [eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native > Method) [na:1.8.0_101] > > ! at sun.reflect.NativeConstructorAccessorImpl.newInstance( > NativeConstructorAccessorImpl.java:62) [na:1.8.0_101] > > ! at sun.reflect.DelegatingConstructorAccessorImpl.newInstance( > DelegatingConstructorAccessorImpl.java:45) [na:1.8.0_101] > > ! at java.lang.reflect.Constructor.newInstance(Constructor.java:423) > [na:1.8.0_101] > > ! at > org.apache.hadoop.hbase.client.ConnectionManager.createConnection(ConnectionManager.java:426) > [eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at > org.apache.hadoop.hbase.client.ConnectionManager.createConnection(ConnectionManager.java:405) > [eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at org.apache.hadoop.hbase.client.ConnectionManager. > getConnectionInternal(ConnectionManager.java:283) [eagle-topology-0.5.0- > SNAPSHOT-assembly.jar:na] > > ! at org.apache.hadoop.hbase.client.HBaseAdmin.<init>(HBaseAdmin.java:191) > [eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at org.apache.eagle.storage.hbase.HBaseEntitySchemaManager.init( > HBaseEntitySchemaManager.java:66) [eagle-storage-hbase-0.5.0- > SNAPSHOT.jar:0.5.0-SNAPSHOT] > > ! at org.apache.eagle.storage.hbase.HBaseStorage.init(HBaseStorage.java:54) > [eagle-storage-hbase-0.5.0-SNAPSHOT.jar:0.5.0-SNAPSHOT] > > ! at > org.apache.eagle.storage.DataStorageManager.newDataStorage(DataStorageManager.java:53) > [eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at org.apache.eagle.storage.DataStorageManager. > getDataStorageByEagleConfig(DataStorageManager.java:81) > [eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at org.apache.eagle.storage.DataStorageManager. > getDataStorageByEagleConfig(DataStorageManager.java:91) > [eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at org.apache.eagle.service.generic.GenericEntityServiceResource.search( > GenericEntityServiceResource.java:438) [eagle-topology-0.5.0- > SNAPSHOT-assembly.jar:na] > > ! at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > ~[na:1.8.0_101] > > > > And this on stop: > > > > WARN [2017-03-18 21:20:27,274] org.apache.hadoop.hbase.client.ClientScanner: > scanner failed to close. Exception follows: java.io.InterruptedIOException > > WARN [2017-03-18 21:20:27,275] org.apache.hadoop.hbase.client.ClientScanner: > scanner failed to close. Exception follows: java.io.InterruptedIOException > > ERROR [2017-03-18 21:20:27,278] > org.apache.eagle.log.entity.GenericEntityScanStreamReader: > Fail reading log > > ! java.nio.channels.ClosedByInterruptException: null > > ! at java.nio.channels.spi.AbstractInterruptibleChannel.end( > AbstractInterruptibleChannel.java:202) ~[na:1.8.0_101] > > ! at sun.nio.ch.SocketChannelImpl.connect(SocketChannelImpl.java:659) > ~[na:1.8.0_101] > > ! at > org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:192) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:530) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:494) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at org.apache.hadoop.hbase.ipc.RpcClient$Connection. > setupConnection(RpcClient.java:612) ~[eagle-topology-0.5.0- > SNAPSHOT-assembly.jar:na] > > ! at org.apache.hadoop.hbase.ipc.RpcClient$Connection. > setupIOstreams(RpcClient.java:920) ~[eagle-topology-0.5.0- > SNAPSHOT-assembly.jar:na] > > ! ... 85 common frames omitted > > ! Causing: java.io.InterruptedIOException: Origin: > ClosedByInterruptException > > ! at > org.apache.hadoop.hbase.util.ExceptionUtil.asInterrupt(ExceptionUtil.java:62) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at org.apache.hadoop.hbase.ipc.RpcClient$Connection. > setupIOstreams(RpcClient.java:974) ~[eagle-topology-0.5.0- > SNAPSHOT-assembly.jar:na] > > ! at org.apache.hadoop.hbase.ipc.RpcClient$Connection. > writeRequest(RpcClient.java:1094) ~[eagle-topology-0.5.0- > SNAPSHOT-assembly.jar:na] > > ! at org.apache.hadoop.hbase.ipc.RpcClient$Connection. > tracedWriteRequest(RpcClient.java:1061) ~[eagle-topology-0.5.0- > SNAPSHOT-assembly.jar:na] > > ! at org.apache.hadoop.hbase.ipc.RpcClient.call(RpcClient.java:1516) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at > org.apache.hadoop.hbase.ipc.RpcClient.callBlockingMethod(RpcClient.java:1724) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at org.apache.hadoop.hbase.ipc.RpcClient$BlockingRpcChannelImplementati > on.callBlockingMethod(RpcClient.java:1777) ~[eagle-topology-0.5.0- > SNAPSHOT-assembly.jar:na] > > ! at org.apache.hadoop.hbase.protobuf.generated. > ClientProtos$ClientService$BlockingStub.get(ClientProtos.java:30373) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at org.apache.hadoop.hbase.protobuf.ProtobufUtil. > getRowOrBefore(ProtobufUtil.java:1604) ~[eagle-topology-0.5.0- > SNAPSHOT-assembly.jar:na] > > ! at org.apache.hadoop.hbase.protobuf.ProtobufUtil. > getRowOrBefore(ProtobufUtil.java:1584) ~[eagle-topology-0.5.0- > SNAPSHOT-assembly.jar:na] > > ! at org.apache.hadoop.hbase.client.ConnectionManager$ > HConnectionImplementation.locateRegionInMeta(ConnectionManager.java:1346) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at org.apache.hadoop.hbase.client.ConnectionManager$ > HConnectionImplementation.locateRegion(ConnectionManager.java:1167) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at org.apache.hadoop.hbase.client.RpcRetryingCallerWithReadRepli > cas.getRegionLocations(RpcRetryingCallerWithReadReplicas.java:294) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at org.apache.hadoop.hbase.client.ScannerCallableWithReplicas.call( > ScannerCallableWithReplicas.java:130) ~[eagle-topology-0.5.0- > SNAPSHOT-assembly.jar:na] > > ! at org.apache.hadoop.hbase.client.ScannerCallableWithReplicas.call( > ScannerCallableWithReplicas.java:55) ~[eagle-topology-0.5.0- > SNAPSHOT-assembly.jar:na] > > ! at > org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithoutRetries(RpcRetryingCaller.java:201) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at org.apache.hadoop.hbase.client.ClientScanner.call(ClientScanner.java:288) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at org.apache.hadoop.hbase.client.ClientScanner. > nextScanner(ClientScanner.java:268) ~[eagle-topology-0.5.0- > SNAPSHOT-assembly.jar:na] > > ! at org.apache.hadoop.hbase.client.ClientScanner. > initializeScannerInConstruction(ClientScanner.java:140) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at > org.apache.hadoop.hbase.client.ClientScanner.<init>(ClientScanner.java:135) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at org.apache.hadoop.hbase.client.HTable.getScanner(HTable.java:802) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at > org.apache.hadoop.hbase.client.HTablePool$PooledHTable.getScanner(HTablePool.java:416) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at > org.apache.eagle.log.entity.HBaseLogReader2.onOpen(HBaseLogReader2.java:56) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at org.apache.eagle.log.entity.AbstractHBaseLogReader.open( > AbstractHBaseLogReader.java:170) ~[eagle-topology-0.5.0- > SNAPSHOT-assembly.jar:na] > > ! at org.apache.eagle.log.entity.GenericEntityScanStreamReader. > readAsStream(GenericEntityScanStreamReader.java:96) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at org.apache.eagle.log.entity.GenericEntityStreamReader.readAsStream( > GenericEntityStreamReader.java:82) [eagle-topology-0.5.0- > SNAPSHOT-assembly.jar:na] > > ! at > org.apache.eagle.query.GenericEntityQuery.result(GenericEntityQuery.java:66) > [eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at org.apache.eagle.storage.hbase.HBaseStorage.query(HBaseStorage.java:171) > [eagle-storage-hbase-0.5.0-SNAPSHOT.jar:0.5.0-SNAPSHOT] > > ! at org.apache.eagle.storage.operation.QueryStatement. > execute(QueryStatement.java:47) [eagle-topology-0.5.0- > SNAPSHOT-assembly.jar:na] > > ! at org.apache.eagle.service.generic.GenericEntityServiceResource.search( > GenericEntityServiceResource.java:444) [eagle-topology-0.5.0- > SNAPSHOT-assembly.jar:na] > > ! at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > ~[na:1.8.0_101] > > > > > > Regards, > > Andrey > > > > *From: *Jayesh Senjaliya <[email protected]> > *Sent: *18 марта 2017 г. 20:49 > *To: *Jayesh Senjaliya <[email protected]>; [email protected] > > *Subject: *Re: Instructions on installing Eagle 0.5 on HDP 2.5 > > > > Do git pull master, the fix is just merged. Eagle-934 > > > > > > > > > > On Sat, Mar 18, 2017 at 2:56 AM <[email protected]> wrote: > > Here is an error in server .yaml > > Do I need to change something there? > > > > $ > > WARN [2017-03-18 12:53:48,828] > org.apache.eagle.app.module.ApplicationExtensionLoader: > Registering modules from HBaseMetricWebApplicationProvider[name=HBase > Metric Monitoring Web , type=HBASE_METRIC_WEB_AP$ > > WARN [2017-03-18 12:53:48,834] > org.apache.eagle.server.module.GuiceBundleLoader: > Loaded 2 modules (scope: metadataStore) > > WARN [2017-03-18 12:53:48,835] > org.apache.eagle.server.module.GuiceBundleLoader: > Loaded 0 modules (scope: global) > > bin/../conf/server.yml has an error: > > * Unrecognized field at: auth.authorization > > Did you mean?: > > - metrics > > - cachePolicy > > - caching > > - logging > > - enabled > > [4 more] > > > > > > Regards, > > Andrey > > *From: *Jayesh Senjaliya <[email protected]> > *Sent: *18 марта 2017 г. 12:32 > *To: *[email protected] > *Cc: *[email protected] > > > *Subject: *Re: Instructions on installing Eagle 0.5 on HDP 2.5 > > > > storage { > > # storage type: ["hbase","jdbc"] > > # default is "hbase" > > type = "jdbc" > > jdbc{ > > adapter="mysql" > > username="root" > > password=basic123 > > database=eagle2 > > connectionUrl="jdbc:mysql://localhost:3306/eagle2" > > connectionProps="encoding=UTF-8" > > driverClass="com.mysql.jdbc.Driver" > > connectionMax=8 > > } > > } > > > > > > On Sat, Mar 18, 2017 at 2:30 AM, <[email protected]> wrote: > > Ok, thank you! Will try it now. > > > > Could you please provide template for jdbc as DeepStorage? > > > > Regards, > > Andrey > > > > *From: *Jayesh Senjaliya <[email protected]> > *Sent: *18 марта 2017 г. 12:27 > *To: *[email protected] > *Subject: *Re: Instructions on installing Eagle 0.5 on HDP 2.5 > > > > i have rebased it now, and build has passed.... > > > > > > about the metadata, > > > > eagle stores all metadata ( all you can create on UI ) in mysql, but for > all the derived or final metrics, it uses hbase (as deep storage) because > it makes the query much faster. > > > > if you dont have big scale metrics, you can use mysql for both. > > > > - Jayesh > > > > > > On Sat, Mar 18, 2017 at 2:14 AM, Markovich <[email protected]> wrote: > > Wow, thank you! > > > > Let me know when u finish please, I will rebuild Eagle and try again. > > > > Also I'm interested about metadata. Why eagle uses Mysql and Hbase? And > where should my site_id configuration be stored? > > > > Regards, > > Andrey > > > > 2017-03-18 11:39 GMT+03:00 Jayesh Senjaliya <[email protected]>: > > aah ok, you can use this patch https://github.com/apache/eagle/pull/812 > > > > wait for sometime, i m doing rebase with master now.. > > > > good luck ! > > Jayesh > > > > > > On Sat, Mar 18, 2017 at 1:24 AM, MyGmail <[email protected]> wrote: > > Hi Jayesh, > > > > I'm sorry, but I can't change storm version. HDP2.5 includes 1.0.1. > > > > Any workaround? > > > > Regards, > > Andrey > > > 18 марта 2017 г., в 5:53, Jayesh Senjaliya <[email protected]> написал(а): > > Hi Markovich, > > > > Eagle 0.5 is well supported with storm 0.9.3 , can you please try using > that ? > > > > or do you have to use it with storm 1.x ? > > > > - Jayesh > > > > > > On Fri, Mar 17, 2017 at 6:50 AM, Markovich <[email protected]> wrote: > > Hello eagle users and dev, > > > > I'm stuck on installing eagle on HDP2.5 cluster with jdk 1.8.0_101. > > > > Here is my service versions and my cluster is secured using kerberos + > ranger. > > > > HDFS 2.7.3 > > Hive 1.2.1000 > > Storm 1.0.1 > > Kafka 0.10.0 > > Kerberos 1.10.3-10 > > > > Here is what I've done already: > > > > 1) Downloaded latest eagle from github (version 0.5 snapshot). > > 2) Builded it using mvn clean package -DskipTests > > [INFO] BUILD SUCCESS > > [INFO] ------------------------------------------------------------ > ------------ > > [INFO] Total time: 16:11 min > > [INFO] Finished at: 2017-03-17T16:00:45+03:00 > > [INFO] Final Memory: 183M/1755M > > [INFO] ------------------------------------------------------------ > ------------ > > > > 3) Moved tarball to /usr/hdp/current/ and extracted to eagle > > 4) Changed conf/eagle.conf for my cluster: > > zkQuorum > > zookeeperZnodeParent = "/hbase-secure" > > metadata -> jdbc -> user, pass and host > > nimbusHost > > 5) Launched bin/eagle-env.sh and bin/eagle-server start > > > > Eagle started on 9090 port. > > 6) In Web UI entered SiteId > > 7) Selected install on HDFS Audit Log Monitor and chenged General settings > (Kafka hosts) and Advanced fs.defaultFS. Execution Mode - cluster > > 8) Created hdfs_audit_log_{SITE_ID}, hdfs_audit_log_enriched_{SITE_ID} > 9) Launched Logstash to write to hdfs_audit_log_{SITE_ID} > 10) Checked, logs are pushed in kafka > > 11) Appliaction is initialized > > 12) Tried to start this appliaction: > > > > INFO [2017-03-17 13:28:45,926] org.apache.eagle.dataproc. > impl.storm.kafka.KafkaSpoutProvider: Use topic : hdfs_audit_log, > zkConnection : localhost:2181 , fetchSize : 1048576 > > WARN [2017-03-17 13:28:46,216] > org.apache.eagle.app.messaging.KafkaStreamProvider: > Using default shared sink topic dataSinkConfig.topic: hdfs_audit_event > > INFO [2017-03-17 13:28:46,331] > org.apache.eagle.app.environment.impl.StormExecutionRuntime: > Starting HDFS_AUDIT_LOG_MONITOR_APP_MYSITE > > (org.apache.eagle.security.auditlog.HdfsAuditLogApplication), mode: > CLUSTER > > INFO [2017-03-17 13:28:46,332] > org.apache.eagle.app.environment.impl.StormExecutionRuntime: > Overriding application.storm.nimbusHost = demo5 > > INFO [2017-03-17 13:28:46,332] > org.apache.eagle.app.environment.impl.StormExecutionRuntime: > Overriding application.storm.nimbusThriftPort = 6627 > > INFO [2017-03-17 13:28:46,332] > org.apache.eagle.app.environment.impl.StormExecutionRuntime: > Submitting as cluster mode ... > > > > Nothing in Storm UI. Nothing in logs. > > Tried to change KafkaSpoutProvider to normal with SITE_ID. -> Nothing > changed. > > > > I think I need to enter some kerberos related configs for storm... > > > > 13) Ok, changed Execution Mode - Local. > > 14) Started. Status in UI changed to starting. In logs a lot of activity, > but also errors: > > > > WARN [2017-03-17 13:33:40,952] storm.kafka.KafkaUtils: there are more > tasks than partitions (tasks: 2; partitions: 1), some tasks will be idle > > INFO [2017-03-17 13:33:40,952] storm.kafka.KafkaUtils: Task [1/2] > assigned [Partition{host=null:-1, partition=0}] > > INFO [2017-03-17 13:33:40,952] storm.kafka.ZkCoordinator: Task [1/2] > Deleted partition managers: [] > > INFO [2017-03-17 13:33:40,952] storm.kafka.ZkCoordinator: Task [1/2] New > partition managers: [Partition{host=null:-1, partition=0}] > > INFO [2017-03-17 13:33:40,990] storm.kafka.PartitionManager: Read > partition information from: /consumers/hdfs_audit_log_ > tuskpro/eagleConsumer/partition_0 --> null > > ERROR [2017-03-17 13:33:41,047] backtype.storm.util: Async loop died! > > ! java.lang.NullPointerException: null > > ! at org.apache.kafka.common.utils.Utils.formatAddress(Utils.java:312) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at > kafka.consumer.SimpleConsumer$$anonfun$disconnect$1.apply(SimpleConsumer.scala:49) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at > kafka.consumer.SimpleConsumer$$anonfun$disconnect$1.apply(SimpleConsumer.scala:49) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at kafka.utils.Logging$class.debug(Logging.scala:52) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at kafka.consumer.SimpleConsumer.debug(SimpleConsumer.scala:30) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at kafka.consumer.SimpleConsumer.disconnect(SimpleConsumer.scala:49) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at kafka.consumer.SimpleConsumer.liftedTree1$1(SimpleConsumer.scala:82) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at kafka.consumer.SimpleConsumer.kafka$consumer$SimpleConsumer$ > $sendRequest(SimpleConsumer.scala:68) ~[eagle-topology-0.5.0- > SNAPSHOT-assembly.jar:na] > > ! at kafka.consumer.SimpleConsumer.getOffsetsBefore(SimpleConsumer.scala:127) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at > kafka.javaapi.consumer.SimpleConsumer.getOffsetsBefore(SimpleConsumer.scala:79) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at storm.kafka.KafkaUtils.getOffset(KafkaUtils.java:74) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at storm.kafka.KafkaUtils.getOffset(KafkaUtils.java:64) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at storm.kafka.PartitionManager.<init>(PartitionManager.java:89) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at storm.kafka.ZkCoordinator.refresh(ZkCoordinator.java:98) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! ... 6 common frames omitted > > ! Causing: java.lang.RuntimeException: java.lang.NullPointerException > > ! at storm.kafka.ZkCoordinator.refresh(ZkCoordinator.java:103) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at storm.kafka.ZkCoordinator.getMyManagedPartitions(ZkCoordinator.java:69) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at storm.kafka.KafkaSpout.nextTuple(KafkaSpout.java:135) > ~[eagle-topology-0.5.0-SNAPSHOT-assembly.jar:na] > > ! at > backtype.storm.daemon.executor$fn__3373$fn__3388$fn__3417.invoke(executor.clj:565) > ~[storm-core-0.9.3.jar:0.9.3] > > ! at backtype.storm.util$async_loop$fn__464.invoke(util.clj:463) > ~[storm-core-0.9.3.jar:0.9.3] > > ! at clojure.lang.AFn.run(AFn.java:24) [eagle-topology-0.5.0- > SNAPSHOT-assembly.jar:na] > > ! at java.lang.Thread.run(Thread.java:745) [na:1.8.0_101] > > > > .... > > > > ERROR [2017-03-17 13:33:41,079] backtype.storm.util: Halting process: > ("Worker died") > > ! java.lang.RuntimeException: ("Worker died") > > ! at backtype.storm.util$exit_process_BANG_.doInvoke(util.clj:325) > [storm-core-0.9.3.jar:0.9.3] > > ! at clojure.lang.RestFn.invoke(RestFn.java:423) [eagle-topology-0.5.0- > SNAPSHOT-assembly.jar:na] > > ! at backtype.storm.daemon.worker$fn__3808$fn__3809.invoke(worker.clj:452) > [storm-core-0.9.3.jar:0.9.3] > > ! at backtype.storm.daemon.executor$mk_executor_data$fn__ > 3274$fn__3275.invoke(executor.clj:240) [storm-core-0.9.3.jar:0.9.3] > > ! at backtype.storm.util$async_loop$fn__464.invoke(util.clj:473) > [storm-core-0.9.3.jar:0.9.3] > > ! at clojure.lang.AFn.run(AFn.java:24) [eagle-topology-0.5.0- > SNAPSHOT-assembly.jar:na] > > ! at java.lang.Thread.run(Thread.java:745) [na:1.8.0_101] > > > > ERROR [2017-03-17 13:34:10,013] > org.apache.eagle.security.enrich.DataEnrichJob: > Fail to load sensitivity data > > ! java.net.ConnectException: Connection refused > > ! at java.net.PlainSocketImpl.socketConnect(Native Method) ~[na:1.8.0_101] > > ! at java.net.AbstractPlainSocketImpl.doConnect( > AbstractPlainSocketImpl.java:350) ~[na:1.8.0_101] > > ! at java.net.AbstractPlainSocketImpl.connectToAddress( > AbstractPlainSocketImpl.java:206) ~[na:1.8.0_101] > > ! at java.net.AbstractPlainSocketImpl.connect( > AbstractPlainSocketImpl.java:188) ~[na:1.8.0_101] > > ! at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:392) > ~[na:1.8.0_101] > > ! at java.net.Socket.connect(Socket.java:589) ~[na:1.8.0_101] > > ! at sun.net.NetworkClient.doConnect(NetworkClient.java:175) > ~[na:1.8.0_101] > > ! at sun.net.www.http.HttpClient.openServer(HttpClient.java:432) > ~[na:1.8.0_101] > > ! at sun.net.www.http.HttpClient.openServer(HttpClient.java:527) > ~[na:1.8.0_101] > > ! at sun.net.www.http.HttpClient.<init>(HttpClient.java:211) > ~[na:1.8.0_101] > > ! at sun.net.www.http.HttpClient.New(HttpClient.java:308) ~[na:1.8.0_101] > > ! at sun.net.www.http.HttpClient.New(HttpClient.java:326) ~[na:1.8.0_101] > > ! at > sun.net.www.protocol.http.HttpURLConnection.getNewHttpClient(HttpURLConnection.java:1169) > ~[na:1.8.0_101] > > ! at > sun.net.www.protocol.http.HttpURLConnection.plainConnect0(HttpURLConnection.java:1105) > ~[na:1.8.0_101] > > ! at > sun.net.www.protocol.http.HttpURLConnection.plainConnect(HttpURLConnection.java:999) > ~[na:1.8.0_101] > > ! at > sun.net.www.protocol.http.HttpURLConnection.connect(HttpURLConnection.java:933) > ~[na:1.8.0_101] > > ! at > sun.net.www.protocol.http.HttpURLConnection.getInputStream0(HttpURLConnection.java:1513) > ~[na:1.8.0_101] > > ! at > sun.net.www.protocol.http.HttpURLConnection.getInputStream(HttpURLConnection.java:1441) > ~[na:1.8.0_101] > > ! at java.net.HttpURLConnection.getResponseCode(HttpURLConnection.java:480) > ~[na:1.8.0_101] > > ! at com.sun.jersey.client.urlconnection.URLConnectionClientHandler._ > invoke(URLConnectionClientHandler.java:253) ~[jersey-client-1.19.1.jar:1. > 19.1] > > ! at com.sun.jersey.client.urlconnection.URLConnectionClientHandler. > handle(URLConnectionClientHandler.java:153) ~[jersey-client-1.19.1.jar:1. > 19.1] > > ! ... 12 common frames omitted > > ! Causing: com.sun.jersey.api.client.ClientHandlerException: > java.net.ConnectException: Connection refused > > ! at com.sun.jersey.client.urlconnection.URLConnectionClientHandler. > handle(URLConnectionClientHandler.java:155) ~[jersey-client-1.19.1.jar:1. > 19.1] > > ! at com.sun.jersey.api.client.filter.GZIPContentEncodingFilter.handle( > GZIPContentEncodingFilter.java:123) ~[jersey-client-1.19.1.jar:1.19.1] > > ! at com.sun.jersey.api.client.Client.handle(Client.java:652) > ~[jersey-client-1.19.1.jar:1.19.1] > > > > Ui wasn't accessable, but server was still running. > > It was failing again and again. > > > > 13) Restarted server > > 14) Nothing was saved. Again fresh install. UI asks for site_id. > > 15) Checked Hbase and MySQL Server - all is empty. > > > > > > So can someone please help me to get started with Eagle on my cluster? > > > > > > Also here is Logstash info: > > logstash-5.2.2 > > > > > > output { > > > > > > if [type] == "hdp-nn-audit" { > > > kafka { > > > codec => plain {format => "%{message}"} > > > bootstrap_servers => "demo4:6667" > > > topic_id => "hdfs_audit_log_tuskpro" > > > security_protocol => "SASL_PLAINTEXT" > > > sasl_kerberos_service_name => "kafka" > > > jaas_path => "/usr/hdp/current/kafka- > broker/config/kafka_client_jaas.conf" > > kerberos_config => "/etc/krb5.conf" > > > client_id => "hdp-nn-audit" > > > message_key => "%{user}" > > > } > > > } > > > } > > > > > > Also I don't know why, but logstash is not using kafka_client_jaas. > > There is info about sasl_kerberos_service_name , but without this property > logstash is not working. > > > > · > > KafkaClient { > > > > com.sun.security.auth.module.Krb5LoginModule required > > > > useTicketCache=true > > > > renewTicket=true > > > > serviceName="kafka"; > > > > }; > > > > Client { > > > > com.sun.security.auth.module.Krb5LoginModule required > > > > useTicketCache=true > > > > renewTicket=true > > > > serviceName="zookeeper"; > > > > }; > > > > > > > > Regards, > Andrey > > > > > > > > > > > > > > > > > > > > > >
