@Siddharth, "17:29:40,698 WARN [main] NativeCodeLoader:62 - Unable to load native-hadoop library for your platform... using builtin-java classes where applicable“
The above message is usually meant to be harmless as it is warning about the use of non-performant java implementations instead of using native code paths. Could you explain why this would affect the functionality? Does this mean that one would never be able to deploy/run AMS on a Mac because hadoop never has had any native libs built for Darwin? thanks — Hitesh On Oct 20, 2015, at 6:50 PM, Siddharth Wagle <[email protected]> wrote: > Hi Stan, > > Based on the col.txt attached, the real problem is: > > 17:29:40,698 WARN [main] NativeCodeLoader:62 - Unable to load native-hadoop > library for your platform... using builtin-java classes where applicable > > This would mean incorrect binaries installed for AMS. Possibly wrong repo url > used to install the components. > Can you please provide the ambari.repo URL used to install the service and > the version and flavor of the OS on which Metrics Collector is installed? > > The hb.txt, looks like a clean log file. > > Here is a link to all info that is useful for debugging: > https://cwiki.apache.org/confluence/display/AMBARI/Troubleshooting+Guide > > Best Regards, > Sid > > > From: [email protected] <[email protected]> > Sent: Monday, October 19, 2015 12:33 PM > To: Siddharth Wagle > Cc: Daryl Heinz > Subject: Ambari Metrics > > Hello Siddharth, > > I am hoping to get your input on an issue that has arisen with the Ambari > Metrics Collector. This is with ambari 2.0.1 an HDP 2.2.6. The error > message received was: > Caused by: java.sql.SQLException: ERROR 1102 (XCL02): Cannot get all table > regions > > Caused by: java.io.IOException: HRegionInfo was null in hbase:meta > > ------- CUT partial collector log ----- > > 11:13:35,203 WARN [main] ConnectionManager$HConnectionImplementation:1228 - > Encountered problems when prefetch hbase:meta table: > java.io.IOException: HRegionInfo was null or empty in Meta for > SYSTEM.CATALOG, row=SYSTEM.CATALOG,,99999999999999 > at > org.apache.hadoop.hbase.client.MetaScanner.metaScan(MetaScanner.java:170) > at > org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.prefetchRegionCache(ConnectionManager.java:1222) > at > org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.locateRegionInMeta(ConnectionManager.java:1286) > at > org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.locateRegion(ConnectionManager.java:1135) > at > org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.locateRegion(ConnectionManager.java:1118) > at > org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.locateRegion(ConnectionManager.java:1075) > at > org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getRegionLocation(ConnectionManager.java:909) > at > org.apache.phoenix.query.ConnectionQueryServicesImpl.getAllTableRegions(ConnectionQueryServicesImpl.java:401) > at > org.apache.phoenix.query.ConnectionQueryServicesImpl.checkClientServerCompatibility(ConnectionQueryServicesImpl.java:853) > at > org.apache.phoenix.query.ConnectionQueryServicesImpl.ensureTableCreated(ConnectionQueryServicesImpl.java:797) > at > org.apache.phoenix.query.ConnectionQueryServicesImpl.createTable(ConnectionQueryServicesImpl.java:1107) > at > org.apache.phoenix.query.DelegateConnectionQueryServices.createTable(DelegateConnectionQueryServices.java:110) > at > org.apache.phoenix.schema.MetaDataClient.createTableInternal(MetaDataClient.java:1527) > at > org.apache.phoenix.schema.MetaDataClient.createTable(MetaDataClient.java:535) > at > org.apache.phoenix.compile.CreateTableCompiler$2.execute(CreateTableCompiler.java:184) > at > org.apache.phoenix.jdbc.PhoenixStatement$2.call(PhoenixStatement.java:260) > at > org.apache.phoenix.jdbc.PhoenixStatement$2.call(PhoenixStatement.java:252) > at org.apache.phoenix.call.CallRunner.run(CallRunner.java:53) > at > org.apache.phoenix.jdbc.PhoenixStatement.executeMutation(PhoenixStatement.java:250) > at > org.apache.phoenix.jdbc.PhoenixStatement.executeUpdate(PhoenixStatement.java:1026) > at > org.apache.phoenix.query.ConnectionQueryServicesImpl$9.call(ConnectionQueryServicesImpl.java:1532) > at > org.apache.phoenix.query.ConnectionQueryServicesImpl$9.call(ConnectionQueryServicesImpl.java:1501) > at > org.apache.phoenix.util.PhoenixContextExecutor.call(PhoenixContextExecutor.java:77) > at > org.apache.phoenix.query.ConnectionQueryServicesImpl.init(ConnectionQueryServicesImpl.java:1501) > at > org.apache.phoenix.jdbc.PhoenixDriver.getConnectionQueryServices(PhoenixDriver.java:162) > at > org.apache.phoenix.jdbc.PhoenixEmbeddedDriver.connect(PhoenixEmbeddedDriver.java:126) > at > org.apache.phoenix.jdbc.PhoenixDriver.connect(PhoenixDriver.java:133) > at java.sql.DriverManager.getConnection(DriverManager.java:571) > at java.sql.DriverManager.getConnection(DriverManager.java:233) > at > org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.DefaultPhoenixDataSource.getConnection(DefaultPhoenixDataSource.java:69) > at > org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.PhoenixHBaseAccessor.getConnection(PhoenixHBaseAccessor.java:149) > at > org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.PhoenixHBaseAccessor.getConnectionRetryingOnException(PhoenixHBaseAccessor.java:127) > at > org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.PhoenixHBaseAccessor.initMetricSchema(PhoenixHBaseAccessor.java:268) > at > org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.HBaseTimelineMetricStore.initializeSubsystem(HBaseTimelineMetricStore.java:64) > at > org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.HBaseTimelineMetricStore.serviceInit(HBaseTimelineMetricStore.java:58) > at > org.apache.hadoop.service.AbstractService.init(AbstractService.java:163) > at > org.apache.hadoop.service.CompositeService.serviceInit(CompositeService.java:107) > at > org.apache.hadoop.yarn.server.applicationhistoryservice.ApplicationHistoryServer.serviceInit(ApplicationHistoryServer.java:84) > at > org.apache.hadoop.service.AbstractService.init(AbstractService.java:163) > at > org.apache.hadoop.yarn.server.applicationhistoryservice.ApplicationHistoryServer.launchAppHistoryServer(ApplicationHistoryServer.java:137) > at > org.apache.hadoop.yarn.server.applicationhistoryservice.ApplicationHistoryServer.main(ApplicationHistoryServer.java:147) > > > The (partial) contents of the embedded hbase and collector logs are in the > attached. Any light that you could shed on this would be appreciated. The > incident I believe started after an upgrade on July 20th at 17:29 pm > > > Thanks in advance, > > Stan > > -- > > -- > Ad Altiora Tendo > > Stanley J. Mlynarczyk - Ph.D. > Chief Technology Officer > > <image001.jpg> > > Mobile: +1 630-607-2223 >
