ambari-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Hitesh Shah <hit...@apache.org>
Subject Re: Ambari Metrics
Date Wed, 21 Oct 2015 17:20:43 GMT
@Siddharth,

"17:29:40,698  WARN [main] NativeCodeLoader:62 - Unable to load native-hadoop library for
your platform... using builtin-java classes where applicable“

The above message is usually meant to be harmless as it is warning about the use of non-performant
java implementations instead of using native code paths. Could you explain why this would
affect the functionality? Does this mean that one would never be able to deploy/run AMS on
a Mac because hadoop never has had any native libs built for Darwin? 

thanks
— Hitesh  


 
On Oct 20, 2015, at 6:50 PM, Siddharth Wagle <swagle@hortonworks.com> wrote:

> Hi Stan,
> 
> Based on the col.txt attached, the real problem is:
> 
> 17:29:40,698  WARN [main] NativeCodeLoader:62 - Unable to load native-hadoop library
for your platform... using builtin-java classes where applicable
> 
> This would mean incorrect binaries installed for AMS. Possibly wrong repo url used to
install the components.
> Can you please provide the ambari.repo URL used to install the service and the version
and flavor of the OS on which Metrics Collector is installed?
> 
> The hb.txt, looks like a clean log file.
> 
> Here is a link to all info that is useful for debugging:
> https://cwiki.apache.org/confluence/display/AMBARI/Troubleshooting+Guide
> 
> Best Regards,
> Sid
> 
> 
> From: smlynarczyk@prognosive.com <smlynarczyk@prognosive.com>
> Sent: Monday, October 19, 2015 12:33 PM
> To: Siddharth Wagle
> Cc: Daryl Heinz
> Subject: Ambari Metrics
>  
> Hello Siddharth,
> 
> I am hoping to get your input on an issue that has arisen with the Ambari Metrics Collector.
   This is with ambari 2.0.1 an HDP 2.2.6.  The error message received was:
> Caused by: java.sql.SQLException: ERROR 1102 (XCL02): Cannot get all table regions
> 
> Caused by: java.io.IOException: HRegionInfo was null in hbase:meta
> 
> ------- CUT partial collector log -----
> 
> 11:13:35,203  WARN [main] ConnectionManager$HConnectionImplementation:1228 - Encountered
problems when prefetch hbase:meta table:
> java.io.IOException: HRegionInfo was null or empty in Meta for SYSTEM.CATALOG, row=SYSTEM.CATALOG,,99999999999999
>         at org.apache.hadoop.hbase.client.MetaScanner.metaScan(MetaScanner.java:170)
>         at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.prefetchRegionCache(ConnectionManager.java:1222)
>         at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.locateRegionInMeta(ConnectionManager.java:1286)
>         at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.locateRegion(ConnectionManager.java:1135)
>         at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.locateRegion(ConnectionManager.java:1118)
>         at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.locateRegion(ConnectionManager.java:1075)
>         at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getRegionLocation(ConnectionManager.java:909)
>         at org.apache.phoenix.query.ConnectionQueryServicesImpl.getAllTableRegions(ConnectionQueryServicesImpl.java:401)
>         at org.apache.phoenix.query.ConnectionQueryServicesImpl.checkClientServerCompatibility(ConnectionQueryServicesImpl.java:853)
>         at org.apache.phoenix.query.ConnectionQueryServicesImpl.ensureTableCreated(ConnectionQueryServicesImpl.java:797)
>         at org.apache.phoenix.query.ConnectionQueryServicesImpl.createTable(ConnectionQueryServicesImpl.java:1107)
>         at org.apache.phoenix.query.DelegateConnectionQueryServices.createTable(DelegateConnectionQueryServices.java:110)
>         at org.apache.phoenix.schema.MetaDataClient.createTableInternal(MetaDataClient.java:1527)
>         at org.apache.phoenix.schema.MetaDataClient.createTable(MetaDataClient.java:535)
>         at org.apache.phoenix.compile.CreateTableCompiler$2.execute(CreateTableCompiler.java:184)
>         at org.apache.phoenix.jdbc.PhoenixStatement$2.call(PhoenixStatement.java:260)
>         at org.apache.phoenix.jdbc.PhoenixStatement$2.call(PhoenixStatement.java:252)
>         at org.apache.phoenix.call.CallRunner.run(CallRunner.java:53)
>         at org.apache.phoenix.jdbc.PhoenixStatement.executeMutation(PhoenixStatement.java:250)
>         at org.apache.phoenix.jdbc.PhoenixStatement.executeUpdate(PhoenixStatement.java:1026)
>         at org.apache.phoenix.query.ConnectionQueryServicesImpl$9.call(ConnectionQueryServicesImpl.java:1532)
>         at org.apache.phoenix.query.ConnectionQueryServicesImpl$9.call(ConnectionQueryServicesImpl.java:1501)
>         at org.apache.phoenix.util.PhoenixContextExecutor.call(PhoenixContextExecutor.java:77)
>         at org.apache.phoenix.query.ConnectionQueryServicesImpl.init(ConnectionQueryServicesImpl.java:1501)
>         at org.apache.phoenix.jdbc.PhoenixDriver.getConnectionQueryServices(PhoenixDriver.java:162)
>         at org.apache.phoenix.jdbc.PhoenixEmbeddedDriver.connect(PhoenixEmbeddedDriver.java:126)
>         at org.apache.phoenix.jdbc.PhoenixDriver.connect(PhoenixDriver.java:133)
>         at java.sql.DriverManager.getConnection(DriverManager.java:571)
>         at java.sql.DriverManager.getConnection(DriverManager.java:233)
>         at org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.DefaultPhoenixDataSource.getConnection(DefaultPhoenixDataSource.java:69)
>         at org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.PhoenixHBaseAccessor.getConnection(PhoenixHBaseAccessor.java:149)
>         at org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.PhoenixHBaseAccessor.getConnectionRetryingOnException(PhoenixHBaseAccessor.java:127)
>         at org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.PhoenixHBaseAccessor.initMetricSchema(PhoenixHBaseAccessor.java:268)
>         at org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.HBaseTimelineMetricStore.initializeSubsystem(HBaseTimelineMetricStore.java:64)
>         at org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.HBaseTimelineMetricStore.serviceInit(HBaseTimelineMetricStore.java:58)
>         at org.apache.hadoop.service.AbstractService.init(AbstractService.java:163)
>         at org.apache.hadoop.service.CompositeService.serviceInit(CompositeService.java:107)
>         at org.apache.hadoop.yarn.server.applicationhistoryservice.ApplicationHistoryServer.serviceInit(ApplicationHistoryServer.java:84)
>         at org.apache.hadoop.service.AbstractService.init(AbstractService.java:163)
>         at org.apache.hadoop.yarn.server.applicationhistoryservice.ApplicationHistoryServer.launchAppHistoryServer(ApplicationHistoryServer.java:137)
>         at org.apache.hadoop.yarn.server.applicationhistoryservice.ApplicationHistoryServer.main(ApplicationHistoryServer.java:147)
> 
> 
> The (partial)  contents of the embedded hbase and collector logs are in the attached.
 Any light that you could shed on this would be appreciated.  The incident I believe started
after an upgrade on July 20th at 17:29 pm
> 
> 
> Thanks in advance,
> 
> Stan
> 
> -- 
> 
> -- 
> Ad Altiora Tendo
> 
> Stanley J. Mlynarczyk - Ph.D.
> Chief Technology Officer
> 
> <image001.jpg>
> 
> Mobile: +1 630-607-2223
> 


Mime
View raw message