ambari-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Siddharth Wagle <swa...@hortonworks.com>
Subject Re: Ambari Metrics
Date Wed, 21 Oct 2015 17:26:16 GMT
AMS uses SNAPPY compression by default. So the service would start up fine but fail when Phoenix
tried to CREATE TABLE.

The work around is to set the compression code property in ams-site to "NONE" instead of SNAPPY.

So, it will work on the MAC just not with compression enabled.

- Sid
________________________________________
From: Hitesh Shah <hitesh@apache.org>
Sent: Wednesday, October 21, 2015 10:20 AM
To: user@ambari.apache.org
Cc: smlynarczyk@prognosive.com; Daryl Heinz
Subject: Re: Ambari Metrics

@Siddharth,

"17:29:40,698  WARN [main] NativeCodeLoader:62 - Unable to load native-hadoop library for
your platform... using builtin-java classes where applicable“

The above message is usually meant to be harmless as it is warning about the use of non-performant
java implementations instead of using native code paths. Could you explain why this would
affect the functionality? Does this mean that one would never be able to deploy/run AMS on
a Mac because hadoop never has had any native libs built for Darwin?

thanks
— Hitesh



On Oct 20, 2015, at 6:50 PM, Siddharth Wagle <swagle@hortonworks.com> wrote:

> Hi Stan,
>
> Based on the col.txt attached, the real problem is:
>
> 17:29:40,698  WARN [main] NativeCodeLoader:62 - Unable to load native-hadoop library
for your platform... using builtin-java classes where applicable
>
> This would mean incorrect binaries installed for AMS. Possibly wrong repo url used to
install the components.
> Can you please provide the ambari.repo URL used to install the service and the version
and flavor of the OS on which Metrics Collector is installed?
>
> The hb.txt, looks like a clean log file.
>
> Here is a link to all info that is useful for debugging:
> https://cwiki.apache.org/confluence/display/AMBARI/Troubleshooting+Guide
>
> Best Regards,
> Sid
>
>
> From: smlynarczyk@prognosive.com <smlynarczyk@prognosive.com>
> Sent: Monday, October 19, 2015 12:33 PM
> To: Siddharth Wagle
> Cc: Daryl Heinz
> Subject: Ambari Metrics
>
> Hello Siddharth,
>
> I am hoping to get your input on an issue that has arisen with the Ambari Metrics Collector.
   This is with ambari 2.0.1 an HDP 2.2.6.  The error message received was:
> Caused by: java.sql.SQLException: ERROR 1102 (XCL02): Cannot get all table regions
>
> Caused by: java.io.IOException: HRegionInfo was null in hbase:meta
>
> ------- CUT partial collector log -----
>
> 11:13:35,203  WARN [main] ConnectionManager$HConnectionImplementation:1228 - Encountered
problems when prefetch hbase:meta table:
> java.io.IOException: HRegionInfo was null or empty in Meta for SYSTEM.CATALOG, row=SYSTEM.CATALOG,,99999999999999
>         at org.apache.hadoop.hbase.client.MetaScanner.metaScan(MetaScanner.java:170)
>         at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.prefetchRegionCache(ConnectionManager.java:1222)
>         at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.locateRegionInMeta(ConnectionManager.java:1286)
>         at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.locateRegion(ConnectionManager.java:1135)
>         at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.locateRegion(ConnectionManager.java:1118)
>         at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.locateRegion(ConnectionManager.java:1075)
>         at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getRegionLocation(ConnectionManager.java:909)
>         at org.apache.phoenix.query.ConnectionQueryServicesImpl.getAllTableRegions(ConnectionQueryServicesImpl.java:401)
>         at org.apache.phoenix.query.ConnectionQueryServicesImpl.checkClientServerCompatibility(ConnectionQueryServicesImpl.java:853)
>         at org.apache.phoenix.query.ConnectionQueryServicesImpl.ensureTableCreated(ConnectionQueryServicesImpl.java:797)
>         at org.apache.phoenix.query.ConnectionQueryServicesImpl.createTable(ConnectionQueryServicesImpl.java:1107)
>         at org.apache.phoenix.query.DelegateConnectionQueryServices.createTable(DelegateConnectionQueryServices.java:110)
>         at org.apache.phoenix.schema.MetaDataClient.createTableInternal(MetaDataClient.java:1527)
>         at org.apache.phoenix.schema.MetaDataClient.createTable(MetaDataClient.java:535)
>         at org.apache.phoenix.compile.CreateTableCompiler$2.execute(CreateTableCompiler.java:184)
>         at org.apache.phoenix.jdbc.PhoenixStatement$2.call(PhoenixStatement.java:260)
>         at org.apache.phoenix.jdbc.PhoenixStatement$2.call(PhoenixStatement.java:252)
>         at org.apache.phoenix.call.CallRunner.run(CallRunner.java:53)
>         at org.apache.phoenix.jdbc.PhoenixStatement.executeMutation(PhoenixStatement.java:250)
>         at org.apache.phoenix.jdbc.PhoenixStatement.executeUpdate(PhoenixStatement.java:1026)
>         at org.apache.phoenix.query.ConnectionQueryServicesImpl$9.call(ConnectionQueryServicesImpl.java:1532)
>         at org.apache.phoenix.query.ConnectionQueryServicesImpl$9.call(ConnectionQueryServicesImpl.java:1501)
>         at org.apache.phoenix.util.PhoenixContextExecutor.call(PhoenixContextExecutor.java:77)
>         at org.apache.phoenix.query.ConnectionQueryServicesImpl.init(ConnectionQueryServicesImpl.java:1501)
>         at org.apache.phoenix.jdbc.PhoenixDriver.getConnectionQueryServices(PhoenixDriver.java:162)
>         at org.apache.phoenix.jdbc.PhoenixEmbeddedDriver.connect(PhoenixEmbeddedDriver.java:126)
>         at org.apache.phoenix.jdbc.PhoenixDriver.connect(PhoenixDriver.java:133)
>         at java.sql.DriverManager.getConnection(DriverManager.java:571)
>         at java.sql.DriverManager.getConnection(DriverManager.java:233)
>         at org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.DefaultPhoenixDataSource.getConnection(DefaultPhoenixDataSource.java:69)
>         at org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.PhoenixHBaseAccessor.getConnection(PhoenixHBaseAccessor.java:149)
>         at org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.PhoenixHBaseAccessor.getConnectionRetryingOnException(PhoenixHBaseAccessor.java:127)
>         at org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.PhoenixHBaseAccessor.initMetricSchema(PhoenixHBaseAccessor.java:268)
>         at org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.HBaseTimelineMetricStore.initializeSubsystem(HBaseTimelineMetricStore.java:64)
>         at org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.HBaseTimelineMetricStore.serviceInit(HBaseTimelineMetricStore.java:58)
>         at org.apache.hadoop.service.AbstractService.init(AbstractService.java:163)
>         at org.apache.hadoop.service.CompositeService.serviceInit(CompositeService.java:107)
>         at org.apache.hadoop.yarn.server.applicationhistoryservice.ApplicationHistoryServer.serviceInit(ApplicationHistoryServer.java:84)
>         at org.apache.hadoop.service.AbstractService.init(AbstractService.java:163)
>         at org.apache.hadoop.yarn.server.applicationhistoryservice.ApplicationHistoryServer.launchAppHistoryServer(ApplicationHistoryServer.java:137)
>         at org.apache.hadoop.yarn.server.applicationhistoryservice.ApplicationHistoryServer.main(ApplicationHistoryServer.java:147)
>
>
> The (partial)  contents of the embedded hbase and collector logs are in the attached.
 Any light that you could shed on this would be appreciated.  The incident I believe started
after an upgrade on July 20th at 17:29 pm
>
>
> Thanks in advance,
>
> Stan
>
> --
>
> --
> Ad Altiora Tendo
>
> Stanley J. Mlynarczyk - Ph.D.
> Chief Technology Officer
>
> <image001.jpg>
>
> Mobile: +1 630-607-2223
>



Mime
View raw message