ambari-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Bryan Bende <bbe...@gmail.com>
Subject Re: Posting Metrics to Ambari
Date Wed, 29 Jul 2015 19:28:05 GMT
FWIW I was finally able to get this to work and the issue seems to be case
sensitivity in the appId field...

Send a metric with APP_ID=*NIFI* then query:
http://localhost:6188/ws/v1/timeline/metrics?metricNames=FlowFiles_Received_Last_5_mins&appId=
*NIFI*&hostname=localhost&startTime=1438193080000&endTime=1438193082000
Gets 0 results.

Send a metric with APP_ID=*nifi* then query:
http://localhost:6188/ws/v1/timeline/metrics?metricNames=FlowFiles_Received_Last_5_mins&appId=
*NIFI*&hostname=localhost&startTime=1438193080000&endTime=1438193082000
Gets results, even though appId=NIFI in the query.

Would this be worthy of a jira? I would expect that if the query side is
always going to search lower case, than the ingest side should be
normalizing to lower case.

Thanks,

Bryan


On Tue, Jul 28, 2015 at 1:07 PM, Bryan Bende <bbende@gmail.com> wrote:

> As an update, I was able to create a new service and get it installed in
> Ambari, and got a widget to display on the metrics panel for the service.
>
> So now it seems like the missing piece is getting the metrics exposed
> through the Ambari REST API, which may or may not be related to not getting
> results from the collector service API. I have a metrics.json with the
> following:
>
> {
>>   "NIFI_MASTER": {
>>     "Component": [{
>>         "type": "ganglia",
>>         "metrics": {
>>           "default": {
>>             "metrics/nifi/FlowFilesReceivedLast5mins": {
>>               "metric": "FlowFiles_Received_Last_5_mins",
>>               "pointInTime": false,
>>               "temporal": true
>>             }
>>           }
>>         }
>>     }]
>>   }
>> }
>
>
> and widgets.json with the following:
>
> {
>>   "layouts": [
>>     {
>>       "layout_name": "default_nifi_dashboard",
>>       "display_name": "Standard NiFi Dashboard",
>>       "section_name": "NIFI_SUMMARY",
>>       "widgetLayoutInfo": [
>>         {
>>           "widget_name": "Flow Files Received Last 5 mins",
>>           "description": "The number of flow files received in the last 5
>> minutes.",
>>           "widget_type": "GRAPH",
>>           "is_visible": true,
>>           "metrics": [
>>             {
>>               "name": "FlowFiles_Received_Last_5_mins",
>>               "metric_path": "metrics/nifi/FlowFilesReceivedLast5mins",
>>               "service_name": "NIFI",
>>               "component_name": "NIFI_MASTER"
>>             }
>>           ],
>>           "values": [
>>             {
>>               "name": "Flow Files Received",
>>               "value": "${FlowFiles_Received_Last_5_mins}"
>>             }
>>           ],
>>           "properties": {
>>             "display_unit": "%",
>>             "graph_type": "LINE",
>>             "time_range": "1"
>>           }
>>         }
>>       ]
>>     }
>>   ]
>> }
>
>
> Hitting this end-point doesn't show any metrics though:
>
> http://localhost:8080/api/v1/clusters/Sandbox/services/NIFI/components/NIFI_MASTER
>
> -Bryan
>
>
> On Tue, Jul 28, 2015 at 9:39 AM, Bryan Bende <bbende@gmail.com> wrote:
>
>> The data is present in the aggregate tables...
>>
>> 0: jdbc:phoenix:localhost:61181:/hbase> SELECT * from *METRIC_RECORD*
>> WHERE METRIC_NAME = 'FlowFiles_Received_Last_5_mins' order by SERVER_TIME
>> desc limit 10;
>>
>>
>> +------------------------------------------+------------------------------------------+------------------------------------------+------------------------------------------+-------------------------------------------+
>>
>> |               METRIC_NAME                |                 HOSTNAME
>>               |               SERVER_TIME                |
>>     APP_ID                  |               INSTANCE_ID                 |
>>
>>
>> +------------------------------------------+------------------------------------------+------------------------------------------+------------------------------------------+-------------------------------------------+
>>
>> | FlowFiles_Received_Last_5_mins           | localhost
>>         | 1438047369541                            | NIFI
>>                       | 5dbaaa80-0760-4241-80aa-b00b52f8efb4      |
>>
>> 0: jdbc:phoenix:localhost:61181:/hbase> SELECT * from
>> *METRIC_RECORD_MINUTE* WHERE METRIC_NAME =
>> 'FlowFiles_Received_Last_5_mins' order by SERVER_TIME desc limit 10;
>>
>>
>> +------------------------------------------+------------------------------------------+------------------------------------------+------------------------------------------+-------------------------------------------+
>>
>> |               METRIC_NAME                |                 HOSTNAME
>>               |                  APP_ID                  |
>> INSTANCE_ID                |               SERVER_TIME                 |
>>
>>
>> +------------------------------------------+------------------------------------------+------------------------------------------+------------------------------------------+-------------------------------------------+
>>
>> | FlowFiles_Received_Last_5_mins           | localhost
>>         | NIFI                                     | 5dbaaa80-0760-4241-80aa-b00b52f8efb4
>>     | 1438047369541                             |
>>
>>
>> 0: jdbc:phoenix:localhost:61181:/hbase> SELECT * from
>> *METRIC_RECORD_HOURLY* WHERE METRIC_NAME =
>> 'FlowFiles_Received_Last_5_mins' order by SERVER_TIME desc limit 10;
>>
>>
>> +------------------------------------------+------------------------------------------+------------------------------------------+------------------------------------------+-------------------------------------------+
>>
>> |               METRIC_NAME                |                 HOSTNAME
>>               |                  APP_ID                  |
>> INSTANCE_ID                |               SERVER_TIME                 |
>>
>>
>> +------------------------------------------+------------------------------------------+------------------------------------------+------------------------------------------+-------------------------------------------+
>>
>> | FlowFiles_Received_Last_5_mins           | localhost
>>         | NIFI                                     | 5dbaaa80-0760-4241-80aa-b00b52f8efb4
>>     | 1438045569276                             |
>>
>>
>> Trying a smaller time range (2 mins surrounding the timestamp from the
>> first record above)....
>>
>>
>> http://localhost:6188/ws/v1/timeline/metrics?metricNames=FlowFiles_Received_Last_5_mins&appId=NIFI&hostname=localhost&startTime=1438047300000&endTime=1438047420000
>>
>>
>> http://localhost:6188/ws/v1/timeline/metrics?metricNames=FlowFiles_Received_Last_5_mins&appId=NIFI&hostname=localhost&startTime=1438047300000&endTime=1438047420000&precision=seconds
>>
>>
>> http://localhost:6188/ws/v1/timeline/metrics?metricNames=FlowFiles_Received_Last_5_mins&appId=NIFI&hostname=localhost&startTime=1438047300000&endTime=1438047420000&precision=minutes
>>
>>
>> http://localhost:6188/ws/v1/timeline/metrics?metricNames=FlowFiles_Received_Last_5_mins&appId=NIFI&hostname=localhost&startTime=1438047300000&endTime=1438047420000&precision=hours
>>
>>
>> Those all get no results. The only time I got a difference response, was
>> this example:
>>
>>
>> http://localhost:6188/ws/v1/timeline/metrics?metricNames=FlowFiles_Received_Last_5_mins&appId=NIFI&hostname=localhost&startTime=143804556927&endTime=1438047420000
>>
>> which returned:
>>
>> {"exception":"BadRequestException","message":"java.lang.Exception: The time range
query for precision table exceeds row count limit, please query aggregate table instead.","javaClassName":"org.apache.hadoop.yarn.webapp.BadRequestException"}
>>
>>
>> On Mon, Jul 27, 2015 at 10:50 PM, Siddharth Wagle <swagle@hortonworks.com
>> > wrote:
>>
>>>  For Step1, when you say exposing metrics through the Ambari REST
>>> API... are you talking about the metrics collector REST API, or through the
>>> Ambari Server REST API?
>>>
>>> Answer: Ambari REST API: Note that this is intended use because this is
>>> what ties the metrics to you your cluster resources, example: You can query
>>> for say give me metrics for the active Namenode only using Ambari's API.
>>>
>>>
>>>
>>>  Is SERVER_TIME the field that has to fall between startTime and
>>> endTime?
>>>
>>> Yes. That is correct
>>>
>>>
>>>  There is nothing special about the query  you seem to have the
>>> fragments right, only this is you are query for a large time window,
>>> AMS would not return data from METRIC_RECORD table for a such a large time
>>> window it would try to find this in the aggregate
>>> table, METRIC_RECORD_MINUTE or HOURLY. Try reducing you time range,
>>> also check the aggregate tables, the data should still be present in those
>>> tables.
>>>
>>>
>>>  Precision params:
>>>
>>>
>>> https://cwiki.apache.org/confluence/display/AMBARI/Metrics+Collector+API+Specification
>>>
>>>
>>>  -Sid
>>>
>>>
>>>  ------------------------------
>>> *From:* Bryan Bende <bbende@gmail.com>
>>> *Sent:* Monday, July 27, 2015 6:21 PM
>>>
>>> *To:* user@ambari.apache.org
>>> *Subject:* Re: Posting Metrics to Ambari
>>>
>>>  Hi Jaimin,
>>>
>>>  For Step1, when you say exposing metrics through the Ambari REST
>>> API... are you talking about the metrics collector REST API, or through the
>>> Ambari Server REST API?
>>>
>>>  I am able to see data through Phoenix, as an example:
>>>
>>> +------------------------------------------+------------------------------------------+------------------------------------------+------------------------------------------+
>>>
>>> |               METRIC_NAME                |                 HOSTNAME
>>>               |               SERVER_TIME                |
>>>     APP_ID                  |
>>>
>>>
>>> +------------------------------------------+------------------------------------------+------------------------------------------+------------------------------------------+
>>>
>>> | FlowFiles_Received_Last_5_mins           | localhost
>>>         | 1438045869329                            | NIFI
>>>                       |
>>>
>>>
>>>  Then I try to use this API call:
>>>
>>>
>>> http://localhost:6188/ws/v1/timeline/metrics?metricNames=FlowFiles_Received_Last_5_mins&appId=NIFI&hostname=localhost&startTime=1437870332000&endTime=1438129532000
>>>
>>> and I get: {"metrics":[]}
>>>
>>> Something must not be lining up with what I am sending over. Is
>>> SERVER_TIME the field that has to fall between startTime and endTime?
>>>
>>> -Bryan
>>>
>>> On Mon, Jul 27, 2015 at 1:40 PM, Jaimin Jetly <jaimin@hortonworks.com>
>>> wrote:
>>>
>>>>  Hi Bryan,
>>>>
>>>>
>>>>  There are 2 steps in this that needs to be achieved.
>>>>
>>>>
>>>>  STEP-1:  Exposing service metrics successfully through Ambari REST API
>>>>
>>>> STEP-2:  Ambari UI displaying widgets comprised from newly exposed
>>>> metrics via Ambari server.
>>>>
>>>>
>>>>
>>>>  As step-1 is pre-requisite to step-2, can you confirm that you were
>>>> able to achieve step-1 (exposing service metrics successfully through
>>>> Ambari REST API) ?
>>>>
>>>>
>>>>  *NOTE:*
>>>> /var/lib/ambari-server/resources/common-services/AMBARI_METRICS/0.1.0/metrics.json
>>>> are the metrics specific to Ambari metrics service. If the new metrics
>>>> that you want to expose are related to any other service then please
>>>> edit/create metrics.json file in that specific service package and not in
Ambari
>>>> metrics service package. widgets.json also needs to be changed/added in the
>>>> same service package and not at
>>>> /var/lib/ambari-server/resources/stacks/HDP/2.0.6/widgets.json (unless
>>>> you want to add system heatmaps for a stack that inherits HDP-2.0.6 stack).
>>>>
>>>>
>>>>
>>>>  -- Thanks
>>>>
>>>>     Jaimin
>>>>  ------------------------------
>>>> *From:* Bryan Bende <bbende@gmail.com>
>>>> *Sent:* Sunday, July 26, 2015 2:10 PM
>>>>
>>>> *To:* user@ambari.apache.org
>>>> *Subject:* Re: Posting Metrics to Ambari
>>>>
>>>>
>>>> Hi Sid,
>>>>
>>>> Thanks for the pointers about how to add a metric to the UI. Based on
>>>> those instructions I modified
>>>> /var/lib/ambari-server/resources/common-services/AMBARI_METRICS/0.1.0/metrics.json
>>>> and added the following based on the test metrics I posted:
>>>>
>>>> "metrics/SmokeTest/FakeMetric": {
>>>>
>>>>               "metric": "AMBARI_METRICS.SmokeTest.FakeMetric",
>>>>
>>>>               "pointInTime": true,
>>>>
>>>>               "temporal": true
>>>>
>>>>             }
>>>>
>>>> From digging around the filesystem there appears to be a widgets.json
>>>> in /var/lib/ambari-server/resources/stacks/HDP/2.0.6/widgets.json. It looks
>>>> like this file only contained the definitions of the heatmaps, so I wasn't
>>>> sure if this was the right place, but just to see what happened I modified
>>>> it as follows:
>>>>
>>>> 1) Added a whole new layout:
>>>>
>>>> http://pastebin.com/KqeT8xfe
>>>>
>>>> 2) Added a heatmap for the test metric:
>>>>
>>>> http://pastebin.com/AQDT7u6v
>>>>
>>>> Then I restarted the HDP VM but I don't see anything in the UI under
>>>> Metric Actions -> Add, or under Heatmaps. Anything that seems completely
>>>> wrong about what I did? Maybe I should be going down the route of defining
>>>> a new service type for system I will be sending metrics from?
>>>>
>>>> Sorry to keep bothering with all these questions, I just don't have any
>>>> previous experience with Ambari.
>>>>
>>>> Thanks,
>>>>
>>>> Bryan
>>>>
>>>> On Sun, Jul 26, 2015 at 12:10 AM, Siddharth Wagle <
>>>> swagle@hortonworks.com> wrote:
>>>>
>>>>>  The AMS API does not allow open ended queries so startTime and
>>>>> endTime are required fields, the curl call should return the error code
>>>>> with the apt response.
>>>>>
>>>>>
>>>>>  If this doesn't happen please go ahead and file a Jira.
>>>>>
>>>>>
>>>>>  Using AMS through Ambari UI after getting the plumbing work with
>>>>> metrics.json completed would be much easier. The AMS API does need some
>>>>> refinement. Jiras / Bugs are welcome.
>>>>>
>>>>>
>>>>>  -Sid
>>>>>
>>>>>
>>>>>
>>>>>  ------------------------------
>>>>> *From:* Siddharth Wagle <swagle@hortonworks.com>
>>>>> *Sent:* Saturday, July 25, 2015 9:01 PM
>>>>>
>>>>> *To:* user@ambari.apache.org
>>>>> *Subject:* Re: Posting Metrics to Ambari
>>>>>
>>>>>
>>>>> No dev work need only need to modify metrics.json file and then add
>>>>> widget from UI.
>>>>>
>>>>>
>>>>>  Stack details:
>>>>>
>>>>>
>>>>> https://cwiki.apache.org/confluence/display/AMBARI/Stack+Defined+Metrics
>>>>>
>>>>>
>>>>>  UI specifics:
>>>>>
>>>>>
>>>>> https://cwiki.apache.org/confluence/display/AMBARI/Enhanced+Service+Dashboard
>>>>>
>>>>>
>>>>>  -Sid
>>>>>
>>>>>
>>>>>  ------------------------------
>>>>> *From:* Bryan Bende <bbende@gmail.com>
>>>>> *Sent:* Saturday, July 25, 2015 7:10 PM
>>>>> *To:* user@ambari.apache.org
>>>>> *Subject:* Re: Posting Metrics to Ambari
>>>>>
>>>>>  Quick update, I was able to connect with the phoenix 4.2.2 client
>>>>> and I did get results querying with:
>>>>> SELECT * from METRIC_RECORD WHERE METRIC_NAME =
>>>>> 'AMBARI_METRICS.SmokeTest.FakeMetric' order by SERVER_TIME desc limit
10;
>>>>>
>>>>>  Now that I know the metrics are posting, I am less concerned about
>>>>> querying through the REST API.
>>>>>
>>>>>  Is there any way to get a custom metric added to the main page of
>>>>> Ambari? or does this require development work?
>>>>>
>>>>>  Thanks,
>>>>>
>>>>>  Bryan
>>>>>
>>>>> On Sat, Jul 25, 2015 at 9:42 PM, Bryan Bende <bbende@gmail.com>
wrote:
>>>>>
>>>>>> Hi Sid,
>>>>>>
>>>>>>  Thanks for the suggestions. I turned on DEBUG for the metrics
>>>>>> collector (had to do this through the Ambari UI configs section)
and now I
>>>>>> can see some activity... When I post a metric I see:
>>>>>>
>>>>>>  01:30:18,372 DEBUG [95266635@qtp-171166092-2 -
>>>>>> /ws/v1/timeline/metrics] TimelineWebServices:270 - Storing metrics:
{
>>>>>>
>>>>>>   "metrics" : [ {
>>>>>>
>>>>>>     "timestamp" : 1432075898000,
>>>>>>
>>>>>>     "metricname" : "AMBARI_METRICS.SmokeTest.FakeMetric",
>>>>>>
>>>>>>     "appid" : "amssmoketestfake",
>>>>>>
>>>>>>     "hostname" : "localhost",
>>>>>>
>>>>>>     "starttime" : 1432075898000,
>>>>>>
>>>>>>     "metrics" : {
>>>>>>
>>>>>>       "1432075898000" : 0.963781711428,
>>>>>>
>>>>>>       "1432075899000" : 1.432075898E12
>>>>>>
>>>>>>     }
>>>>>>
>>>>>>   } ]
>>>>>>
>>>>>> }
>>>>>>
>>>>>> 01:30:18,372 DEBUG [95266635@qtp-171166092-2 -
>>>>>> /ws/v1/timeline/metrics] DefaultPhoenixDataSource:67 - Metric store
>>>>>> connection url: jdbc:phoenix:localhost:61181:/hbase
>>>>>>
>>>>>> 01:30:18,376 DEBUG [95266635@qtp-171166092-2 -
>>>>>> /ws/v1/timeline/metrics] MutationState:361 - Sending 2 mutations
for
>>>>>> METRIC_RECORD with 8 key values of total size 925 bytes
>>>>>>
>>>>>> 01:30:18,380 DEBUG [95266635@qtp-171166092-2 -
>>>>>> /ws/v1/timeline/metrics] MutationState:436 - Total time for batch
call of
>>>>>> 2 mutations into METRIC_RECORD: 3 ms
>>>>>>
>>>>>> 01:30:18,381 DEBUG [95266635@qtp-171166092-2 -
>>>>>> /ws/v1/timeline/metrics] log:40 - RESPONSE /ws/v1/timeline/metrics
 200
>>>>>>
>>>>>>
>>>>>>  So it looks like it posted successfully. Then I hit:
>>>>>>
>>>>>>
>>>>>> http://localhost:6188/ws/v1/timeline/metrics?metricNames=AMBARI_METRICS.SmokeTest.FakeMetric
>>>>>>
>>>>>> and I see...
>>>>>>
>>>>>> 01:31:16,952 DEBUG [95266635@qtp-171166092-2 -
>>>>>> /ws/v1/timeline/metrics?metricNames=AMBARI_METRICS.SmokeTest.FakeMetric]
>>>>>> ParallelIterators:412 - Guideposts: ]
>>>>>>
>>>>>> 01:31:16,953 DEBUG [95266635@qtp-171166092-2 -
>>>>>> /ws/v1/timeline/metrics?metricNames=AMBARI_METRICS.SmokeTest.FakeMetric]
>>>>>> ParallelIterators:481 - The parallelScans:
>>>>>> [[{"timeRange":[0,1437874276946],"batch":-1,"startRow":"AMBARI_METRICS.SmokeTest.FakeMetric","stopRow":"AMBARI_METRICS.SmokeTest.FakeMetric\\x01","loadColumnFamiliesOnDemand":null,"totalColumns":1,"cacheBlocks":true,"families":{"0":["ALL"]},"maxResultSize":-1,"maxVersions":1,"filter":"","caching":-1}]]
>>>>>>
>>>>>> 01:31:16,953 DEBUG [95266635@qtp-171166092-2 -
>>>>>> /ws/v1/timeline/metrics?metricNames=AMBARI_METRICS.SmokeTest.FakeMetric]
>>>>>> BaseQueryPlan:243 - Iterator ready: MergeSortTopNResultIterator [limit=1,
>>>>>> count=0, orderByColumns=[METRIC_NAME DESC, SERVER_TIME DESC], ptr1=,
ptr2=]
>>>>>>
>>>>>> 01:31:16,957 DEBUG [phoenix-1-thread-171] ParallelIterators:629 -
Id:
>>>>>> d0c9c381-f35f-48e6-b970-8b6d5997684b, Time: 3ms, Scan:
>>>>>> {"timeRange":[0,1437874276946],"batch":-1,"startRow":"AMBARI_METRICS.SmokeTest.FakeMetric","stopRow":"AMBARI_METRICS.SmokeTest.FakeMetric\\x01","loadColumnFamiliesOnDemand":null,"totalColumns":1,"cacheBlocks":true,"families":{"0":["ALL"]},"maxResultSize":-1,"maxVersions":1,"filter":"","caching":10000}
>>>>>>
>>>>>> 01:31:16,959 DEBUG [95266635@qtp-171166092-2 -
>>>>>> /ws/v1/timeline/metrics?metricNames=AMBARI_METRICS.SmokeTest.FakeMetric]
>>>>>> PhoenixHBaseAccessor:552 - Aggregate records size: 0
>>>>>>
>>>>>> I'll see if I can get the phoenix client working and see what that
>>>>>> returns.
>>>>>>
>>>>>> Thanks,
>>>>>>
>>>>>> Bryan
>>>>>>
>>>>>> On Fri, Jul 24, 2015 at 5:44 PM, Siddharth Wagle <
>>>>>> swagle@hortonworks.com> wrote:
>>>>>>
>>>>>>>  Hi Bryan,
>>>>>>>
>>>>>>>
>>>>>>>  Few things you can do:
>>>>>>>
>>>>>>>
>>>>>>>  1. Turn on DEBUG mode by changing log4j.properties at,
>>>>>>> /etc/ambari-metrics-collector/conf/
>>>>>>>
>>>>>>> This might reveal more info, I don't think we print every metrics
>>>>>>> received to the log in 2.0 or 2.1, I did add this option if TRACE
is
>>>>>>> enabled to trunk recently.
>>>>>>>
>>>>>>>
>>>>>>>  2. Connect using Phoenix directly and you can do a SELECT query
>>>>>>> like this:
>>>>>>>
>>>>>>> SELECT * from METRIC_RECORD WHERE METRIC_NAME = '<your-metric-name>'
>>>>>>> order by SERVER_TIME desc limit 10;
>>>>>>>
>>>>>>>
>>>>>>>  Instructions for connecting to Phoenix:
>>>>>>>
>>>>>>> https://cwiki.apache.org/confluence/display/AMBARI/Phoenix+Schema
>>>>>>>
>>>>>>>
>>>>>>>  3. What API call are you making to get metrics?
>>>>>>>
>>>>>>> E.g.: http://
>>>>>>> <ams-collector>:6188/ws/v1/timeline/metrics?metricNames=<your-metric-name>&startTime=<epoch>&endTime=<epoch>&hostname=<hostname>
>>>>>>>
>>>>>>>
>>>>>>>  -Sid
>>>>>>>
>>>>>>>
>>>>>>>  ------------------------------
>>>>>>> *From:* Bryan Bende <bbende@gmail.com>
>>>>>>> *Sent:* Friday, July 24, 2015 2:03 PM
>>>>>>> *To:* user@ambari.apache.org
>>>>>>> *Subject:* Posting Metrics to Ambari
>>>>>>>
>>>>>>>   I'm interested in sending metrics to Ambari and I've been looking
>>>>>>> at the Metrics Collector REST API described here:
>>>>>>>
>>>>>>> https://cwiki.apache.org/confluence/display/AMBARI/Metrics+Collector+API+Specification
>>>>>>>
>>>>>>>  I figured the easiest way to test it would be to get the latest
>>>>>>> HDP Sandbox... so I downloaded and started it up. The Metrics
Collector
>>>>>>> service wasn't running so I started it, and also added port 6188
to the VM
>>>>>>> port forwarding. From there I used the example POST on the Wiki
page and
>>>>>>> made a successful POST which got a 200 response. After that I
tried the
>>>>>>> query, but could never get any results to come back.
>>>>>>>
>>>>>>>  I know this list is not specific to HDP, but I was wondering
if
>>>>>>> anyone has any suggestions as to what I can look at to figure
out what is
>>>>>>> happening with the data I am posting.
>>>>>>>
>>>>>>>  I was watching the metrics collector log while posting and
>>>>>>> querying and didn't see any activity besides the periodic aggregation.
>>>>>>>
>>>>>>>  Any suggestions would be greatly appreciated.
>>>>>>>
>>>>>>>  Thanks,
>>>>>>>
>>>>>>>  Bran
>>>>>>>
>>>>>>
>>>>>>
>>>>>
>>>>
>>>
>>
>

Mime
View raw message