activemq-dev mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Bruce Snyder (JIRA)" <>
Subject [jira] Commented: (AMQ-1974) MemoryUsage stays at 100% used forever after all messages are sent
Date Tue, 14 Oct 2008 19:43:52 GMT


Bruce Snyder commented on AMQ-1974:

I've created Maven projects to run each of these clients in their own VM using the exec-maven-plugin.
Each one has now been running for a while (nearly 4 million messages have been sent) and I'm
not seeing any errors at all. Am I missing something? 

I'm seeing the following output from the TryJmsManager and it doesn't seem to be changing
at all: 

ActiveMQ memeory 0 1024
ActiveMQ message store 0
ActiveMQ temp space 0
sent message 3946000 (id=ID:mongoose.local-50509-1224010696716-3:0:1:1:3946001) after 2630643
sent message 3947000 (id=ID:mongoose.local-50509-1224010696716-3:0:1:1:3947001) after 2631294
sent message 3948000 (id=ID:mongoose.local-50509-1224010696716-3:0:1:1:3948001) after 2631974
sent message 3949000 (id=ID:mongoose.local-50509-1224010696716-3:0:1:1:3949001) after 2632658
sent message 3950000 (id=ID:mongoose.local-50509-1224010696716-3:0:1:1:3950001) after 2633326
sent message 3951000 (id=ID:mongoose.local-50509-1224010696716-3:0:1:1:3951001) after 2633965
sent message 3952000 (id=ID:mongoose.local-50509-1224010696716-3:0:1:1:3952001) after 2634663

Here is the output I'm seeing from the TryJmsClient: 

ot message ID:mongoose.local-50509-1224010696716-3:0:1:1:3994202
got message ID:mongoose.local-50509-1224010696716-3:0:1:1:3994203
got message ID:mongoose.local-50509-1224010696716-3:0:1:1:3994204
got message ID:mongoose.local-50509-1224010696716-3:0:1:1:3994205
got message ID:mongoose.local-50509-1224010696716-3:0:1:1:3994206
got message ID:mongoose.local-50509-1224010696716-3:0:1:1:3994207
got message ID:mongoose.local-50509-1224010696716-3:0:1:1:3994208
got message ID:mongoose.local-50509-1224010696716-3:0:1:1:3994209
got message ID:mongoose.local-50509-1224010696716-3:0:1:1:3994210
got message ID:mongoose.local-50509-1224010696716-3:0:1:1:3994211
got message ID:mongoose.local-50509-1224010696716-3:0:1:1:3994212
got message ID:mongoose.local-50509-1224010696716-3:0:1:1:3994213
got message ID:mongoose.local-50509-1224010696716-3:0:1:1:3994214
got message ID:mongoose.local-50509-1224010696716-3:0:1:1:3994215
got message ID:mongoose.local-50509-1224010696716-3:0:1:1:3994216
got message ID:mongoose.local-50509-1224010696716-3:0:1:1:3994217

> MemoryUsage stays at 100% used forever after all messages are sent
> ------------------------------------------------------------------
>                 Key: AMQ-1974
>                 URL:
>             Project: ActiveMQ
>          Issue Type: Bug
>    Affects Versions: 5.1.0
>         Environment: Linux with 2.6.18 kernel
> JDK 1.5.0_15
>            Reporter: Neal Yin
>            Priority: Critical
>         Attachments: MsgsPerSecond.diff,,
> We  use a network of brokers and topic to do messaging. All our message producers and
consumers talk to the local VM broker. We reply on bridges between network of brokers to move
messages between brokers. We found a case that MemoryUsage stays at 100% forever after all
messages are sent.
> I attached two files to show this memory leak problem. TryJmsClient keeps sending persistent
messages to a Topic. TryJmsManager has a message consumer. Start TryJmsManager first and then
start TryJmsClient. After running about 10 seconds, TryJmsClient stops sending because of
getting "SystemUsage memory limit reached" error. About 24K messages are sent and consumer
shows it get all messages. Now examine ActiveMQ MemoryUsage printout of TryJmsClient, MemoryUsage
percentage stays at 100% or 99% forever. Producer can never send any more messages.
> Note1: If I slow down sending by adding a sleep, I don't see memory leak immediately.
I am not sure for longer running time.
> Note2: If I use Queue instead of Topic, there is no memory leak. 
> Note3: This looks like similar to AMQ-1833 that is fixed in 5.2. But after testing with
5.2 snapshot, I still get the problem.

This message is automatically generated by JIRA.
You can reply to this email to add a comment to the issue online.

View raw message