jackrabbit-users mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Vijay Pandey <VPan...@mdes.ms.gov>
Subject RE: Large DMS - 43 GB - Out Of Memory Errors
Date Thu, 08 Jan 2009 18:08:23 GMT
We got the jackrabbit 1.5 upgraded without any issue and did the test on 1.5

on a smaller repository of size around 20 GB.

This is how our content is setup

a) Root Node
		Node 1
			This node has got around 80000 flat(non-hierchical)
child 			nodes and each of these child nodes have around 5 to
20 			set of PDF docs (size varies from 15kb to 70kb)
along with 			several other properties (around 12 - string
type 				properties)
		Node 2 -- 
		Node 3
		Node 4
		Node 5

We are using JackRabbit through RMI -- we have 2 separate applications
accessing JackRabbit.

This is what we found out through YourKit Java Profiler. We ran this in test
environment with only 1 user connected to JackRabbit.

a) When we execute this piece of code on client side through RMI
	Just on the above operation we saw a spike of around 37 MB memory on
the heap.

After doing the force garbage collection (through yourkit profiler) – we saw
around 18 MB memory being reclaimed, but still there was around 19 MB memory
being used in heap, we ran the snapshot to see which objects have claimed so
much of heap. This is what we found out.

		Sun.rmi.transport.WeakRef --- has the size of around 19 MB
with the following structure.

Org.apache.jackrabbit.core.state.ChildNodeEntries ----- this one occupied 19

After doing some more force GC, around 10 mints later this memory was

Is this issue just because we have too many nodes (non-hierarchical) or it’s
a combination of too many non-hierarchical nodes and the JackRabbit RMI

Any hint how we can get over this problem will be great.


-----Original Message-----
From: Jukka Zitting [mailto:jukka.zitting@gmail.com] 
Sent: Tuesday, January 06, 2009 3:13 AM
To: users@jackrabbit.apache.org
Subject: Re: Large DMS - 43 GB - Out Of Memory Errors


On Tue, Jan 6, 2009 at 9:32 AM, Thomas Müller <thomas.mueller@day.com>
> Upgrading is a good idea, you will have to do a full export/import

Only if you decide to start using the DataStore or bundle persistence
features, both of which should provide nice performance benefits in
this case.

But you should be able to upgrade to Jackrabbit 1.5 simply by
replacing all the relevant jar files (note that 1.5 uses quite a bit
more jars than 1.0.1). Your existing configuration files and
repository data should work fine with Jackrabbit 1.5.


Jukka Zitting



We are using Jackrabbit 1.0.1 now for more than 2 years. Currently we are
facing a lot of issues related to the Out Of Memory exception, goes down
around 3 times a day. These are the relevant details of the DMS.


a)       JackRabbit 1.0.1 - set up as a standalone RMI server

b)       Its primarily used to store PDF documents ( size vary from 50 KB to
100 KB for each document) in binary format

c)       We have a nightly java batch programs that inserts around 5000 to
6000 of these documents every day

d)       During the day time these documents are accessed by the user from a
web application - we connect through RMI to the jackrabbit server and fetch
the document

e)       The database used is DB2 and the documents ( blobs) are stored on
the file system ( setting for externalblobs is true)

f)         The total size of the repository has reached around 43 GB


Will really appreciate if you all can throw some light on this and what can
be done to resolve this issue?


I am also thinking of migrating to 1.5 Jackrabbit, what I have read is other
than various jar file changes; it should be compatible to the 1.x release? I
also read about the DataStore, but to use it do I have to do the full




View raw message