couchdb-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Diogo Moitinho de Almeida <>
Subject Considering CouchDB
Date Tue, 19 Nov 2013 21:16:44 GMT

Based on the research that I've done, CouchDB seems like a very good fit 
for the problem I'm trying to solve, but when talking to people from 
within the company, they're expressed that there are some unadvertised 
down sides to it (though they tried using it 3 years ago) and that we 
would have to migrate fairly quickly off of it.

The worries that they expressed were:
1. Scalability (with a stop the world lock when writing to disk).
2. Erlang stack traces not being the easiest to deal with. (I assumed 
with the "crash only" design, that common problems could be solved by 
restarting a node).
3. Garbage collection being a very slow process.

And questions that I had were:
1. Would anyone have approximate numbers on how it scales? In terms of 
writes, size of the data, and cluster size? How big of a cluster have 
people gotten working without issues? Have people tested this out in the 
petabyte range?
2. Would anyone have approximate times on how long a map reduce query 
could take (with whatever sized data, I'm just curious)?

The use that I'd be looking at is about 200 writes/second of documents 
<50 kB each with updates fairly rarely, though if everything goes well, 
the documents could be a bit larger with a more writes and a lot more 

Thanks in advance for any help,

View raw message