storageSize
is the sum of all extents for that data, excluding indexes.
So that collection takes up 2 extents, they are ~2GB each, hence ~4GB. size
includes indexes and I believe a couple of other things which inflate the number. Neither really represents the proper on-disk size. For disk size, db.stats()
has a filesize field which is closer to what you want I think you're looking for.
The manual is somewhat better at outlining what the various fields mean, see here for collections:
http://docs.mongodb.org/manual/reference/collection-statistics/
And here for database stats:
http://docs.mongodb.org/manual/reference/database-statistics/
Some other potentially relevant information:
The compact command does not shrink any datafiles; it only defragments deleted space so that larger objects might reuse it. The compact command will never delete or shrink database files, and in general requires extra space to do its work, usually a minimum of one extra extent.
If you repair the database it will essentially rewrite the data files from scratch, which will remove padding and store them on disk as efficiently as you are going to get. However you will need to have ~2x the size on disk to do so (actually less, but it's a decent guide).
One other thing to bear in mind here - repair and compact remove padding. The padding factor varies between 1 (no moves of documents caused by documents growing), to 2 (lots of moves caused by documents growing). Your padding factor of ~1.67 would indicate you are growing (and hence causing moves) quite a bit.
When you compact or repair a database you remove that padding - subsequent document growth is therefore going to trigger even more moves than before. Because moves are relatiely expensive operations, this can have a serious impact on your performance. More info here:
http://www.mongodb.org/display/DOCS/Padding+Factor
OK, we solved this.
Issue is that mongo auto creates the collection when incoming data comes in.
So the problem was:
- Drop collection
- // transactions are coming in
- Create collection as capped
But #3 comes too late, as the collection is already present, as not capped, because of #2.
Solution:
- stop mongo
- bring up mango on non standard port // to prevent auto creation of collection
- create collection as capped
- stop mongo, bring it back up on regular port
Now I have a capped collection!
Best Answer
Check out https://jira.mongodb.org/browse/SERVER-5615
You might be running into a similar issue. Do you have several indexes on your data? Can you do a db.coll.stats()? And maybe show some output of mongostat while it's running slower?
I've heard similar complaints in the past, but haven't actually used capped collections myself.