Hi,
I wondered what was a good setup for AWS/MongoDB in terms of machines and the sizes of their disks.
Current setup
- 3 micro machines for the config servers, 1 mongos and arbiters. The 8Gb limit is almost reached. (and I ran the arbiters with --nojournal)
- per shard : a replica set of 2 machines m1.large with 8Gb for system + 20Gb for data
- everything is on EBS.
Questions
- is 20Gb too big or too small ? Should I go with 100Gb for example ?
- Am I supposed to inform mongodb about the 20Gb (or other) disk limit ?
- Do you see anything wrong that I don't see ? Im new to mongodb and aws but Im an ok experienced SWE
Plan of use
My database should allow a 100 qps (mostly writes), and should grow up to 1Tb over the next 3 years. The plan is to add as many shards as needed, more or less manually (with scripts), when we see that more memory is needed on the database.
We will also run a few mapreduce over this and have some scripts that do aggregates with the data over the past 15 minutes, every 15 minutes.
We are a very small company, spending up to a few hundred $ per month on our servers would be ok but we can't go crazy on cash.
We hope that we won't have to manually take care of too many machine failures, manually taking care of things once a month would be fine.
Thanks for telling me what you think about that.
Thomas