I'm using couchdb since a while. I love that, but I was always working with a stand alone configuration.
Now I'm trying to set up a serious system made up of a "federation" of coucdbs. (i.e. a cluster). I understood that the most proper tool to use is coudhb-lounge but now I've to understand if such architecture is sustainable. In particular in terms of amount of data.
I tryed to get as much information from the following sources One way replication with CouchDB, What is the CouchDB replication protocol? Is it like Git?, CouchDB Replication Protocol , but I still feel I need some support.
So, as a first set of question, I'm trying to understand:
- I searched a bit around and from this post and this post I understood that it works over HTTP. Is it really efficient for huge amounts of data?
- is it possible to replicate 10Gb of data each day without encountering any serious load issue?
- In case it is slow, how much erlang programming would be needed to change couchdb replication protocol? Would I have to seriously hack the code or couchdb envisage also to use different protocols?
thanks