acd
New Member
I was talking to Fran this evening about statistics and logging for his ... pony... thing and we discussed some of the technical problems involved and solutions thereof; right now, he's considering a custom solution and I suggested building it off a nosql solution like leveldb. A major issue is write throughput to the tune of ~25 parameters per sample, 65+ samples per second. He tried running it backed by mongodb but trying with just a million samples, it choked and would need some fo-serious ssd backing to make it happen. Leveldb didn't have that problem (1M inserts in about 40 seconds inc. json parsing), but has limitations on the number of processes that can have the db open; namely just 1, unless you write a daemon on top of it.
We discussed the advantages and disadvantages of rrdtool & mtrg based solutions, and their derivatives like Graphite (which, if you haven't looked at it, is actually pretty sweet). Before one of us dives off the deep end and starts building a custom solution from scratch I want to know what you guys use for this sort of thing and what do you store? How long do you retain data and how frequently do you collect it? Maybe there's an off-the-shelf solution that's a good fit for one or the other of us.
We discussed the advantages and disadvantages of rrdtool & mtrg based solutions, and their derivatives like Graphite (which, if you haven't looked at it, is actually pretty sweet). Before one of us dives off the deep end and starts building a custom solution from scratch I want to know what you guys use for this sort of thing and what do you store? How long do you retain data and how frequently do you collect it? Maybe there's an off-the-shelf solution that's a good fit for one or the other of us.