Does anyone have any formulas, or maybe some sample data from their environment that can help me estimate how much disk space will be used by graphite per datapoint?
Disk Capacity Planning for Whisper / Graphite
graphitemonitoring
graphitemonitoring
Does anyone have any formulas, or maybe some sample data from their environment that can help me estimate how much disk space will be used by graphite per datapoint?
Best Answer
whisper-info.py
gives you a lot of insight into what and how each file is aggregated, including the file's size.However it's only useful for existing whisper files.
When you want to see predictive sizing of a schema before putting it in place, try a Whisper Calculator, such as the one available at https://gist.github.com/jjmaestro/5774063
EDIT:
When asked for an example...
storage_schema:
Looking at my file
applied-in-last-hour.wsp
,ls -l
yieldsand
whisper-info.py ./applied-in-last-hour.wsp
yieldsSo, basically you combine your hosts per retention-match per retention-period-segment per stat, multiply by a factor of systems that you intend to apply this too, factor in the number of new stats that you're going to track. Then you take whatever amount of storage that is and at least double it (because we're buying storage, and we know we'll use it...)