Addshore added a comment. > I think there's value in a single shared instance for ease of use
Well, this was also my initial thought. Until Joe said: > any opsen if given the choice between a 10 minutes downtime of a monitoring > tool and dropping old data will choose the latter @JanZerebecki I imagine that wherever we put the data, we will backup. Right now they are mainly stored in SQL tables... which I backup / export. If they were started in graphite I would likely run a daily / weekly export too. In-fact even if they were primarily in HDFS I would still want them backed up elsewhere. > Which one should we use for the Wikidata related metrics that are not > generated in Hadoop? Well, as it stands basically no metrics are generated in hadoop. We simply want to store time series data / numbers and timestamps. Have them easily writeable, accessible, backed up, integrated into other solutions. Graphite + a backup script / weekly / daily export does all of this. If months down the line we decide on a potential better solution that the analytics team may also like migrating to it should be trivial. TASK DETAIL https://phabricator.wikimedia.org/T117732 EMAIL PREFERENCES https://phabricator.wikimedia.org/settings/panel/emailpreferences/ To: Addshore Cc: Joe, Lydia_Pintscher, fgiunchedi, Christopher, JanZerebecki, Nuria, Ottomata, Aklapper, Addshore, StudiesWorld, Wikidata-bugs, aude, Mbch331 _______________________________________________ Wikidata-bugs mailing list Wikidata-bugs@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikidata-bugs