Addshore added a comment.

> I think there's value in a single shared instance for ease of use


Well, this was also my initial thought. Until Joe said:

> any opsen if given the choice between a 10 minutes downtime of a monitoring 
> tool and dropping old data will choose the latter


@JanZerebecki I imagine that wherever we put the data, we will backup.
Right now they are mainly stored in SQL tables... which I backup / export.
If they were started in graphite I would likely run a daily / weekly export too.
In-fact even if they were primarily in HDFS I would still want them backed up 
elsewhere.

> Which one should we use for the Wikidata related metrics that are not 
> generated in Hadoop?


Well, as it stands basically no metrics are generated in hadoop.

We simply want to store time series data / numbers and timestamps. Have them 
easily writeable, accessible, backed up, integrated into other solutions. 
Graphite + a backup script / weekly / daily export does all of this.

If months down the line we decide on a potential better solution that the 
analytics team may also like migrating to it should be trivial.


TASK DETAIL
  https://phabricator.wikimedia.org/T117732

EMAIL PREFERENCES
  https://phabricator.wikimedia.org/settings/panel/emailpreferences/

To: Addshore
Cc: Joe, Lydia_Pintscher, fgiunchedi, Christopher, JanZerebecki, Nuria, 
Ottomata, Aklapper, Addshore, StudiesWorld, Wikidata-bugs, aude, Mbch331



_______________________________________________
Wikidata-bugs mailing list
Wikidata-bugs@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata-bugs

Reply via email to