Ottomata added subscribers: JAllemandou, Ottomata.
Ottomata added a comment.


  COOL! :)
  
  > it's important to note that the state of step 3 is tightly coupled with its 
dump and thus we will have to instantiate a new stream per imported dump. In 
other words a wdqs system imported using dump Y will have to consume the RDF 
stream generated from an initial state based on this same dump. This means that 
the RDF stream will be named against a particular dump instance.
  
  Hm.  Would it be possible instead to lambda architecture this part?  Instead 
of having to reload from a full dump and then recreate a new stream, could 
accomplish the same cleanups by backfilling from a batch job in Hadoop?  I'm 
not sure I fully understand the 'cleanups' here.  Are they not do-able with the 
stream because events representing some of the state changes don't exist (yet)?

TASK DETAIL
  https://phabricator.wikimedia.org/T244590

EMAIL PREFERENCES
  https://phabricator.wikimedia.org/settings/panel/emailpreferences/

To: Ottomata
Cc: Ottomata, JAllemandou, Aklapper, Zbyszko, Gehel, dcausse, darthmon_wmde, 
Nandana, Lahi, Gq86, Lucas_Werkmeister_WMDE, GoranSMilovanovic, QZanden, 
EBjune, merbst, LawExplorer, _jensen, rosalieper, Scott_WUaS, Jonas, Xmlizer, 
jkroll, Smalyshev, Wikidata-bugs, Jdouglas, aude, Tobias1984, Dinoguy1000, 
Manybubbles, Mbch331
_______________________________________________
Wikidata-bugs mailing list
[email protected]
https://lists.wikimedia.org/mailman/listinfo/wikidata-bugs

Reply via email to