That's very cool! To get an idea, how big is your dataset? On Tue Sep 30 2014 at 12:06:56 PM Daniel Kinzler < [email protected]> wrote:
> What makes it so slow? > > Note that you can use wbeditentity to perform complex edits with a single > api > call. It's not as streight forward to use as, say, wbaddclaim, but much > more > powerfull and efficient. > > -- daniel > > Am 30.09.2014 19:00, schrieb Andra Waagmeester: > > Hi All, > > > > I have joined the development team of the ProteinBoxBot > > (https://www.wikidata.org/wiki/User:ProteinBoxBot) . Our goal is to make > > Wikidata the canonical resource for referencing and translating > identifiers for > > genes and proteins from different species. > > > > Currently adding all genes from the human genome and their related > identifiers > > to Wikidata takes more then a month to complete. With the objective to > add other > > species, as well as having frequent updates for each of the genomes, it > would be > > convenient if we could increase this throughput. > > > > Would it be accepted if we increase the throughput by running multiple > instances > > of ProteinBoxBot in parallel. If so, what would be an accepted number of > > parallel instances of a bot to run? We can run multiple instances from > different > > geographical locations if necessary. > > > > Kind regards, > > > > > > Andra > > > > > > > > > > _______________________________________________ > > Wikidata-l mailing list > > [email protected] > > https://lists.wikimedia.org/mailman/listinfo/wikidata-l > > > > > -- > Daniel Kinzler > Senior Software Developer > > Wikimedia Deutschland > Gesellschaft zur Förderung Freien Wissens e.V. > > _______________________________________________ > Wikidata-l mailing list > [email protected] > https://lists.wikimedia.org/mailman/listinfo/wikidata-l >
_______________________________________________ Wikidata-l mailing list [email protected] https://lists.wikimedia.org/mailman/listinfo/wikidata-l
