What makes it so slow?

Note that you can use wbeditentity to perform complex edits with a single api
call. It's not as streight forward to use as, say, wbaddclaim, but much more
powerfull and efficient.

-- daniel

Am 30.09.2014 19:00, schrieb Andra Waagmeester:
> Hi All,
> 
>       I have joined the development team of the ProteinBoxBot
> (https://www.wikidata.org/wiki/User:ProteinBoxBot) . Our goal is to make
> Wikidata the canonical resource for referencing and translating identifiers 
> for
> genes and proteins from different species.
> 
> Currently adding all genes from the human genome and their related identifiers
> to Wikidata takes more then a month to complete. With the objective to add 
> other
> species, as well as having frequent updates for each of the genomes, it would 
> be
> convenient if we could increase this throughput.
> 
> Would it be accepted if we increase the throughput by running multiple 
> instances
> of ProteinBoxBot in parallel. If so, what would be an accepted number of
> parallel instances of a bot to run? We can run multiple instances from 
> different
> geographical locations if necessary.  
> 
> Kind regards,
> 
> 
> Andra
> 
> 
> 
> 
> _______________________________________________
> Wikidata-l mailing list
> Wikidata-l@lists.wikimedia.org
> https://lists.wikimedia.org/mailman/listinfo/wikidata-l
> 


-- 
Daniel Kinzler
Senior Software Developer

Wikimedia Deutschland
Gesellschaft zur Förderung Freien Wissens e.V.

_______________________________________________
Wikidata-l mailing list
Wikidata-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikidata-l

Reply via email to