It should be relatively easy to catch a significant percentage of those
copyright violations with the assistance of automated search tools. The
trick is to do it at a large scale in near-realtime, which might require
some computationally intensive and bandwidth intensive work. James, can I
suggest
Isn't that what Corenbot does/did? I always found it very confusing though
whenever I ran into it, and the false positives are huge (so many sites
copy Wikimedia content these days)
On Mon, Jul 21, 2014 at 9:11 AM, Pine W wiki.p...@gmail.com wrote:
It should be relatively easy to catch a
This has nothing to do with wiki-research-l, it's mostly about who can
pay for the API keys. https://wikitech.wikimedia.org/wiki/Web_search
Nemo
___
Wikimedia-l mailing list, guidelines at:
https://meta.wikimedia.org/wiki/Mailing_lists/Guidelines
hi james, this sounds interesting. the only obstacle i see is that neither
the service nor the software nor the database are open. by using turnitin
we would give up on the principle of opening contents, at least in this
area. this might be perceived bad for opening content, editor retention,
and
I dont know if we have to work with Turn-it-in per se. Given that most
copy-paste is from digital sources, we could start with a program/tool to
search the web, which could lessen the burden on editors significantly by
itself. (Speaking as a writing teacher, I can tell you this is how I find
The report on Wikimedia Nederland activities in June is available:
https://meta.wikimedia.org/wiki/Wikimedia_chapters/Reports/Wikimedia_Nederland/201406
It is also included as text in this message.
Sandra Rientjes
Directeur/Executive Director Wikimedia Nederland
*COMMUNITY: supporting and