Re: [Wikitech-l] Using mediawiki from within the Social networks?
On Wed, Oct 3, 2012 at 7:43 AM, Yury Katkov katkov.ju...@gmail.com wrote: I'm not sure that '''editing''' can be made more easy with the help of social network client. Any ideas on that? any ideas on what else can be made more engaging with the power of social networks? Well what if people can click on a bit of text and comment on it, they could suggest in that comment that the text is replaced. mike -- James Michael DuPont Member of Free Libre Open Source Software Kosova http://flossk.org Saving wikipedia(tm) articles from deletion http://SpeedyDeletion.wikia.com Contributor FOSM, the CC-BY-SA map of the world http://fosm.org Mozilla Rep https://reps.mozilla.org/u/h4ck3rm1k3 Free Software Foundation Europe Fellow http://fsfe.org/support/?h4ck3rm1k3 ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l
Re: [Wikitech-l] Using mediawiki from within the Social networks?
On Wed, Oct 3, 2012 at 1:03 PM, Yury Katkov katkov.ju...@gmail.com wrote: hmmm, much like the blips on Google Wave? like line comments in github : https://github.com/h4ck3rm1k3/wikiteam/commit/4da7f7f4a813b53be13bff7e29a1e5325bb68a30#L0R58 We need a way to track and rate comments and then we can resolve them with changes after people have stormed over them. -- James Michael DuPont Member of Free Libre Open Source Software Kosova http://flossk.org Saving wikipedia(tm) articles from deletion http://SpeedyDeletion.wikia.com Contributor FOSM, the CC-BY-SA map of the world http://fosm.org Mozilla Rep https://reps.mozilla.org/u/h4ck3rm1k3 Free Software Foundation Europe Fellow http://fsfe.org/support/?h4ck3rm1k3 ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l
Re: [Wikitech-l] Github replication
On Wed, Oct 3, 2012 at 6:27 PM, Chad innocentkil...@gmail.com wrote: Hi everyone, Just letting everyone know: mediawiki/core is now replicating from gerrit to github. https://github.com/mediawiki/core that is great news. mike ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l
Re: [Wikitech-l] Using mediawiki from within the Social networks?
well what about the user granting you the rights to the comments etc under certain conditions and they just use fb and pull that data in via the api. so you would have some pages that are autogenerated via fb, you might have Talk:page for normal users and FB:Page for fb users. mike On Tue, Oct 2, 2012 at 1:21 AM, Yury Katkov katkov.ju...@gmail.com wrote: Maybe we can brainstorm this idea a little bit together? I have several crowdsourcing projects and I always suspect that the pure wiki editing requires too much motivation and it's not the best approach in many cases. - Yury Katkov On Tue, Oct 2, 2012 at 1:16 AM, Mike Dupont jamesmikedup...@googlemail.com wrote: Yuri, I agree with you 100% this is a good idea and would help out alot. I have also had similar ideas! mike On Mon, Oct 1, 2012 at 11:09 PM, Yury Katkov katkov.ju...@gmail.com wrote: There is no such client to the best of my knowledge. But here I don't think much about the big-text-projects like wikipedia: more about the projects where users need to - vote for the articles or - add really small fragments of text - or link page together via Semantic MediaWiki properties or - categorize and tagged the articles... these are all very small activities. The idea is that those activities can be shown in the social network MW-client in the most easy-to-consume way to the very broad audience. Or maybe they can be even gamified! - Yury Katkov On Tue, Oct 2, 2012 at 1:02 AM, Mike Dupont jamesmikedup...@googlemail.com wrote: On Mon, Oct 1, 2012 at 10:37 PM, Yury Katkov katkov.ju...@gmail.com wrote: MediaWiki social network client tell me more, links? -- James Michael DuPont Member of Free Libre Open Source Software Kosova http://flossk.org Saving wikipedia(tm) articles from deletion http://SpeedyDeletion.wikia.com Contributor FOSM, the CC-BY-SA map of the world http://fosm.org Mozilla Rep https://reps.mozilla.org/u/h4ck3rm1k3 Free Software Foundation Europe Fellow http://fsfe.org/support/?h4ck3rm1k3 ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l -- James Michael DuPont Member of Free Libre Open Source Software Kosova http://flossk.org Saving wikipedia(tm) articles from deletion http://SpeedyDeletion.wikia.com Contributor FOSM, the CC-BY-SA map of the world http://fosm.org Mozilla Rep https://reps.mozilla.org/u/h4ck3rm1k3 Free Software Foundation Europe Fellow http://fsfe.org/support/?h4ck3rm1k3 ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l -- James Michael DuPont Member of Free Libre Open Source Software Kosova http://flossk.org Saving wikipedia(tm) articles from deletion http://SpeedyDeletion.wikia.com Contributor FOSM, the CC-BY-SA map of the world http://fosm.org Mozilla Rep https://reps.mozilla.org/u/h4ck3rm1k3 Free Software Foundation Europe Fellow http://fsfe.org/support/?h4ck3rm1k3 ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l
Re: [Wikitech-l] Using mediawiki from within the Social networks?
have you seen the customization that wikia.com has done? it is a step in the right direction for social integration. On Mon, Oct 1, 2012 at 8:24 PM, Chris Steipp cste...@wikimedia.org wrote: It's possible, but it would be a little ugly. You could write a web frontend to what would basically be a proxy handling the work on the backend over the api. However, there is no secure way for a user to login through something like that, so all of the edits would have to come from your app and not the user. Alternatively, you could write a javascript (or flash) application, and get it approved to make CORS/crossdomain requests to the target MediaWiki api, so the user would be making api calls directly. Getting the wiki admin to approve domains, or add a crossdomain.xml file may or may not be a problem for you. OAuth will make the delegated authorization easier, but unfortunately the priority for getting OAuth integrated with MediaWiki is lower than some other critical projects right now, so it will be a few months before this is possible. On Mon, Oct 1, 2012 at 10:44 AM, Yury Katkov katkov.ju...@gmail.com wrote: Hi everyone! Is it possible to use MediaWiki as a service whereas the UI is located on a Facebook app? So all the editing and viewing is take place on a Facebook and MediaWiki provide the storage, revision control and lots of extensions? - Yury Katkov ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l -- James Michael DuPont Member of Free Libre Open Source Software Kosova http://flossk.org Saving wikipedia(tm) articles from deletion http://SpeedyDeletion.wikia.com Contributor FOSM, the CC-BY-SA map of the world http://fosm.org Mozilla Rep https://reps.mozilla.org/u/h4ck3rm1k3 Free Software Foundation Europe Fellow http://fsfe.org/support/?h4ck3rm1k3 ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l
Re: [Wikitech-l] Using mediawiki from within the Social networks?
On Mon, Oct 1, 2012 at 10:37 PM, Yury Katkov katkov.ju...@gmail.com wrote: MediaWiki social network client tell me more, links? -- James Michael DuPont Member of Free Libre Open Source Software Kosova http://flossk.org Saving wikipedia(tm) articles from deletion http://SpeedyDeletion.wikia.com Contributor FOSM, the CC-BY-SA map of the world http://fosm.org Mozilla Rep https://reps.mozilla.org/u/h4ck3rm1k3 Free Software Foundation Europe Fellow http://fsfe.org/support/?h4ck3rm1k3 ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l
Re: [Wikitech-l] Using mediawiki from within the Social networks?
Yuri, I agree with you 100% this is a good idea and would help out alot. I have also had similar ideas! mike On Mon, Oct 1, 2012 at 11:09 PM, Yury Katkov katkov.ju...@gmail.com wrote: There is no such client to the best of my knowledge. But here I don't think much about the big-text-projects like wikipedia: more about the projects where users need to - vote for the articles or - add really small fragments of text - or link page together via Semantic MediaWiki properties or - categorize and tagged the articles... these are all very small activities. The idea is that those activities can be shown in the social network MW-client in the most easy-to-consume way to the very broad audience. Or maybe they can be even gamified! - Yury Katkov On Tue, Oct 2, 2012 at 1:02 AM, Mike Dupont jamesmikedup...@googlemail.com wrote: On Mon, Oct 1, 2012 at 10:37 PM, Yury Katkov katkov.ju...@gmail.com wrote: MediaWiki social network client tell me more, links? -- James Michael DuPont Member of Free Libre Open Source Software Kosova http://flossk.org Saving wikipedia(tm) articles from deletion http://SpeedyDeletion.wikia.com Contributor FOSM, the CC-BY-SA map of the world http://fosm.org Mozilla Rep https://reps.mozilla.org/u/h4ck3rm1k3 Free Software Foundation Europe Fellow http://fsfe.org/support/?h4ck3rm1k3 ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l -- James Michael DuPont Member of Free Libre Open Source Software Kosova http://flossk.org Saving wikipedia(tm) articles from deletion http://SpeedyDeletion.wikia.com Contributor FOSM, the CC-BY-SA map of the world http://fosm.org Mozilla Rep https://reps.mozilla.org/u/h4ck3rm1k3 Free Software Foundation Europe Fellow http://fsfe.org/support/?h4ck3rm1k3 ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l
Re: [Wikitech-l] Lua deployed to www.mediawiki.org
On Thu, Aug 23, 2012 at 2:00 AM, Tim Starling tstarl...@wikimedia.orgwrote: I think there will be some changes to template invocations. For example, a typical coord invocation looks like this: {{Coord|33|51|35.9|S|151|12|40|E}} With the string processing facilities that Lua provides, that might change to: {{Coord|33°51'35.9S 151°12'40E}} Of course, backwards compatibility would need to be maintained, but that's easy enough. HI all, I can understand that some people might be upset about a new language and change. But in the defense of lua proponents, it is not like the old templates are being deleted here, there are millions of copies of them and they are not going away. Also media wiki as a language is difficult to process, using a standard language for tempaltes might help out. So a migration path that I would think is reasonable and I am sure has been proposed is to make first a compatibility layer that allows old templates to be used in lua and make some evalold function to process old template code exactly as it was. thanks mike -- James Michael DuPont Member of Free Libre Open Source Software Kosova http://flossk.org http://flossk.orgSaving wikipedia(tm) articles from deletion http://SpeedyDeletion.wikia.com Contributor FOSM, the CC-BY-SA map of the world http://fosm.org Mozilla Rep https://reps.mozilla.org/u/h4ck3rm1k3 ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l
Re: [Wikitech-l] Lua deployed to www.mediawiki.org
On Thu, Aug 23, 2012 at 4:18 AM, MZMcBride z...@mzmcbride.com wrote: This is my concern with Lua modules, in a nutshell. So please tell me, what are the options to fix this? Is there going to be a common code repo and maybe an easy way to sync in a git filesystem of template code into the wiki? Maybe a git extension? or is there one already? mike -- James Michael DuPont Member of Free Libre Open Source Software Kosova http://flossk.org http://flossk.orgSaving wikipedia(tm) articles from deletion http://SpeedyDeletion.wikia.com Contributor FOSM, the CC-BY-SA map of the world http://fosm.org Mozilla Rep https://reps.mozilla.org/u/h4ck3rm1k3 ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l
Re: [Wikitech-l] [Xmldatadumps-l] XML dumps/Media mirrors update
I have run cron archiving now every 30 minutes, http://ia700802.us.archive.org/34/items/wikipedia-delete-2012-06/ it is amazing how fast the stuff gets deleted on wikipedia. what about the proposed deletes are there categories for that? thanks mike On Wed, May 30, 2012 at 6:26 AM, Mike Dupont jamesmikedup...@googlemail.com wrote: https://github.com/h4ck3rm1k3/wikiteam code here On Wed, May 30, 2012 at 6:26 AM, Mike Dupont jamesmikedup...@googlemail.com wrote: Ok, I merged the code from wikteam and have a full history dump script that uploads to archive.org, next step is to fix the bucket metadata in the script mike On Tue, May 29, 2012 at 3:08 AM, Mike Dupont jamesmikedup...@googlemail.com wrote: Well, I have now updated the script to include the xml dump in raw format. I will have to add more information the achive.org item, at least a basic readme. other thing is that the wikipybot does not support the full history it seems, so that I will have to move over to the wikiteam version and rework it, I just spent 2 hours on this so i am pretty happy for the first version. mike On Tue, May 29, 2012 at 1:52 AM, Hydriz Wikipedia ad...@alphacorp.tk wrote: This is quite nice, though the item's metadata is too little :) On Tue, May 29, 2012 at 3:40 AM, Mike Dupont jamesmikedup...@googlemail.com wrote: first version of the Script is ready , it gets the versions, puts them in a zip and puts that on archive.org https://github.com/h4ck3rm1k3/pywikipediabot/blob/master/export_deleted.py here is an example output : http://archive.org/details/wikipedia-delete-2012-05 http://ia601203.us.archive.org/24/items/wikipedia-delete-2012-05/archive2012-05-28T21:34:02.302183.zip I will cron this, and it should give a start of saving deleted data. Articles will be exported once a day, even if they they were exported yesterday as long as they are in one of the categories. mike On Mon, May 21, 2012 at 7:21 PM, Mike Dupont jamesmikedup...@googlemail.com wrote: Thanks! and run that 1 time per day, they dont get deleted that quickly. mike On Mon, May 21, 2012 at 9:11 PM, emijrp emi...@gmail.com wrote: Create a script that makes a request to Special:Export using this category as feed https://en.wikipedia.org/wiki/Category:Candidates_for_speedy_deletion More info https://www.mediawiki.org/wiki/Manual:Parameters_to_Special:Export 2012/5/21 Mike Dupont jamesmikedup...@googlemail.com Well I whould be happy for items like this : http://en.wikipedia.org/wiki/Template:Db-a7 would it be possible to extract them easily? mike On Thu, May 17, 2012 at 2:23 PM, Ariel T. Glenn ar...@wikimedia.org wrote: There's a few other reasons articles get deleted: copyright issues, personal identifying data, etc. This makes maintaning the sort of mirror you propose problematic, although a similar mirror is here: http://deletionpedia.dbatley.com/w/index.php?title=Main_Page The dumps contain only data publically available at the time of the run, without deleted data. The articles aren't permanently deleted of course. The revisions texts live on in the database, so a query on toolserver, for example, could be used to get at them, but that would need to be for research purposes. Ariel Στις 17-05-2012, ημέρα Πεμ, και ώρα 13:30 +0200, ο/η Mike Dupont έγραψε: Hi, I am thinking about how to collect articles deleted based on the not notable criteria, is there any way we can extract them from the mysql binlogs? how are these mirrors working? I would be interested in setting up a mirror of deleted data, at least that which is not spam/vandalism based on tags. mike On Thu, May 17, 2012 at 1:09 PM, Ariel T. Glenn ar...@wikimedia.org wrote: We now have three mirror sites, yay! The full list is linked to from http://dumps.wikimedia.org/ and is also available at http://meta.wikimedia.org/wiki/Mirroring_Wikimedia_project_XML_dumps#Current_Mirrors Summarizing, we have: C3L (Brazil) with the last 5 good known dumps, Masaryk University (Czech Republic) with the last 5 known good dumps, Your.org (USA) with the complete archive of dumps, and for the latest version of uploaded media, Your.org with http/ftp/rsync access. Thanks to Carlos, Kevin and Yenya respectively at the above sites for volunteering space, time and effort to make this happen. As people noticed earlier, a series of media tarballs per-project (excluding commons) is being generated. As soon as the first run of these is complete we'll announce its location and start generating them on a semi-regular basis. As we've been getting the bugs out of the mirroring setup, it is getting easier to add new locations. Know anyone interested? Please let us know; we would love to have them. Ariel
Re: [Wikitech-l] [Xmldatadumps-l] XML dumps/Media mirrors update
Ok, I merged the code from wikteam and have a full history dump script that uploads to archive.org, next step is to fix the bucket metadata in the script mike On Tue, May 29, 2012 at 3:08 AM, Mike Dupont jamesmikedup...@googlemail.com wrote: Well, I have now updated the script to include the xml dump in raw format. I will have to add more information the achive.org item, at least a basic readme. other thing is that the wikipybot does not support the full history it seems, so that I will have to move over to the wikiteam version and rework it, I just spent 2 hours on this so i am pretty happy for the first version. mike On Tue, May 29, 2012 at 1:52 AM, Hydriz Wikipedia ad...@alphacorp.tk wrote: This is quite nice, though the item's metadata is too little :) On Tue, May 29, 2012 at 3:40 AM, Mike Dupont jamesmikedup...@googlemail.com wrote: first version of the Script is ready , it gets the versions, puts them in a zip and puts that on archive.org https://github.com/h4ck3rm1k3/pywikipediabot/blob/master/export_deleted.py here is an example output : http://archive.org/details/wikipedia-delete-2012-05 http://ia601203.us.archive.org/24/items/wikipedia-delete-2012-05/archive2012-05-28T21:34:02.302183.zip I will cron this, and it should give a start of saving deleted data. Articles will be exported once a day, even if they they were exported yesterday as long as they are in one of the categories. mike On Mon, May 21, 2012 at 7:21 PM, Mike Dupont jamesmikedup...@googlemail.com wrote: Thanks! and run that 1 time per day, they dont get deleted that quickly. mike On Mon, May 21, 2012 at 9:11 PM, emijrp emi...@gmail.com wrote: Create a script that makes a request to Special:Export using this category as feed https://en.wikipedia.org/wiki/Category:Candidates_for_speedy_deletion More info https://www.mediawiki.org/wiki/Manual:Parameters_to_Special:Export 2012/5/21 Mike Dupont jamesmikedup...@googlemail.com Well I whould be happy for items like this : http://en.wikipedia.org/wiki/Template:Db-a7 would it be possible to extract them easily? mike On Thu, May 17, 2012 at 2:23 PM, Ariel T. Glenn ar...@wikimedia.org wrote: There's a few other reasons articles get deleted: copyright issues, personal identifying data, etc. This makes maintaning the sort of mirror you propose problematic, although a similar mirror is here: http://deletionpedia.dbatley.com/w/index.php?title=Main_Page The dumps contain only data publically available at the time of the run, without deleted data. The articles aren't permanently deleted of course. The revisions texts live on in the database, so a query on toolserver, for example, could be used to get at them, but that would need to be for research purposes. Ariel Στις 17-05-2012, ημέρα Πεμ, και ώρα 13:30 +0200, ο/η Mike Dupont έγραψε: Hi, I am thinking about how to collect articles deleted based on the not notable criteria, is there any way we can extract them from the mysql binlogs? how are these mirrors working? I would be interested in setting up a mirror of deleted data, at least that which is not spam/vandalism based on tags. mike On Thu, May 17, 2012 at 1:09 PM, Ariel T. Glenn ar...@wikimedia.org wrote: We now have three mirror sites, yay! The full list is linked to from http://dumps.wikimedia.org/ and is also available at http://meta.wikimedia.org/wiki/Mirroring_Wikimedia_project_XML_dumps#Current_Mirrors Summarizing, we have: C3L (Brazil) with the last 5 good known dumps, Masaryk University (Czech Republic) with the last 5 known good dumps, Your.org (USA) with the complete archive of dumps, and for the latest version of uploaded media, Your.org with http/ftp/rsync access. Thanks to Carlos, Kevin and Yenya respectively at the above sites for volunteering space, time and effort to make this happen. As people noticed earlier, a series of media tarballs per-project (excluding commons) is being generated. As soon as the first run of these is complete we'll announce its location and start generating them on a semi-regular basis. As we've been getting the bugs out of the mirroring setup, it is getting easier to add new locations. Know anyone interested? Please let us know; we would love to have them. Ariel ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l -- James Michael DuPont Member of Free Libre Open Source Software Kosova http://flossk.org Contributor FOSM, the CC-BY-SA map
Re: [Wikitech-l] [Xmldatadumps-l] XML dumps/Media mirrors update
https://github.com/h4ck3rm1k3/wikiteam code here On Wed, May 30, 2012 at 6:26 AM, Mike Dupont jamesmikedup...@googlemail.com wrote: Ok, I merged the code from wikteam and have a full history dump script that uploads to archive.org, next step is to fix the bucket metadata in the script mike On Tue, May 29, 2012 at 3:08 AM, Mike Dupont jamesmikedup...@googlemail.com wrote: Well, I have now updated the script to include the xml dump in raw format. I will have to add more information the achive.org item, at least a basic readme. other thing is that the wikipybot does not support the full history it seems, so that I will have to move over to the wikiteam version and rework it, I just spent 2 hours on this so i am pretty happy for the first version. mike On Tue, May 29, 2012 at 1:52 AM, Hydriz Wikipedia ad...@alphacorp.tk wrote: This is quite nice, though the item's metadata is too little :) On Tue, May 29, 2012 at 3:40 AM, Mike Dupont jamesmikedup...@googlemail.com wrote: first version of the Script is ready , it gets the versions, puts them in a zip and puts that on archive.org https://github.com/h4ck3rm1k3/pywikipediabot/blob/master/export_deleted.py here is an example output : http://archive.org/details/wikipedia-delete-2012-05 http://ia601203.us.archive.org/24/items/wikipedia-delete-2012-05/archive2012-05-28T21:34:02.302183.zip I will cron this, and it should give a start of saving deleted data. Articles will be exported once a day, even if they they were exported yesterday as long as they are in one of the categories. mike On Mon, May 21, 2012 at 7:21 PM, Mike Dupont jamesmikedup...@googlemail.com wrote: Thanks! and run that 1 time per day, they dont get deleted that quickly. mike On Mon, May 21, 2012 at 9:11 PM, emijrp emi...@gmail.com wrote: Create a script that makes a request to Special:Export using this category as feed https://en.wikipedia.org/wiki/Category:Candidates_for_speedy_deletion More info https://www.mediawiki.org/wiki/Manual:Parameters_to_Special:Export 2012/5/21 Mike Dupont jamesmikedup...@googlemail.com Well I whould be happy for items like this : http://en.wikipedia.org/wiki/Template:Db-a7 would it be possible to extract them easily? mike On Thu, May 17, 2012 at 2:23 PM, Ariel T. Glenn ar...@wikimedia.org wrote: There's a few other reasons articles get deleted: copyright issues, personal identifying data, etc. This makes maintaning the sort of mirror you propose problematic, although a similar mirror is here: http://deletionpedia.dbatley.com/w/index.php?title=Main_Page The dumps contain only data publically available at the time of the run, without deleted data. The articles aren't permanently deleted of course. The revisions texts live on in the database, so a query on toolserver, for example, could be used to get at them, but that would need to be for research purposes. Ariel Στις 17-05-2012, ημέρα Πεμ, και ώρα 13:30 +0200, ο/η Mike Dupont έγραψε: Hi, I am thinking about how to collect articles deleted based on the not notable criteria, is there any way we can extract them from the mysql binlogs? how are these mirrors working? I would be interested in setting up a mirror of deleted data, at least that which is not spam/vandalism based on tags. mike On Thu, May 17, 2012 at 1:09 PM, Ariel T. Glenn ar...@wikimedia.org wrote: We now have three mirror sites, yay! The full list is linked to from http://dumps.wikimedia.org/ and is also available at http://meta.wikimedia.org/wiki/Mirroring_Wikimedia_project_XML_dumps#Current_Mirrors Summarizing, we have: C3L (Brazil) with the last 5 good known dumps, Masaryk University (Czech Republic) with the last 5 known good dumps, Your.org (USA) with the complete archive of dumps, and for the latest version of uploaded media, Your.org with http/ftp/rsync access. Thanks to Carlos, Kevin and Yenya respectively at the above sites for volunteering space, time and effort to make this happen. As people noticed earlier, a series of media tarballs per-project (excluding commons) is being generated. As soon as the first run of these is complete we'll announce its location and start generating them on a semi-regular basis. As we've been getting the bugs out of the mirroring setup, it is getting easier to add new locations. Know anyone interested? Please let us know; we would love to have them. Ariel ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l
Re: [Wikitech-l] Desktop upload tool
How will it be different than the commonist? geotagging would be one big feature and the ability to group photos with common tags would be useful! mike On Mon, May 28, 2012 at 7:16 AM, Petr Bena benap...@gmail.com wrote: I would name it Snuggle, so that it's in harmony with Huggle :) On Fri, May 25, 2012 at 5:55 PM, John phoenixoverr...@gmail.com wrote: One thing I have done quite a bit of was Image modification. Everything from watermark removal to coping to rotating. When Im on a spur ofH watermark removal, it would be awesome if I could just point the program to a directory of fixed files, set a upload summary and have the bot upload those files keeping the existing file information. (that is the biggest drawback of http://commons.wikimedia.org/wiki/Commons:Tools/Commonist ) John On Fri, May 25, 2012 at 11:47 AM, Platonides platoni...@gmail.com wrote: As some of you are already aware, I'm doing for this GSoC a Desktop tool for conveniently uploading images to commons. http://thread.gmane.org/gmane.org.wikimedia.wikilovesmonuments/2641 How can you help with this? a) Propose a cool name for this tool b) Suggest new functionalities/requisites Either on http://www.mediawiki.org/wiki/User:Platonides/GSOC_proposal or in this thread. c) Volunteer for testing the program and giving early feedback (please reply directly to avoid spamming the mailing list with I can test it, too) ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l -- James Michael DuPont Member of Free Libre Open Source Software Kosova http://flossk.org Contributor FOSM, the CC-BY-SA map of the world http://fosm.org Mozilla Rep https://reps.mozilla.org/u/h4ck3rm1k3 ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l
Re: [Wikitech-l] [Xmldatadumps-l] XML dumps/Media mirrors update
first version of the Script is ready , it gets the versions, puts them in a zip and puts that on archive.org https://github.com/h4ck3rm1k3/pywikipediabot/blob/master/export_deleted.py here is an example output : http://archive.org/details/wikipedia-delete-2012-05 http://ia601203.us.archive.org/24/items/wikipedia-delete-2012-05/archive2012-05-28T21:34:02.302183.zip I will cron this, and it should give a start of saving deleted data. Articles will be exported once a day, even if they they were exported yesterday as long as they are in one of the categories. mike On Mon, May 21, 2012 at 7:21 PM, Mike Dupont jamesmikedup...@googlemail.com wrote: Thanks! and run that 1 time per day, they dont get deleted that quickly. mike On Mon, May 21, 2012 at 9:11 PM, emijrp emi...@gmail.com wrote: Create a script that makes a request to Special:Export using this category as feed https://en.wikipedia.org/wiki/Category:Candidates_for_speedy_deletion More info https://www.mediawiki.org/wiki/Manual:Parameters_to_Special:Export 2012/5/21 Mike Dupont jamesmikedup...@googlemail.com Well I whould be happy for items like this : http://en.wikipedia.org/wiki/Template:Db-a7 would it be possible to extract them easily? mike On Thu, May 17, 2012 at 2:23 PM, Ariel T. Glenn ar...@wikimedia.org wrote: There's a few other reasons articles get deleted: copyright issues, personal identifying data, etc. This makes maintaning the sort of mirror you propose problematic, although a similar mirror is here: http://deletionpedia.dbatley.com/w/index.php?title=Main_Page The dumps contain only data publically available at the time of the run, without deleted data. The articles aren't permanently deleted of course. The revisions texts live on in the database, so a query on toolserver, for example, could be used to get at them, but that would need to be for research purposes. Ariel Στις 17-05-2012, ημέρα Πεμ, και ώρα 13:30 +0200, ο/η Mike Dupont έγραψε: Hi, I am thinking about how to collect articles deleted based on the not notable criteria, is there any way we can extract them from the mysql binlogs? how are these mirrors working? I would be interested in setting up a mirror of deleted data, at least that which is not spam/vandalism based on tags. mike On Thu, May 17, 2012 at 1:09 PM, Ariel T. Glenn ar...@wikimedia.org wrote: We now have three mirror sites, yay! The full list is linked to from http://dumps.wikimedia.org/ and is also available at http://meta.wikimedia.org/wiki/Mirroring_Wikimedia_project_XML_dumps#Current_Mirrors Summarizing, we have: C3L (Brazil) with the last 5 good known dumps, Masaryk University (Czech Republic) with the last 5 known good dumps, Your.org (USA) with the complete archive of dumps, and for the latest version of uploaded media, Your.org with http/ftp/rsync access. Thanks to Carlos, Kevin and Yenya respectively at the above sites for volunteering space, time and effort to make this happen. As people noticed earlier, a series of media tarballs per-project (excluding commons) is being generated. As soon as the first run of these is complete we'll announce its location and start generating them on a semi-regular basis. As we've been getting the bugs out of the mirroring setup, it is getting easier to add new locations. Know anyone interested? Please let us know; we would love to have them. Ariel ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l -- James Michael DuPont Member of Free Libre Open Source Software Kosova http://flossk.org Contributor FOSM, the CC-BY-SA map of the world http://fosm.org Mozilla Rep https://reps.mozilla.org/u/h4ck3rm1k3 ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l -- Emilio J. Rodríguez-Posada. E-mail: emijrp AT gmail DOT com Pre-doctoral student at the University of Cádiz (Spain) Projects: AVBOT | StatMediaWiki | WikiEvidens | WikiPapers | WikiTeam Personal website: https://sites.google.com/site/emijrp/ ___ Xmldatadumps-l mailing list xmldatadump...@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/xmldatadumps-l -- James Michael DuPont Member of Free Libre Open Source Software Kosova http://flossk.org Contributor FOSM, the CC-BY-SA map of the world http://fosm.org Mozilla Rep https://reps.mozilla.org/u/h4ck3rm1k3 -- James Michael DuPont Member of Free Libre Open Source Software Kosova http://flossk.org Contributor
Re: [Wikitech-l] [Xmldatadumps-l] XML dumps/Media mirrors update
Well, I have now updated the script to include the xml dump in raw format. I will have to add more information the achive.org item, at least a basic readme. other thing is that the wikipybot does not support the full history it seems, so that I will have to move over to the wikiteam version and rework it, I just spent 2 hours on this so i am pretty happy for the first version. mike On Tue, May 29, 2012 at 1:52 AM, Hydriz Wikipedia ad...@alphacorp.tk wrote: This is quite nice, though the item's metadata is too little :) On Tue, May 29, 2012 at 3:40 AM, Mike Dupont jamesmikedup...@googlemail.com wrote: first version of the Script is ready , it gets the versions, puts them in a zip and puts that on archive.org https://github.com/h4ck3rm1k3/pywikipediabot/blob/master/export_deleted.py here is an example output : http://archive.org/details/wikipedia-delete-2012-05 http://ia601203.us.archive.org/24/items/wikipedia-delete-2012-05/archive2012-05-28T21:34:02.302183.zip I will cron this, and it should give a start of saving deleted data. Articles will be exported once a day, even if they they were exported yesterday as long as they are in one of the categories. mike On Mon, May 21, 2012 at 7:21 PM, Mike Dupont jamesmikedup...@googlemail.com wrote: Thanks! and run that 1 time per day, they dont get deleted that quickly. mike On Mon, May 21, 2012 at 9:11 PM, emijrp emi...@gmail.com wrote: Create a script that makes a request to Special:Export using this category as feed https://en.wikipedia.org/wiki/Category:Candidates_for_speedy_deletion More info https://www.mediawiki.org/wiki/Manual:Parameters_to_Special:Export 2012/5/21 Mike Dupont jamesmikedup...@googlemail.com Well I whould be happy for items like this : http://en.wikipedia.org/wiki/Template:Db-a7 would it be possible to extract them easily? mike On Thu, May 17, 2012 at 2:23 PM, Ariel T. Glenn ar...@wikimedia.org wrote: There's a few other reasons articles get deleted: copyright issues, personal identifying data, etc. This makes maintaning the sort of mirror you propose problematic, although a similar mirror is here: http://deletionpedia.dbatley.com/w/index.php?title=Main_Page The dumps contain only data publically available at the time of the run, without deleted data. The articles aren't permanently deleted of course. The revisions texts live on in the database, so a query on toolserver, for example, could be used to get at them, but that would need to be for research purposes. Ariel Στις 17-05-2012, ημέρα Πεμ, και ώρα 13:30 +0200, ο/η Mike Dupont έγραψε: Hi, I am thinking about how to collect articles deleted based on the not notable criteria, is there any way we can extract them from the mysql binlogs? how are these mirrors working? I would be interested in setting up a mirror of deleted data, at least that which is not spam/vandalism based on tags. mike On Thu, May 17, 2012 at 1:09 PM, Ariel T. Glenn ar...@wikimedia.org wrote: We now have three mirror sites, yay! The full list is linked to from http://dumps.wikimedia.org/ and is also available at http://meta.wikimedia.org/wiki/Mirroring_Wikimedia_project_XML_dumps#Current_Mirrors Summarizing, we have: C3L (Brazil) with the last 5 good known dumps, Masaryk University (Czech Republic) with the last 5 known good dumps, Your.org (USA) with the complete archive of dumps, and for the latest version of uploaded media, Your.org with http/ftp/rsync access. Thanks to Carlos, Kevin and Yenya respectively at the above sites for volunteering space, time and effort to make this happen. As people noticed earlier, a series of media tarballs per-project (excluding commons) is being generated. As soon as the first run of these is complete we'll announce its location and start generating them on a semi-regular basis. As we've been getting the bugs out of the mirroring setup, it is getting easier to add new locations. Know anyone interested? Please let us know; we would love to have them. Ariel ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l -- James Michael DuPont Member of Free Libre Open Source Software Kosova http://flossk.org Contributor FOSM, the CC-BY-SA map of the world http://fosm.org Mozilla Rep https://reps.mozilla.org/u/h4ck3rm1k3 ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l
Re: [Wikitech-l] XML dumps/Media mirrors update
Well I whould be happy for items like this : http://en.wikipedia.org/wiki/Template:Db-a7 would it be possible to extract them easily? mike On Thu, May 17, 2012 at 2:23 PM, Ariel T. Glenn ar...@wikimedia.org wrote: There's a few other reasons articles get deleted: copyright issues, personal identifying data, etc. This makes maintaning the sort of mirror you propose problematic, although a similar mirror is here: http://deletionpedia.dbatley.com/w/index.php?title=Main_Page The dumps contain only data publically available at the time of the run, without deleted data. The articles aren't permanently deleted of course. The revisions texts live on in the database, so a query on toolserver, for example, could be used to get at them, but that would need to be for research purposes. Ariel Στις 17-05-2012, ημέρα Πεμ, και ώρα 13:30 +0200, ο/η Mike Dupont έγραψε: Hi, I am thinking about how to collect articles deleted based on the not notable criteria, is there any way we can extract them from the mysql binlogs? how are these mirrors working? I would be interested in setting up a mirror of deleted data, at least that which is not spam/vandalism based on tags. mike On Thu, May 17, 2012 at 1:09 PM, Ariel T. Glenn ar...@wikimedia.org wrote: We now have three mirror sites, yay! The full list is linked to from http://dumps.wikimedia.org/ and is also available at http://meta.wikimedia.org/wiki/Mirroring_Wikimedia_project_XML_dumps#Current_Mirrors Summarizing, we have: C3L (Brazil) with the last 5 good known dumps, Masaryk University (Czech Republic) with the last 5 known good dumps, Your.org (USA) with the complete archive of dumps, and for the latest version of uploaded media, Your.org with http/ftp/rsync access. Thanks to Carlos, Kevin and Yenya respectively at the above sites for volunteering space, time and effort to make this happen. As people noticed earlier, a series of media tarballs per-project (excluding commons) is being generated. As soon as the first run of these is complete we'll announce its location and start generating them on a semi-regular basis. As we've been getting the bugs out of the mirroring setup, it is getting easier to add new locations. Know anyone interested? Please let us know; we would love to have them. Ariel ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l -- James Michael DuPont Member of Free Libre Open Source Software Kosova http://flossk.org Contributor FOSM, the CC-BY-SA map of the world http://fosm.org Mozilla Rep https://reps.mozilla.org/u/h4ck3rm1k3 ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l
Re: [Wikitech-l] [Xmldatadumps-l] XML dumps/Media mirrors update
Thanks! and run that 1 time per day, they dont get deleted that quickly. mike On Mon, May 21, 2012 at 9:11 PM, emijrp emi...@gmail.com wrote: Create a script that makes a request to Special:Export using this category as feed https://en.wikipedia.org/wiki/Category:Candidates_for_speedy_deletion More info https://www.mediawiki.org/wiki/Manual:Parameters_to_Special:Export 2012/5/21 Mike Dupont jamesmikedup...@googlemail.com Well I whould be happy for items like this : http://en.wikipedia.org/wiki/Template:Db-a7 would it be possible to extract them easily? mike On Thu, May 17, 2012 at 2:23 PM, Ariel T. Glenn ar...@wikimedia.org wrote: There's a few other reasons articles get deleted: copyright issues, personal identifying data, etc. This makes maintaning the sort of mirror you propose problematic, although a similar mirror is here: http://deletionpedia.dbatley.com/w/index.php?title=Main_Page The dumps contain only data publically available at the time of the run, without deleted data. The articles aren't permanently deleted of course. The revisions texts live on in the database, so a query on toolserver, for example, could be used to get at them, but that would need to be for research purposes. Ariel Στις 17-05-2012, ημέρα Πεμ, και ώρα 13:30 +0200, ο/η Mike Dupont έγραψε: Hi, I am thinking about how to collect articles deleted based on the not notable criteria, is there any way we can extract them from the mysql binlogs? how are these mirrors working? I would be interested in setting up a mirror of deleted data, at least that which is not spam/vandalism based on tags. mike On Thu, May 17, 2012 at 1:09 PM, Ariel T. Glenn ar...@wikimedia.org wrote: We now have three mirror sites, yay! The full list is linked to from http://dumps.wikimedia.org/ and is also available at http://meta.wikimedia.org/wiki/Mirroring_Wikimedia_project_XML_dumps#Current_Mirrors Summarizing, we have: C3L (Brazil) with the last 5 good known dumps, Masaryk University (Czech Republic) with the last 5 known good dumps, Your.org (USA) with the complete archive of dumps, and for the latest version of uploaded media, Your.org with http/ftp/rsync access. Thanks to Carlos, Kevin and Yenya respectively at the above sites for volunteering space, time and effort to make this happen. As people noticed earlier, a series of media tarballs per-project (excluding commons) is being generated. As soon as the first run of these is complete we'll announce its location and start generating them on a semi-regular basis. As we've been getting the bugs out of the mirroring setup, it is getting easier to add new locations. Know anyone interested? Please let us know; we would love to have them. Ariel ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l -- James Michael DuPont Member of Free Libre Open Source Software Kosova http://flossk.org Contributor FOSM, the CC-BY-SA map of the world http://fosm.org Mozilla Rep https://reps.mozilla.org/u/h4ck3rm1k3 ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l -- Emilio J. Rodríguez-Posada. E-mail: emijrp AT gmail DOT com Pre-doctoral student at the University of Cádiz (Spain) Projects: AVBOT | StatMediaWiki | WikiEvidens | WikiPapers | WikiTeam Personal website: https://sites.google.com/site/emijrp/ ___ Xmldatadumps-l mailing list xmldatadump...@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/xmldatadumps-l -- James Michael DuPont Member of Free Libre Open Source Software Kosova http://flossk.org Contributor FOSM, the CC-BY-SA map of the world http://fosm.org Mozilla Rep https://reps.mozilla.org/u/h4ck3rm1k3 ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l
Re: [Wikitech-l] XML dumps/Media mirrors update
Hi, I am thinking about how to collect articles deleted based on the not notable criteria, is there any way we can extract them from the mysql binlogs? how are these mirrors working? I would be interested in setting up a mirror of deleted data, at least that which is not spam/vandalism based on tags. mike On Thu, May 17, 2012 at 1:09 PM, Ariel T. Glenn ar...@wikimedia.org wrote: We now have three mirror sites, yay! The full list is linked to from http://dumps.wikimedia.org/ and is also available at http://meta.wikimedia.org/wiki/Mirroring_Wikimedia_project_XML_dumps#Current_Mirrors Summarizing, we have: C3L (Brazil) with the last 5 good known dumps, Masaryk University (Czech Republic) with the last 5 known good dumps, Your.org (USA) with the complete archive of dumps, and for the latest version of uploaded media, Your.org with http/ftp/rsync access. Thanks to Carlos, Kevin and Yenya respectively at the above sites for volunteering space, time and effort to make this happen. As people noticed earlier, a series of media tarballs per-project (excluding commons) is being generated. As soon as the first run of these is complete we'll announce its location and start generating them on a semi-regular basis. As we've been getting the bugs out of the mirroring setup, it is getting easier to add new locations. Know anyone interested? Please let us know; we would love to have them. Ariel ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l -- James Michael DuPont Member of Free Libre Open Source Software Kosova http://flossk.org Contributor FOSM, the CC-BY-SA map of the world http://fosm.org Mozilla Rep https://reps.mozilla.org/u/h4ck3rm1k3 ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l
Re: [Wikitech-l] Mark Hershberger departing Wikimedia Foundation in May
Yeah Mark, It was great to have worked with you, I am sorry that I did not have more time for wikipedia myself, but such is life. mike On Fri, Mar 16, 2012 at 8:01 PM, Bináris wikipo...@gmail.com wrote: Mark, thanks for helping Wikipedia and good luck whereever you go! -- Bináris ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l -- James Michael DuPont Member of Free Libre Open Source Software Kosova http://flossk.org ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l
Re: [Wikitech-l] Moving forward with Lua
very cool. how can we see the lua errors? how can we load wiki pages as data in lua? thanks, mike On Tue, Feb 28, 2012 at 11:01 PM, Owen Davis o...@wikia-inc.com wrote: For anyone interested in playing around with Lua on a wiki, I've enabled the Lua extension on http://lua.wikia.com It's just a test wiki right now with no real content. It will probably remain a test wiki, although if users want to flesh it out into a more authoritative site, that would be cool too. I am using http://www.mediawiki.org/wiki/Extension:Lua -- James Michael DuPont Member of Free Libre Open Source Software Kosova http://flossk.org ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l
Re: [Wikitech-l] Using computer vision to categorize images at Commons
On Tue, Feb 21, 2012 at 8:09 AM, Antoine Musso hashar+...@free.fr wrote: Le 20/02/12 22:08, Maarten Dammers a écrit : Some time ago I played around with computer vision to get images categorized on Commons. I documented this at https://commons.wikimedia.org/wiki/User:Multichill/Using_OpenCV_to_categorize_files . I don't think I'm going to spend time on it soon, but the results were quite promising, so maybe someone else feels like working on this? Would probably be a pretty nice student project or just fun to do. Google used a game to add keywords to images. Two people (of of them could be a robot) were presented two pictures. The aim is to enter a keyword that best describe the picture and that the other people will choose too. If both people use the same keyword, you gain points :-) Imagine, you are being shown a picture whose main subject is a monkey, you enter monkey, the other one monkey too. You both earn points. The game host now know that the picture is of a monkey :-D The Google Image Labeler, there are (funny and offensive) videos of that online. It is a good idea. Mediawiki does not need a person to check it live, it has users that are reviewing the changes anyway. It would be nice to have a *simple* way to put people down and have them label images, I could find people if it was really simple. Right now it is not really simple. Most people just dont have the concentration to read the docs or learn how to use the wiki. mike ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l
Re: [Wikitech-l] Common photo management plugins for MediaWiki
Thats great news! Any code to find? digikam is great, but also huge. I am using shotwell all the time. mike On Sun, Feb 12, 2012 at 3:32 PM, Guillaume Paumier gpaum...@wikimedia.orgwrote: FYI, there's also a group of French students working on a (partially developed) export plugin to Wikimedia Commons from digiKam ( http://www.digikam.org ), and they may also create an export plugin for Shotwell too ( http://yorba.org/shotwell/ ), depending on their progress. -- James Michael DuPont Member of Free Libre Open Source Software Kosova http://flossk.org ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l
Re: [Wikitech-l] Common photo management plugins for MediaWiki
This works, http://en.wikipedia.org/wiki/Wikipedia:Multimedia_beta#Add_Media_Wizard the new wizard is great! Has outdated information : http://www.mediawiki.org/wiki/Extension:Add_Media_Wizard mike On Sun, Feb 12, 2012 at 3:42 PM, Helder helder.w...@gmail.com wrote: http://www.mediawiki.org/wiki/Extension:Add_Media_Wizard http://en.wikipedia.org/wiki/Wikipedia:Multimedia_beta#Add_Media_Wizard -- James Michael DuPont Member of Free Libre Open Source Software Kosova http://flossk.org ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l
Re: [Wikitech-l] Common photo management plugins for MediaWiki
Yes, I would like to be able to select a few pictures, and be able to tag them together, add them to categories and such. The upload media wizard is great, and I would love if I could apply a category to many items. mike On Sat, Feb 11, 2012 at 12:48 PM, lakmal padmakumara lakmalpadmakum...@gmail.com wrote: GUI way of managing pictures . -- James Michael DuPont Member of Free Libre Open Source Software Kosova http://flossk.org ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l
Re: [Wikitech-l] Common photo management plugins for MediaWiki
Hi there It would be great to have a better photo management tool. Currently I am using google to host my creative commons photos, and using the easy upload tools from google plus to load into the albums, from there they can be imported into panoramio and imported into mediawiki. So, really I find that that best way to upload photos is via google, and then import them into commons. the things that are really needed is 1. GUI way to manage the geotagging of photos. 2. an easy way to import existing photos. 3. a better way to manage many photos. 4. Import of files from archive.org would also be good. the entire system of processing photos for commons is too complex for beginners. mike On Sat, Feb 11, 2012 at 7:54 AM, lakmal padmakumara lakmalpadmakum...@gmail.com wrote: Hey Thanks for the quick reply . It is the third idea which is mentioned in this suggested ideashttp://www.mediawiki.org/wiki/Summer_of_Code_2012#Project_ideaslist . At the moment I'm doing a research on the availability of APIs of these photo management Softwares .I am hoping to come up with a draft proposal which includes possible plugin list and basic implementation details of the project in next few days .I will keep community updated about that . Meanwhile any guidelines ,tips or comments from the community is more than welcome and highly appreciated . On Sat, Feb 11, 2012 at 3:30 AM, Platonides platoni...@gmail.com wrote: Hello lakmal, Good to meet you. This is indeed the right place. I don't see where it was suggested, but I haven't heard anyone to be working on that idea, so it seems perfectly available. I encourage you to prepare a draft proposal based on that. ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l -- Thanks Kind Regards Lakmal Padmakumara Undergraduate Computer Science and Engineering Department University of Moratuwa Sri Lanka http://srilankatechnology.blogspot.com/ ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l -- James Michael DuPont Member of Free Libre Open Source Software Kosova http://flossk.org ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l
Re: [Wikitech-l] SOPA commercial social networks
ON the german site, they have identi.ca as well. diaspora is missing on all. It would be nice to have these social tools as opt in tools for wikipedia in general. I dont see why we can have them for the fundraising or when wp wants to be known, but not for articles. mike On Wed, Jan 18, 2012 at 10:59 AM, David Gerard dger...@gmail.com wrote: On 18 January 2012 06:19, Thomas Schmidt schm...@netaction.de wrote: Clicking on the Google+ button I am forced to accept the tracking of my location and some new general business terms. Yeah - is there any sort of G+ link possible that *doesn't* demand you sign up for the mobile TOS (even on a desktop)? Linking to a commercial data-gathering exercise is less than ideal ... - d. ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l -- James Michael DuPont Member of Free Libre Open Source Software Kosova http://flossk.org ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l
Re: [Wikitech-l] SOPA commercial social networks
well, when it serves the purpose, then wp shows no shame in promoting some other sites. It is just inconsistent. but I guess I dont have to say that? Sometimes it is ok, sometimes it is not ok. mike On Wed, Jan 18, 2012 at 11:12 AM, David Gerard dger...@gmail.com wrote: On 18 January 2012 10:03, Mike Dupont jamesmikedup...@googlemail.com wrote: ON the german site, they have identi.ca as well. diaspora is missing on all. It would be nice to have these social tools as opt in tools for wikipedia in general. I dont see why we can have them for the fundraising or when wp wants to be known, but not for articles. The usual barrier on en:wp has been neutrality concerns, and being seen to advocate particular services over others. - d. ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l -- James Michael DuPont Member of Free Libre Open Source Software Kosova http://flossk.org ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l
Re: [Wikitech-l] SOPA commercial social networks
Ok, thanks mark. that is good to know. Will look into this some time. mike On Wed, Jan 18, 2012 at 1:53 PM, Mark A. Hershberger mhershber...@wikimedia.org wrote: Mike Dupont jamesmikedup...@googlemail.com writes: It would be nice to have these social tools as opt in tools for wikipedia in general. Unless I'm mistaken, Admins on-wiki could do this without consulting the WMF by installing a gadget which people could then opt-in to use. As far as built-in functionality, the first thing to do is to get an extension that provides this reviewed. https://www.mediawiki.org/wiki/Review_queue does not show anything that would provide this. Mark. ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l -- James Michael DuPont Member of Free Libre Open Source Software Kosova http://flossk.org ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l
Re: [Wikitech-l] Performance roadmap update
Well thanks for sharing this, I will also put my efforts on HPHP on hold for now as well. mike On Thu, Dec 15, 2011 at 3:26 PM, Jay Ashworth j...@baylink.com wrote: - Original Message - From: Tim Starling tstarl...@wikimedia.org So we've decided to defer our HipHop deployment until hhvm is at a suitable level of maturity. We don't know exactly when that will be, but Jason Evans says in the note linked above that the first 90% is done; now we're on to the second 90% as we make it really shine. It's good to hear that Jason understands the realities of tool development, isn't it? Cheers, -- jra -- Jay R. Ashworth Baylink j...@baylink.com Designer The Things I Think RFC 2100 Ashworth Associates http://baylink.pitas.com 2000 Land Rover DII St Petersburg FL USA http://photo.imageinc.us +1 727 647 1274 ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l -- James Michael DuPont Member of Free Libre Open Source Software Kosova http://flossk.org ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l
Re: [Wikitech-l] Access to the Deleted Archives
I have also wanted this for a long time. http://www.petitiononline.com/urmwpnow/petition.html http://undeletewikipedia.blogspot.com/2009/10/clarity-in-petition.html mike On Sun, Dec 11, 2011 at 1:32 PM, Katja Müller katja_muel...@lavabit.com wrote: Hi there, we are doing information retrieval research on the Wikipedia history. Currently we are thinking about including the archive of Deleted Articles in the analysis. What is the current regulation on access to the Deleted Archive? According to this page, admins can permit access to single articles on request: http://en.wikipedia.org/wiki/Wikipedia:Deletion_policy#Access_to_deleted_pages However, is there a way (for researchers) to either: a) access (API) or download the whole archive of Deleted Articles b) get statistics or meta data about the Deleted Archive (article counts, revision meta information, logs) ? Kind regards, Katja Mueller ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l -- James Michael DuPont Member of Free Libre Open Source Software Kosova http://flossk.org ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l
Re: [Wikitech-l] Access to the Deleted Archives
On Sun, Dec 11, 2011 at 10:43 PM, Platonides platoni...@gmail.com wrote: On 11/12/11 13:38, Mike Dupont wrote: I have also wanted this for a long time. http://www.petitiononline.com/urmwpnow/petition.html http://undeletewikipedia.blogspot.com/2009/10/clarity-in-petition.html mike I don't think any sysop would reject a reasonable petition of a good wikipedian of the content of an article he previously wrote. (good wikipedian meaning it's not an instance of I want that copyvio content to recreate it from sockpuppet accounts using proxies) There are many articles that have been deleted, and we dont even know the full list of them. That's not true. The list of deleted pages is available at http://en.wikipedia.org/wiki/Special:Log/delete For very old deletions (before December 23, 2004) see http://en.wikipedia.org/wiki/Wikipedia:Deletion_log However, for most articles the deleted text would (should?) be in the lines of Article: John Smith Content: He's fat child in classroom. In other cases they may be well-written articles which violate the copyright of eg. Encarta. So they can't be shown either. You seem to be targetting good articles deleted due to the target being non-notable, and you indeed have a point for them. You could launch a project to host those files if you wish to (I think there was already one doing it? at least many wikis have spun off to their own wiki about their topic). Well, I dont know what to say, except that I put work into my articles, and having them deleted hurts. I would like to have them in my userspace at least. mike ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l
Re: [Wikitech-l] Access to the Deleted Archives
I can understand that about the user space. good point about the blog. mike On Sun, Dec 11, 2011 at 11:38 PM, Platonides platoni...@gmail.com wrote: On 11/12/11 22:44, Mike Dupont wrote: Well, I dont know what to say, except that I put work into my articles, and having them deleted hurts. I would like to have them in my userspace at least. mike I understand you. Having them in your user space may be a bit controversial (I see arguments for both for and against) but there would certainly be no problem in you publishing them in eg. your own blog. ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l -- James Michael DuPont Member of Free Libre Open Source Software Kosova http://flossk.org ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l
Re: [Wikitech-l] Extracting text from Wikipedia
you want from all articles or just one? see here http://stackoverflow.com/questions/627594/is-there-a-wikipedia-api mike On Sun, Nov 27, 2011 at 6:02 PM, Khalida BEN SIDI AHMED send.to.khal...@gmail.com wrote: Hello! I don't know if the subject of this question belongs to the scope of this group. Anyway, I will be pleased if I find an aswer to my question. I'm writing some Java code in order to realize NLP tasks upon texts using Wikipedia. What can I do in order to extract the first paragraph of a Wikipedia article? Thanks a lot. Truly yours Ben Sidi Ahmed ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l -- James Michael DuPont Member of Free Libre Open Source Software Kosova http://flossk.org ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l
Re: [Wikitech-l] Extracting text from Wikipedia
look, for 100 articles, just create a list of them, and export them as xml. or use the book creator. http://en.wikipedia.org/wiki/Help:Books also there is a json api to pull single articles. http://www.barattalo.it/2010/08/29/php-bot-to-get-wikipedia-definitions/ mike On Sun, Nov 27, 2011 at 6:15 PM, Khalida BEN SIDI AHMED send.to.khal...@gmail.com wrote: I have already read the responses given in this post. I want to the extract the first paragraph (or the first sentence) for a list of 100 articles. I could not use JWPL beacause I don't have a big hard disk space to create the DB. I try to use JSoup but I need examples. ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l -- James Michael DuPont Member of Free Libre Open Source Software Kosova http://flossk.org ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l
Re: [Wikitech-l] Extracting text from Wikipedia
http://code.google.com/p/jwpl/ this looks also good JWPL (Java Wikipedia Library) is a free, Java-based application programming interface that allows to access all information contained in Wikipedia. have not tried that , but you said you wanted to do it in java. mike On Sun, Nov 27, 2011 at 6:20 PM, Mike Dupont jamesmikedup...@googlemail.com wrote: look, for 100 articles, just create a list of them, and export them as xml. or use the book creator. http://en.wikipedia.org/wiki/Help:Books also there is a json api to pull single articles. http://www.barattalo.it/2010/08/29/php-bot-to-get-wikipedia-definitions/ mike On Sun, Nov 27, 2011 at 6:15 PM, Khalida BEN SIDI AHMED send.to.khal...@gmail.com wrote: I have already read the responses given in this post. I want to the extract the first paragraph (or the first sentence) for a list of 100 articles. I could not use JWPL beacause I don't have a big hard disk space to create the DB. I try to use JSoup but I need examples. ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l -- James Michael DuPont Member of Free Libre Open Source Software Kosova http://flossk.org -- James Michael DuPont Member of Free Libre Open Source Software Kosova http://flossk.org ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l
Re: [Wikitech-l] Extracting text from Wikipedia
have you seen wordnet,http://en.wikipedia.org/wiki/WordNet wikitionary? in any case you can call the export routine as you need it, even for single articles. I think it would not cause too much load on the server. anyway, good luck mike On Sun, Nov 27, 2011 at 6:36 PM, Khalida BEN SIDI AHMED send.to.khal...@gmail.com wrote: The list of the articles I will need is not known from the beggining. Through my project, I will find a list of words (50). I try to find for them definitions in Wikipedia. After that I will extract the hyperonym of each word. I will have a new list for which I then retrieve the respective articles...etc. So the the list of articles I need grows through the evolution of the algorithm' execution. ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l -- James Michael DuPont Member of Free Libre Open Source Software Kosova http://flossk.org ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l
Re: [Wikitech-l] Extracting text from Wikipedia
If you are focused on jsoup, then best to ask there on the mailing list http://jsoup.org/discussion On Sun, Nov 27, 2011 at 7:51 PM, Khalida BEN SIDI AHMED send.to.khal...@gmail.com wrote: Thank you Hoehrmann. I will try to apply the options you've mentionned. However, if someone can help me in using JSoup, his ideas are welcomed. ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l -- James Michael DuPont Member of Free Libre Open Source Software Kosova http://flossk.org ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l
Re: [Wikitech-l] Progress made on packaging hiphop for debian
I've been working on a replacement for the libcurl patch. -- Tim Starling Thats Good news, Last night I removed all traces of the curl_multi_select from the code, no idea if it runs properly, but it produces a package. Also I commented out all the non standard libevent code. The package has a compiler that produces code, and you can at least experiment with it. We need more people testing, I am getting help from somone from the debian team as well. Will continue working on this as I find time. we are organizing a conference on Nov 12 if anyone is interested, www.flossk.org/en/blog/sfk11 that is also taking up time. thanks, mike On Tue, Sep 20, 2011 at 11:43 PM, Tim Starling tstarl...@wikimedia.orgwrote: On 20/09/11 11:05, Mark A. Hershberger wrote: At that point, I contacted Mike and asked for his help. He suggested we attempt to get HipHop compiled *without* the patches since the upstream developers weren't too enthusiastic about them. ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l -- James Michael DuPont Member of Free Libre Open Source Software Kosova http://flossk.org ___ Wikitech-l mailing list Wikitech-l@lists.wikimedia.org https://lists.wikimedia.org/mailman/listinfo/wikitech-l