Re: [Wikitech-l] Using mediawiki from within the Social networks?

2012-10-03 Thread Mike Dupont
On Wed, Oct 3, 2012 at 7:43 AM, Yury Katkov katkov.ju...@gmail.com wrote:
 I'm not sure that '''editing''' can be made more easy with the help of
 social network client. Any ideas on that? any ideas on what else can
 be made  more engaging with the power of social networks?

Well what if people can click on a bit of text and comment on it, they
could suggest in that comment that the text is replaced.

mike


-- 
James Michael DuPont
Member of Free Libre Open Source Software Kosova http://flossk.org
Saving wikipedia(tm) articles from deletion http://SpeedyDeletion.wikia.com
Contributor FOSM, the CC-BY-SA map of the world http://fosm.org
Mozilla Rep https://reps.mozilla.org/u/h4ck3rm1k3
Free Software Foundation Europe Fellow http://fsfe.org/support/?h4ck3rm1k3

___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l


Re: [Wikitech-l] Using mediawiki from within the Social networks?

2012-10-03 Thread Mike Dupont
On Wed, Oct 3, 2012 at 1:03 PM, Yury Katkov katkov.ju...@gmail.com wrote:
 hmmm, much like the blips on Google Wave?
like line comments in github :
https://github.com/h4ck3rm1k3/wikiteam/commit/4da7f7f4a813b53be13bff7e29a1e5325bb68a30#L0R58

We need a way to track and rate comments and then we can resolve them
with changes after people have stormed over them.


-- 
James Michael DuPont
Member of Free Libre Open Source Software Kosova http://flossk.org
Saving wikipedia(tm) articles from deletion http://SpeedyDeletion.wikia.com
Contributor FOSM, the CC-BY-SA map of the world http://fosm.org
Mozilla Rep https://reps.mozilla.org/u/h4ck3rm1k3
Free Software Foundation Europe Fellow http://fsfe.org/support/?h4ck3rm1k3

___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l


Re: [Wikitech-l] Github replication

2012-10-03 Thread Mike Dupont
On Wed, Oct 3, 2012 at 6:27 PM, Chad innocentkil...@gmail.com wrote:
 Hi everyone,

 Just letting everyone know: mediawiki/core is now replicating from
 gerrit to github.

 https://github.com/mediawiki/core

that is great news.
mike

___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l


Re: [Wikitech-l] Using mediawiki from within the Social networks?

2012-10-02 Thread Mike Dupont
well what about the user granting you the rights to the comments etc
under certain conditions and they just use fb and pull that data in
via the api.
so you would have some pages that are autogenerated via fb, you might
have Talk:page for normal users and FB:Page for fb users.

mike

On Tue, Oct 2, 2012 at 1:21 AM, Yury Katkov katkov.ju...@gmail.com wrote:
 Maybe we can brainstorm this idea a little bit together? I have
 several crowdsourcing projects and I always suspect that the pure wiki
 editing requires too much motivation and it's not the best approach in
 many cases.
 -
 Yury Katkov



 On Tue, Oct 2, 2012 at 1:16 AM, Mike  Dupont
 jamesmikedup...@googlemail.com wrote:
 Yuri,
 I agree with you 100% this is a good idea and would help out alot. I
 have also had similar ideas!
 mike

 On Mon, Oct 1, 2012 at 11:09 PM, Yury Katkov katkov.ju...@gmail.com wrote:
 There is no such client to the best of my knowledge. But here I don't
 think much about the big-text-projects like wikipedia: more about the
 projects where users need to
 -  vote for the articles or
 - add really small fragments of text
 - or link page together via Semantic MediaWiki properties or
 - categorize and tagged the articles...
  these are all very small activities. The idea is that those
 activities can be shown in the social network MW-client in the most
 easy-to-consume way to the very broad audience. Or maybe they can be
 even gamified!
 -
 Yury Katkov



 On Tue, Oct 2, 2012 at 1:02 AM, Mike  Dupont
 jamesmikedup...@googlemail.com wrote:
 On Mon, Oct 1, 2012 at 10:37 PM, Yury Katkov katkov.ju...@gmail.com 
 wrote:
 MediaWiki social network client
 tell me more, links?


 --
 James Michael DuPont
 Member of Free Libre Open Source Software Kosova http://flossk.org
 Saving wikipedia(tm) articles from deletion http://SpeedyDeletion.wikia.com
 Contributor FOSM, the CC-BY-SA map of the world http://fosm.org
 Mozilla Rep https://reps.mozilla.org/u/h4ck3rm1k3
 Free Software Foundation Europe Fellow http://fsfe.org/support/?h4ck3rm1k3

 ___
 Wikitech-l mailing list
 Wikitech-l@lists.wikimedia.org
 https://lists.wikimedia.org/mailman/listinfo/wikitech-l

 ___
 Wikitech-l mailing list
 Wikitech-l@lists.wikimedia.org
 https://lists.wikimedia.org/mailman/listinfo/wikitech-l



 --
 James Michael DuPont
 Member of Free Libre Open Source Software Kosova http://flossk.org
 Saving wikipedia(tm) articles from deletion http://SpeedyDeletion.wikia.com
 Contributor FOSM, the CC-BY-SA map of the world http://fosm.org
 Mozilla Rep https://reps.mozilla.org/u/h4ck3rm1k3
 Free Software Foundation Europe Fellow http://fsfe.org/support/?h4ck3rm1k3

 ___
 Wikitech-l mailing list
 Wikitech-l@lists.wikimedia.org
 https://lists.wikimedia.org/mailman/listinfo/wikitech-l

 ___
 Wikitech-l mailing list
 Wikitech-l@lists.wikimedia.org
 https://lists.wikimedia.org/mailman/listinfo/wikitech-l



-- 
James Michael DuPont
Member of Free Libre Open Source Software Kosova http://flossk.org
Saving wikipedia(tm) articles from deletion http://SpeedyDeletion.wikia.com
Contributor FOSM, the CC-BY-SA map of the world http://fosm.org
Mozilla Rep https://reps.mozilla.org/u/h4ck3rm1k3
Free Software Foundation Europe Fellow http://fsfe.org/support/?h4ck3rm1k3

___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l


Re: [Wikitech-l] Using mediawiki from within the Social networks?

2012-10-01 Thread Mike Dupont
have you seen the customization that wikia.com has done? it is a step
in the right direction for social integration.

On Mon, Oct 1, 2012 at 8:24 PM, Chris Steipp cste...@wikimedia.org wrote:
 It's possible, but it would be a little ugly. You could write a web
 frontend to what would basically be a proxy handling the work on the
 backend over the api. However, there is no secure way for a user to
 login through something like that, so all of the edits would have to
 come from your app and not the user. Alternatively, you could write a
 javascript (or flash) application, and get it approved to make
 CORS/crossdomain requests to the target MediaWiki api, so the user
 would be making api calls directly. Getting the wiki admin to approve
 domains, or add a crossdomain.xml file may or may not be a problem for
 you.

 OAuth will make the delegated authorization easier, but unfortunately
 the priority for getting OAuth integrated with MediaWiki is lower than
 some other critical projects right now, so it will be a few months
 before this is possible.


 On Mon, Oct 1, 2012 at 10:44 AM, Yury Katkov katkov.ju...@gmail.com wrote:
 Hi everyone!

 Is it possible to use MediaWiki as a service whereas the UI is located
 on a Facebook app? So all the editing and viewing is take place on a
 Facebook and MediaWiki provide the storage, revision control and lots
 of extensions?
 -
 Yury Katkov

 ___
 Wikitech-l mailing list
 Wikitech-l@lists.wikimedia.org
 https://lists.wikimedia.org/mailman/listinfo/wikitech-l

 ___
 Wikitech-l mailing list
 Wikitech-l@lists.wikimedia.org
 https://lists.wikimedia.org/mailman/listinfo/wikitech-l



-- 
James Michael DuPont
Member of Free Libre Open Source Software Kosova http://flossk.org
Saving wikipedia(tm) articles from deletion http://SpeedyDeletion.wikia.com
Contributor FOSM, the CC-BY-SA map of the world http://fosm.org
Mozilla Rep https://reps.mozilla.org/u/h4ck3rm1k3
Free Software Foundation Europe Fellow http://fsfe.org/support/?h4ck3rm1k3

___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l


Re: [Wikitech-l] Using mediawiki from within the Social networks?

2012-10-01 Thread Mike Dupont
On Mon, Oct 1, 2012 at 10:37 PM, Yury Katkov katkov.ju...@gmail.com wrote:
 MediaWiki social network client
tell me more, links?


-- 
James Michael DuPont
Member of Free Libre Open Source Software Kosova http://flossk.org
Saving wikipedia(tm) articles from deletion http://SpeedyDeletion.wikia.com
Contributor FOSM, the CC-BY-SA map of the world http://fosm.org
Mozilla Rep https://reps.mozilla.org/u/h4ck3rm1k3
Free Software Foundation Europe Fellow http://fsfe.org/support/?h4ck3rm1k3

___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l


Re: [Wikitech-l] Using mediawiki from within the Social networks?

2012-10-01 Thread Mike Dupont
Yuri,
I agree with you 100% this is a good idea and would help out alot. I
have also had similar ideas!
mike

On Mon, Oct 1, 2012 at 11:09 PM, Yury Katkov katkov.ju...@gmail.com wrote:
 There is no such client to the best of my knowledge. But here I don't
 think much about the big-text-projects like wikipedia: more about the
 projects where users need to
 -  vote for the articles or
 - add really small fragments of text
 - or link page together via Semantic MediaWiki properties or
 - categorize and tagged the articles...
  these are all very small activities. The idea is that those
 activities can be shown in the social network MW-client in the most
 easy-to-consume way to the very broad audience. Or maybe they can be
 even gamified!
 -
 Yury Katkov



 On Tue, Oct 2, 2012 at 1:02 AM, Mike  Dupont
 jamesmikedup...@googlemail.com wrote:
 On Mon, Oct 1, 2012 at 10:37 PM, Yury Katkov katkov.ju...@gmail.com wrote:
 MediaWiki social network client
 tell me more, links?


 --
 James Michael DuPont
 Member of Free Libre Open Source Software Kosova http://flossk.org
 Saving wikipedia(tm) articles from deletion http://SpeedyDeletion.wikia.com
 Contributor FOSM, the CC-BY-SA map of the world http://fosm.org
 Mozilla Rep https://reps.mozilla.org/u/h4ck3rm1k3
 Free Software Foundation Europe Fellow http://fsfe.org/support/?h4ck3rm1k3

 ___
 Wikitech-l mailing list
 Wikitech-l@lists.wikimedia.org
 https://lists.wikimedia.org/mailman/listinfo/wikitech-l

 ___
 Wikitech-l mailing list
 Wikitech-l@lists.wikimedia.org
 https://lists.wikimedia.org/mailman/listinfo/wikitech-l



-- 
James Michael DuPont
Member of Free Libre Open Source Software Kosova http://flossk.org
Saving wikipedia(tm) articles from deletion http://SpeedyDeletion.wikia.com
Contributor FOSM, the CC-BY-SA map of the world http://fosm.org
Mozilla Rep https://reps.mozilla.org/u/h4ck3rm1k3
Free Software Foundation Europe Fellow http://fsfe.org/support/?h4ck3rm1k3

___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l


Re: [Wikitech-l] Lua deployed to www.mediawiki.org

2012-08-22 Thread Mike Dupont
On Thu, Aug 23, 2012 at 2:00 AM, Tim Starling tstarl...@wikimedia.orgwrote:

 I think there will be some changes to template invocations. For
 example, a typical coord invocation looks like this:

 {{Coord|33|51|35.9|S|151|12|40|E}}

 With the string processing facilities that Lua provides, that might
 change to:

 {{Coord|33°51'35.9S 151°12'40E}}

 Of course, backwards compatibility would need to be maintained, but
 that's easy enough.


HI all,
I can understand that some people might be upset about a new language and
change. But in the defense of lua proponents, it is not like the old
templates are being deleted here, there are millions of copies of them and
they are not going away. Also media wiki as a language is difficult to
process, using a standard language for tempaltes might help out. So a
migration path that I would think is reasonable and I am sure has been
proposed is to make first a compatibility layer that allows old templates
to be used in lua and make some evalold function to process old template
code exactly as it was.

thanks
mike

-- 
James Michael DuPont
Member of Free Libre Open Source Software Kosova http://flossk.org
 http://flossk.orgSaving wikipedia(tm) articles from deletion
http://SpeedyDeletion.wikia.com
Contributor FOSM, the CC-BY-SA map of the world http://fosm.org
Mozilla Rep https://reps.mozilla.org/u/h4ck3rm1k3
___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l


Re: [Wikitech-l] Lua deployed to www.mediawiki.org

2012-08-22 Thread Mike Dupont
On Thu, Aug 23, 2012 at 4:18 AM, MZMcBride z...@mzmcbride.com wrote:

 This is my concern with Lua modules, in a nutshell.


So please tell me, what are the options to fix this? Is there going to be a
common code repo and maybe an easy way to sync in a git filesystem of
template code into the wiki?

Maybe a git extension? or is there one already?

mike


-- 
James Michael DuPont
Member of Free Libre Open Source Software Kosova http://flossk.org
 http://flossk.orgSaving wikipedia(tm) articles from deletion
http://SpeedyDeletion.wikia.com
Contributor FOSM, the CC-BY-SA map of the world http://fosm.org
Mozilla Rep https://reps.mozilla.org/u/h4ck3rm1k3
___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l


Re: [Wikitech-l] [Xmldatadumps-l] XML dumps/Media mirrors update

2012-06-01 Thread Mike Dupont
I have run cron archiving now every 30 minutes,
http://ia700802.us.archive.org/34/items/wikipedia-delete-2012-06/
it is amazing how fast the stuff gets deleted on wikipedia.
what about the proposed deletes are there categories for that?
thanks
mike

On Wed, May 30, 2012 at 6:26 AM, Mike  Dupont
jamesmikedup...@googlemail.com wrote:
 https://github.com/h4ck3rm1k3/wikiteam code here

 On Wed, May 30, 2012 at 6:26 AM, Mike  Dupont
 jamesmikedup...@googlemail.com wrote:
 Ok, I merged the code from wikteam and have a full history dump script
 that uploads to archive.org,
 next step is to fix the bucket metadata in the script
 mike

 On Tue, May 29, 2012 at 3:08 AM, Mike  Dupont
 jamesmikedup...@googlemail.com wrote:
 Well, I have now updated the script to include  the xml dump in raw
 format. I will have to add more information the achive.org item, at
 least a basic readme.
 other thing is that the wikipybot does not support the full history it
 seems, so that I will have to move over to the wikiteam version and
 rework it,
 I just spent 2 hours on this so i am pretty happy for the first version.

 mike

 On Tue, May 29, 2012 at 1:52 AM, Hydriz Wikipedia ad...@alphacorp.tk 
 wrote:
 This is quite nice, though the item's metadata is too little :)

 On Tue, May 29, 2012 at 3:40 AM, Mike Dupont 
 jamesmikedup...@googlemail.com
 wrote:

 first version of the Script is ready , it gets the versions, puts them
 in a zip and puts that on archive.org
 https://github.com/h4ck3rm1k3/pywikipediabot/blob/master/export_deleted.py

 here is an example output :
 http://archive.org/details/wikipedia-delete-2012-05

 http://ia601203.us.archive.org/24/items/wikipedia-delete-2012-05/archive2012-05-28T21:34:02.302183.zip

 I will cron this, and it should give a start of saving deleted data.
 Articles will be exported once a day, even if they they were exported
 yesterday as long as they are in one of the categories.

 mike

 On Mon, May 21, 2012 at 7:21 PM, Mike  Dupont
 jamesmikedup...@googlemail.com wrote:
  Thanks! and run that 1 time per day, they dont get deleted that quickly.
  mike
 
  On Mon, May 21, 2012 at 9:11 PM, emijrp emi...@gmail.com wrote:
  Create a script that makes a request to Special:Export using this
 category
  as feed
  https://en.wikipedia.org/wiki/Category:Candidates_for_speedy_deletion
 
  More info
 https://www.mediawiki.org/wiki/Manual:Parameters_to_Special:Export
 
 
  2012/5/21 Mike Dupont jamesmikedup...@googlemail.com
 
  Well I whould be happy for items like this :
  http://en.wikipedia.org/wiki/Template:Db-a7
  would it be possible to extract them easily?
  mike
 
  On Thu, May 17, 2012 at 2:23 PM, Ariel T. Glenn ar...@wikimedia.org
  wrote:
   There's a few other reasons articles get deleted: copyright issues,
   personal identifying data, etc.  This makes maintaning the sort of
   mirror you propose problematic, although a similar mirror is here:
   http://deletionpedia.dbatley.com/w/index.php?title=Main_Page
  
   The dumps contain only data publically available at the time of the
 run,
   without deleted data.
  
   The articles aren't permanently deleted of course.  The revisions
 texts
   live on in the database, so a query on toolserver, for example,
 could be
   used to get at them, but that would need to be for research 
   purposes.
  
   Ariel
  
   Στις 17-05-2012, ημέρα Πεμ, και ώρα 13:30 +0200, ο/η Mike Dupont
 έγραψε:
   Hi,
   I am thinking about how to collect articles deleted based on the
 not
   notable criteria,
   is there any way we can extract them from the mysql binlogs? how 
   are
   these mirrors working? I would be interested in setting up a mirror
 of
   deleted data, at least that which is not spam/vandalism based on
 tags.
   mike
  
   On Thu, May 17, 2012 at 1:09 PM, Ariel T. Glenn 
 ar...@wikimedia.org
   wrote:
We now have three mirror sites, yay!  The full list is linked to
 from
http://dumps.wikimedia.org/ and is also available at
   
   
 http://meta.wikimedia.org/wiki/Mirroring_Wikimedia_project_XML_dumps#Current_Mirrors
   
Summarizing, we have:
   
C3L (Brazil) with the last 5 good known dumps,
Masaryk University (Czech Republic) with the last 5 known good
 dumps,
Your.org (USA) with the complete archive of dumps, and
   
for the latest version of uploaded media, Your.org with
http/ftp/rsync
access.
   
Thanks to Carlos, Kevin and Yenya respectively at the above sites
 for
volunteering space, time and effort to make this happen.
   
As people noticed earlier, a series of media tarballs per-project
(excluding commons) is being generated.  As soon as the first run
 of
these is complete we'll announce its location and start 
generating
them
on a semi-regular basis.
   
As we've been getting the bugs out of the mirroring setup, it is
getting
easier to add new locations.  Know anyone interested?  Please let
 us
know; we would love to have them.
   
Ariel

Re: [Wikitech-l] [Xmldatadumps-l] XML dumps/Media mirrors update

2012-05-30 Thread Mike Dupont
Ok, I merged the code from wikteam and have a full history dump script
that uploads to archive.org,
next step is to fix the bucket metadata in the script
mike

On Tue, May 29, 2012 at 3:08 AM, Mike  Dupont
jamesmikedup...@googlemail.com wrote:
 Well, I have now updated the script to include  the xml dump in raw
 format. I will have to add more information the achive.org item, at
 least a basic readme.
 other thing is that the wikipybot does not support the full history it
 seems, so that I will have to move over to the wikiteam version and
 rework it,
 I just spent 2 hours on this so i am pretty happy for the first version.

 mike

 On Tue, May 29, 2012 at 1:52 AM, Hydriz Wikipedia ad...@alphacorp.tk wrote:
 This is quite nice, though the item's metadata is too little :)

 On Tue, May 29, 2012 at 3:40 AM, Mike Dupont jamesmikedup...@googlemail.com
 wrote:

 first version of the Script is ready , it gets the versions, puts them
 in a zip and puts that on archive.org
 https://github.com/h4ck3rm1k3/pywikipediabot/blob/master/export_deleted.py

 here is an example output :
 http://archive.org/details/wikipedia-delete-2012-05

 http://ia601203.us.archive.org/24/items/wikipedia-delete-2012-05/archive2012-05-28T21:34:02.302183.zip

 I will cron this, and it should give a start of saving deleted data.
 Articles will be exported once a day, even if they they were exported
 yesterday as long as they are in one of the categories.

 mike

 On Mon, May 21, 2012 at 7:21 PM, Mike  Dupont
 jamesmikedup...@googlemail.com wrote:
  Thanks! and run that 1 time per day, they dont get deleted that quickly.
  mike
 
  On Mon, May 21, 2012 at 9:11 PM, emijrp emi...@gmail.com wrote:
  Create a script that makes a request to Special:Export using this
 category
  as feed
  https://en.wikipedia.org/wiki/Category:Candidates_for_speedy_deletion
 
  More info
 https://www.mediawiki.org/wiki/Manual:Parameters_to_Special:Export
 
 
  2012/5/21 Mike Dupont jamesmikedup...@googlemail.com
 
  Well I whould be happy for items like this :
  http://en.wikipedia.org/wiki/Template:Db-a7
  would it be possible to extract them easily?
  mike
 
  On Thu, May 17, 2012 at 2:23 PM, Ariel T. Glenn ar...@wikimedia.org
  wrote:
   There's a few other reasons articles get deleted: copyright issues,
   personal identifying data, etc.  This makes maintaning the sort of
   mirror you propose problematic, although a similar mirror is here:
   http://deletionpedia.dbatley.com/w/index.php?title=Main_Page
  
   The dumps contain only data publically available at the time of the
 run,
   without deleted data.
  
   The articles aren't permanently deleted of course.  The revisions
 texts
   live on in the database, so a query on toolserver, for example,
 could be
   used to get at them, but that would need to be for research purposes.
  
   Ariel
  
   Στις 17-05-2012, ημέρα Πεμ, και ώρα 13:30 +0200, ο/η Mike Dupont
 έγραψε:
   Hi,
   I am thinking about how to collect articles deleted based on the
 not
   notable criteria,
   is there any way we can extract them from the mysql binlogs? how are
   these mirrors working? I would be interested in setting up a mirror
 of
   deleted data, at least that which is not spam/vandalism based on
 tags.
   mike
  
   On Thu, May 17, 2012 at 1:09 PM, Ariel T. Glenn 
 ar...@wikimedia.org
   wrote:
We now have three mirror sites, yay!  The full list is linked to
 from
http://dumps.wikimedia.org/ and is also available at
   
   
 http://meta.wikimedia.org/wiki/Mirroring_Wikimedia_project_XML_dumps#Current_Mirrors
   
Summarizing, we have:
   
C3L (Brazil) with the last 5 good known dumps,
Masaryk University (Czech Republic) with the last 5 known good
 dumps,
Your.org (USA) with the complete archive of dumps, and
   
for the latest version of uploaded media, Your.org with
http/ftp/rsync
access.
   
Thanks to Carlos, Kevin and Yenya respectively at the above sites
 for
volunteering space, time and effort to make this happen.
   
As people noticed earlier, a series of media tarballs per-project
(excluding commons) is being generated.  As soon as the first run
 of
these is complete we'll announce its location and start generating
them
on a semi-regular basis.
   
As we've been getting the bugs out of the mirroring setup, it is
getting
easier to add new locations.  Know anyone interested?  Please let
 us
know; we would love to have them.
   
Ariel
   
   
___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l
  
  
  
  
  
  
   ___
   Wikitech-l mailing list
   Wikitech-l@lists.wikimedia.org
   https://lists.wikimedia.org/mailman/listinfo/wikitech-l
 
 
 
  --
  James Michael DuPont
  Member of Free Libre Open Source Software Kosova http://flossk.org
  Contributor FOSM, the CC-BY-SA map

Re: [Wikitech-l] [Xmldatadumps-l] XML dumps/Media mirrors update

2012-05-30 Thread Mike Dupont
https://github.com/h4ck3rm1k3/wikiteam code here

On Wed, May 30, 2012 at 6:26 AM, Mike  Dupont
jamesmikedup...@googlemail.com wrote:
 Ok, I merged the code from wikteam and have a full history dump script
 that uploads to archive.org,
 next step is to fix the bucket metadata in the script
 mike

 On Tue, May 29, 2012 at 3:08 AM, Mike  Dupont
 jamesmikedup...@googlemail.com wrote:
 Well, I have now updated the script to include  the xml dump in raw
 format. I will have to add more information the achive.org item, at
 least a basic readme.
 other thing is that the wikipybot does not support the full history it
 seems, so that I will have to move over to the wikiteam version and
 rework it,
 I just spent 2 hours on this so i am pretty happy for the first version.

 mike

 On Tue, May 29, 2012 at 1:52 AM, Hydriz Wikipedia ad...@alphacorp.tk wrote:
 This is quite nice, though the item's metadata is too little :)

 On Tue, May 29, 2012 at 3:40 AM, Mike Dupont jamesmikedup...@googlemail.com
 wrote:

 first version of the Script is ready , it gets the versions, puts them
 in a zip and puts that on archive.org
 https://github.com/h4ck3rm1k3/pywikipediabot/blob/master/export_deleted.py

 here is an example output :
 http://archive.org/details/wikipedia-delete-2012-05

 http://ia601203.us.archive.org/24/items/wikipedia-delete-2012-05/archive2012-05-28T21:34:02.302183.zip

 I will cron this, and it should give a start of saving deleted data.
 Articles will be exported once a day, even if they they were exported
 yesterday as long as they are in one of the categories.

 mike

 On Mon, May 21, 2012 at 7:21 PM, Mike  Dupont
 jamesmikedup...@googlemail.com wrote:
  Thanks! and run that 1 time per day, they dont get deleted that quickly.
  mike
 
  On Mon, May 21, 2012 at 9:11 PM, emijrp emi...@gmail.com wrote:
  Create a script that makes a request to Special:Export using this
 category
  as feed
  https://en.wikipedia.org/wiki/Category:Candidates_for_speedy_deletion
 
  More info
 https://www.mediawiki.org/wiki/Manual:Parameters_to_Special:Export
 
 
  2012/5/21 Mike Dupont jamesmikedup...@googlemail.com
 
  Well I whould be happy for items like this :
  http://en.wikipedia.org/wiki/Template:Db-a7
  would it be possible to extract them easily?
  mike
 
  On Thu, May 17, 2012 at 2:23 PM, Ariel T. Glenn ar...@wikimedia.org
  wrote:
   There's a few other reasons articles get deleted: copyright issues,
   personal identifying data, etc.  This makes maintaning the sort of
   mirror you propose problematic, although a similar mirror is here:
   http://deletionpedia.dbatley.com/w/index.php?title=Main_Page
  
   The dumps contain only data publically available at the time of the
 run,
   without deleted data.
  
   The articles aren't permanently deleted of course.  The revisions
 texts
   live on in the database, so a query on toolserver, for example,
 could be
   used to get at them, but that would need to be for research purposes.
  
   Ariel
  
   Στις 17-05-2012, ημέρα Πεμ, και ώρα 13:30 +0200, ο/η Mike Dupont
 έγραψε:
   Hi,
   I am thinking about how to collect articles deleted based on the
 not
   notable criteria,
   is there any way we can extract them from the mysql binlogs? how are
   these mirrors working? I would be interested in setting up a mirror
 of
   deleted data, at least that which is not spam/vandalism based on
 tags.
   mike
  
   On Thu, May 17, 2012 at 1:09 PM, Ariel T. Glenn 
 ar...@wikimedia.org
   wrote:
We now have three mirror sites, yay!  The full list is linked to
 from
http://dumps.wikimedia.org/ and is also available at
   
   
 http://meta.wikimedia.org/wiki/Mirroring_Wikimedia_project_XML_dumps#Current_Mirrors
   
Summarizing, we have:
   
C3L (Brazil) with the last 5 good known dumps,
Masaryk University (Czech Republic) with the last 5 known good
 dumps,
Your.org (USA) with the complete archive of dumps, and
   
for the latest version of uploaded media, Your.org with
http/ftp/rsync
access.
   
Thanks to Carlos, Kevin and Yenya respectively at the above sites
 for
volunteering space, time and effort to make this happen.
   
As people noticed earlier, a series of media tarballs per-project
(excluding commons) is being generated.  As soon as the first run
 of
these is complete we'll announce its location and start generating
them
on a semi-regular basis.
   
As we've been getting the bugs out of the mirroring setup, it is
getting
easier to add new locations.  Know anyone interested?  Please let
 us
know; we would love to have them.
   
Ariel
   
   
___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l
  
  
  
  
  
  
   ___
   Wikitech-l mailing list
   Wikitech-l@lists.wikimedia.org
   https://lists.wikimedia.org/mailman/listinfo/wikitech-l

Re: [Wikitech-l] Desktop upload tool

2012-05-28 Thread Mike Dupont
How will it be different than the commonist? geotagging would be one
big feature and the ability to group photos with common tags would be
useful!
mike

On Mon, May 28, 2012 at 7:16 AM, Petr Bena benap...@gmail.com wrote:
 I would name it Snuggle, so that it's in harmony with Huggle :)

 On Fri, May 25, 2012 at 5:55 PM, John phoenixoverr...@gmail.com wrote:
 One thing I have done quite a bit of was Image modification. Everything
 from watermark removal to coping to rotating. When Im on a spur ofH 
 watermark removal, it would be awesome if I could just point the program to
 a directory of fixed files, set a upload summary and have the bot upload
 those files keeping the existing file information. (that is the biggest
 drawback of  http://commons.wikimedia.org/wiki/Commons:Tools/Commonist )

 John

 On Fri, May 25, 2012 at 11:47 AM, Platonides platoni...@gmail.com wrote:

 As some of you are already aware, I'm doing for this GSoC a Desktop tool
 for conveniently uploading images to commons.
  http://thread.gmane.org/gmane.org.wikimedia.wikilovesmonuments/2641

 How can you help with this?

 a) Propose a cool name for this tool


 b) Suggest new functionalities/requisites
 Either on http://www.mediawiki.org/wiki/User:Platonides/GSOC_proposal or
 in this thread.


 c) Volunteer for testing the program and giving early feedback
 (please reply directly to avoid spamming the mailing list with I can
 test it, too)


 ___
 Wikitech-l mailing list
 Wikitech-l@lists.wikimedia.org
 https://lists.wikimedia.org/mailman/listinfo/wikitech-l

 ___
 Wikitech-l mailing list
 Wikitech-l@lists.wikimedia.org
 https://lists.wikimedia.org/mailman/listinfo/wikitech-l

 ___
 Wikitech-l mailing list
 Wikitech-l@lists.wikimedia.org
 https://lists.wikimedia.org/mailman/listinfo/wikitech-l



-- 
James Michael DuPont
Member of Free Libre Open Source Software Kosova http://flossk.org
Contributor FOSM, the CC-BY-SA map of the world http://fosm.org
Mozilla Rep https://reps.mozilla.org/u/h4ck3rm1k3

___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l


Re: [Wikitech-l] [Xmldatadumps-l] XML dumps/Media mirrors update

2012-05-28 Thread Mike Dupont
first version of the Script is ready , it gets the versions, puts them
in a zip and puts that on archive.org
https://github.com/h4ck3rm1k3/pywikipediabot/blob/master/export_deleted.py

here is an example output :
http://archive.org/details/wikipedia-delete-2012-05
http://ia601203.us.archive.org/24/items/wikipedia-delete-2012-05/archive2012-05-28T21:34:02.302183.zip

I will cron this, and it should give a start of saving deleted data.
Articles will be exported once a day, even if they they were exported
yesterday as long as they are in one of the categories.

mike

On Mon, May 21, 2012 at 7:21 PM, Mike  Dupont
jamesmikedup...@googlemail.com wrote:
 Thanks! and run that 1 time per day, they dont get deleted that quickly.
 mike

 On Mon, May 21, 2012 at 9:11 PM, emijrp emi...@gmail.com wrote:
 Create a script that makes a request to Special:Export using this category
 as feed
 https://en.wikipedia.org/wiki/Category:Candidates_for_speedy_deletion

 More info https://www.mediawiki.org/wiki/Manual:Parameters_to_Special:Export


 2012/5/21 Mike Dupont jamesmikedup...@googlemail.com

 Well I whould be happy for items like this :
 http://en.wikipedia.org/wiki/Template:Db-a7
 would it be possible to extract them easily?
 mike

 On Thu, May 17, 2012 at 2:23 PM, Ariel T. Glenn ar...@wikimedia.org
 wrote:
  There's a few other reasons articles get deleted: copyright issues,
  personal identifying data, etc.  This makes maintaning the sort of
  mirror you propose problematic, although a similar mirror is here:
  http://deletionpedia.dbatley.com/w/index.php?title=Main_Page
 
  The dumps contain only data publically available at the time of the run,
  without deleted data.
 
  The articles aren't permanently deleted of course.  The revisions texts
  live on in the database, so a query on toolserver, for example, could be
  used to get at them, but that would need to be for research purposes.
 
  Ariel
 
  Στις 17-05-2012, ημέρα Πεμ, και ώρα 13:30 +0200, ο/η Mike Dupont έγραψε:
  Hi,
  I am thinking about how to collect articles deleted based on the not
  notable criteria,
  is there any way we can extract them from the mysql binlogs? how are
  these mirrors working? I would be interested in setting up a mirror of
  deleted data, at least that which is not spam/vandalism based on tags.
  mike
 
  On Thu, May 17, 2012 at 1:09 PM, Ariel T. Glenn ar...@wikimedia.org
  wrote:
   We now have three mirror sites, yay!  The full list is linked to from
   http://dumps.wikimedia.org/ and is also available at
  
   http://meta.wikimedia.org/wiki/Mirroring_Wikimedia_project_XML_dumps#Current_Mirrors
  
   Summarizing, we have:
  
   C3L (Brazil) with the last 5 good known dumps,
   Masaryk University (Czech Republic) with the last 5 known good dumps,
   Your.org (USA) with the complete archive of dumps, and
  
   for the latest version of uploaded media, Your.org with
   http/ftp/rsync
   access.
  
   Thanks to Carlos, Kevin and Yenya respectively at the above sites for
   volunteering space, time and effort to make this happen.
  
   As people noticed earlier, a series of media tarballs per-project
   (excluding commons) is being generated.  As soon as the first run of
   these is complete we'll announce its location and start generating
   them
   on a semi-regular basis.
  
   As we've been getting the bugs out of the mirroring setup, it is
   getting
   easier to add new locations.  Know anyone interested?  Please let us
   know; we would love to have them.
  
   Ariel
  
  
   ___
   Wikitech-l mailing list
   Wikitech-l@lists.wikimedia.org
   https://lists.wikimedia.org/mailman/listinfo/wikitech-l
 
 
 
 
 
 
  ___
  Wikitech-l mailing list
  Wikitech-l@lists.wikimedia.org
  https://lists.wikimedia.org/mailman/listinfo/wikitech-l



 --
 James Michael DuPont
 Member of Free Libre Open Source Software Kosova http://flossk.org
 Contributor FOSM, the CC-BY-SA map of the world http://fosm.org
 Mozilla Rep https://reps.mozilla.org/u/h4ck3rm1k3

 ___
 Wikitech-l mailing list
 Wikitech-l@lists.wikimedia.org
 https://lists.wikimedia.org/mailman/listinfo/wikitech-l




 --
 Emilio J. Rodríguez-Posada. E-mail: emijrp AT gmail DOT com
 Pre-doctoral student at the University of Cádiz (Spain)
 Projects: AVBOT | StatMediaWiki | WikiEvidens | WikiPapers | WikiTeam
 Personal website: https://sites.google.com/site/emijrp/


 ___
 Xmldatadumps-l mailing list
 xmldatadump...@lists.wikimedia.org
 https://lists.wikimedia.org/mailman/listinfo/xmldatadumps-l




 --
 James Michael DuPont
 Member of Free Libre Open Source Software Kosova http://flossk.org
 Contributor FOSM, the CC-BY-SA map of the world http://fosm.org
 Mozilla Rep https://reps.mozilla.org/u/h4ck3rm1k3



-- 
James Michael DuPont
Member of Free Libre Open Source Software Kosova http://flossk.org
Contributor

Re: [Wikitech-l] [Xmldatadumps-l] XML dumps/Media mirrors update

2012-05-28 Thread Mike Dupont
Well, I have now updated the script to include  the xml dump in raw
format. I will have to add more information the achive.org item, at
least a basic readme.
other thing is that the wikipybot does not support the full history it
seems, so that I will have to move over to the wikiteam version and
rework it,
I just spent 2 hours on this so i am pretty happy for the first version.

mike

On Tue, May 29, 2012 at 1:52 AM, Hydriz Wikipedia ad...@alphacorp.tk wrote:
 This is quite nice, though the item's metadata is too little :)

 On Tue, May 29, 2012 at 3:40 AM, Mike Dupont jamesmikedup...@googlemail.com
 wrote:

 first version of the Script is ready , it gets the versions, puts them
 in a zip and puts that on archive.org
 https://github.com/h4ck3rm1k3/pywikipediabot/blob/master/export_deleted.py

 here is an example output :
 http://archive.org/details/wikipedia-delete-2012-05

 http://ia601203.us.archive.org/24/items/wikipedia-delete-2012-05/archive2012-05-28T21:34:02.302183.zip

 I will cron this, and it should give a start of saving deleted data.
 Articles will be exported once a day, even if they they were exported
 yesterday as long as they are in one of the categories.

 mike

 On Mon, May 21, 2012 at 7:21 PM, Mike  Dupont
 jamesmikedup...@googlemail.com wrote:
  Thanks! and run that 1 time per day, they dont get deleted that quickly.
  mike
 
  On Mon, May 21, 2012 at 9:11 PM, emijrp emi...@gmail.com wrote:
  Create a script that makes a request to Special:Export using this
 category
  as feed
  https://en.wikipedia.org/wiki/Category:Candidates_for_speedy_deletion
 
  More info
 https://www.mediawiki.org/wiki/Manual:Parameters_to_Special:Export
 
 
  2012/5/21 Mike Dupont jamesmikedup...@googlemail.com
 
  Well I whould be happy for items like this :
  http://en.wikipedia.org/wiki/Template:Db-a7
  would it be possible to extract them easily?
  mike
 
  On Thu, May 17, 2012 at 2:23 PM, Ariel T. Glenn ar...@wikimedia.org
  wrote:
   There's a few other reasons articles get deleted: copyright issues,
   personal identifying data, etc.  This makes maintaning the sort of
   mirror you propose problematic, although a similar mirror is here:
   http://deletionpedia.dbatley.com/w/index.php?title=Main_Page
  
   The dumps contain only data publically available at the time of the
 run,
   without deleted data.
  
   The articles aren't permanently deleted of course.  The revisions
 texts
   live on in the database, so a query on toolserver, for example,
 could be
   used to get at them, but that would need to be for research purposes.
  
   Ariel
  
   Στις 17-05-2012, ημέρα Πεμ, και ώρα 13:30 +0200, ο/η Mike Dupont
 έγραψε:
   Hi,
   I am thinking about how to collect articles deleted based on the
 not
   notable criteria,
   is there any way we can extract them from the mysql binlogs? how are
   these mirrors working? I would be interested in setting up a mirror
 of
   deleted data, at least that which is not spam/vandalism based on
 tags.
   mike
  
   On Thu, May 17, 2012 at 1:09 PM, Ariel T. Glenn 
 ar...@wikimedia.org
   wrote:
We now have three mirror sites, yay!  The full list is linked to
 from
http://dumps.wikimedia.org/ and is also available at
   
   
 http://meta.wikimedia.org/wiki/Mirroring_Wikimedia_project_XML_dumps#Current_Mirrors
   
Summarizing, we have:
   
C3L (Brazil) with the last 5 good known dumps,
Masaryk University (Czech Republic) with the last 5 known good
 dumps,
Your.org (USA) with the complete archive of dumps, and
   
for the latest version of uploaded media, Your.org with
http/ftp/rsync
access.
   
Thanks to Carlos, Kevin and Yenya respectively at the above sites
 for
volunteering space, time and effort to make this happen.
   
As people noticed earlier, a series of media tarballs per-project
(excluding commons) is being generated.  As soon as the first run
 of
these is complete we'll announce its location and start generating
them
on a semi-regular basis.
   
As we've been getting the bugs out of the mirroring setup, it is
getting
easier to add new locations.  Know anyone interested?  Please let
 us
know; we would love to have them.
   
Ariel
   
   
___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l
  
  
  
  
  
  
   ___
   Wikitech-l mailing list
   Wikitech-l@lists.wikimedia.org
   https://lists.wikimedia.org/mailman/listinfo/wikitech-l
 
 
 
  --
  James Michael DuPont
  Member of Free Libre Open Source Software Kosova http://flossk.org
  Contributor FOSM, the CC-BY-SA map of the world http://fosm.org
  Mozilla Rep https://reps.mozilla.org/u/h4ck3rm1k3
 
  ___
  Wikitech-l mailing list
  Wikitech-l@lists.wikimedia.org
  https://lists.wikimedia.org/mailman/listinfo/wikitech-l

Re: [Wikitech-l] XML dumps/Media mirrors update

2012-05-21 Thread Mike Dupont
Well I whould be happy for items like this :
http://en.wikipedia.org/wiki/Template:Db-a7
would it be possible to extract them easily?
mike

On Thu, May 17, 2012 at 2:23 PM, Ariel T. Glenn ar...@wikimedia.org wrote:
 There's a few other reasons articles get deleted: copyright issues,
 personal identifying data, etc.  This makes maintaning the sort of
 mirror you propose problematic, although a similar mirror is here:
 http://deletionpedia.dbatley.com/w/index.php?title=Main_Page

 The dumps contain only data publically available at the time of the run,
 without deleted data.

 The articles aren't permanently deleted of course.  The revisions texts
 live on in the database, so a query on toolserver, for example, could be
 used to get at them, but that would need to be for research purposes.

 Ariel

 Στις 17-05-2012, ημέρα Πεμ, και ώρα 13:30 +0200, ο/η Mike Dupont έγραψε:
 Hi,
 I am thinking about how to collect articles deleted based on the not
 notable criteria,
 is there any way we can extract them from the mysql binlogs? how are
 these mirrors working? I would be interested in setting up a mirror of
 deleted data, at least that which is not spam/vandalism based on tags.
 mike

 On Thu, May 17, 2012 at 1:09 PM, Ariel T. Glenn ar...@wikimedia.org wrote:
  We now have three mirror sites, yay!  The full list is linked to from
  http://dumps.wikimedia.org/ and is also available at
  http://meta.wikimedia.org/wiki/Mirroring_Wikimedia_project_XML_dumps#Current_Mirrors
 
  Summarizing, we have:
 
  C3L (Brazil) with the last 5 good known dumps,
  Masaryk University (Czech Republic) with the last 5 known good dumps,
  Your.org (USA) with the complete archive of dumps, and
 
  for the latest version of uploaded media, Your.org with http/ftp/rsync
  access.
 
  Thanks to Carlos, Kevin and Yenya respectively at the above sites for
  volunteering space, time and effort to make this happen.
 
  As people noticed earlier, a series of media tarballs per-project
  (excluding commons) is being generated.  As soon as the first run of
  these is complete we'll announce its location and start generating them
  on a semi-regular basis.
 
  As we've been getting the bugs out of the mirroring setup, it is getting
  easier to add new locations.  Know anyone interested?  Please let us
  know; we would love to have them.
 
  Ariel
 
 
  ___
  Wikitech-l mailing list
  Wikitech-l@lists.wikimedia.org
  https://lists.wikimedia.org/mailman/listinfo/wikitech-l






 ___
 Wikitech-l mailing list
 Wikitech-l@lists.wikimedia.org
 https://lists.wikimedia.org/mailman/listinfo/wikitech-l



-- 
James Michael DuPont
Member of Free Libre Open Source Software Kosova http://flossk.org
Contributor FOSM, the CC-BY-SA map of the world http://fosm.org
Mozilla Rep https://reps.mozilla.org/u/h4ck3rm1k3

___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l

Re: [Wikitech-l] [Xmldatadumps-l] XML dumps/Media mirrors update

2012-05-21 Thread Mike Dupont
Thanks! and run that 1 time per day, they dont get deleted that quickly.
mike

On Mon, May 21, 2012 at 9:11 PM, emijrp emi...@gmail.com wrote:
 Create a script that makes a request to Special:Export using this category
 as feed
 https://en.wikipedia.org/wiki/Category:Candidates_for_speedy_deletion

 More info https://www.mediawiki.org/wiki/Manual:Parameters_to_Special:Export


 2012/5/21 Mike Dupont jamesmikedup...@googlemail.com

 Well I whould be happy for items like this :
 http://en.wikipedia.org/wiki/Template:Db-a7
 would it be possible to extract them easily?
 mike

 On Thu, May 17, 2012 at 2:23 PM, Ariel T. Glenn ar...@wikimedia.org
 wrote:
  There's a few other reasons articles get deleted: copyright issues,
  personal identifying data, etc.  This makes maintaning the sort of
  mirror you propose problematic, although a similar mirror is here:
  http://deletionpedia.dbatley.com/w/index.php?title=Main_Page
 
  The dumps contain only data publically available at the time of the run,
  without deleted data.
 
  The articles aren't permanently deleted of course.  The revisions texts
  live on in the database, so a query on toolserver, for example, could be
  used to get at them, but that would need to be for research purposes.
 
  Ariel
 
  Στις 17-05-2012, ημέρα Πεμ, και ώρα 13:30 +0200, ο/η Mike Dupont έγραψε:
  Hi,
  I am thinking about how to collect articles deleted based on the not
  notable criteria,
  is there any way we can extract them from the mysql binlogs? how are
  these mirrors working? I would be interested in setting up a mirror of
  deleted data, at least that which is not spam/vandalism based on tags.
  mike
 
  On Thu, May 17, 2012 at 1:09 PM, Ariel T. Glenn ar...@wikimedia.org
  wrote:
   We now have three mirror sites, yay!  The full list is linked to from
   http://dumps.wikimedia.org/ and is also available at
  
   http://meta.wikimedia.org/wiki/Mirroring_Wikimedia_project_XML_dumps#Current_Mirrors
  
   Summarizing, we have:
  
   C3L (Brazil) with the last 5 good known dumps,
   Masaryk University (Czech Republic) with the last 5 known good dumps,
   Your.org (USA) with the complete archive of dumps, and
  
   for the latest version of uploaded media, Your.org with
   http/ftp/rsync
   access.
  
   Thanks to Carlos, Kevin and Yenya respectively at the above sites for
   volunteering space, time and effort to make this happen.
  
   As people noticed earlier, a series of media tarballs per-project
   (excluding commons) is being generated.  As soon as the first run of
   these is complete we'll announce its location and start generating
   them
   on a semi-regular basis.
  
   As we've been getting the bugs out of the mirroring setup, it is
   getting
   easier to add new locations.  Know anyone interested?  Please let us
   know; we would love to have them.
  
   Ariel
  
  
   ___
   Wikitech-l mailing list
   Wikitech-l@lists.wikimedia.org
   https://lists.wikimedia.org/mailman/listinfo/wikitech-l
 
 
 
 
 
 
  ___
  Wikitech-l mailing list
  Wikitech-l@lists.wikimedia.org
  https://lists.wikimedia.org/mailman/listinfo/wikitech-l



 --
 James Michael DuPont
 Member of Free Libre Open Source Software Kosova http://flossk.org
 Contributor FOSM, the CC-BY-SA map of the world http://fosm.org
 Mozilla Rep https://reps.mozilla.org/u/h4ck3rm1k3

 ___
 Wikitech-l mailing list
 Wikitech-l@lists.wikimedia.org
 https://lists.wikimedia.org/mailman/listinfo/wikitech-l




 --
 Emilio J. Rodríguez-Posada. E-mail: emijrp AT gmail DOT com
 Pre-doctoral student at the University of Cádiz (Spain)
 Projects: AVBOT | StatMediaWiki | WikiEvidens | WikiPapers | WikiTeam
 Personal website: https://sites.google.com/site/emijrp/


 ___
 Xmldatadumps-l mailing list
 xmldatadump...@lists.wikimedia.org
 https://lists.wikimedia.org/mailman/listinfo/xmldatadumps-l




-- 
James Michael DuPont
Member of Free Libre Open Source Software Kosova http://flossk.org
Contributor FOSM, the CC-BY-SA map of the world http://fosm.org
Mozilla Rep https://reps.mozilla.org/u/h4ck3rm1k3

___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l

Re: [Wikitech-l] XML dumps/Media mirrors update

2012-05-17 Thread Mike Dupont
Hi,
I am thinking about how to collect articles deleted based on the not
notable criteria,
is there any way we can extract them from the mysql binlogs? how are
these mirrors working? I would be interested in setting up a mirror of
deleted data, at least that which is not spam/vandalism based on tags.
mike

On Thu, May 17, 2012 at 1:09 PM, Ariel T. Glenn ar...@wikimedia.org wrote:
 We now have three mirror sites, yay!  The full list is linked to from
 http://dumps.wikimedia.org/ and is also available at
 http://meta.wikimedia.org/wiki/Mirroring_Wikimedia_project_XML_dumps#Current_Mirrors

 Summarizing, we have:

 C3L (Brazil) with the last 5 good known dumps,
 Masaryk University (Czech Republic) with the last 5 known good dumps,
 Your.org (USA) with the complete archive of dumps, and

 for the latest version of uploaded media, Your.org with http/ftp/rsync
 access.

 Thanks to Carlos, Kevin and Yenya respectively at the above sites for
 volunteering space, time and effort to make this happen.

 As people noticed earlier, a series of media tarballs per-project
 (excluding commons) is being generated.  As soon as the first run of
 these is complete we'll announce its location and start generating them
 on a semi-regular basis.

 As we've been getting the bugs out of the mirroring setup, it is getting
 easier to add new locations.  Know anyone interested?  Please let us
 know; we would love to have them.

 Ariel


 ___
 Wikitech-l mailing list
 Wikitech-l@lists.wikimedia.org
 https://lists.wikimedia.org/mailman/listinfo/wikitech-l



-- 
James Michael DuPont
Member of Free Libre Open Source Software Kosova http://flossk.org
Contributor FOSM, the CC-BY-SA map of the world http://fosm.org
Mozilla Rep https://reps.mozilla.org/u/h4ck3rm1k3

___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l


Re: [Wikitech-l] Mark Hershberger departing Wikimedia Foundation in May

2012-03-16 Thread Mike Dupont
Yeah Mark, It was great to have worked with you, I am sorry that I did not
have more time for wikipedia myself,
but such is life.

mike

On Fri, Mar 16, 2012 at 8:01 PM, Bináris wikipo...@gmail.com wrote:

 Mark, thanks for helping Wikipedia and good luck whereever you go!

 --
 Bináris
 ___
 Wikitech-l mailing list
 Wikitech-l@lists.wikimedia.org
 https://lists.wikimedia.org/mailman/listinfo/wikitech-l




-- 
James Michael DuPont
Member of Free Libre Open Source Software Kosova http://flossk.org
___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l


Re: [Wikitech-l] Moving forward with Lua

2012-02-28 Thread Mike Dupont
very cool.

how can we see the lua errors?
how can we load wiki pages as data in lua?

thanks,
mike

On Tue, Feb 28, 2012 at 11:01 PM, Owen Davis o...@wikia-inc.com wrote:

 For anyone interested in playing around with Lua on a wiki, I've enabled
 the Lua extension on http://lua.wikia.com
 It's just a test wiki right now with no real content.  It will probably
 remain a test wiki, although if users want to flesh it out into a more
 authoritative site, that would be cool too.  I am using
 http://www.mediawiki.org/wiki/Extension:Lua




-- 
James Michael DuPont
Member of Free Libre Open Source Software Kosova http://flossk.org
___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l


Re: [Wikitech-l] Using computer vision to categorize images at Commons

2012-02-21 Thread Mike Dupont
On Tue, Feb 21, 2012 at 8:09 AM, Antoine Musso hashar+...@free.fr wrote:
 Le 20/02/12 22:08, Maarten Dammers a écrit :

 Some time ago I played around with computer vision to get images
 categorized on Commons. I documented this at

 https://commons.wikimedia.org/wiki/User:Multichill/Using_OpenCV_to_categorize_files
 . I don't think I'm going to spend time on it soon, but the results were
 quite promising, so maybe someone else feels like working on this? Would
 probably be a pretty nice student project or just fun to do.


 Google used a game to add keywords to images. Two people (of of them could
 be a robot) were presented two pictures. The aim is to enter a keyword that
 best describe the picture and that the other people will choose too. If both
 people use the same keyword, you gain points :-)

 Imagine, you are being shown a picture whose main subject is a monkey, you
 enter monkey, the other one monkey too. You both earn points. The game
 host now know that the picture is of a monkey :-D

The Google Image Labeler, there are (funny and offensive) videos of
that online.
It is a good idea. Mediawiki does not need a person to check it live,
it has users that are reviewing the changes anyway.

It would be nice to have a *simple* way to put people down and have
them label images, I could find people if it was really simple.
Right now it is not really simple.

Most people just dont have the concentration to read the docs or learn
how to use the wiki.

mike

___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l


Re: [Wikitech-l] Common photo management plugins for MediaWiki

2012-02-12 Thread Mike Dupont
Thats great news! Any code to find? digikam is great, but also huge. I am
using shotwell all the time.
mike

On Sun, Feb 12, 2012 at 3:32 PM, Guillaume Paumier
gpaum...@wikimedia.orgwrote:


 FYI, there's also a group of French students working on a (partially
 developed) export plugin to Wikimedia Commons from digiKam (
 http://www.digikam.org ), and they may also create an export plugin
 for Shotwell too ( http://yorba.org/shotwell/ ), depending on their
 progress.




-- 
James Michael DuPont
Member of Free Libre Open Source Software Kosova http://flossk.org
___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l


Re: [Wikitech-l] Common photo management plugins for MediaWiki

2012-02-12 Thread Mike Dupont
This works,
http://en.wikipedia.org/wiki/Wikipedia:Multimedia_beta#Add_Media_Wizard
the new wizard is great!

Has outdated information :
http://www.mediawiki.org/wiki/Extension:Add_Media_Wizard

mike


On Sun, Feb 12, 2012 at 3:42 PM, Helder helder.w...@gmail.com wrote:

 http://www.mediawiki.org/wiki/Extension:Add_Media_Wizard
 http://en.wikipedia.org/wiki/Wikipedia:Multimedia_beta#Add_Media_Wizard




-- 
James Michael DuPont
Member of Free Libre Open Source Software Kosova http://flossk.org
___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l


Re: [Wikitech-l] Common photo management plugins for MediaWiki

2012-02-12 Thread Mike Dupont
Yes, I would like to be able to select a few pictures, and be able to tag
them together, add them to categories and such. The upload media wizard is
great, and I would love if I could apply a category to many items.
mike
On Sat, Feb 11, 2012 at 12:48 PM, lakmal padmakumara 
lakmalpadmakum...@gmail.com wrote:

 GUI way of managing pictures .




-- 
James Michael DuPont
Member of Free Libre Open Source Software Kosova http://flossk.org
___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l


Re: [Wikitech-l] Common photo management plugins for MediaWiki

2012-02-10 Thread Mike Dupont
Hi there
It would be great to have a better photo management tool.
Currently I am using google to host my creative commons photos, and using
the easy upload tools from google plus to load into the albums,
from there they can be imported into panoramio and imported into mediawiki.

So, really I find that that best way to upload photos is via google, and
then import them into commons.

the things that are really needed is
1. GUI way to manage the geotagging of photos.
2. an easy way to import existing photos.
3. a better way to manage many photos.
4. Import of files from archive.org would also be good.

the entire system of processing photos for commons is too complex for
beginners.

mike

On Sat, Feb 11, 2012 at 7:54 AM, lakmal padmakumara 
lakmalpadmakum...@gmail.com wrote:

 Hey Thanks for the quick reply .
 It is the third idea which is mentioned in this suggested
 ideashttp://www.mediawiki.org/wiki/Summer_of_Code_2012#Project_ideaslist
 .
 At the moment I'm doing a research on the availability of APIs of these
 photo management Softwares .I am hoping to come up with  a draft proposal
 which includes possible plugin list and basic implementation details of the
 project in next few days .I will keep community updated about that .

 Meanwhile any guidelines ,tips or comments from the community is more than
 welcome and highly appreciated .

 On Sat, Feb 11, 2012 at 3:30 AM, Platonides platoni...@gmail.com wrote:

  Hello lakmal,
  Good to meet you. This is indeed the right place.
  I don't see where it was suggested, but I haven't heard anyone to be
  working on that idea, so it seems perfectly available.
  I encourage you to prepare a draft proposal based on that.
 
 
  ___
  Wikitech-l mailing list
  Wikitech-l@lists.wikimedia.org
  https://lists.wikimedia.org/mailman/listinfo/wikitech-l
 



 --
 Thanks  Kind Regards

 Lakmal Padmakumara
 Undergraduate
 Computer Science and Engineering Department
 University of Moratuwa
 Sri Lanka
 http://srilankatechnology.blogspot.com/
 ___
 Wikitech-l mailing list
 Wikitech-l@lists.wikimedia.org
 https://lists.wikimedia.org/mailman/listinfo/wikitech-l




-- 
James Michael DuPont
Member of Free Libre Open Source Software Kosova http://flossk.org
___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l


Re: [Wikitech-l] SOPA commercial social networks

2012-01-18 Thread Mike Dupont
ON the german site, they have identi.ca as well. diaspora is missing on all.

It would be nice to have these social tools as opt in tools for
wikipedia in general.
I dont see why we can have them for the fundraising or when wp wants
to be known, but not for articles.

mike

On Wed, Jan 18, 2012 at 10:59 AM, David Gerard dger...@gmail.com wrote:
 On 18 January 2012 06:19, Thomas Schmidt schm...@netaction.de wrote:

 Clicking on the Google+ button I am forced to accept the tracking of
 my location and some new general business terms.


 Yeah - is there any sort of G+ link possible that *doesn't* demand you
 sign up for the mobile TOS (even on a desktop)? Linking to a
 commercial data-gathering exercise is less than ideal ...


 - d.

 ___
 Wikitech-l mailing list
 Wikitech-l@lists.wikimedia.org
 https://lists.wikimedia.org/mailman/listinfo/wikitech-l



-- 
James Michael DuPont
Member of Free Libre Open Source Software Kosova http://flossk.org

___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l


Re: [Wikitech-l] SOPA commercial social networks

2012-01-18 Thread Mike Dupont
well, when it serves the purpose, then wp shows no shame in promoting
some other sites.

It is just inconsistent. but I guess I dont have to say that?

Sometimes it is ok, sometimes it is not ok.

mike

On Wed, Jan 18, 2012 at 11:12 AM, David Gerard dger...@gmail.com wrote:
 On 18 January 2012 10:03, Mike  Dupont jamesmikedup...@googlemail.com wrote:

 ON the german site, they have identi.ca as well. diaspora is missing on all.
 It would be nice to have these social tools as opt in tools for
 wikipedia in general.
 I dont see why we can have them for the fundraising or when wp wants
 to be known, but not for articles.


 The usual barrier on en:wp has been neutrality concerns, and being
 seen to advocate particular services over others.


 - d.

 ___
 Wikitech-l mailing list
 Wikitech-l@lists.wikimedia.org
 https://lists.wikimedia.org/mailman/listinfo/wikitech-l



-- 
James Michael DuPont
Member of Free Libre Open Source Software Kosova http://flossk.org

___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l


Re: [Wikitech-l] SOPA commercial social networks

2012-01-18 Thread Mike Dupont
Ok, thanks mark. that is good to know.
Will look into this some time.
mike

On Wed, Jan 18, 2012 at 1:53 PM, Mark A. Hershberger
mhershber...@wikimedia.org wrote:
 Mike  Dupont jamesmikedup...@googlemail.com writes:

 It would be nice to have these social tools as opt in tools for
 wikipedia in general.

 Unless I'm mistaken, Admins on-wiki could do this without consulting the
 WMF by installing a gadget which people could then opt-in to use.

 As far as built-in functionality, the first thing to do is to get an
 extension that provides this reviewed.
 https://www.mediawiki.org/wiki/Review_queue does not show anything that
 would provide this.

 Mark.


 ___
 Wikitech-l mailing list
 Wikitech-l@lists.wikimedia.org
 https://lists.wikimedia.org/mailman/listinfo/wikitech-l



-- 
James Michael DuPont
Member of Free Libre Open Source Software Kosova http://flossk.org

___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l


Re: [Wikitech-l] Performance roadmap update

2011-12-20 Thread Mike Dupont
Well thanks for sharing this, I will also put my efforts on HPHP on
hold for now as well.
mike

On Thu, Dec 15, 2011 at 3:26 PM, Jay Ashworth j...@baylink.com wrote:
 - Original Message -
 From: Tim Starling tstarl...@wikimedia.org

 So we've decided to defer our HipHop deployment until hhvm is at a
 suitable level of maturity. We don't know exactly when that will be,
 but Jason Evans says in the note linked above that the first 90% is
 done; now we're on to the second 90% as we make it really shine.

 It's good to hear that Jason understands the realities of tool development,
 isn't it?

 Cheers,
 -- jra
 --
 Jay R. Ashworth                  Baylink                       
 j...@baylink.com
 Designer                     The Things I Think                       RFC 2100
 Ashworth  Associates     http://baylink.pitas.com         2000 Land Rover DII
 St Petersburg FL USA      http://photo.imageinc.us             +1 727 647 1274

 ___
 Wikitech-l mailing list
 Wikitech-l@lists.wikimedia.org
 https://lists.wikimedia.org/mailman/listinfo/wikitech-l



-- 
James Michael DuPont
Member of Free Libre Open Source Software Kosova http://flossk.org

___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l


Re: [Wikitech-l] Access to the Deleted Archives

2011-12-11 Thread Mike Dupont
I have also wanted this for a long time.
http://www.petitiononline.com/urmwpnow/petition.html
http://undeletewikipedia.blogspot.com/2009/10/clarity-in-petition.html
mike

On Sun, Dec 11, 2011 at 1:32 PM, Katja Müller katja_muel...@lavabit.com wrote:
 Hi there,

 we are doing information retrieval research on the Wikipedia history.
 Currently we are thinking about including the archive of Deleted
 Articles in the analysis.

 What is the current regulation on access to the Deleted Archive?
 According to this page, admins can permit access to single articles on
 request:
 http://en.wikipedia.org/wiki/Wikipedia:Deletion_policy#Access_to_deleted_pages

 However, is there a way (for researchers) to either:
 a) access (API) or download the whole archive of Deleted Articles
 b) get statistics or meta data about the Deleted Archive (article
 counts, revision meta information, logs)
 ?


 Kind regards,
 Katja Mueller


 ___
 Wikitech-l mailing list
 Wikitech-l@lists.wikimedia.org
 https://lists.wikimedia.org/mailman/listinfo/wikitech-l



-- 
James Michael DuPont
Member of Free Libre Open Source Software Kosova http://flossk.org

___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l


Re: [Wikitech-l] Access to the Deleted Archives

2011-12-11 Thread Mike Dupont
On Sun, Dec 11, 2011 at 10:43 PM, Platonides platoni...@gmail.com wrote:
 On 11/12/11 13:38, Mike Dupont wrote:
 I have also wanted this for a long time.
 http://www.petitiononline.com/urmwpnow/petition.html
 http://undeletewikipedia.blogspot.com/2009/10/clarity-in-petition.html
 mike

 I don't think any sysop would reject a reasonable petition of a good
 wikipedian of the content of an article he previously wrote.

 (good wikipedian meaning it's not an instance of I want that copyvio
 content to recreate it from sockpuppet accounts using proxies)

 There are many articles that have been deleted, and we dont even know the 
 full list of them.

 That's not true. The list of deleted pages is available at
 http://en.wikipedia.org/wiki/Special:Log/delete
 For very old deletions (before December 23, 2004) see
 http://en.wikipedia.org/wiki/Wikipedia:Deletion_log

 However, for most articles the deleted text would (should?) be in the
 lines of
 Article: John Smith
 Content: He's fat child in classroom.

 In other cases they may be well-written articles which violate the
 copyright of eg. Encarta. So they can't be shown either.

 You seem to be targetting good articles deleted due to the target being
 non-notable, and you indeed have a point for them.
 You could launch a project to host those files if you wish to (I think
 there was already one doing it? at least many wikis have spun off to
 their own wiki about their topic).


Well, I dont know what to say, except that I put work into my
articles, and having them deleted hurts. I would like to have them in
my userspace at least.

mike

___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l


Re: [Wikitech-l] Access to the Deleted Archives

2011-12-11 Thread Mike Dupont
I can understand that about the user space. good point about the blog.

mike

On Sun, Dec 11, 2011 at 11:38 PM, Platonides platoni...@gmail.com wrote:
 On 11/12/11 22:44, Mike Dupont wrote:
 Well, I dont know what to say, except that I put work into my
 articles, and having them deleted hurts. I would like to have them in
 my userspace at least.

 mike

 I understand you. Having them in your user space may be a bit
 controversial (I see arguments for both for and against) but there would
 certainly be no problem in you publishing them in eg. your own blog.



 ___
 Wikitech-l mailing list
 Wikitech-l@lists.wikimedia.org
 https://lists.wikimedia.org/mailman/listinfo/wikitech-l



-- 
James Michael DuPont
Member of Free Libre Open Source Software Kosova http://flossk.org

___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l


Re: [Wikitech-l] Extracting text from Wikipedia

2011-11-27 Thread Mike Dupont
you want from all articles or just one?
see here
http://stackoverflow.com/questions/627594/is-there-a-wikipedia-api

mike

On Sun, Nov 27, 2011 at 6:02 PM, Khalida BEN SIDI AHMED
send.to.khal...@gmail.com wrote:
 Hello!
 I don't know if the subject of this question belongs to the scope of this
 group. Anyway, I will be pleased if I find an aswer to my question.
 I'm writing some Java code in order to realize NLP tasks upon texts using
 Wikipedia. What can I do in order to extract the first paragraph of a
 Wikipedia article? Thanks a lot.

 Truly yours
 Ben Sidi Ahmed
 ___
 Wikitech-l mailing list
 Wikitech-l@lists.wikimedia.org
 https://lists.wikimedia.org/mailman/listinfo/wikitech-l




-- 
James Michael DuPont
Member of Free Libre Open Source Software Kosova http://flossk.org

___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l


Re: [Wikitech-l] Extracting text from Wikipedia

2011-11-27 Thread Mike Dupont
look,
for 100 articles, just create a list of them, and export them as xml.
or use the book creator.
http://en.wikipedia.org/wiki/Help:Books

also there is a json api to pull single articles.
http://www.barattalo.it/2010/08/29/php-bot-to-get-wikipedia-definitions/

mike

On Sun, Nov 27, 2011 at 6:15 PM, Khalida BEN SIDI AHMED
send.to.khal...@gmail.com wrote:
 I have already read the responses given in this post.

 I want to the extract the first paragraph (or the first sentence) for a
 list of 100 articles.
 I could not use JWPL beacause I don't have a big hard disk space to create
 the DB. I try to use JSoup but I need examples.
 ___
 Wikitech-l mailing list
 Wikitech-l@lists.wikimedia.org
 https://lists.wikimedia.org/mailman/listinfo/wikitech-l




-- 
James Michael DuPont
Member of Free Libre Open Source Software Kosova http://flossk.org

___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l


Re: [Wikitech-l] Extracting text from Wikipedia

2011-11-27 Thread Mike Dupont
http://code.google.com/p/jwpl/ this looks also good
JWPL (Java Wikipedia Library) is a free, Java-based application
programming interface that allows to access all information contained
in Wikipedia.

have not tried that , but you said you wanted to do it in java.
mike

On Sun, Nov 27, 2011 at 6:20 PM, Mike  Dupont
jamesmikedup...@googlemail.com wrote:
 look,
 for 100 articles, just create a list of them, and export them as xml.
 or use the book creator.
 http://en.wikipedia.org/wiki/Help:Books

 also there is a json api to pull single articles.
 http://www.barattalo.it/2010/08/29/php-bot-to-get-wikipedia-definitions/

 mike

 On Sun, Nov 27, 2011 at 6:15 PM, Khalida BEN SIDI AHMED
 send.to.khal...@gmail.com wrote:
 I have already read the responses given in this post.

 I want to the extract the first paragraph (or the first sentence) for a
 list of 100 articles.
 I could not use JWPL beacause I don't have a big hard disk space to create
 the DB. I try to use JSoup but I need examples.
 ___
 Wikitech-l mailing list
 Wikitech-l@lists.wikimedia.org
 https://lists.wikimedia.org/mailman/listinfo/wikitech-l




 --
 James Michael DuPont
 Member of Free Libre Open Source Software Kosova http://flossk.org




-- 
James Michael DuPont
Member of Free Libre Open Source Software Kosova http://flossk.org

___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l


Re: [Wikitech-l] Extracting text from Wikipedia

2011-11-27 Thread Mike Dupont
have you seen wordnet,http://en.wikipedia.org/wiki/WordNet
wikitionary?
in any case you can call the export routine as you need it, even for
single articles. I think it would not cause too much load on the
server.
anyway, good luck

mike

On Sun, Nov 27, 2011 at 6:36 PM, Khalida BEN SIDI AHMED
send.to.khal...@gmail.com wrote:
 The list of the articles I will need is not known from the beggining.
 Through my project, I will find a list of words (50). I try to find for
 them definitions in Wikipedia. After that I will extract the hyperonym of
 each word. I will have a new list for which I then retrieve the respective
 articles...etc. So the the list of articles I need grows through the
 evolution of the algorithm' execution.
 ___
 Wikitech-l mailing list
 Wikitech-l@lists.wikimedia.org
 https://lists.wikimedia.org/mailman/listinfo/wikitech-l




-- 
James Michael DuPont
Member of Free Libre Open Source Software Kosova http://flossk.org

___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l


Re: [Wikitech-l] Extracting text from Wikipedia

2011-11-27 Thread Mike Dupont
If you are focused on jsoup, then best to ask there on the mailing
list http://jsoup.org/discussion

On Sun, Nov 27, 2011 at 7:51 PM, Khalida BEN SIDI AHMED
send.to.khal...@gmail.com wrote:
 Thank you Hoehrmann. I will try to apply the options you've mentionned.
 However, if someone can help me in using JSoup, his ideas are welcomed.
 ___
 Wikitech-l mailing list
 Wikitech-l@lists.wikimedia.org
 https://lists.wikimedia.org/mailman/listinfo/wikitech-l




-- 
James Michael DuPont
Member of Free Libre Open Source Software Kosova http://flossk.org

___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l


Re: [Wikitech-l] Progress made on packaging hiphop for debian

2011-09-20 Thread Mike Dupont
I've been working on a replacement for the libcurl patch.
-- Tim Starling
Thats Good news,
Last night I removed all traces of the curl_multi_select from the code, no
idea if it runs properly, but it produces a package. Also I commented out
all the non standard libevent code.
The package has a compiler that produces code, and you can at least
experiment with it.
We need more people testing, I am getting help from somone from the debian
team as well.
Will continue working on this as I find time.


we are organizing a conference on Nov 12 if anyone is interested,
www.flossk.org/en/blog/sfk11 that is also taking up time.

thanks,

mike


On Tue, Sep 20, 2011 at 11:43 PM, Tim Starling tstarl...@wikimedia.orgwrote:

 On 20/09/11 11:05, Mark A. Hershberger wrote:
  At that point, I contacted Mike and asked for his help.  He suggested we
  attempt to get HipHop compiled *without* the patches since the upstream
  developers weren't too enthusiastic about them.



 ___
 Wikitech-l mailing list
 Wikitech-l@lists.wikimedia.org
 https://lists.wikimedia.org/mailman/listinfo/wikitech-l




-- 
James Michael DuPont
Member of Free Libre Open Source Software Kosova http://flossk.org
___
Wikitech-l mailing list
Wikitech-l@lists.wikimedia.org
https://lists.wikimedia.org/mailman/listinfo/wikitech-l