In particular, we are using this dataset: http://dumps.wikimedia.org/enwiki/latest/enwiki-latest-pages-articles.xml.bz2
Ankur <http://www.ankurdave.com/> On Sun, Mar 30, 2014 at 12:45 AM, Ankur Dave <ankurd...@gmail.com> wrote: > The GraphX team has been using Wikipedia dumps from > http://dumps.wikimedia.org/enwiki/. Unfortunately, these are in a less > convenient format than the Freebase dumps. In particular, an article may > span multiple lines, so more involved input parsing is required. > > Dan Crankshaw (cc'd) wrote a driver that uses a Hadoop InputFormat XML > parser from Mahout: see > WikiPipelineBenchmark.scala<https://github.com/amplab/graphx/blob/860918486a81cb4c88a056a9b64b1f7d8b0ed5ff/graphx/src/main/scala/org/apache/spark/graphx/WikiPipelineBenchmark.scala#L157>and > WikiArticle.scala<https://github.com/amplab/graphx/blob/860918486a81cb4c88a056a9b64b1f7d8b0ed5ff/graphx/src/main/scala/org/apache/spark/graphx/WikiArticle.scala> > . > > However, we plan to upload a parsed version of this dataset to S3 for > easier access from Spark and GraphX. > > Ankur <http://www.ankurdave.com/> > > On 27 Mar, 2014, at 9:45 pm, Niko Stahl <r.niko.st...@gmail.com> wrote: > > I would like to run the > WikipediaPageRank<https://github.com/amplab/graphx/blob/f8544981a6d05687fa950639cb1eb3c31e9b6bf5/examples/src/main/scala/org/apache/spark/examples/bagel/WikipediaPageRank.scala>example, > but the Wikipedia dump XML files are no longer available on >> Freebase. Does anyone know an alternative source for the data? >> > >