Hi Raymond,

Don't know Bixo, how does it compare Nutch.

Very early project to create a web crawler toolkit (versus a system like Nutch).

What do you mean by copy to S3...

Amazon's S3 (Simple Storage Service - http://s3.amazonaws.com/) is what I use for persistent storage, since the clusters go up/down as needed. The cost of storage in S3 is almost negligible, though each transfer has a small price.

A potential major drawback is the time & cost of getting content out of S3 to some other, non-AWS set of servers. E.g. if you want to run your search servers in your own data center, but crawl using EC2, the time & cost of moving the content could be excessive. Though Amazon recently introduced AWS Import/Export to help address this issue.

-- Ken


2009/6/5 Ken Krugler <[email protected]>

 how long does it take for your 6 millions URLs to be
 crawled/parsed/indexed... I'm curious to know because I'm about to shoot
 in
 this area but I have no idea how long it will take.


 Not sure this helps, but below are some general stats from a fresh crawl I
 just did using Bixo, which should be similar to Nutch times.

 This is for a 5 server cluster (1 master, 4 slaves) of lowest-end boxes at
 Amazon's EC2.

 * 1.2M URLs from 42K domains
 * Fetch took 3.5 hours, mostly due to long tail
 * Parse took 2.5 hours
 * Index took 25 minutes
 * Copy to S3 took 1 hour

 Adding more servers would have made some things faster, but it wouldn't
 have significantly increased the speed of the fetch, due to the limited
 number of domains.

 So total time was 7.5 hours. 8 hours * 5 servers = 40 compute hours, or $4
 using EC2 pricing of $0.10/hr.

 Amazon's Elastic MapReduce (EMR) is slightly more expensive, but you could
 avoid dealing with Hadoop config/setup time.

 -- Ken

  2009/6/5 John Martyniak <[email protected]>

   Arkady,

   > I think that is beauty of nutch I have built a index of a little more
 6

  million urls with "out of the box" Nutch.  I would say that is pretty
 good
  for most situations before you have to start getting into hadoop and

  > multiple machines.
  >

  -John


  On Jun 4, 2009, at 5:19 PM, <[email protected]> wrote:

  Hi Andrzej,


  -----Original Message-----

  From: Andrzej Bialecki [mailto:[email protected]]
  Sent: Thursday, June 04, 2009 9:47 PM
  To: [email protected]
  Subject: Re: Merge taking forever

  Bartosz Gadzimski wrote:

   As Arkadi said, your hdd is to slow for 2 x quad core processor. I
 have
  the same problem and now thinking of using more boxes or very fast
  drives (sas 15k).

  Raymond Balmãs pisze:


   Well I suspect the sort function is mono-threaded as usually they
 are

   so


   only one core is used 25% is the max you will get.

  I have a dual core and it only goes to 50% CPU in many of the steps
 ...

   I


   assumed that some phases are mono-threaded.



   Folks,

  From your conversation I suspect that you are running Hadoop with
  LocalJobtracker, i.e. in a single JVM - correct?

  While this works ok for small datasets, you don't really benefit from
  map-reduce parallelism (and you still pay the penalty for the
  overheads). As your dataset grows, you will quickly reach the
  scalability limits - in this case, the limit of IO throughput of a
  single drive, during the sort phase of a large dataset. The excessive
 IO
  demands can be solved by distributing the load (over many drives, and
  over many machines), which is what HDFS is designed to do well.

  Hadoop tasks are usually single-threaded, and additionally
  LocalJobTracker implements only a primitive non-parallel model of task
 >>>>>  execution - i.e. each task is scheduled to run sequentially in turn.
 >>>>> If
  you run the regular distributed JobTracker, Hadoop splits the load
 among
  many tasks running in parallel.

  So, the solution is this: set up a distributed Hadoop cluster, even if
  it's going to consist of a single node - because then the data will be
  split and processed in parallel by several JVM instances. This will
 also
  help the operating system to schedule these processes over multiple
  CPU-s. Additionally, if you still experience IO contention, consider
  moving to HDFS as the filestystem, and spread it over more than 1

  >>> machine and more than 1 disk in each machine.
  >>

  Thank you for these recommendations.

  I think that there is a large group of users (perhaps limited by budget
 or
  time they are willing to spend) that will give up on trying to use
 Nutch
  unless they can run it on a single box with simple configuration.

  Regards,

   >> Arkadi



 --
 Ken Krugler
 +1 530-210-6378


--
Ken Krugler
+1 530-210-6378

Reply via email to