JavaScript code that creates dynamic URLs is always a problem for web crawlers.

Most web sites try to make their content crawlable by creating alternative static links to the content.

I think Google now does some analysis/execution of JS code, but it's a tricky problem.

I would suggest modifying the HTML parser to explicitly look for calls being made to your function, and generate appropriate outlinks.

-- Ken


On Sep 14, 2009, at 8:04am, Mohamed Parvez wrote:

Can anyone please through some light on this

Thanks/Regards,
Parvez


On Fri, Sep 11, 2009 at 3:23 PM, Mohamed Parvez <[email protected]> wrote:

We have a JavaScript function, which takes some prams and builds an URL and
then uses  window.location to send the user to that URL.

Our website uses this feature a lot and most of the urls are built using
this function.

I am trying to crawl using Nutch and I am also using the parse-js plugin.

But it does not look like Nautch is able to crawl these URLs.

Am I doing something wrong or Nutch is not able to crawl URLs build by
JavaScript function.

----
Thanks/Regards,
Parvez



--------------------------
Ken Krugler
TransPac Software, Inc.
<http://www.transpac.com>
+1 530-210-6378

Reply via email to