Re: Mulitple-site question

2001-10-03 Thread CJ Kucera

I said:
> I'd like to use wget to take a snapshot of www.theonion.com.  On that
[snip!]

I forgot to mention I'm using wget 1.7.  Sorry 'bout that.

-CJ

WOW: Rapacious   | A priest advised Voltaire on his death bed to
apocalyptech.com/wow |  renounce the devil.  Replied Voltaire, "This
[EMAIL PROTECTED] |  is no time to make new enemies."



Mulitple-site question

2001-10-03 Thread CJ Kucera

Greetings, humans!

I'd like to use wget to take a snapshot of www.theonion.com.  On that
site, all of the graphics are served from "graphics.theonion.com",
and there's a bunch of other sub-domains as well.  Also, it links over
to "www.theonionavclub.com", which I would also like to "mirror."
How can I do this in a way that allows me to basically get any domain
"*.theonion.com" and "*.theonionavclub.com", but not get anything
else?

The closest I've come is (and there's lots of extraneous stuff in there):

> wget -r -l inf -k -p --wait=1 -H 
>--domains=theonion.com,graphics.theonion.com,www.theonion.com,theonionavclub.com,www.theonionavclub.com
> http://www.theonion.com

Could someone help me out with this?

Thanks much,
CJ

WOW: Rapacious   | A priest advised Voltaire on his death bed to
apocalyptech.com/wow |  renounce the devil.  Replied Voltaire, "This
[EMAIL PROTECTED] |  is no time to make new enemies."



Re: Bug in wget 1.7

2001-10-03 Thread Daniel Stenberg

On Wed, 3 Oct 2001, Thomas Preymesser wrote:

> The file index.html is saved an complete in directory www.lehele.de. If I
> call wget without recursion then everything is ok, but when i try to go
> deeper wget is crashing.

It would probably help a lot if you could do

'gdb /path/to/wget core'

and then type

'where' to display the stack trace.

If you've built without the -g (debug) option, please rebuild it and try
again, as that will give lots of more details in the stack trace dump.

-- 
  Daniel Stenberg - http://daniel.haxx.se - +46-705-44 31 77
   ech`echo xiun|tr nu oc|sed 'sx\([sx]\)\([xoi]\)xo un\2\1 is xg'`ol




Bug in wget 1.7

2001-10-03 Thread Thomas Preymesser

Hello.

I have discovered a bug in wget 1.7

When I try to get thist page: http://www.lehele.de/

this error occurs:
-
wget -d -r -l 1 www.lehele.de 
DEBUG output created by Wget 1.7 on linux-gnu.

parseurl ("www.lehele.de") -> host www.lehele.de -> opath  -> dir  -> 
file  -> ndir
newpath: /
Checking for www.lehele.de in host_name_address_map.
Checking for www.lehele.de in host_slave_master_map.
First time I hear about www.lehele.de by that name; looking it up.
Caching www.lehele.de <-> 212.227.118.88
Checking again for www.lehele.de in host_slave_master_map.
--19:59:51--  http://www.lehele.de/
   => `www.lehele.de/index.html'
Verbindungsaufbau zu www.lehele.de:80... Found www.lehele.de in 
host_name_address_map: 212.227.118.88
Created fd 3.
verbunden!
---request begin---
GET / HTTP/1.0
User-Agent: Wget/1.7
Host: www.lehele.de
Accept: */*
Connection: Keep-Alive

---request end---
HTTP Anforderung gesendet, auf Antwort wird gewartet... HTTP/1.1 200 OK
Date: Wed, 03 Oct 2001 18:03:52 GMT
Server: Apache/1.3.14 (Unix)
Connection: close
Content-Type: text/html


Länge: nicht spezifiziert [text/html]

0K ...@ 332.21 B/s

Closing fd 3
20:00:15 (332.21 B/s) - »www.lehele.de/index.html« gespeichert [3830]

parseurl ("www.lehele.de") -> host www.lehele.de -> opath  -> dir  -> 
file  -> ndir
newpath: /
Loaded www.lehele.de/index.html (size 3830).
Speicherzugriffsfehler (core dumped)


The file index.html is saved an complete in directory www.lehele.de.
If I call wget without recursion then everything is ok, but when i try 
to go deeper wget is crashing.

-Thomas






RE: Segfault on Linux/390 for wget 1.6 and 1.7

2001-10-03 Thread Post, Mark K

Jan,

Did you ever make any progress on this?

Mark Post

-Original Message-
From: Jan Prikryl [mailto:[EMAIL PROTECTED]]
Sent: Thursday, July 19, 2001 1:53 PM
To: Post, Mark K
Cc: Wget mailing list
Subject: Re: Segfault on Linux/390 for wget 1.6 and 1.7


Quoting Post, Mark K ([EMAIL PROTECTED]):

> When I compile wget with -O0 to turn off optimization, wget works, but I
get
> some garbage in the output as follows:

Could you please try 

(1) to run wget with the -d parameter to switch on the debugging
output 

(2) compile wget using -O2 -g and have a look what

  gdb wget core

reports? It shall be able to provide us with the content of the
call stack in the moment of crash that in turn would reveal the
place where wget crashes.

Thanks,

-- jan

---+
  Dr. Jan Prikryl  icq | vr|vis center for virtual reality and
  <[EMAIL PROTECTED]>  83242638 | visualisation http://www.vrvis.at
---+



Re: Wget on DG/UX

2001-10-03 Thread Ian Abbott

On 3 Oct 2001, at 10:00, Sebastien Mougey wrote:

> I've compiled WGET on Data General DG/UX (R4.11), and it works fine, but
> DG/UX doesn't know the define MAP_FAILED, used in utils.c, defined in
> sys/mman.h. So it should be replaced by (-1) on this unix...

Yes, the current CVS version (a.k.a. wget-1.7.1-pre1) already copes 
with systems that do not define MAP_FAILED (-1).



Wget on DG/UX

2001-10-03 Thread Sebastien Mougey

Hello,

I've compiled WGET on Data General DG/UX (R4.11), and it works fine, but
DG/UX doesn't know the define MAP_FAILED, used in utils.c, defined in
sys/mman.h. So it should be replaced by (-1) on this unix...

Sébastien

+---+
Sebastien Mougey  -  Pixtel
eMail : [EMAIL PROTECTED]
+---+