Re: getting the correct links
Christopher Stone wrote: Thank you all. Now the issue seems to be that it only gets the root directory. I ran 'wget -km -nd http://www.mywebsite.com -r Max.
Re: getting the correct links
Jens Rösner wrote: Hi! Max' hint is incorrect I think, as -m includes -N (timestamps) and -r (recursive) Ooops, you're right. I tend not to use -m much myself. I should pay more attention! Max.
getting the correct links
Hi. I am new to wget, and although it doesn't seem to difficult, I am unable to get the desired results that I am looking for. I currently have a web site hosted by a web hosting site. I would like to take this web site as is and bring it to my local web server. Obviously, the ip address and all the links point back to this web server. When I ran wget and sucked the site to my local box, it pulled all the pages down and the index page comes up fine, but when I click on a link, it goes back to the remote server. What switch(s) do I use, so that when I pull the pages to my box, that all of the links are changed also? Thank you. Chris please cc to me, as i am not a list subscriber. __ Do You Yahoo!? Yahoo! Finance - Get real-time stock quotes http://finance.yahoo.com
Re: getting the correct links
Christopher Stone wrote: When I ran wget and sucked the site to my local box, it pulled all the pages down and the index page comes up fine, but when I click on a link, it goes back to the remote server. What switch(s) do I use, so that when I pull the pages to my box, that all of the links are changed also? -k Also look at -K and -E. Max.
Re: getting the correct links
Hi Chris! Using the -k switch (convert local files to relative links) should do what you want. CU Jens Christopher Stone wrote: Hi. I am new to wget, and although it doesn't seem to difficult, I am unable to get the desired results that I am looking for. I currently have a web site hosted by a web hosting site. I would like to take this web site as is and bring it to my local web server. Obviously, the ip address and all the links point back to this web server. When I ran wget and sucked the site to my local box, it pulled all the pages down and the index page comes up fine, but when I click on a link, it goes back to the remote server. What switch(s) do I use, so that when I pull the pages to my box, that all of the links are changed also? Thank you. Chris please cc to me, as i am not a list subscriber. __ Do You Yahoo!? Yahoo! Finance - Get real-time stock quotes http://finance.yahoo.com
Re: getting the correct links
Thank you all. Now the issue seems to be that it only gets the root directory. I ran 'wget -km -nd http://www.mywebsite.com and it doesn't get anything from the subdirectories. Now what am i doing wrong? Thank you Chris --- Jens Rösner [EMAIL PROTECTED] wrote: Hi Chris! Using the -k switch (convert local files to relative links) should do what you want. CU Jens Christopher Stone wrote: Hi. I am new to wget, and although it doesn't seem to difficult, I am unable to get the desired results that I am looking for. I currently have a web site hosted by a web hosting site. I would like to take this web site as is and bring it to my local web server. Obviously, the ip address and all the links point back to this web server. When I ran wget and sucked the site to my local box, it pulled all the pages down and the index page comes up fine, but when I click on a link, it goes back to the remote server. What switch(s) do I use, so that when I pull the pages to my box, that all of the links are changed also? Thank you. Chris please cc to me, as i am not a list subscriber. __ Do You Yahoo!? Yahoo! Finance - Get real-time stock quotes http://finance.yahoo.com __ Do You Yahoo!? Yahoo! Finance - Get real-time stock quotes http://finance.yahoo.com