how to download a complete website

How to do things, solutions, recipes, tutorials
Message
Author
User avatar
alienjeff
Posts: 2265
Joined: Sat 08 Jul 2006, 20:19
Location: Winsted, CT - USA

#21 Post by alienjeff »

Image
[size=84][i]hangout:[/i] ##b0rked on irc.freenode.net
[i]diversion:[/i] [url]http://alienjeff.net[/url] - visit The Fringe
[i]quote:[/i] "The foundation of authority is based upon the consent of the people." - Thomas Hooker[/size]

sky king
Posts: 72
Joined: Wed 08 Apr 2009, 08:42

pmirrorget

#22 Post by sky king »

I just downloaded my wife's whole website, very cool. She has a front page that you have to click, so the very first page is all I got on the first try. But the next page URL did the trick. Amazingly easy, haven't checked all the pages of course, it's got more than a thousand. But I'm just thrilled that it is so convenient to do in the latest Puppy.

I'll have to convert her to Puppy use now, she just ended up with malware programs, been agonizing over that.

seaside
Posts: 934
Joined: Thu 12 Apr 2007, 00:19

#23 Post by seaside »

My friend found out that her blog was about to be taken down and the backup on her harddisk had been lost in a crash. Panic time :D

Well, I thought, Puppy Pmirrorget to the rescue!

I downloaded 90M of blog very nicely into a directory with subdirectories and was feeling very heroic for a bit... until..

Now, the on-line blog is gone, and a very nicely structured directory is in place on a new harddisk with all the internal links pointing back to the former on-line site.

Is there any tool or easy method to change these links, so that it can be used as an off-line viewer.

Thanks for any ideas,
s
( I can't imagine doing search and replaces for all these links :) )

seaside
Posts: 934
Joined: Thu 12 Apr 2007, 00:19

#24 Post by seaside »

....Ahhh....

Discovered what the problem is in downloading a site with PMirrorget.

I looked at the code in "/usr/bin/pmwget" and found it did not use the "--convert-links" option which means the links will be left pointing back to the on-line website.

This can be fixed by adding "-k" switch to this line-

Code: Select all

 rxvt -name PMirrorget -bg "#F3F2DF" -e wget -m -c -r -np -P "$DEST" $SOURCE
so it looks like this-

Code: Select all

 rxvt -name PMirrorget -bg "#F3F2DF" -e wget -m -c -k -r -np -P "$DEST" $SOURCE
Cheers,
s
(Come to think of it, it should probably be changed to this in the official Puppy, because I can't offhand see why you'd normally want the links to be absolute on downloaded material to be used off-line) :D

User avatar
gposil
Posts: 1300
Joined: Mon 06 Apr 2009, 10:00
Location: Stanthorpe (The Granite Belt), QLD, Australia
Contact:

#25 Post by gposil »

[img]http://gposil.netne.net/images/tlp80.gif[/img] [url=http://www.dpup.org][b]Dpup Home[/b][/url]

seaside
Posts: 934
Joined: Thu 12 Apr 2007, 00:19

#26 Post by seaside »

gposil,

Wow, thanks, that was fast.

Cheers,
s

User avatar
GustavoYz
Posts: 883
Joined: Wed 07 Jul 2010, 05:11
Location: .ar

#27 Post by GustavoYz »

If you're running Firefox, check out the Scrap-book add-on...good stuff! 8) .
I made it run on SeaMonkey but i don't know why crashes on the "export" part. :roll:

//Edited : If you add the xSideBar add-on, the 1.37 version (Firefox add-on) will run on SeaMonkey 2xx.

Post Reply