From: <La...@us...> - 2004-10-21 19:24:32
|
Hello list, I would like to use wget to make a static snapshot of my PhpWiki, but wget does not follow the links. Do I have to configure anything in PhpWiki or do I do anything wrong ? Thanxs. phpwiki: 1.3.4 wget -r -l 0 http://my/wiki/ |
From: Reini U. <ru...@x-...> - 2004-10-21 22:05:26
|
La...@us... schrieb: > I would like to use wget to make a static snapshot of my PhpWiki, > but wget does not follow the links. > Do I have to configure anything in PhpWiki or do I do anything wrong ? > Thanxs. auth? > phpwiki: 1.3.4 > wget -r -l 0 http://my/wiki/ or PhpWikiAdministration => Dump Pages. For dumphtml you need a newer version than 1.3.4. This will create a real static snapshot. -- Reini Urban http://xarch.tu-graz.ac.at/home/rurban/ |
From: LaFambe <La...@us...> - 2004-10-22 12:40:19
|
Reini Urban wrote: > > auth? > What do you mean, auth ? Cookies not handled by wget ? On the install I have, there are no auth restrictions (AFAIK) except for editing, but wget is not going to go there, is it ? >> phpwiki: 1.3.4 >> wget -r -l 0 http://my/wiki/ Still will not dump more than one page. Is this path a cul-de-sac ;) ? > > or PhpWikiAdministration => Dump Pages. > > For dumphtml you need a newer version than 1.3.4. This will create a > real static snapshot. So I should upgrade you say ? A+O. |
From: Reini U. <ru...@x-...> - 2004-10-22 08:22:43
|
LaFambe schrieb: > Reini Urban wrote: > > > > auth? > > > > What do you mean, auth ? Cookies not handled by wget ? > On the install I have, there are no auth restrictions (AFAIK) except for > editing, but wget is not going to go there, is it ? > > >> phpwiki: 1.3.4 > >> wget -r -l 0 http://my/wiki/ wget works fine if you will call it correctly. even timestamping is supported. wget -nd -r -l1 -nH http://my/wiki/ > Still will not dump more than one page. > Is this path a cul-de-sac ;) ? > > or PhpWikiAdministration => Dump Pages. > > > > For dumphtml you need a newer version than 1.3.4. This will create a > > real static snapshot. > > So I should upgrade you say ? If you can live with the memory limitations after the upgrade. until 1.3.4 it did no output buffering. After it needs more than 8MB. http://phpwiki.sourceforge.net/phpwiki/PhpMemoryExhausted/Testresults -- Reini Urban http://xarch.tu-graz.ac.at/home/rurban/ |
From: <La...@us...> - 2004-10-22 12:56:51
|
Reini Urban [Fri, 22 Oct 2004 10:22:35 +0200] wrote: > > wget -nd -r -l1 -nH http://my/wiki/ > This will not work either for me. (sorry to insist) (anyway the additional flags -nH -nd are essentially cosmetic right ?) May be I did something wrong in the install (PhpWiki *or* Apache) that prevents wget from crawling through the wiki. Or wget is too old ? wget --version GNU Wget 1.8.2 http://phpwiki.sourceforge.net/phpwiki/BackupStrategies does say things about backuping with wget, but uses the zip-dump interface. http://amphi-gouri.org/blog/2004/09/16/73-LeConvertisseurWikiDuPauvreConvertirUnSiteSimpleEnSyntaxeMoinmoinEnQuelquesLignes uses --no-parent => same result, only one page dumped. I read somthing about protection from bots: http://phpwiki.sourceforge.net/phpwiki/HowToHandleRobots "Only action=browse and action=index is allowed for statically identified robots, but authorized action must be allowed, e.g. for my daily backups with Wget." Is that an issue ? allowing the read action ? I have no clue what/where this is ... > > If you can live with the memory limitations after the upgrade. > until 1.3.4 it did no output buffering. After it needs more than 8MB. > > http://phpwiki.sourceforge.net/phpwiki/PhpMemoryExhausted/Testresults Yep, I saw that in you post, thanxs. A+O. |
From: Reini U. <ru...@x-...> - 2004-10-22 13:20:19
|
La...@us... schrieb: > Reini Urban [Fri, 22 Oct 2004 10:22:35 +0200] wrote: > >>wget -nd -r -l1 -nH http://my/wiki/ > > This will not work either for me. (sorry to insist) > > (anyway the additional flags -nH -nd are essentially cosmetic right ?) nd and nH dont create the host and dir subdirs. > May be I did something wrong in the install (PhpWiki *or* Apache) that > prevents wget from crawling through the wiki. Or wget is too old ? > > wget --version > GNU Wget 1.8.2 wget cannot be too old, it is dumb on purpose. if can click through your wiki, wget can "click" through it also. $ wget --version GNU Wget 1.9.1 There exist faster wget versions (using a hash instead of a list internally). > http://phpwiki.sourceforge.net/phpwiki/BackupStrategies does say > things about backuping with wget, but uses the zip-dump interface. > > http://amphi-gouri.org/blog/2004/09/16/73-LeConvertisseurWikiDuPauvreConvertirUnSiteSimpleEnSyntaxeMoinmoinEnQuelquesLignes > uses --no-parent => same result, only one page dumped. sure. one zip, which is your whole wiki. all pages zipped. > I read somthing about protection from bots: > http://phpwiki.sourceforge.net/phpwiki/HowToHandleRobots > > "Only action=browse and action=index is allowed for statically > identified robots, but authorized action must be allowed, e.g. for my > daily backups with Wget." this is for a very old wiki version of mine. there's no action=index anymore. > Is that an issue ? allowing the read action ? I have no clue > what/where this is ... no, currently we don't block wget robots. but maybe your global /robots.txt disallows wget? >>If you can live with the memory limitations after the upgrade. >>until 1.3.4 it did no output buffering. After it needs more than 8MB. >> >>http://phpwiki.sourceforge.net/phpwiki/PhpMemoryExhausted/Testresults -- Reini Urban http://xarch.tu-graz.ac.at/home/rurban/ |
From: <la...@us...> - 2004-10-22 15:09:20
|
La...@us... [Fri, 22 Oct 2004 14:56:41 +0200] wrote: > > Reini Urban [Fri, 22 Oct 2004 10:22:35 +0200] wrote: >> >> wget -nd -r -l1 -nH http://my/wiki/ >> > > This will not work either for me. (sorry to insist) > > (anyway the additional flags -nH -nd are essentially cosmetic right ?) > > May be I did something wrong in the install (PhpWiki *or* Apache) that > prevents wget from crawling through the wiki. Or wget is too old ? > Ok, it's a dns or proxy problem. Using a fqdn on the url works as expected. Sorry for the fuss. A+O. |