https://www.gnu.org/software/wget/manua ... tions.htmlKappes Buur wrote:However, how can I prevent it from downloading files other than WIKI items ?
For example only files from https://zdoom.org/wiki/ but not from https://zdoom.org/files/
I tried the level option --level= with 0, 1, 2 etc. but that did not work.
Would somebody have some insight with that?
Telling it to reject something like ^((?!wiki).)*$ might work. Gross hack stolen from here.
If you are trying to dump the wiki syntax, it's probably better to download the "source view" pages, e.g. https://zdoom.org/w/index.php?title=Mai ... action=rawKappes Buur wrote:Does anybody know of a HTML to MediaWiki editor that actually works ? For Win 10 and preferably free.
Not sure how to script this properly though, it may as well be impossible using simple wget commands.
You might be able to do it in two passes, first collect a list of links by getting HTML, then in the second pass fetching each page as raw. That, however, is double load on the server.
I'd just write my own Python script here to parse the [[]] links (and in-wiki file/image references as well) and construct the final download list based on those recursively.