There is a webcomic called strong female protagonist that i want to persevere(in case the website is ever lost) but not sure how.
The image you see above is not a webpage of the site but rather a drop-down like menu. There is a web crawler called WFDownloader(that i am using the window’s exe file inside bottles)that can grab images and can follow links, grab images “N” number of pages down but since this a drop-down menu i am not sure it will work
There also the issue of organizing the images. WFDownloader doesn’t have options for organizing.
What i am thinking about, is somehow translating the html for the drop-down menu into separate xml file based on issues/titles, run a script to download the images, have each image named after its own hyperlink and have each issue in its own folder. Later on i can create a stitch-up version of the each issues.
If you view the source of the homepage, you’ll see some HTML that starts with this:
<div class="archive-dropdown-wrap"> <ul class="archive-dropdown"> <li><span class="chapter-label">Issue 1</span><ul><li><a href="https://strongfemaleprotagonist.com/issue-1/page-0/">CoverThat’s the HTML for the drop-down. Although if I were you, I’d look into taking advantage of WordPress’ JSON API, since that website uses WordPress.
For example, here’s a list of the images uploaded to the site in JSON format: https://strongfemaleprotagonist.com/wp-json/wp/v2/media?per_page=100&page=1 (Limited to 100 entries per page)
i’m presuming that you’ve tried something like curl or wget wrapped in a for loop to iterate through each page to do this and that it didn’t work somehow.
robots.txt would probably put a stop to that


