browser extension inspired by httrack that crawls websites and saves pages as self-contained html files for offline viewing.
- saves pages with inlined assets (images, css, fonts as base64) or separate files
- by default rewrites links between pages for offline navigation (like httrack)
- static mode strips javascript for clean archives of js-heavy sites (sveltekit, next.js, etc.)
- url filtering with contains/path-starts/regex
- screenshots via native scroll-stitch or html2canvas
- two storage backends: browser downloads or filesystem (via dashboard)
- indexeddb queue survives browser crashes
- cross-browser compatible
- clone repo
- open
chrome://extensions(orabout:addonsin firefox) - enable developer mode
- load unpacked from
pagespider/folder
- click extension icon
- enter url or click "use current tab"
- set depth (how many links deep to crawl)
- optionally set url filter to limit scope
- choose storage method (filesystem requires opening dashboard first)
- start crawl
downloads. saves to browser downloads folder. simple but no real folder structure.
filesystem. saves to folder you choose with proper structure. requires keeping the dashboard tab open during crawl.
| option | description |
|---|---|
| url filter | only crawl urls matching pattern |
| depth | how many links deep (0 = single page, -1 = unlimited) |
| delay | ms between requests |
| assets | inline (base64) or separate files |
| screenshots | native (scroll-stitch) and/or html2canvas |
| static mode | strip scripts for clean static pages |
| rewrite links | convert links to local files |
| same origin | stay on same domain |
- buttons using js navigation (e.g.
router.push()) won't work in static mode - only<a href>links get rewritten - html2canvas may miss some css effects
- native screenshots may duplicate fixed/sticky elements

