Best way to back up entire website on a schedule

This page summarizes the projects mentioned and recommended in the original post on /r/DataHoarder

Our great sponsors
  • WorkOS - The modern identity platform for B2B SaaS
  • InfluxDB - Power Real-Time Data Analytics at Scale
  • SaaSHub - Software Alternatives and Reviews
  • fetchurls

    A bash script to spider a site, follow links, and fetch urls (with built-in filtering) into a generated text file.

  • You could also look into something like archivebox.io, but it doesn't really mirror so great. fetchurls can make an URL list though which could in turn be fed into archivebox. Archivebox would maybe be handy if you wanted the wget download along with a PDF print + maybe sending to Wayback Machine.

  • ArchiveBox

    🗃 Open source self-hosted web archiving. Takes URLs/browser history/bookmarks/Pocket/Pinboard/etc., saves HTML, JS, PDFs, media, and more...

  • You could also look into something like archivebox.io, but it doesn't really mirror so great. fetchurls can make an URL list though which could in turn be fed into archivebox. Archivebox would maybe be handy if you wanted the wget download along with a PDF print + maybe sending to Wayback Machine.

  • WorkOS

    The modern identity platform for B2B SaaS. The APIs are flexible and easy-to-use, supporting authentication, user identity, and complex enterprise features like SSO and SCIM provisioning.

    WorkOS logo
NOTE: The number of mentions on this list indicates mentions on common posts plus user suggested alternatives. Hence, a higher number means a more popular project.

Suggest a related project

Related posts