_ _ _
(_) | | | |
__ ___ ___ | | ___| |_
\ \ / / |/ _ \| |/ _ \ __|
\ V /| | (_) | | __/ |_
\_/ |_|\___/|_|\___|\__|
we archive the web
instead of relying on the internet archive or archive.is, we manually scrape web pages off of neocities and nekoweb and archive them.
our mission is to archive the indie web as best as we can and help recover websites when they're lost or deleted or whatever
see our archivemake a new file named "NOCONSENT.txt" in the root and put the following content: "noconsent". every website we go through, we always check for that single file. if we see that "NOCONSENT.txt" doesnt exist in the site, we will scrape your site and archive it here. if you have robots.txt and it disallows everything (or disallows whatever bot you didnt specify) then we will respect it as well.
our volunteers scrape your site using curl in a command prompt. we then will scrape your site every single month. learn more about curl here.
yes of course! you can submit a site link to us here!
awesomee! go through this form here!!!
thats alright. you can email me anytime!
$ts=Get-Date -Format "yyyy-MM-dd"; curl.exe https://example.com -o "C:\Users\rearw\Downloads\$ts.html