diff --git a/README.md b/README.md index c5a3d5f7..4c58a194 100644 --- a/README.md +++ b/README.md @@ -33,19 +33,19 @@ curl -sSL 'https://get.archivebox.io' | sh # (or see pip/brew/Docker instruct **ArchiveBox is a powerful, self-hosted internet archiving solution to collect, save, and view websites offline.** -Without active preservation effort, everything on the internet eventually dissapears or gets taken down. Archive.org does a great job, but as a large public central archive they can't save anything that requires a login. +Without active preservation effort, everything on the internet eventually dissapears or degrades. Archive.org does a great job as a free public central archive, but they can't save anything that requires a login. *ArchiveBox helps you save web content offline for a variety of situations: download old family photos off Flickr, preserve evidence for legal cases, backup an old Soundcloud mix, snapshot papers for academic research, and more...* > ➡️ *Use ArchiveBox as a [command-line package](#quickstart) and/or [self-hosted web app](#quickstart) on Linux, macOS, or in [Docker](#quickstart).* -

+
-**You can feed ArchiveBox URLs one at a time, or schedule regular imports** from browser bookmarks or history, feeds like RSS, bookmark services like Pocket/Pinboard, and more. See input formats for a full list. +📥 **You can feed ArchiveBox URLs one at a time, or schedule regular imports** from browser bookmarks or history, feeds like RSS, bookmark services like Pocket/Pinboard, and more. See input formats for a full list. snapshot detail page -**It saves offline-viewable snapshots of the URLs you feed it in a few redundant formats.** +💾 **It saves offline-viewable snapshots of the URLs you feed it in a few redundant formats.** It also auto-detects the content featured *inside* each webpage extracts it out to common, easy file formats: - `HTML/Generic Websites -> HTML/PDF/PNG/WARC` - `YouTube/SoundCloud/etc. -> mp3/mp4`,