mirror of
https://github.com/ArchiveBox/ArchiveBox.git
synced 2026-04-25 17:16:00 +03:00
[GH-ISSUE #303] Question: RSSMix useless? #3242
Labels
No labels
expected: maybe someday
expected: next release
expected: release after next
expected: unlikely unless contributed
good first ticket
help wanted
pull-request
scope: all users
scope: windows users
size: easy
size: hard
size: medium
size: medium
status: backlog
status: blocked
status: done
status: idea-phase
status: needs followup
status: wip
status: wontfix
touches: API/CLI/Spec
touches: configuration
touches: data/schema/architecture
touches: dependencies/packaging
touches: docs
touches: js
touches: views/replayers/html/css
why: correctness
why: functionality
why: performance
why: security
No milestone
No project
No assignees
1 participant
Notifications
Due date
No due date set.
Dependencies
No dependencies set.
Reference
starred/ArchiveBox#3242
Loading…
Add table
Add a link
Reference in a new issue
No description provided.
Delete branch "%!s()"
Deleting a branch is permanent. Although the deleted branch may continue to exist for a short time before it actually gets removed, it CANNOT be undone in most cases. Continue?
Originally created by @dataarchivist on GitHub (Nov 20, 2019).
Original GitHub issue: https://github.com/ArchiveBox/ArchiveBox/issues/303
I didnt try if AB can do it by itself, but for now, I use http://rssmix.com/ to merge my rss feeds into one to cronjob a lot of websites at once. Its fully free btw.
If its not already possible to catch them easily via txt file or something else, it would be nice to have.
@pirate commented on GitHub (Nov 20, 2019):
Archivebox never re-downloads urls that have been already archived, so you can already pipe in all your RSS feeds without merging / deduplicating them first, so this isn't needed.
@dataarchivist commented on GitHub (Nov 20, 2019):
Is an option to redownload a site again planned?
@pirate commented on GitHub (Nov 20, 2019):
Yes, see https://github.com/pirate/ArchiveBox/issues/179
@dataarchivist commented on GitHub (Nov 21, 2019):
But its does for rss feeds.
So, if you merge the feeds, you normaly get every x hour an fresh rss with new links.
@pirate commented on GitHub (Nov 21, 2019):
It downloads the feed itself, but it doesn't re-download the pages within it after the first time.
@dataarchivist commented on GitHub (Nov 22, 2019):
A normal feed didnt have the same link a second time. ;)
Anyway, I did this with another cron, it was just interesting to know.