101010.pl is one of the many independent Mastodon servers you can use to participate in the fediverse.
101010.pl czyli najstarszy polski serwer Mastodon. Posiadamy wpisy do 2048 znaków.

Server stats:

542
active users

#datahoarding

0 posts0 participants0 posts today
Continued thread

also hijacking my own post here

"downloading the entirety of Wikipedia" feels right to me given the current political climate

it's easy to download copies of Wikipedia and similar sites for Kiwix, and older copies of them are available on the Internet Archive

there are also programs you can use to download a single webpage (SingleFile) and mirror static websites (wget and HTTrack)

(inspired by devine's post: merveilles.town/@neauoire/1142)

MerveillesDevine Lu Linvega (@neauoire@merveilles.town)Found something valuable on a .edu domain? Make a copy as you can be sure it'll be the first thing to bitrot into oblivion.

I have a mini Intel Atom-powered home server in my house.

However, I’ve overlooked two things:

How do I back up data and keep it safe (from both security and quality perspectives)?

I’m still a newbie at GnuPG Privacy Guard. How do I secure the backup of my private keys?

#gpg#gnupg#privacy
Replied in thread

mastodon.ml/@mintbug/114086408
Написал скрипт, который принимает https-ссылку на пост, получает ссылки на вложения с помощью #toot, преобразует https-ссылку в fedi-ссылку и сохраняет вложения в директорию с именами вида `<fedi-ссылка> <номер вложения>.ext` (см. иллюстрацию). Слеши в fedi-ссылке заменены на знак `∫`, потому что ~~а почему бы и нет~~ я ещё не настолько сошёл с ума, чтобы использовать слеши в именах файлов, поэтому их нужно на что-то заменить.

Запуск `fedi-save (wl-paste)` повесил на хоткей в #sway; открытие ссылки, содержащейся в имени файла (`fedi-save --open <path>`) повесил на хоткей в #yazi.

Just discovered ArchiveBox — FOSS, self-hosted internet archiving.

The way the web is going, with the US government redacting and outright erasing historic content, publishers segmenting content by region (and also sometimes redacting/censoring it), and CloudFlare shitting all over everything, I think it's time for me to start my #archiving and #DataHoarding journey.

#SelfHosting #SelfHosted #DataHoarder

github.com/ArchiveBox/ArchiveB

GitHubGitHub - ArchiveBox/ArchiveBox: 🗃 Open source self-hosted web archiving. Takes URLs/browser history/bookmarks/Pocket/Pinboard/etc., saves HTML, JS, PDFs, media, and more...🗃 Open source self-hosted web archiving. Takes URLs/browser history/bookmarks/Pocket/Pinboard/etc., saves HTML, JS, PDFs, media, and more... - ArchiveBox/ArchiveBox

- I've setup a VNET thick jail on my FreeBSD NAS.
- The jail has its own IP address on my LAN.
- I declared a devfs ruleset to unhide /dev/tun* for the VNET jail.
- I installed Wireguard in the jail.
- I enabled Wireguard with a ProtonVPN configuration.
- I installed qbittorrent-nox and configured it to use the Wireguard interface.

I now have a home ISP-proof qBittorrent setup with which to torrent Anna's Archive.

Data hoarding question from an ally: What things on the internet are in most need of saving?

I have backup tapes and hard drives and the capability of providing more. I want my resources to be available for the greater good.

I’m aware of people backing up YouTube channels that are at risk as well as archive.org and government data.

What else is at risk?

I’m also open to do tape backups if someone want to collab. I have an LTO-4 drive.

How do #windows users handle incremental backups?

A friend is asking for help. He doesn't want to use a cloud, just an external HDD. What he did till now is just copy paste everything once every other month and hope that nothing is missing.

Any non-proprietary windows software that I can recommend to him that can handle incremental backups? Encryption would also be nice imo 😄

Is there some kind of file #storage that does delta compression? I'd love to have a file #archive with

• revisions (i.e. changes to files are tracked)
• deduplication (identical files only take up space once)
• deltas (small changes to large files only take up the size of the difference)

Git would cover this, but doesn't work well with TBs of data. Backup software with delta support like Borg or restic can do it too, but have bad UX for tracking file revisions.