No description
Find a file
2022-05-04 23:21:06 +01:00
.github remove github actions 2020-11-07 10:32:33 +00:00
.vscode add vscode tasks 2018-12-22 16:08:49 +00:00
benches make sure bench does run on CI correctly 2021-05-16 21:23:45 +01:00
scripts make sure bench does run on CI correctly 2021-05-16 21:23:45 +01:00
src update, start adding cli 2022-03-07 23:05:20 +00:00
tests fix clippy 2021-10-31 10:45:18 +00:00
.build.yml wip on error correction 2021-05-15 20:10:26 +01:00
.envrc migrate to flakes & new rust 2021-10-22 21:35:10 +01:00
.gitignore migrate to flakes & new rust 2021-10-22 21:35:10 +01:00
Cargo.lock update, start adding cli 2022-03-07 23:05:20 +00:00
Cargo.toml update, start adding cli 2022-03-07 23:05:20 +00:00
flake.lock update, start adding cli 2022-03-07 23:05:20 +00:00
flake.nix reformat flake 2022-05-04 23:21:06 +01:00
LICENSE Initial commit 2018-08-14 19:59:05 +01:00
README.md add todo 2021-11-25 11:11:21 +00:00
rust-toolchain add toolchain 2018-12-20 16:47:22 +00:00
rustfmt.toml rustfmt config 2018-10-04 13:42:39 +01:00
shell.nix update, start adding cli 2022-03-07 23:05:20 +00:00

Bakare: modern and simple, yet efficient backup solution

This is pre-alpha, contributions welcome

It does not even have a standalone binary yet, but the engine looks promising. Contributions are welcome :)

Goals for bakare:

  • simple deduplication of file data - don't store the same exact file data twice
  • advanced deduplication - store files that only changed slightly more efficiently
  • survive index corruption
  • resilient towards interrupting the backup (create index checkpoints etc)
  • fast
  • using max bandwidth
  • use max cpu
  • use max disk I/O
  • memory usage limit
  • encryption by default - asymmetric, creates a keypair for you
  • fuzzy find by file name in stored files
  • failure to process one file should not affect any other files
  • intermittent network failures should not make the whole process fail (test with random packet drop)
  • system suspend/resume should not make the repo become corrupted, even in the presence of other active backup processes running on other computers, targeting same repo - this is where restic fails

Nice to haves:

  • daemon that listens for file events and updates a list of files to be backed up on the next backup run - or a continous backup mode - the daemon uploads the file whenever it sees the change
  • peer2peer mode - people storing encrypted backups for each other
  • relay mode, where daemon works on one or more central points with local storage (e.g. NAS) and various computers sync with that central location. Then though the central locaiton uploads everything to the other location, typically the cloud.

Implementation notes

  • automatic node discovery - two roles: data publisher and data persister - should be able to figure out which node is which automatically
  • test with randomly created dirs and files, with property based tests and fuzzer
  • see if we can use salsa for recomputation
  • index corruption tests - mutate random byte and see if everything is readable
  • network packet drop tests
  • use bevy for decoupling ?
  • remove all unwraps

Motivation: All the backup systems I tried are either slow or crashing or both for me

duply: works but is very slow

--------------[ Backup Statistics ]--------------
StartTime 1547198362.85 (Fri Jan 11 09:19:22 2019)
EndTime 1547209509.04 (Fri Jan 11 12:25:09 2019)
ElapsedTime 11146.19 (3 hours 5 minutes 46.19 seconds)
SourceFiles 3065438
SourceFileSize 585041709586 (545 GB)
NewFiles 0
NewFileSize 0 (0 bytes)
DeletedFiles 0
ChangedFiles 0
ChangedFileSize 0 (0 bytes)
ChangedDeltaSize 0 (0 bytes)
DeltaEntries 0
RawDeltaSize 0 (0 bytes)
TotalDestinationSizeChange 111 (111 bytes)
Errors 0
-------------------------------------------------

--- Finished state OK at 12:25:15.000 - Runtime 03:06:43.000 ---

restic:

  • crashes with OOM sometimes
  • corrupts repo if you suspend one backup process and launch another from a different computer

TODOs

  • compare with Kopia