Hi, I’m hoping someone can help me figure out what settings I should be using for this, because I suspect that I am doing it wrong.
What I want to back up are several large files (full disk image incremental backups in the 10-20GB range) and one even larger file (~100GB). I do not want versioning (the files are already versioned), just a single backup of this data. The target is OneDrive business.
The nature of the data means that daily, one new file will be uploaded and one old one deleted, and some subset of the larger file changed. I believe Duplicati can back this up without needing to re-upload the entire 100GB file, sending only the changed bits - that’s why I want to use it.
I set the remote volume size to 1GB, and left the block size at the default.
What I’m finding is that very soon Duplicati wants to do a Compact, and to do this, for some reason, starts downloading gigabytes worth of data. I don’t want it to download anything (other than the verification file after a backup), unless I’m doing a restore. It shouldn’t need to, it’s got all the data locally, I just want it to send the new data and delete the old.
If I disable compact with --no-auto-compact, is my backup just going to keep growing in size without deleting old data? Should I have chosen different block or volume sizes? What’s the proper way of backing up large files without wasting a lot of remote space and without unnecessarily downloading a lot of data that I just uploaded, only to re-upload it again?