I’m sorry if this question has already been asked, I couldn’t find any similar topic.
Is it possible to set up a complex backup sequence, with dependencies?
I mean, a set of backed-up directories, each with their own set of filters, that need to be performed in a specific order. E.g.:
C . . D
.\ . . /
. .\ ./
. . E
dir A is backed-up first and only then dir B is backed-up, then dir C.
In parallel to that sequence, dir D (located on another drive) is archived, and finally, only after all of them, dir E is processed.
I know I can set up multiple back-ups, but I have no control in which order they’re executed.
OR I can set up a single backup with all those directories included and all the filters in a single huge bunch, but, again, I can’t control in which order they’re executed or make some folders to be processed in parallel.
The best I can do is assume how long would it take for each backup to finish, and set up their start time accordingly. But if there are a bunch of folders, each of which is variable in size (even if they have about the same size in total), I need to plan it for the worst-case scenario for each folder. Say, each folder takes somewhere from 10 minutes to 4h to back up. If I have 6 of those, the backups are performed for the whole day and PC is always “could be busy”, even if the actual total time is somewhere about 6h.
I guess, it’s a feature request.
Or, maybe, this feature is already implemented? If so, could you point me to the doc page on how to accomplish that?