The SSD hard drive in my laptop crashed - I put a new SSD and reinstalled operating system and duplicati.
I download the backup files from the cloud and copied them directly to the C drive (about 200GB)
I pointed duplicati gui to restore from local folder (where the files have been downloaded from cloud), gave it the encryption key and began the restoration process.
for about 5 days duplicati has been slowing progressing “Running task: Recreating database” a nd finally one day it must have finished that part and it was back to normal screen WITHOUT having restored the files!!
Pls help - i have very important files stuck in there. What do i do next?
This should regenerate just enough of the database to recover the last version of your files. If you have a particular folder that you want restored “faster”, you can add it at the end of the command, and the database should only be built for the files you add (wildcards also work).
Finally, there is the option to no use a database, which may or may not be faster. To do this, use the recovery tool as explained here:
Since you have already downloaded the files, you may be able to skip step 1, unless they are encrypted.
Yes in that the database restored from your backup would speed up your data restore, but no in that the actual restore would be more difficult due to the additional steps of:
Recreating the database backup job
Restoring the database
Pointing your recreated data backup job (which would have a new database name) to the restored database
What would work really well for a scenario like this would be if the above steps could be automated as part of the Duplicati installer. “Oh crap, my drive died - that’s OK, I’ll just install Duplicati which will ask if I need to restore from an existing backup with (or without) a backed up database.”
Yeah I figured those additional steps would be needed. But it doesn’t seem so bad. I’d much rather do that than wait possibly DAYS for a local database to be rebuilt! I am thinking of doing local db backups for my larger backups (one is 375GB and another 520GB).
Since having both a local AND remote backup is “the best” solution, it would make sense to leverage each other’s destination. In other words:
Have remote backup run and include local backup’s database
Have local backup run and include remote backup’s database
Of course if your local back works for your restore, then you don’t really need the remote backup database… but at least this way you’re better covered if the local backup is only partially restorable due to something silly like, oh I don’t know - let’s say your dog’s giant wagging tail knocked the USB drive on the floor.
I’m using 184.108.40.206_beta_2017-08-01 on Windows 10 64 with SSD intel i3 cpu
Here is the procedure i had followed:
I installed Duplicati, went to restore, chose “directly restore from configuration”, pointed it to a folder on the drive where i saved the filed from cloud, typed in the encryption key, and after it started the recreating database - fetching path information.
After some time asked me to choose files that i want to restore - i selected the important files, Chose restore and i get "recreating database - building partial temporary database…
It made progress very slowly but eventually after 5 it finally came to an end but it never restored any files - just went back to home screen!
So the 5 days spent recreating database was a waste? can i not use that to continue to restore the files? im just scared to start a new process and loose what has already been done.
I’m not worried about doing any backups at the moment, just need to restore the files. Am i doing something wrong?
Sorry for the delay - hopefully you’ve already gotten this resolved, but in case you haven’t…
I tried a “Directly restore from configuration” process in 220.127.116.11_beta and noticed that in step 2 (Encryption) of the restore some of my Advanced Options had too many dashes in front of them. Did you happen to notice this at all during your process?
Part of the difficulties described above comes from using Duplicati itself to back up the DBs. Since there is no need to save versions (I think), would not it be better to simply use another tool to save a copy of the [\AppData\Local\Duplicati] folder? Something like Freefilesync, for example? The execution of freefilesync could be scheduled for soon after Duplicati.
Thank you all for the support - I managed to get the files.
Here is what i did:
After letting the “recreating database” phase run for 5+ days, I went back to GUI and started the restore option all over again and this time it recovered all the files within 24 hours - i guess because it had already created a database.
In my opinion, this program is not ready to be used as a backup program due to lack of features and very time consuming restore process.
Thanks for your comment - I guess that’s part of why Duplicati 2 is still in beta and canary versions. And I agree that there are definitely some scenarios (such as large backup sets) where using Duplicati is harder than we’d like to be, but with input from users like you we can hopefully get those experiences improved.
The time it takes to rebuild the database varies. But I agree that it is not feasible to wait 5 days. I will think again and see if I can come up with some scheme that makes it faster to restore a database.
Yes, this is what happens behind the scenes. If you just go “restore directly”, it will create the database but keep it temporary. If your browser session times out during this time, it will delete the temporary database requiring you to start over, which is why I recommend doing the database rebuild on the commandline if it takes a long time.
I have been doing this for a couple weeks now and it seems to be working well.
I created the separate task in Duplicati to back up the primary task’s database, exported the command line, and saved to a batch file/script. Then the post-backup option was configured on the primary backup task to trigger the script.
On the Windows platform you have to double up the percentage characters for the command line to work correctly.
I would prefer an option to trigger another backup task via the web interface process instead of running the separate command line/batch file to be honest. This way I could see the progress and stats of the database backup task right in the web GUI.
There has been a few requests for something like this. I think it might make sense to add an additional file type that contains the non-redundant data for quick database recreate at the expense of more storage and upload data.
Hello, I am experiencing the same issue. I tried your step of repairing. I got no print outs except for it asking for my encryption password. I have gotten no errors. The CPU is still running at 100%. It has been running for 4 weeks now. Is this normal? What should I do. I still have not been able to recover my backup since January.