Hi,
On the last couple of runs of one of my backup jobs, it was throwing various database errors. As I result I used the option to delete and rebuild the database.
This has currently been running for about 11 days. Looking at the sqlite database for this particular backup job, it appears to have grown to a similar size as it was before I cleared it. The timestamp on the files show they haven’t changed since 11:23 on 14th July (over two days ago). It seems to have been in this state for probably about a week. The update time on the sqlite databases does change occasionally. The Duplicati process also does appear to be consuming small amounts of CPU on a regular basis.
The backup job is around 1.8TB in size, containing around 2500 files. The backup destination is Jottacloud if it’s of any relevance.
Is this behaviour expected? Is there anything I can do to speed up the rebuild process? Is there any way to work out what it’s actually doing at the moment?
Thanks
Andy
1 Like
In my tests of a database recreate (My experience with Database Rebuild / Recreate) I found it took about 4.25 days to recreate an 853 MB sqlite file (built on ~1TB of source content). If what I suspect is correct, the time it takes for a rebuild will get progressively longer the larger the resulting database should be.
I believe most of the issues performance issue with a rebuild is in the local database reads checking if blocks already exist or still need to be added. If that’s correct, the only way I can think of to speed it up would be to have the database stored on an SSD or ram drive - neither of which is something you can do once the recreate is started.
Since an interrupted rebuild can’t be continued (you’d have to start over) I’d suggest just waiting it out.
If you want to make sure things are stalled, you could take a look at the “lastPgEvent” line of the “System info” page or use a filter like “Profiling” on the Live tab of the the main menu “About” -> “Show log” page.
But that won’t make it go any faster.
This is the lastPgEvent:
lastPgEvent : {"BackupID":"7","TaskID":36,"BackendAction":"Get","BackendPath":"duplicati-b2cfc6caf74574ed4ad9d408212c6e2ea.dblock.zip.aes","BackendFileSize":524305101,"BackendFileProgress":524305101,"BackendSpeed":12256,"BackendIsBlocking":false,"CurrentFilename":null,"CurrentFilesize":0,"CurrentFileoffset":0,"Phase":"Recreate_Running","OverallProgress":0.714285731,"ProcessedFileCount":0,"ProcessedFileSize":0,"TotalFileCount":0,"TotalFileSize":0,"StillCounting":false}
I guess that means it’s about 70% through?
Andy
I believe that’s correct - of course it’s 2 days later now, hopefully it has progressed a little more…
Hi,
Nope, still showing exactly the same progress.
Database file hasn’t been written to since morning of July 19th…
My backups haven’t been running as a result since 3rd July. Not ideal.
Andy
Absolutely agree there - unfortunately, while we have heard from a small number of people about this issue we haven’t been able to figure out exactly what causes it.
Even worse, an aborted rebuild can’t be continued so you end up just starting over again.
@kenkendk, is there a more appropriate way to check on the status of a database rebuild than OverallProgress in the lastPgEvent property on the System Info page?
Still going. Progress is now 0.7214286, so in a week it’s only gone up by about 0.7%.
This is going to take a long time to complete…
Andy
I’m almost tempted just to bin this backup and start again. However it’s a lot of data so would take a long time to complete.
Is there any reason to expect this rebuild to complete soon? If it’s managing 0.7% a week, then extrapolating that suggests it could take something like 40 weeks to complete!
Andy
That’s strange, I did big rebuild recently and it “only” took two days for 600GB. More info here Rebuilding Database Extremely Slow
Well, there was some surprise that it had completed that quick. My backup is more like 2TB I think.
Progress is still 0.7214286, so no progress in the last 19 hours.
Andy
Progress is now 0.7285714.
Is there any information / debugging I can get to you to try to work out why this is taking so long? I’m a developer myself, so if I need to do anything like provide an strace output or similar, I should be able to do that.
Is it even worth allowing this rebuild to complete? Would it be easier just to delete the backup and start again? I could probably live with that, as it’s just a lot of information that doesn’t change often, so I don’t need access to old versions etc.
I assume I can abandon the rebuild and then delete the backup set?
Thanks
Andy
Yes, that’s an option - though the backup set stored at the destination can still be used to restore FROM even if the local database doesn’t exist, so if you have the space you might want to keep it around for a bit while the new job is “filling up”.
It’s possible @kenkendk might have some input or testing requests… We know this is an issue in some instances, but have yet to figure out why it happens only some of the time.
Ok
Best get those requests in son, because it’s now been bout a month with no backups. Can’t really allow this to continue for weeks more.
Andy
I have seen this too. What is even more surprising, sometimes during recreate duplicati seems to be doing literally nothing. It doesn’t use CPU and windows shows neither disk nor network, seemingly for minutes or even hours…
Looking at some of the queries (using joins and other not so fast approaches) in the log I also wondered if dumb simple batch inserts wouldn’t be possible…
Giving up on this now. Can’t afford to have it blocking all my other backups for any longer.
Will start again but keep the old files in place just in case.
Andy
Thanks for letting us know. Good luck with the new backup - at least now you can review your processes and decide if you want to split or join things into smaller or larger backup sets.
Also, remember the files in the destination can still be used as a restore source via “Direct restore from backup files” so if you don’t need the space right away, it might be a good idea to keep those old backup files around while the new backup “fills in”.
For what it’s worth: I’m on my third try to delete/rebuild a huge database (>4TB target files on Google Drive) on Mac OSX (was High Sierra, now Mojave)… I takes about three weeks (or more), and has crashed twice with a “Panic” OSX failure. By the third week (now), my MacbookPro 15" (retina, mid 2014) w/1TB SSD drive is crawling, with some functions (printing, copying, activity monitor) failing intermittently. I’m only seeing about two transaction sets per day in the log now (a download and a bunch of SQL commands, then nothing for another 12 hours). I hope that, since I’ve upgraded to Mojave, this one runs to completion without crashing my system. Meanwhile, I haven’t been able to use Duplicati for weeks, and my non-Duplicati work on my Macbook have been very slow (e.g., taking an hour or more to start printing).
What would really help right now would be the ability to checkpoint the database rebuild, so that I could reboot. I’ve seen that suggested elsewhere.
Any thoughts on how to manage this situation better are welcome.
…Steve
2 Likes
If your main worry is that you need to leave your mac running, why don’t you rent a virtual server for a couple of days or weeks? You can probably find some not too expensive powerful virtual machines out there. Just make sure you don’t change the backup files while it’s running.
Until we get the slow recreate issue resolved, if a checkpoint would let recreates be resumable that sounds like a good idea to me.
Of course I have no idea how much effort that would take so my support doesn’t mean much.
Update: I finally (December 4th) got a completed database rebuild and am running Duplicati normally again. The rebuild that finally completed took just about a month! Since the rebuilds cannot be checkpointed, whenever my MAC OSX Mojave locked up and had to be rebooted, I lost all progress and had to start again.
During the database rebuild, my system progressively became more sluggish due to high I/O traffic, and several subsystems and applications became unusable (e.g., printing worked intermittently and very slowly, other apps timed out accessing data and quit), which I believe contributed to the lockups. Also, since I couldn’t really use my system for normal workload (other than browsing and email), I lost several months of productivity.
Once I got a clean rebuild, I added the “auto-vacuum” and “smart backup retention” options to cull my database(s), which was over 63GB for a 4TB archive. The next backup took a week, and got caught when my system locked up and rebooted. The following backup proceeded to rebuild the database again, taking over a week, and then on the next backup resumed culling the database for another 4-5 days. Finally, the database is down to 5GB and backup (on Google Drive) about 800GB.
My observations: (1) “Rebuild” needs a checkpoint option, either automatic or manual; (2) Large databases are unwieldy and take too long to do anything; (3) it is impossible to “keep all backups” for a few years for an active drive, because the database becomes too large, so either use “smart backup retention” (and risk losing something you may want) or cull your backups manually.
Thank you. I appreciate all of the help and support from the Duplicati developers and community.