Continuing the discussion from One-Time synchronization with constant file changes:
I have a very similar scenario. Yet i don’t care so much about the two phases Edelf mentioned. I just want all files to get synced to the second machine.
The problem in my scenario is: We have some big files (>1GB) which are constantly changing. Even so i want them to get synced to the other machine in order to have a failover system in case the other one fails.
Currently “37 Objects, ~4.69 GiB” won’t sync, because they get modified during transfer over and over again.
Is there any way or mode to enable transfer of such changing files?
Maybe making a snapshot copy before syncing it, or something alike?
I was thinking about other combined solutions, like using a cronjob to copy the folder to a secondary folder once in a while and syncing that one, but this causes other problems, like deletions do not get synced anymore and in general added complexity.