Centos – More efficient method to synchronize very large # of files

centosscriptingsynchronizationwindows 7

I have a large directory that needs to be synced from a local server to my web server and am looking for the most efficient method of doing this. This directory contains 113k files in 14k directories and is roughly 5GB in size. Local to remote comparison of each file/directory is taking several hours to complete even with little changes.

Local machine is Win7, remote is CentOS 5.5

My current setep is using a scripted synchronize with WinSCP, but as said, the crawling through the directories over a single SCP connection is taking hours. The number of files that require updating should be much smaller than the overall set and I'd like to find a way of scripting the sync locally, logging which files were changed, and then only hitting the web server for the upload of the new files.

Any suggestions?

Best Answer

Have a look at Deltacopy or Syncrify which are both based on the rsync protocol. They will only transfer files that have changed or are new etc. More importantly they will only transfer the changed blocks from large files. Rsync will probably already be installed on your Centos machine