3

Short story : We have a need for a rock-solid reliable file mover process. We have source directories that are often being written to that we need to move files from. The files come in pairs - a big binary, and a small XML index. We get a CTL file that defines these file bundles. There is a process that operates on the files once they are in the destination directory; that gets rid of them when it's done. Would rsync do the best job, or do we need to get more complex? Long story as follows :

We have multiple sources to pull from : one set of directories are on a Windows machine (that does have Cygwin and an SSH daemon), and a whole pile of directories are on a set of SFTP servers (Most of these are also Windows.) Our destinations are a list of directories on AIX servers.

We used to use a very reliable Perl script on the Windows/Cygwin machine when it was our only source. However, we're working on getting rid of that machine, and there are other sources now, the SFTP servers, that we cannot presently run our own scripts on.

For security reasons, we can't run the copy jobs on our AIX servers - they have no access to the source servers. We currently have a homegrown Java program on a Linux machine that uses SFTP to pull from the various new SFTP source directories, copies to a local tmp directory, verifies that everything is present, then copies that to the AIX machines, and then deletes the files from the source. However, we're finding any number of bugs or poorly-handled error checking. None of us are Java experts, so fixing/improving this may be difficult.

Concerns for us are:

  • With a remote source (SFTP), will rsync leave alone any file still being written? Some of these files are large.
  • From reading the docs, it seems like rysnc will be very good about not removing the source until the destination is reliably written. Does anyone have experience confirming or disproving this?
  • Additional info We will be concerned about the ingestion process that operates on the files once they are in the destination directory. We don't want it operating on files while we are in the process of copying them; it waits until the small XML index file is present. Our current copy job are supposed to copy the XML file last.
  • Sometimes the network has problems, sometimes the SFTP source servers crap out on us. Sometimes we typo the config files and a destination directory doesn't exist. We never want to lose a file due to this sort of error.
  • We need good logs

If you were presented with this, would you just script up some rsync? Or would you build or buy a tool, and if so, what would it be (or what technologies would it use?) I (and others on my team) are decent with Perl.

mfinni
  • 35,711
  • 3
  • 50
  • 86
  • You say that you have a number of SFTP servers on which you "cannot presently run [your] own processes on". If that is the case, then unfortunately, rsync is likely not the right tool. rsync needs to run on both the source and destination hosts. You might still be able to force rsync into the solution if you are able to use something like sshfs to mount your remote SFTP servers as local filesystems, but such an approach would need to be tested for feasibility. – Steven Monday Dec 30 '10 at 16:24
  • Incorrect - rsync can use SFTP for either (or both) source and/or destination. Plus, we probably could get it installed as a daemon/service, they just don't want us running our own scripts on their servers. – mfinni Dec 30 '10 at 16:26
  • Edited my question :s/processes/scripts/ – mfinni Dec 30 '10 at 16:27
  • Regarding rsync over sftp, I refer you to [this page](http://serverfault.com/questions/135618/is-it-possible-to-use-rsync-over-sftp-without-an-ssh-shell). Although given in April 2010, I believe David Spillett's answer still holds true today. – Steven Monday Dec 30 '10 at 16:36
  • Steven - in my environment, I have used rsync exactly as I described. Our SFTP servers will allow us to run rsync over SSH. – mfinni Dec 30 '10 at 16:44
  • Aargh. I have to eat some crow here. I swear that I remembered doing that, but it was SCP that can do both remote source and destination. Regardless, we can definitely get rsync daemons where we need them. – mfinni Dec 30 '10 at 18:31

1 Answers1

3

Edit: Rsync does an end-to-end check: after the file is transfered it calculates the checksum of that file on the destination and compares it to the checksum on the source. When the the checksums match, only then it declares the transfer successful. This is reflected in the final exit status code - if ALL transfered files passed the test then the exit code will be 0 (Success).

In a similar setup I scripted-up my own solution based on rsync. It was for nightly backups and we do not delete files automatically.

To address some of your concerns:

  • Rsync never modifies anything on the source side (unless you use the --remove-source-files option).
  • If the network goes down for a long time Rsync will give up and give an appropriate exit status. I check this in my script and for specific exit codes (which I observed in practice by logging) I have the script re-try the rsync command up to 3 times.
  • Yes, your script should log as much as possible. Timestamp, Total Running time, Rsync exist status, Rsync --stats output (amount transmitted). I also run find at the end of the transfer to count the number of files and du * to get sizes of directories and log that.

Basically you need to take care of a few things in the script. Mainly: Gathering the exit status, some statistics, and removing source files on a successful transfer.

You can trust rsync's exit status that all the requested files were transfered but you should think about how much do you trust your script to provide rsync with the right files (source directory) before you delete them on the source machine. Maybe counting the files with find on the source and then on the destination (and then checking if those numbers match) would be a good final check before your script deletes the files automatically.

Give it a 10 to 20 tries to develop and test your script. You would need to install Cygwin with rsync and ssh clients on the Windows machines.

It's good to feel confidant about a application like that by knowing exactly how it works. I never used a commercial backup software - but if you can find a rock-solid one and trust it - then go for that - it could save you a lot of time.

Aleksandr Levchuk
  • 2,415
  • 3
  • 21
  • 41
  • Well, not never. There is the `--remove-source-files` option, which I would be using for this purpose. That removes the files after they are done. – mfinni Dec 30 '10 at 16:35
  • I added some more info about the files being moved and the overall process. I hadn't realized how relevant it would be, but in hindsight it's very important. Neither the source nor destination directory are archives - files put in the destination are processed by yet something else. Made edits to my question. – mfinni Dec 30 '10 at 16:43
  • @Aleksander - the points you make about being sure before you delete the source files is important to us. That's why I want to know how reliable rsync's option `--remove-source-files` is. – mfinni Dec 30 '10 at 16:46
  • Looks like `--remove-source-files` has been around for more than 4 years (http://gitweb.samba.org/?p=rsync.git&a=search&s=--remove-source-files) and from the description in the manual page I would trust it. I'm 99% sure that it does not modify the file until the transfer of that file is over and successful. But I would always test a few times before using in production. – Aleksandr Levchuk Dec 30 '10 at 16:54
  • `--remove-source-files` used to be called `--remove-sent-files` (http://gitweb.samba.org/?p=rsync.git&a=search&s=--remove-sent-files) – Aleksandr Levchuk Dec 30 '10 at 17:03
  • @Aleksandr, thanks for the edits. The checksum is definitely something we need. – mfinni Dec 30 '10 at 18:33