How to speed up rsync?

50

18

I'm running rsync to sync a directory onto my external USB HDD. It's about 150 gigs of data. 50000+ files I would guess.

It's running it's first sync at the moment, but its copying files at a rate of only 1-5 MB/s. That seems incredibly slow for a USB 2.0 enclosure. There are no other transfers happening on the drive either.

Here are the options I used:

rsync -avz --progress /mysourcefolder /mytargetfolder

I'm running Ubuntu Server 9.10.

Jake Wilson

Posted 2010-02-16T23:52:41.490

Reputation: 3 044

2are you sure you're getting a USB2 connection? does a (non-rsync) copy or other write operation run at normal speeds? if not, have you tried a copy/other write op with another USB port/cable? – quack quixote – 2010-02-17T00:14:17.290

See also http://serverfault.com/questions/43014/copying-a-large-directory-tree-locally-cp-or-rsync - there people also propose using two piped tar commands or cpio.

– Blaisorblade – 2013-02-23T17:32:41.137

Answers

39

For the first sync just use

cp -a  /mysourcefolder /mytargetfolder

rsync only adds overhead when the destination is empty.

also.. the -z option is probably killing your performance, you shouldn't be using it if you are not transfering data over a slow link.

user23307

Posted 2010-02-16T23:52:41.490

Reputation: 5 915

1

rsync is also a one way sync. Very good for backing up to a server or from a server. However, if you want local TWO-Way sync to a removable drive, you may want to check out csync https://www.csync.org/get-it/ not to be confused with csync2 which is a completely different project.

– Jesse the Wind Wanderer – 2014-12-31T13:32:45.037

3rsync is so called because it's for remote synchronization and is not really appropriate for a locally-connected volume for this very reason. – msanford – 2010-06-27T22:30:57.533

3rsync -avz --progress /mysourcefolder/ /mytargetfolder or you'll get a copy of mysourcefolder inside of mytargetfolder rather than mirroring the contents – editor – 2017-12-01T08:16:36.663

4This answer does not answer the question. The question was about how to optimize rsync - not replace it with the cp command. – oemb1905 – 2018-11-25T04:08:00.987

6It's supposed to be usable also for local transfers, and it's much more flexible. It's only possibly overkill for the first sync. – Blaisorblade – 2013-02-23T17:31:15.943

42

If you're using rsync with a fast network or disk to disk in the same machine,

not using compression -z

and using --inplace

speeds it up to the performance of the harddrives or network

compression uses lots of CPU

not using inplace makes the harddrive thrash alot (it uses a temp file before creating the final)

compression & not using inplace is better for doing it over the internet (slow network)

NEW: Be aware of the destination... if there is NTFS "compression" enabled... this severely slows down large files (I'd say 200MB+) rsync almost seems stalled, it's caused by this.

Scott Kramer

Posted 2010-02-16T23:52:41.490

Reputation: 520

NTFS compression is slowwww on big files – Scott Kramer – 2016-06-03T03:51:32.097

I don't see anything about '--inline' on the man page – Anthony – 2017-11-18T21:52:26.420

1It's '--inplace' – Scott Kramer – 2017-12-01T17:17:21.557

--inplace will only help if you already have a file there, and it's mostly the same as the source. otherwise, rsync defaults to an atomically safe method (destination tmp named, then mv into place.) – markhahn – 2020-02-06T18:45:54.930

28

Use the -W option. This disables delta/diff comparisons. When the file time/sizes differ, rsync copies the whole file.

Also remove the -z option. This is only useful for compressing network traffic.

Now rsync should be as fast as cp.

vdboor

Posted 2010-02-16T23:52:41.490

Reputation: 438

2but note that according to the man page says for -W: "This is the default when both the source and destination are specified as local paths, but only if no batch-writing option is in effect." – GuoLiang Oon – 2017-07-05T08:15:15.700

6Minor note: -z is only useful for low speed network traffic. If your network is fast enough, it'll slow things down, since you'll be limited by CPU. – WhyNotHugo – 2013-07-08T02:28:12.323

3These tips vastly sped up the transfer of my files between two NAS devices, thanks! – djhworld – 2013-09-22T13:30:13.897

19

First - the number of files in this case is going to be a major factor. It's an average size of 3MB each. There's probably an io bottleneck influencing the speed in the OP's case. More here - that's a pretty dry read, but the cover picture is worth it.

So, using rsync to copy to an empty directory? Here are some ways to speed it up:

  1. No -z - definitely don't use -z as in the OP.
  2. --no-compress might speed you up. This could have the biggest impact... my test was 13,000 files, total size 200MB, and using rsync 3.1.3. I synced to a different partition on the same internal SSD drive. With --no-compress, I get 18 MBps, and without it I get 15 MBps. cp, by the way, gets 16 MBps. That's a much smaller average file size though. Also - I can't find any documentation for --no-compress. I learned about it from this post on stackexchange.com.
  3. -W to copy files whole - always use this if you don't want it to compare differences; never mind that the point of rsync is to compare differences and only update the changes.
  4. -S to handle sparse files well - can't hurt if you don't have sparse files.
  5. --exclude-from or something similar to exclude files you might not need will cut down the time, but it won't increase your transfer speed.
  6. It's possible if you send the output to a file like this rsync -a /source /destination >/somewhere/rsync.out 2>/somewhere/rsync.err - the first > basically prints a file with all the stuff you would normally see, and the 2> refers to error messages.
  7. Finally, running multiple instances of rsync for different parts of your transfer could be a big help.

My command would be:

rsync -avAXEWSlHh /source /destination --no-compress --info=progress2 --dry-run

If all looked well, I'd delete "--dry-run" and let it go. A, X, and E cover extended attributes and permissions not covered by -a, l is for soft links, H is for hard links, and h is for human readable.

Updating an already synced directory on a USB drive, or the same drive, or over a network, will all require different rsync commands to maximize transfer speed.

Bonus - here's the rsync man page, and if you want to test your hard drive speed, bonnie++ is a good option, and for your network speed, try iperf.


*The post is almost ten years old, but search engines sure like it, and I keep seeing it. It's a good question, and I don't think the top answer to "how to speed up rsync" should be "use cp instead."

Fin Hirschoff

Posted 2010-02-16T23:52:41.490

Reputation: 191

1In respect of item 7) I could improve a lot more the performance using a traditional HDD as the source, by extending read-ahead buffer using "blockdev --setra 8192 /dev/sdX". So I intend to reduce head seeks. – user2480144 – 2019-08-10T09:08:29.183

1I've found that S (handling Sparse files) is actually slow. Removing it made it much faster – Sammy Guergachi – 2019-11-27T16:26:43.973

--no-compress is not mentioned in the rsync man page. – Tunisia – 2020-02-26T10:59:22.373

2

You don't say what size distribution your files have. If there are many small files then this will reduce overall transfer rate by increasing head movement latency in both the source and destination drives as the tool opens new files and the OS keeps directory entries and other metadata (such as the filesystem's journal if you are using meta-data journaling like ext3/ext4 and NTFS do by default) up to date during the transfer. A file copy proces will only "get into its stride" for larger objects, when a simple bulk transfer is happening.

David Spillett

Posted 2010-02-16T23:52:41.490

Reputation: 22 424

2

Avoid

  • -z/--compress: compression will only load up the CPU as the transfer isn't over a network but over RAM.
  • --append-verify: resume an interrupted transfer. This sounds like a good idea, but it has the dangerous failure case: any destination file the same size (or greater) than the source will be IGNORED. Also, it checksums the whole file at the end, meaning no significant speed up over --no-whole-file while adding a dangerous failure case.

Use

  • -S/--sparse: turn sequences of nulls into sparse blocks
  • --partial or -P which is --partial --progress: save any partially transferred files for future resuming. Note: files won't have a temporary name, so ensure that nothing else is expecting to use the destination until the whole copy has completed.
  • --no-whole-file so that anything that needs to be resent uses delta transfer. Reading half of a partially transferred file is often much quicker than writing it again.
  • --inplace to avoid file copy (but only if nothing is reading the destination until the whole transfer completes)

Tom Hale

Posted 2010-02-16T23:52:41.490

Reputation: 1 348

1

You definitely want to give rclone a try. This thing is crazy fast :

$ tree /usr [...] 26105 directories, 293208 files

$ sudo rclone sync /usr /home/fred/temp -P -L --transfers 64

Transferred: 17.929G / 17.929 GBytes, 100%, 165.692 MBytes/s, ETA 0s Errors: 75 (retrying may help) Checks: 691078 / 691078, 100% Transferred: 345539 / 345539, 100% Elapsed time: 1m50.8s

This is a local copy from and to a LITEONIT LCS-256 (256GB) SSD.

You can add --ignore-checksum on the first run to make it even more faster.

Frédéric N.

Posted 2010-02-16T23:52:41.490

Reputation: 11