They ought not to; files are on/off patterns, and all operations such as move and copy are designed to move data whole, without losing any. It's not like the human game of "telephone" where copying the file over and over makes it more mumbled and confused every time until all that's left is nonsense.
But, every time you do something, there's a tiny chance of it going wrong. The more things you do, the more chance you will run into one of them going wrong. One of them, for instance, is a powercut in the middle of a move - what will you end up with? A partial file? No file? Two files?
Nothing is perfect; here, in 2005, a study looks at disk drive error rates, and finds:
We observed 3 loss events while reading 1.4 PB, This is 3 lost files.
In the datacenter environment we lost 2 of 32,000 10GB files, In the
office setting (System 1) we lost one 10GB file in 35,000 tries, and
no 100GB files in 7,560 tries.
That's just disk errors, they also "experienced at least 4 controller firmware or driver errors"
.
On top of that, there's memory errors - the chance of corruption between loading information into computer memory and then writing it out again. There's an infamous Google study where they track errors in memory accross their thousands of servers, cited at Wikipedia here which states:
"A very large-scale study based on Google's very large number of servers was presented at the SIGMETRICS/Performance’09 conference.[4] The actual error rate [..] 25,000 to 70,000 errors per billion device hours per megabit [..] more than 8% of DIMM memory modules affected by errors per year."
And of course there's a chance of error, a misread, a corruption by power spike or cosmic ray, a design flaw, pushing outside a low tolerance of cheap hardware, in the magnetic media on the disk, in the transfer over cables, in the motherboard, the chipsets, the processor, memory, the firmware or software, and all the way back again.
Incidentally, here is a fantastic article based on the idea that if a computing device corrupts data in memory while looking up an internet address, it might lookup the wrong address. So the author registered some "wrong addresses" (what you'd get if you corrupted a bit of a popular internet site name) to see if any devices connected to them - he had 52,000 requests in almost 6 months.
The study of how to make things /more/ reliable in computing is a busy area. From CRCs and checksums, to RAID with parity, the checksum-tree based filesystem ZFS, to automatic replication, to ECC memory, and so on.
short answer: no. files become corrupted when not transferred in full or parts of it become unreadable/changed due to hard drive malfunction/bad habits involving flash drives. – Lorenzo Von Matterhorn – 2013-04-27T00:56:28.287
The other short answer: maybe (but not likely). Any time you read from the disk, there’s a possibility that the hardware will glitch and give you wrong data; any time you write to the disk, there’s a possibility that the hardware will glitch and save wrong data. But there are mechanisms to help the computer detect that this has happened and take remedial action (as a last resort, reporting an error). – Scott – 2013-04-27T01:15:34.227