cuda v100 and RTX2080 bandwidth fluctuations

2

I have a simple cuda kernel (adding two vectors of size N) pretty similar to to this cuda blog here. I only changed a few things, e.g. running the measurement over various sample. So, let this run for, lets say, 1000 times and writing this measurement to a txt afterwards. If I plot now the measurements for transfering a vector to the device I get the following:

enter image description here

Now, if we take a look at the stddev drawn as vertical errorbars, then it should be clear, that for some reason, the data movements fluctuation scale with the size, because the errorbars are kinda constant in a log-log plot. This can be validated when only the stddev is plotted

enter image description here

If I take the very same programm from the cuda blog, then I get for every 10-th run or so also bandwidth fluctuations. Where does this come from? I observed the same behaviour on two different GPUs, a V100 and a RTX2080. Sorry for the inconvenience regarding the images, but I don't have enough reputation points.

Maxi Köhler

Posted 2019-12-11T16:01:41.587

Reputation: 31

Answers

1

as a nvidia dev forum user suggested I tried pinned memory instead of non-pinned and this worked out pretty well. The fluctuations are almost completely gone.

Maxi Köhler

Posted 2019-12-11T16:01:41.587

Reputation: 31