Numerical dispersion

In applied computational mathematics, numerical dispersion is a difficulty with computer simulations of continua (such as fluids) wherein the simulated medium exhibits a higher dispersivity than the true medium. This phenomenon can be particularly egregious when the system should not be dispersive at all, for example a fluid acquiring some spurious dispersion in a numerical model.

It occurs whenever the dispersion relation for the finite difference approximation is nonlinear.[1][2] For these reasons, it is often seen as a numerical error.

Numerical dispersion is often identified, linked and compared with numerical diffusion,[3] another artifact of similar origin.

Explanation

In simulations, time and space are divided into discrete grids and the continuous differential equations of motion (such as the Navier–Stokes equation) are discretized into finite-difference equations;[4] these discrete equations are in general unidentical to the original differential equations, so the simulated system behaves differently than the intended physical system. The amount and character of the difference depends on the system being simulated and the type of discretization that is used.

gollark: You have a currency API thing? Interesting.
gollark: It also now communicates with icecast to harvest a current listener count.
gollark: Among many other things it fetches information from the MPD instance running my internet radio system, but after several hours it just stops working for ??? reasons.
gollark: This is the one which is partly broken for no reason.
gollark: My miscellaneous API stuff goes into one big 500 line python file, for economies of scale.

See also

References


This article is issued from Wikipedia. The text is licensed under Creative Commons - Attribution - Sharealike. Additional terms may apply for the media files.