Jump to content

Signal-to-noise ratio (imaging)

From Wikipedia, the free encyclopedia

Signal-to-noise ratio (SNR) is used in imaging to characterize image quality. The sensitivity of a (digital or film) imaging system is typically described in the terms of the signal level that yields a threshold level of SNR.

Industry standards define sensitivity in terms of the ISO film speed equivalent, using SNR thresholds (at average scene luminance) of 40:1 for "excellent" image quality and 10:1 for "acceptable" image quality.[1]

SNR is sometimes quantified in decibels (dB) of signal power relative to noise power, though in the imaging field the concept of "power" is sometimes taken to be the power of a voltage signal proportional to optical power; so a 20 dB SNR may mean either 10:1 or 100:1 optical power, depending on which definition is in use.

Definition of SNR

[edit]

Traditionally, SNR is defined to be the ratio of the average signal value to the standard deviation of the signal :[2][3]

when the signal is an optical intensity, or as the square of this value if the signal and noise are viewed as amplitudes (field quantities).[further explanation needed]

See also

[edit]

References

[edit]
  1. ^ ISO 12232: 1998 Photography – Electronic Still Picture Cameras – Determining ISO Speed
  2. ^ Janesick, James R. (2007). 9780819478382/10.1117/3.725073 Photon Transfer. doi:10.1117/3.725073. ISBN 978-0-8194-7838-2. {{cite book}}: Check |url= value (help)
  3. ^ Rowlands, Andy (April 2017). Physics of Digital Photography. IOP Publishing. ISBN 978-0-7503-1243-1.

Further reading

[edit]