next up previous
Next: Results Up: Comparing image resamplers via Previous: Image resampling problems

Computing a perceptual score

The model [18] used here was designed for colour images and uses the opponent colour representation [13] but in this paper we restrict the discussion to monochrome images and use only the B/W channel (which is extremely close to the luminance (Y) channel). A schematic of the system is shown in Figure 2.

   figure32
Figure 2: Schematic of the human vision model

Both the original image and the error are filtered into perceptual channels. The contrast of the original image is then evaluated and used to mask the error. This gives a distortion measure that is averaged in a manner that crudely models the fovea. The blocks labelled ``Perceptual decomposition'' consist of a set of Gabor filters. The first band-pass filter in the set is isotropic with zero response at wavenumber tex2html_wrap_inline617 (to model insensitivity to global luminance level),

equation39

where k is the wavenumber measured in radians per degree of visual of angle and tex2html_wrap_inline621 rad deg tex2html_wrap_inline623 , tex2html_wrap_inline625 rad deg tex2html_wrap_inline623 .

   figure47
Figure 3: Response of the Gabor filter set in Fourier space. Axes are labelled in cycles per degree of visual angle.

The other filters have a bandpass response centred on wavenumber tex2html_wrap_inline629 ,

equation53

where tex2html_wrap_inline631 , tex2html_wrap_inline633 . The filters, shown in Figure 3, are chosen to model the visual channels [8]. Each channel of the distorted image is compared to the same channel from the original image and a masking model applied [5].

The masking model used here allows only within-channel masking and uses masking weights computed as the inverse the normalised detection threshold:

equation62

where tex2html_wrap_inline635 is the detection threshold of the error in the absence of the masker. C is the error contrast and tex2html_wrap_inline639 is the contrast sensitivity function,

equation74

where a=0.0192, c=1.1, d = 2.6 and tex2html_wrap_inline643 rad deg tex2html_wrap_inline623 are experimentally determined constants [9]. tex2html_wrap_inline647 is the contrast of the original image (the masker).

The masked error contrast is averaged using a disc shaped filter. The disc is chosen to subtend 2 tex2html_wrap_inline649 so as to approximate the fovea. The final distortion is computed as

  equation82

where there are N channels, tex2html_wrap_inline655 is the set of M pixels in the foveal disc and e(x,y) is the masked error signal at position x, y. The Minkowski sum in (5) is an attempt to weight errors in the same way as human observers [18]. E(x,y) is called the Visual Difference Score.


next up previous
Next: Results Up: Comparing image resamplers via Previous: Image resampling problems

Stephen King ESE PG
Thu Jul 10 15:27:29 BST 1997