**Processing of Multichannel Remote-Sensing Images with Prediction of Performance Parameters**

Benoit Vozel, Oleksiy Rubel, Alexander Zemliachenko, Sergey Abramov, Sergey Krivenko, Ruslan Kozhemiakin, Vladimir Lukin and Kacem Chehdi

Additional information is available at the end of the chapter

http://dx.doi.org/10.5772/61853

#### **Abstract**

In processing of multichannel remote sensing data, there is a need in automation of basic operations as filtering and compression. Automation presumes undertaking a decision on expedience of image filtering. Automation also deals with obtaining of information based on which certain decisions can be undertaken or parameters of processing algorithms can be chosen. For the considered operations of denoising and lossy compression, it is shown that their basic performance characteristics can be quite easily predicted based on easily calculated local statistics in discrete cosine transform (DCT) domain. The described meth‐ odology of prediction is shown to be general and applicable to different types of noise under condition that its basic characteristics are known in advance or pre-estimated accu‐ rately.

**Keywords:** Multichannel remote sensing data, automatic processing, denoising, lossy compression, performance prediction, DCT

#### **1. Introduction**

Remote-sensing (RS) data are widely used for numerous applications [1], [2]. Primary RS images acquired onboard of airborne or spaceborne carriers and intended for Earth surface monitoring are usually not ready for direct use and, thus, are subject to a certain preprocessing. This preprocessing can be carried out in several stages and includes the following operations: geo-referencing and calibration, blind estimation of noise/distortion characteristics, prefiltering, lossless or lossy compression, [1], [2], etc. These operations can be distributed between onboard and on-land computer means (processors) in different ways depending upon many factors [3-5].

© 2016 The Author(s). Licensee InTech. This chapter is distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/3.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Regardless of the distribution of functions, the operations onboard are usually performed in a fully automatic manner (although there can be some changes in algorithm parameters by command passed from Earth). In turn, the operations carried out on land can be, in general, performed in an interactive manner and labor of highly qualified experts is exploited for this purpose. However, a certain degree of automation of on-land data processing is required as well. The need in processing automation is especially high if one deals with multichannel (e.g., hyperspectral) RS data [6], where the number of channels (components, sub-bands) can reach hundreds. Such RS images have become popular and widespread (available) currently due to their (potential) ability to provide rich information for various applications [6], [7].

Meanwhile, the multichannel nature of RS data results in new problems in their processing [3], [8]. The main problems and actual questions are the following:


The latter question can be mainly addressed as mentioned below. It is strictly connected with other questions as follows:


This chapter will focus on two typical operations of multichannel RS data processing, namely, filtering and lossy compression. While considering them, the fact that the acquired images are noisy is taken into account. One can argue that noise is not seen in many RS images (or components of these images). This is true, and noise cannot be observed in approximately 80% of the visualized sub-band images of hyperspectral data. This is explained by the peculiarities of human vision, which does not see noise if peak signal-to-noise ratio (PSNR) in a given singlechannel (component) image exceeds 32–38 dB. However, recent studies [7], [10-12] have demonstrated that noise is present in all sub-band images and this is due to the principle of operation of hyperspectral imagers.

Moreover, it has been shown in [10], [11] that noise is (can be) of quite a complex nature and the noise acquired in multichannel RS images has specific properties. First, it is signaldependent [10], [11], [13]. Second, it is of essentially a different intensity (see Abramov et al., 2015 in [14]). More precisely, the wide variation of dynamic range and noise intensity in subband images jointly leads to wide limits of signal-to-noise ratio (SNR) in components of multichannel images. This has led to the use of the term "junk bands" [15] and different strategies of coping with noisy channels in multichannel data. Some researchers prefer to use these sub-bands in further processing while others propose to remove them; it is also discussed whether they can be filtered or not [15]. It has been shown that if filtering of these junk bands is efficient, this can improve the classification of hyperspectral data [16]. However, the aforementioned questions concern the efficiency of image preprocessing and its prediction.

Regardless of the distribution of functions, the operations onboard are usually performed in a fully automatic manner (although there can be some changes in algorithm parameters by command passed from Earth). In turn, the operations carried out on land can be, in general, performed in an interactive manner and labor of highly qualified experts is exploited for this purpose. However, a certain degree of automation of on-land data processing is required as well. The need in processing automation is especially high if one deals with multichannel (e.g., hyperspectral) RS data [6], where the number of channels (components, sub-bands) can reach hundreds. Such RS images have become popular and widespread (available) currently due to

Meanwhile, the multichannel nature of RS data results in new problems in their processing [3],

**•** How to manage large volumes of acquired data with maximal or appropriate efficiency

**•** Is it possible to skip some operations of data processing if their efficiency is not high and,

The latter question can be mainly addressed as mentioned below. It is strictly connected with

**•** Is it possible to predict the performance of some standard operations of RS data (image)

**•** What is the accuracy of such a prediction and is this accuracy high enough to undertake a decision to skip carrying out an operation or to set a certain value of some parameter used

This chapter will focus on two typical operations of multichannel RS data processing, namely, filtering and lossy compression. While considering them, the fact that the acquired images are noisy is taken into account. One can argue that noise is not seen in many RS images (or components of these images). This is true, and noise cannot be observed in approximately 80% of the visualized sub-band images of hyperspectral data. This is explained by the peculiarities of human vision, which does not see noise if peak signal-to-noise ratio (PSNR) in a given singlechannel (component) image exceeds 32–38 dB. However, recent studies [7], [10-12] have demonstrated that noise is present in all sub-band images and this is due to the principle of

Moreover, it has been shown in [10], [11] that noise is (can be) of quite a complex nature and the noise acquired in multichannel RS images has specific properties. First, it is signaldependent [10], [11], [13]. Second, it is of essentially a different intensity (see Abramov et al., 2015 in [14]). More precisely, the wide variation of dynamic range and noise intensity in subband images jointly leads to wide limits of signal-to-noise ratio (SNR) in components of multichannel images. This has led to the use of the term "junk bands" [15] and different strategies of coping with noisy channels in multichannel data. Some researchers prefer to use these sub-bands in further processing while others propose to remove them; it is also discussed whether they can be filtered or not [15]. It has been shown that if filtering of these junk bands

their (potential) ability to provide rich information for various applications [6], [7].

[8]. The main problems and actual questions are the following:

(here, different criteria of efficiency can be used)?

consequently, if it is not worth performing them?

other questions as follows:

374 Environmental Applications of Remote Sensing

in the image-processing chain [9]?

operation of hyperspectral imagers.

processing?

The questions raised can be partly answered with the results obtained in recent research. The objective is to show that important performance parameters of image denoising and/or lossy compression can be quickly and quite accurately predicted using simple input parameter(s) and dependences obtained in advance. The obtained results are divided into two parts. The first part deals with the prediction of filtering efficiency. This research has started in 2013 [17] and has its history in a study conducted in [18]. The second part relates to the compression of noisy images [19], [20]. In fact, the results obtained for predicting the parameters of lossy compression can be treated as based on the same principle as that for image filtering and for further research.

Before taking the image performance criteria and preprocessing techniques into consideration, it is important to note the following: first, there are two hypotheses. It is supposed that noise type is known or determined in advance. It is also assumed that its parameters are either known or accurately pre-estimated. It is to be noted that, currently, there are quite a few efficient methods for estimating the parameters of pure additive noise [8], [21-25], speckle noise [26], and different types of signal-dependent noise [10-12], [27], [28]. The noise parameters are taken into account by the most modern filtering techniques that belong to the families of orthogonaltransform-based filters [29-33] and nonlocal filters, for example, block-matching and threedimensional filtering (BM3D) [34]. The same relates to modern methods of lossy compression of noisy images [19], [35].

Second, we restrict ourselves to consider the image- filtering and compression techniques based on discrete cosine transform (DCT). This is explained using several reasons. DCT is a powerful orthogonal transform widely exploited in image processing. Filters and compression techniques based on DCT are currently among the best [34]. They can be quite easily adapted to the signal-dependent noise directly [32], [36] or equipped with proper variance-stabilizing transformations (VST) [19], [32], [37]. This restriction does not mean that the approach to prediction cannot be applied to other filtering and lossy compression techniques. This approach should be applicable (with certain modifications) but is yet to be thoroughly checked.

Third, in the analysis of the prediction approach, traditional quality metrics are employed such as mean square error (MSE) and peak signal-to-noise ratio (PSNR), as well as some visual quality metrics such as PSNR human visual system masking metric (PSNR-HVS-M) [38]. Behavior and properties of traditional metrics are understood well by those dealing with image processing. Although PSNR-HVS-M is less popular, this is one of the best metrics that takes into account the peculiarities of human visual system (HVS) and that can be calculated for either one component of a multichannel image or a group of components of a multichannel image. It is expressed in dB, and it is usually either slightly smaller than PSNR (for annoying types of distortions like spatially correlated noise) or larger than PSNR (if distortions are masked by texture). This is important since we assume that the processing of multichannel images is carried out either component-wise or in groups of channel images, where a group includes the entire image in marginal case.

Fourth, other criteria of image-processing efficiency, such as classification accuracy, object detectability, etc., are important for the preprocessed RS data. We are unable to predict them, but recent research shows [39] that these criteria are connected with the traditional criteria of image processing. Thus, it is expected that if good values of conventional and HVS metrics are provided due to preprocessing, appropriate classification accuracy and other criteria will be attained.

### **2. The considered image-performance criteria and preprocessing techniques**

This chapter considers the following model of an observed multichannel image:

$$I\_{\rm kij}^{\rm nasy} = I\_{\rm kij}^{\rm true} + n\_{\rm k,j}(I\_{\rm kij}^{\rm true}), \; i = 1, \dots, I, \; j = 1, \dots, I, \& k = 1, \dots, K \tag{1}$$

where *Ik ij* noisy is *ij* -th sample of noisy (original) *k*-th component of a multichannel image, *nk ij* denotes the *ij*-th value of the noise in *k*-th component statistic, which is, in general, supposed to be dependent on the true image value *Ik ij* true in this voxel (3D pixel), *I* and *J* define the image size, and *K* denotes the number of channels. It is also assumed that the images {*Ik ij* true} and {*Ik* +1 *ij* true } are strongly correlated and they have similar dynamic ranges *Dk* and *Dk* +1 determined as *Dk* = *Ik* max <sup>−</sup> *Ik* min, where *Ik* max and *Ik* min are maximal and minimal values in *k*-th channel image, respectively. It is also possible to assume that noise is of the same type and neighbor channels have quite close values of input MSEs (equal to noise variance *σ<sup>k</sup>* <sup>2</sup> if the noise is pure additive) as follows:

$$MSE\_k^{\text{loop}} = \sum\_{l=1}^{l} \sum\_{j=1}^{l} (I\_{k\bar{j}}^{\text{noise}} - I\_{k\bar{j}}^{\text{true}})^2 \;/\ (II) \,, k = 1 \dots \text{.} K \tag{2}$$

and input PSNR

$$PSNR\_k^{\text{top}} = 10\log\_{10}(D\_k^2 \;/\, MSE\_k^{\text{top}}), k = 1, \dots, K. \tag{3}$$

The same assumptions are valid for input *PSNR* −*HVS* −*Mk* inp determined similarly to expres‐ sion (3) with the difference that *MSEk* inp is replaced by *MSEHVS <sup>k</sup>* inp , which is a special kind of weighted MSE calculated in spectral (DCT) domain considering the masking effects [38]. The aforementioned assumptions are valid for color red, green, blue (RGB) images [27], multi‐ spectral and hyperspectral RS images [14], [40], dual polarization, and multifrequency radar images [2]. These properties can be effectively exploited in multichannel image preprocessing [39].

After applying a considered filter, one obtains a filtered image {*Ik ij <sup>f</sup>* }, *i* =1, ..., *I*, *j* =1, ..., *J*, *k* =1, ..., *K* that is supposed to be closer to {*Ik ij* true}, *i* =1, ..., *I*, *j* =1, ..., *J*, *k* =1, ..., *K* according to a chosen metric (a quantitative criterion). These output metrics are calculated as

$$MSE\_k^{\text{out}} = \sum\_{l=1}^{l} \sum\_{j=1}^{l} (I\_{k|j}^f - I\_{k|j}^{\text{true}})^2 \;/\ \{II\} \,, k = 1, \ldots, K,\tag{4}$$

$$PSNR\_{k}^{\text{top}} = 10\log\_{10}(D\_{k}^{2} / MSE\_{k}^{\text{out}}), k = 1, \ldots, K. \tag{5}$$

Output *PSNR* −*HVS* −*Mk* out is determined similarly to (5).

Fourth, other criteria of image-processing efficiency, such as classification accuracy, object detectability, etc., are important for the preprocessed RS data. We are unable to predict them, but recent research shows [39] that these criteria are connected with the traditional criteria of image processing. Thus, it is expected that if good values of conventional and HVS metrics are provided due to preprocessing, appropriate classification accuracy and other criteria will be

**2. The considered image-performance criteria and preprocessing**

This chapter considers the following model of an observed multichannel image:

size, and *K* denotes the number of channels. It is also assumed that the images {*Ik ij*

inp noise true 2 1 1

*MSE I I IJ k K*

inp 2 inp

noisy true true ( ), 1,..., , 1,..., , 1,..., *k ij k ij k ij kij I I n I i Ij Jk K* =+ = = = (1)

true in this voxel (3D pixel), *I* and *J* define the image

min are maximal and minimal values in *k*-th channel image,

=- = åå (2)

<sup>10</sup> 10log ( / ), 1,..., . *<sup>k</sup> k k PSNR* = = *D MSE k K* (3)

true} and {*Ik* +1 *ij* true }

<sup>2</sup> if the noise is pure additive)

inp determined similarly to expres‐

inp , which is a special kind of

noisy is *ij* -th sample of noisy (original) *k*-th component of a multichannel image, *nk ij*

denotes the *ij*-th value of the noise in *k*-th component statistic, which is, in general, supposed

are strongly correlated and they have similar dynamic ranges *Dk* and *Dk* +1 determined as

respectively. It is also possible to assume that noise is of the same type and neighbor channels

( ) / ( ), 1,...,

inp is replaced by *MSEHVS <sup>k</sup>*

weighted MSE calculated in spectral (DCT) domain considering the masking effects [38]. The aforementioned assumptions are valid for color red, green, blue (RGB) images [27], multi‐ spectral and hyperspectral RS images [14], [40], dual polarization, and multifrequency radar

attained.

376 Environmental Applications of Remote Sensing

**techniques**

where *Ik ij*

*Dk* = *Ik*

as follows:

max <sup>−</sup> *Ik*

and input PSNR

to be dependent on the true image value *Ik ij*

max and *Ik*

have quite close values of input MSEs (equal to noise variance *σ<sup>k</sup>*

*I J k k ij kij i j*

= =

The same assumptions are valid for input *PSNR* −*HVS* −*Mk*

sion (3) with the difference that *MSEk*

min, where *Ik*

Then, one has to characterize the efficiency of filtering. One way to do this is to use

out inp / , *MSE MSE k k* k= (6)

$$IPSNR\_k = PSNR\_k^{\text{out}} - PSNR\_k^{\text{inp}} \,\, \, \, \, \tag{7}$$

$$\text{IPHVSUM}\_{k} = \text{PSNR-HVS-M}\_{k}^{\text{out}} - \text{PSNR-HVS-M}\_{k}^{\text{inp}}.\tag{8}$$

Small values of the ratio in expression (6) and large values of expressions (7) and (8), both expressed in dB, are evidence in favor of efficient filtering.

Similarly, after lossy compression, one obtains {*Ik ij <sup>c</sup>* }, *i* =1, ..., *I*, *j* =1, ..., *J*, *k* =1, ..., *K*. It is usually supposed that for a larger compression ratio (CR), the quality of compressed image is worse. This is true for lossy compression of noise-free images where more distortions are introduced for a larger CR. However, in lossy compression of noisy images, the situation is specific [41]. Lossy compression results in certain filtering (noise removal) effect under certain conditions. Due to this filtering effect, it is possible that

$$MSE\_k^c = \sum\_{i=1}^l \sum\_{j=1}^l (I\_{kij}^c - I\_{kij}^{\text{true}})^2 \mid \{II\} \_\prime k = 1, \ldots, K \tag{9}$$

(Zemliachenko et al., 2014b).

occurs to be less than *MS Ek* inp. Then, the compression method parameter (quantization step (QS), scaling factor (SF) or bits per pixel (bpp) depending upon a coder used) for which *MS Ek c* falls into global minimum is called optimal operation point (OOP). This parameter is important and needs additional explanation. Fig. 1(a) presents the dependences of occurs to be less than inp *M <sup>k</sup> SE* . Then, the compression method parameter (quantization step (QS), scaling factor (SF) or bits per pixel (bpp) depending upon a coder used) for which *<sup>c</sup> M <sup>k</sup> SE* falls into global minimum is called optimal operation point (OOP). This parameter is important and needs additional explanation. Fig. 1(a) presents the dependences of

1 1

*k k ij kij i j*

Similarly, after lossy compression, one obtains { }, 1,..., , 1,..., , 1,..., *<sup>c</sup>*

removal) effect under certain conditions. Due to this filtering effect, it is possible that

usually supposed that for a larger compression ratio (CR), the quality of compressed image is worse. This is true for lossy compression of noise-free images where more distortions are introduced for a larger CR. However, in lossy compression of noisy images, the situation is specific (Al-Shaykh & Mersereau, 1998). Lossy compression results in certain filtering (noise

true 2

( ) / ( ), 1,...,

(9)

*k ij I i I j Jk K* . It is

$$PSNR\_k^\epsilon = 10\log\_{10}(D\_k^2 \;/\, MSE\_k^\epsilon), k = 1, \dots, K \tag{10}$$

on QS for the lossy DCT-based coder AGU [42] for two known gray-scale test images Airfield (Fig. 1(b)) and Frisco (Fig. 1(c)) corrupted by additive white Gaussian noise (AWGN) with variance *σ* <sup>2</sup> =100. The test image Frisco has a simpler structure – it contains more homogeneous image regions that correspond to sea surface. Due to this, the filtering effect of lossy compres‐ sion is larger and the dependence has an obvious global maximum (i.e., the OOP), according to *PSN R <sup>c</sup>* , since maximum of *PSN R <sup>c</sup>* corresponds to minimum of *MS E <sup>c</sup>* . Formally, there is no OOP for the other test image Airfield, but the dependence *PSN R <sup>c</sup>* (*QS*) has local maximum. Both aforementioned maxima take place for *QS*OOP ≈4*σ*, which is a recommended choice for the coder AGU [43]. on QS for the lossy DCT-based coder AGU (Ponomarenko et al., 2005) for two known grayscale test images Airfield (Fig. 1(b)) and Frisco (Fig. 1(c)) corrupted by additive white Gaussian noise (AWGN) with variance <sup>2</sup> 100 . The test image Frisco has a simpler structure – it contains more homogeneous image regions that correspond to sea surface. Due to this, the filtering effect of lossy compression is larger and the dependence has an obvious global maximum (i.e., the OOP), according to *<sup>c</sup> PSNR* , since maximum of *<sup>c</sup> PSNR* corresponds to minimum of *<sup>c</sup> MSE* . Formally, there is no OOP for the other test image Airfield, but the dependence ( ) *<sup>c</sup> PSNR QS* has local maximum. Both aforementioned maxima take place for OOP *QS* 4, which is a recommended choice for the coder AGU

**Figure 1.** Dependences *PSN R <sup>c</sup>* (*QS*) for the coder AGU (a) and test images Airfield (b) and Frisco (c) corrupted by AWGN with noise variance equal to 100

The lossy compression in the neighborhood of OOP has obvious advantages. Compressed images have high quality, and, at the same time, they have CR considerably larger than for lossless compression [9], [44]. Because of these benefits, the lossy compression of noisy images in the OOP neighborhood is considered. If OOP does not exist, nevertheless, the recommended setting *QS*OOP ≈4*σ* can be considered. If noise is signal dependent and VST is not used, the setting is *QS*OOP ≈4*σ*equiv where *σ*equiv <sup>2</sup> =*MS E* inp. Then, in OOP, one has parameters *MSEk* OOP, *PSN Rk* OOP, *PSNR*-*HVS*-*Mk* OOP and it is possible to determine for them the following metrics (parameters characterizing compression performance):

occurs to be less than *MS Ek*

378 Environmental Applications of Remote Sensing

, since maximum of *PSN R <sup>c</sup>*

maxima take place for OOP *QS* 4

(Zemliachenko et al., 2014b).

Gaussian noise (AWGN) with variance <sup>2</sup>

Airfield, but the dependence ( )

to *PSN R <sup>c</sup>*

the coder AGU [43].

**Figure 1.** Dependences *PSN R <sup>c</sup>*

AWGN with noise variance equal to 100

setting is *QS*OOP ≈4*σ*equiv where *σ*equiv

inp. Then, the compression method parameter (quantization step

( ) / ( ), 1,...,

(9)

<sup>10</sup> 10log ( / ), 1,..., *c c PSNR D MSE k K k kk* (10)

true 2

*MSE I I IJ k K*

occurs to be less than inp *M <sup>k</sup> SE* . Then, the compression method parameter (quantization step (QS), scaling factor (SF) or bits per pixel (bpp) depending upon a coder used) for which *<sup>c</sup> M <sup>k</sup> SE* falls into global minimum is called optimal operation point (OOP). This parameter is

2

on QS for the lossy DCT-based coder AGU (Ponomarenko et al., 2005) for two known grayscale test images Airfield (Fig. 1(b)) and Frisco (Fig. 1(c)) corrupted by additive white

*k kk PSNR D MSE k K* = = (10)

corresponds to minimum of *MS E <sup>c</sup>*

(a) (c)

The lossy compression in the neighborhood of OOP has obvious advantages. Compressed images have high quality, and, at the same time, they have CR considerably larger than for lossless compression [9], [44]. Because of these benefits, the lossy compression of noisy images in the OOP neighborhood is considered. If OOP does not exist, nevertheless, the recommended setting *QS*OOP ≈4*σ* can be considered. If noise is signal dependent and VST is not used, the

(*QS*) for the coder AGU (a) and test images Airfield (b) and Frisco (c) corrupted by

<sup>2</sup> =*MS E* inp. Then, in OOP, one has parameters

structure – it contains more homogeneous image regions that correspond to sea surface. Due to this, the filtering effect of lossy compression is larger and the dependence has an obvious global maximum (i.e., the OOP), according to *<sup>c</sup> PSNR* , since maximum of *<sup>c</sup> PSNR* corresponds to minimum of *<sup>c</sup> MSE* . Formally, there is no OOP for the other test image

*c*

. Formally, there is no

(*QS*) has local maximum.

(b)

100 . The test image Frisco has a simpler

*<sup>c</sup> PSNR QS* has local maximum. Both aforementioned

, which is a recommended choice for the coder AGU

*k ij I i I j Jk K* . It is

(QS), scaling factor (SF) or bits per pixel (bpp) depending upon a coder used) for which *MS Ek*

Similarly, after lossy compression, one obtains { }, 1,..., , 1,..., , 1,..., *<sup>c</sup>*

removal) effect under certain conditions. Due to this filtering effect, it is possible that

usually supposed that for a larger compression ratio (CR), the quality of compressed image is worse. This is true for lossy compression of noise-free images where more distortions are introduced for a larger CR. However, in lossy compression of noisy images, the situation is specific (Al-Shaykh & Mersereau, 1998). Lossy compression results in certain filtering (noise

falls into global minimum is called optimal operation point (OOP). This parameter is important

important and needs additional explanation. Fig. 1(a) presents the dependences of

on QS for the lossy DCT-based coder AGU [42] for two known gray-scale test images Airfield (Fig. 1(b)) and Frisco (Fig. 1(c)) corrupted by additive white Gaussian noise (AWGN) with variance *σ* <sup>2</sup> =100. The test image Frisco has a simpler structure – it contains more homogeneous image regions that correspond to sea surface. Due to this, the filtering effect of lossy compres‐ sion is larger and the dependence has an obvious global maximum (i.e., the OOP), according

Both aforementioned maxima take place for *QS*OOP ≈4*σ*, which is a recommended choice for

2 <sup>10</sup> 10log ( / ), 1,..., *c c*

1 1

*I J c c k k ij kij i j*

and needs additional explanation. Fig. 1(a) presents the dependences of

OOP for the other test image Airfield, but the dependence *PSN R <sup>c</sup>*

$$\kappa = MSE\_k^{\text{COP}} / MSE\_k^{\text{top}} \text{ \,\,\,\,\,\tag{11}$$

$$\text{INSNR}\_{\text{k}} = \text{PSNR}\_{\text{k}}^{\text{OOP}} - \text{PSNR}\_{\text{k}}^{\text{inp}} \prime \tag{12}$$

$$\text{I-IPHVSM}\_{k} = \text{PSNR-HVS-M}\_{k}^{\text{COP}} - \text{PSNR-HVS-M}\_{k}^{\text{inp}} \text{ \tag{13}$$

where *MSEk* OOP / *MSEk* inp <1 and positive *IPSN Rk* or *IPHVSMk* mean that OOP exists according to the corresponding metric.

Certainly, there are also other valuable performance criteria. For image pre-filtering, it is important to know the computational efficiency of the denoising method and how easily it can be implemented, especially onboard. For image lossy compression, it is important to know CR provided and how easily it can be attained. To partly address these issues, the filtering and compression techniques are briefly described.

DCT-based filtering [18], [30] is performed in a block-wise manner, where 8 × 8 pixels are a typically set block size. Filtering can be performed with nonoverlapping, partly overlapping, and fully overlapping blocks. In the latter case, filtering efficiency (expressed in improvement of PSNR (*IPSNR*) or improvement of PSNR-HVS-M (*IPHVSM* )) is the highest but more computations are needed. Nevertheless, the filter is very fast since it is possible to use fast algorithms and to parallelize computations.

There are three main steps in processing: direct 2D DCT in each block; thresholding of DCT coefficients; inverse DCT applied to thresholded DCT coefficients; then, the filtered data from overlapping blocks are aggregated. Within this structure, different variants of thresholding are possible but employing hard thresholding is preferred, where DCT coefficient values remain unchanged if their amplitudes exceed a threshold or are assigned zero values other‐ wise. If one deals with AWGN, the threshold is set fixed as

$$T = \beta \sigma.\tag{14}$$

For spatially uncorrelated signal-dependent noise with *a priori* known or accurately preestimated dependence of local standard deviation on local (block) mean *σ*loc= *f* (*I* ¯ bl), one has to set a locally adaptive threshold:

$$T\_{\rm bl} = \mathcal{J}f(\overline{I}\_{\rm bl}).\tag{15}$$

Finally, for spatially correlated and signal-dependent noise with *a priori* known or properly pre-estimated normalized DCT spectrum *Wqs* norm, *q* =0, ..., 7, *s* =0, ..., 7 , where *qs* are indices of DCT coefficients in blocks [33], the thresholds are locally adaptive and frequency dependent:

$$T\_{\rm bl}(q,s) = \beta f(\overline{I}\_{\rm bl}) \sqrt{\mathcal{W}\_{qs}^{\rm norm}}.\tag{16}$$

In expressions (14–16), *β* is the parameter. Depending upon the image complexity and noise intensity, its optimal value can vary a little [18], but the recommended choices are *β* = 2.6 to provide good filtering according to *IPSNR* and *β* = 2.3 to ensure quasi-optimal denoising according to *IPHVSM* . In further studies, *β* =2.6 will be used. A 3D version of the DCT–based filter [39] performs similarly. The difference is that the blocks are 3D, of size 8 × 8 × *Kgr*, where *Kgr* ≤ *K* denotes a channel group size.

Conventional BM3D [34] is a more sophisticated denoising method. It presumes search for similar patches (blocks), with their joint processing in a 3D manner using DCT and Haar transform, and post-processing stage. This filtering principle, originally designed to cope with AWGN in gray-scale images, has been later adapted to the cases of signal-dependent noise after a proper VST [37], spatially correlated noise [45] and color (three-channel) images corrupted by AWGN [46]. The BM3D and its modifications provide a slightly better perform‐ ance than the corresponding modifications of the conventional DCT-based denoising by the expense of considerably more extensive computations.

The lossy compression technique called AGU [42] is based on DCT in 32 × 32 pixel blocks, a more efficient (compared to JPEG) coding of quantized DCT coefficients and post-processing to remove the blocking artifacts after decompression. This coder is quite simple but slightly more efficient than JPEG 2000) or set partitioning in hierarchical trees (SPIHT) in rate/distortion sense. This coder has 3D version [19] and CR for both 2D and 3D versions is controlled (changed) by QS.

#### **3. Prediction of filtering efficiency**

The main idea of filtering efficiency prediction is the following [17]. Suppose there is some input parameter(s) able to jointly characterize image complexity and noise intensity and also there is some output parameter(s) capable of adequately describing the image denoising efficiency. Assume that there is a rather strict connection between these input and output parameters that allows predicting output value(s) having input value(s).

An additional assumption (and requirement to prediction) is that input parameter(s) have to be calculated easily and quickly enough, faster than denoising itself (otherwise, the prediction becomes useless). If all these assumptions are valid, it becomes possible to determine a predicted output value before starting image filtering and to decide whether it is worth filtering a given image (component) or not. Another decision can relate to setting parameter(s) of a used filter. For example, if a processed image seems to be textural (having high complexity), parameter(s) of a used filter can be adjusted to provide better edge/detail/texture preservation. For example, the parameter *β* for the DCT-based filter can be set equal to 2.3.

Keeping these general principles in mind, we have to address several tasks:

Finally, for spatially correlated and signal-dependent noise with *a priori* known or properly

DCT coefficients in blocks [33], the thresholds are locally adaptive and frequency dependent:

In expressions (14–16), *β* is the parameter. Depending upon the image complexity and noise intensity, its optimal value can vary a little [18], but the recommended choices are *β* = 2.6 to provide good filtering according to *IPSNR* and *β* = 2.3 to ensure quasi-optimal denoising according to *IPHVSM* . In further studies, *β* =2.6 will be used. A 3D version of the DCT–based filter [39] performs similarly. The difference is that the blocks are 3D, of size 8 × 8 × *Kgr*, where

Conventional BM3D [34] is a more sophisticated denoising method. It presumes search for similar patches (blocks), with their joint processing in a 3D manner using DCT and Haar transform, and post-processing stage. This filtering principle, originally designed to cope with AWGN in gray-scale images, has been later adapted to the cases of signal-dependent noise after a proper VST [37], spatially correlated noise [45] and color (three-channel) images corrupted by AWGN [46]. The BM3D and its modifications provide a slightly better perform‐ ance than the corresponding modifications of the conventional DCT-based denoising by the

The lossy compression technique called AGU [42] is based on DCT in 32 × 32 pixel blocks, a more efficient (compared to JPEG) coding of quantized DCT coefficients and post-processing to remove the blocking artifacts after decompression. This coder is quite simple but slightly more efficient than JPEG 2000) or set partitioning in hierarchical trees (SPIHT) in rate/distortion sense. This coder has 3D version [19] and CR for both 2D and 3D versions is controlled

The main idea of filtering efficiency prediction is the following [17]. Suppose there is some input parameter(s) able to jointly characterize image complexity and noise intensity and also there is some output parameter(s) capable of adequately describing the image denoising efficiency. Assume that there is a rather strict connection between these input and output

An additional assumption (and requirement to prediction) is that input parameter(s) have to be calculated easily and quickly enough, faster than denoising itself (otherwise, the prediction becomes useless). If all these assumptions are valid, it becomes possible to determine a predicted output value before starting image filtering and to decide whether it is worth filtering a given image (component) or not. Another decision can relate to setting parameter(s) of a used

parameters that allows predicting output value(s) having input value(s).

bl bl (,) ( ) . *qs T qs f I W* = b

norm

norm, *q* =0, ..., 7, *s* =0, ..., 7 , where *qs* are indices of

(16)

pre-estimated normalized DCT spectrum *Wqs*

380 Environmental Applications of Remote Sensing

*Kgr* ≤ *K* denotes a channel group size.

(changed) by QS.

expense of considerably more extensive computations.

**3. Prediction of filtering efficiency**


These questions are partly answered below and the outcomes obtained in design and per‐ formance analysis of prediction techniques are described. We believe that a partial answer to the second question is the following. The ratio in expression (6) as well as the parameters *IPSN Rk* and *IPHVSMk* (especially if analyzed jointly) are able to provide the initial insights (characterization) of filtering efficiency. Note that expressions (6) and (7) are mutually dependent metrics and *IPSN Rk* =10log10(*MSEk inp* / *MSEk out* ). Thus, they can be used as output parameter(s) at the current stage of research.

#### **3.1. Input and output parameter sets testing and comparison**

Based on the outcomes of the study [18], Abramov et al. in 2013 [17] observed that there is dependence between efficiency of filtering expressed by (6) and simple statistics of DCTcoefficients determined in 8 × 8 blocks. Two probability parameters have been considered. The first one denoted as *P*2*<sup>σ</sup>* is the mean probability that the amplitudes of DCT coefficients are not larger than 2*σ,* where *σ* denotes the standard deviation of additive white Gaussian noise. This parameter originated from analogies with known sigma filter [47]. The second parameter denoted as *P*2.7*<sup>σ</sup>* is the mean probability that the amplitudes of DCT coefficients are larger than 2.7*σ*. Here, there is an obvious analogy with hard thresholding in DCT-based filter, where the recommended *β* = 2.7. At the starting point, Abramov et al., 2013 had no idea on the optimality of input parameters. The objective was just to check whether the prediction is possible, in principle, using a restricted set of test gray-scale images (18) and standard deviations of AWGN (5, 10, 15). The data have been presented as scatterplots, where the *Y*-axis reflects the ratio in expression (6) and *X*-axis corresponds to a considered statistical (input) parameter (either *P2σ* or *P2.7σ*). These scatterplots are represented in Fig. 2. Obviously, the scatterplots' points are clustered well along the fitted lines (for easy fitting, second-order polynomials were used). Interestingly, small *P2σ* and large *P*2.7*<sup>σ</sup>* correspond to complex structure images corrupted by low-intensity noise. In this case, efficiency of image filtering is low (the ratio in expression (6) is close to unity, see Fig. 2). Note that this is in agreement with the theory of filtering [48], [49]. It shows that efficient filtering of textural images is problematic for any existing filters including the most sophisticated nonlocal ones [34].

**Figure 2.** Examples of scatterplots and curve fitting into them for *P*2*<sup>σ</sup>* (a) and *P*2.7*<sup>σ</sup>* (b)

even to conventional DCT-based filtering.

The results of the study conducted by Abramov et al. (2013) have also shown the following. First, quality of fitting has to be characterized quantitatively. For this purpose, the approach of Cameron et al. (1997) works well. It provides the parameter (coefficient of determination) *R*2 that tends to unity for perfectly fitted curves and root mean square error (RMSE) of fitting that should be as small as possible. These parameters are strictly connected with prediction accuracy. For perfectly determined *P*2*<sup>σ</sup>* or *P*2.7*<sup>σ</sup>*, RMSE of fitting directly describes The results of the study conducted in [17] have also shown the following. First, quality of fitting has to be characterized quantitatively. For this purpose, the approach [50] works well. It provides the parameter (coefficient of determination) *R*<sup>2</sup> that tends to unity for perfectly fitted curves and root mean square error (RMSE) of fitting that should be as small as possible. These parameters are strictly connected with prediction accuracy. For perfectly determined *P*2*<sup>σ</sup>* or *P*2.7*<sup>σ</sup>*, RMSE of fitting directly describes the accuracy of prediction.

Fig. 2. Examples of scatterplots and curve fitting into them for *P*2*<sup>σ</sup>* (a) and *P*2.7*<sup>σ</sup>* (b)

the accuracy of prediction. The conclusions drawn in Abramov et al. (2013) can be recalled here. First, the prediction of filtering efficiency for BM3D is less accurate than for the conventional DCT-based filter. This conclusion has been confirmed in later studies. This is associated with the use of two denoising mechanisms (DCT denoising and similar block search with their joint processing), where the latter mechanism has no connection to DCT statistics. Second, although the prediction accuracy for both *P*2*<sup>σ</sup>* and *P*2.7*<sup>σ</sup>* is quite good (*R*2 > 0.9 and RMSE < 1.0), the probability *P*2*<sup>σ</sup>* provides sufficiently better prediction (quality of fitting) than *P*2.7*<sup>σ</sup>*. This shows that the use of other input parameters is possible. Third, different types of fitting functions (polynomials, power and exponential functions) were able to provide approximately the same quality of fitting (for example, the fitted curve in Fig. 2(a) is 2 2 2 -2.63 + 2.15P + 0.38 *P* ; for the BM3D filter, the obtained function of *P*2.70.73 = 1.86*P*2.7 ). Thus, certain reserves in improving the fitting accuracy "are hidden" in choosing an approximating curve and its parameters. Fourth, it has also been shown that the The conclusions drawn in [17] can be recalled here. First, the prediction of filtering efficiency for BM3D is less accurate than for the conventional DCT-based filter. This conclusion has been confirmed in later studies. This is associated with the use of two denoising mechanisms (DCT denoising and similar block search with their joint processing), where the latter mechanism has no connection to DCT statistics. Second, although the prediction accuracy for both *P*2*<sup>σ</sup>* and *P*2.7*σ* is quite good (*R*<sup>2</sup> > 0.9 and RMSE < 1.0), the probability *P*2*<sup>σ</sup>* provides sufficiently better prediction (quality of fitting) than *P*2.7*<sup>σ</sup>*. This shows that the use of other input parameters is possible. Third, different types of fitting functions (polynomials, power and exponential functions) were able to provide approximately the same quality of fitting (for example, the fitted curve in Fig. 2(a) is *κ* = - 2.63*P*2*<sup>σ</sup>* <sup>2</sup> <sup>+</sup> 2.15P2*σ*<sup>+</sup> 0.38 ; for the BM3D filter, the obtained function of *P*2.7*<sup>σ</sup>* is *κ*= 1.86*P*2.7*<sup>σ</sup>* 0.73). Thus, certain reserves in improving the fitting accuracy "are hidden" in choosing an approximating curve and its parameters. Fourth, it has also been shown that the probabilities *P*2*<sup>σ</sup>* and *P*2.7*<sup>σ</sup>* can be determined with appropriate accuracy from analysis of not all possible overlapping blocks but from partly or even nonoverlapping blocks if their total number is not less than 300...500. This additionally accelerates the prediction compared even to conventional DCT-based filtering.

probabilities *P*2*<sup>σ</sup>* and *P*2.7*<sup>σ</sup>* can be determined with appropriate accuracy from analysis of not all possible overlapping blocks but from partly or even nonoverlapping blocks if their total number is not less than 300…500. This additionally accelerates the prediction compared

There are also observations understood later (in two recent years). First, there should be some restrictions imposed on the approximating function. For example, it is clear that the ratio in expression (6) cannot be negative. It is also clear that an approximating (fitting) function should be determined for all possible values of its arguments. Since the probabilities serve as arguments, they can vary from zero to unity. Meanwhile, arguments

is There are also observations understood later (in two recent years). First, there should be some restrictions imposed on the approximating function. For example, it is clear that the ratio in expression (6) cannot be negative. It is also clear that an approximating (fitting) function should be determined for all possible values of its arguments. Since the probabilities serve as argu‐ ments, they can vary from zero to unity. Meanwhile, arguments in both scatterplots in Fig. 2 vary in narrower limits. Besides, it could be good for curve fitting to have point arguments with approximately uniform density.

These requirements have been satisfied by using considerably more test images (including highly textural ones) and a wider set of noise standard deviations (including quite small ones). This has allowed obtaining scatterplot points for small *P*2*<sup>σ</sup>* and large *P*2.7*<sup>σ</sup>*.

Examples of the obtained scatterplots and fitted curves for the DCT-based denoising are shown in Fig. 3. As it is seen, fitting is rather good and coefficient of determination is approximately 0.95 (see the details below). We believe these are already good results that allow practical recommendations. For example, it is clearly seen that there is no reason to carry out filtering if *P*2*<sup>σ</sup>* is smaller than 0.5 since the benefit obtained due to denoising is negligible (approximately 1 dB or less). Prediction itself is carried out as follows. Having the fitted curves obtained in advance as described above, it is needed to calculate *P*2σ or *P*2,7σ for a given image before filtering and to substitute it as argument into the approximating function to calculate a desired metric that characterizes the predicted denoising efficiency. a given image before filtering and to substitute it as argument into the approximating

function to calculate a desired metric that characterizes the predicted denoising efficiency.

Expressions for the obtained approximations for the DCT filter are as follows (we give only **Figure 3.** Scatterplots of *κ* (a) and IPSNR (b) on *P*2σ and the fitted curves

the prediction of *IPHVSM* is worth improving.

Table 1. Goodness of fit (*R*2) of the obtained approximations

is

0.5 0.55 0.6 0.65 0.7 0.75 0.8 0.85 0.9 0.95 0.1

<sup>0</sup> 0.05 0.1 0.15 0.2 0.25 0.3 0.35 0.1

 *P2.7*

that tends to unity for perfectly fitted

The results of the study conducted by Abramov et al. (2013) have also shown the following. First, quality of fitting has to be characterized quantitatively. For this purpose, the approach of Cameron et al. (1997) works well. It provides the parameter (coefficient of determination) *R*2 that tends to unity for perfectly fitted curves and root mean square error (RMSE) of fitting that should be as small as possible. These parameters are strictly connected with prediction accuracy. For perfectly determined *P*2*<sup>σ</sup>* or *P*2.7*<sup>σ</sup>*, RMSE of fitting directly describes

The conclusions drawn in Abramov et al. (2013) can be recalled here. First, the prediction of filtering efficiency for BM3D is less accurate than for the conventional DCT-based filter. This conclusion has been confirmed in later studies. This is associated with the use of two denoising mechanisms (DCT denoising and similar block search with their joint processing), where the latter mechanism has no connection to DCT statistics. Second, although the prediction accuracy for both *P*2*<sup>σ</sup>* and *P*2.7*<sup>σ</sup>* is quite good (*R*2 > 0.9 and RMSE < 1.0), the probability *P*2*<sup>σ</sup>* provides sufficiently better prediction (quality of fitting) than *P*2.7*<sup>σ</sup>*. This shows that the use of other input parameters is possible. Third, different types of fitting functions (polynomials, power and exponential functions) were able to provide approximately the same quality of fitting (for example, the fitted curve in Fig. 2(a) is

> 0.9 and RMSE < 1.0), the probability *P*2*<sup>σ</sup>* provides sufficiently better

0.73). Thus, certain reserves in improving the fitting accuracy "are hidden"

; for the BM3D filter, the obtained function of *P*2.7

). Thus, certain reserves in improving the fitting accuracy "are hidden" in

choosing an approximating curve and its parameters. Fourth, it has also been shown that the probabilities *P*2*<sup>σ</sup>* and *P*2.7*<sup>σ</sup>* can be determined with appropriate accuracy from analysis of not all possible overlapping blocks but from partly or even nonoverlapping blocks if their total number is not less than 300…500. This additionally accelerates the prediction compared

<sup>2</sup> <sup>+</sup> 2.15P2*σ*<sup>+</sup> 0.38 ; for the BM3D filter, the obtained function

There are also observations understood later (in two recent years). First, there should be some restrictions imposed on the approximating function. For example, it is clear that the ratio in expression (6) cannot be negative. It is also clear that an approximating (fitting) function should be determined for all possible values of its arguments. Since the probabilities serve as arguments, they can vary from zero to unity. Meanwhile, arguments

0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9

Fig. 2. Examples of scatterplots and curve fitting into them for *P*2*<sup>σ</sup>* (a) and *P*2.7*<sup>σ</sup>* (b)

(a) (b)

The results of the study conducted in [17] have also shown the following. First, quality of fitting has to be characterized quantitatively. For this purpose, the approach [50] works well. It

curves and root mean square error (RMSE) of fitting that should be as small as possible. These parameters are strictly connected with prediction accuracy. For perfectly determined *P*2*<sup>σ</sup>* or

The conclusions drawn in [17] can be recalled here. First, the prediction of filtering efficiency for BM3D is less accurate than for the conventional DCT-based filter. This conclusion has been confirmed in later studies. This is associated with the use of two denoising mechanisms (DCT denoising and similar block search with their joint processing), where the latter mechanism has no connection to DCT statistics. Second, although the prediction accuracy for both *P*2*<sup>σ</sup>* and

prediction (quality of fitting) than *P*2.7*<sup>σ</sup>*. This shows that the use of other input parameters is possible. Third, different types of fitting functions (polynomials, power and exponential functions) were able to provide approximately the same quality of fitting (for example, the

in choosing an approximating curve and its parameters. Fourth, it has also been shown that the probabilities *P*2*<sup>σ</sup>* and *P*2.7*<sup>σ</sup>* can be determined with appropriate accuracy from analysis of not all possible overlapping blocks but from partly or even nonoverlapping blocks if their total number is not less than 300...500. This additionally accelerates the prediction compared even

 *MSEout / 2*

 *P2*

**Figure 2.** Examples of scatterplots and curve fitting into them for *P*2*<sup>σ</sup>* (a) and *P*2.7*<sup>σ</sup>* (b)

*P*2.7*<sup>σ</sup>*, RMSE of fitting directly describes the accuracy of prediction.

provides the parameter (coefficient of determination) *R*<sup>2</sup>

the accuracy of prediction.

2

fitted curve in Fig. 2(a) is *κ* = - 2.63*P*2*<sup>σ</sup>*

0.73

to conventional DCT-based filtering.

= 1.86*P*2.7

*P*2.7*σ* is quite good (*R*<sup>2</sup>

of *P*2.7*<sup>σ</sup>* is *κ*= 1.86*P*2.7*<sup>σ</sup>*

2 2

 

even to conventional DCT-based filtering.

 -2.63 + 2.15P + 0.38 *P*

0.2

0.3

0.4

 *M SEout / 2*

0.5

0.6

0.7

0.8

382 Environmental Applications of Remote Sensing

the functions of P2 , more details can be found in Rubel & Lukin (2014): 2 = 1.45 + 0.45P + 0.96 *P*2 2 , (17) Expressions for the obtained approximations for the DCT filter are as follows (we give only the functions of P2*σ*, more details can be found in [51]):

<sup>2</sup> 2.08 IPHVSM= 100\*exp - 0.67

The values of *R*2 are presented in Table 1. The analysis confirms that it is better to use *P*2σ than *P*2.7σ. Prediction of *κ* is slightly more accurate than the prediction of *IPSNR.* However,

> Metric *P*2σ *P*2.7σ *Κ* 0.978 0.955 *IPSNR* 0.963 0.935 *IPHVSM* 0.82 0.78

It has been discovered that not only the mean of local (block) estimates of probability *P*2σ is connected with predicted metrics (Rubel & Lukin, 2014), but the other statistical parameters of the distribution of local estimates can also be exploited to improve prediction. The

$$\kappa = -1.45P\_{2\sigma}^2 + 0.45P\_{2\sigma} + 0.96\,,\tag{17}$$

 *P*

2

. (19)

 

$$\text{IPSNNR} = 100 \ast \exp\left(\cdot \left(\frac{P\_{2\sigma} - 1.92}{0.63}\right)^2\right) \tag{18}$$

$$\text{IPHVSM} = 100^{\text{a}} \exp\left(\cdot \left(\frac{P\_{2\sigma} - 2.08}{0.67}\right)^{2}\right) \tag{19}$$

The values of *R*<sup>2</sup> are presented in Table 1. The analysis confirms that it is better to use *P*2σ than *P*2.7σ. Prediction of *κ* is slightly more accurate than the prediction of *IPSNR.* However, the prediction of *IPHVSM* is worth improving.


**Table 1.** Goodness of fit (*R*<sup>2</sup> ) of the obtained approximations

It has been discovered that not only the mean of local (block) estimates of probability *P*2σ is connected with predicted metrics [51], but the other statistical parameters of the distribution of local estimates can also be exploited to improve prediction. The general framework to obtain an estimate of a predicted metric by multiparameter fitting is described by the following formula:

$$\mathbf{Metric}\_{\text{est}} = \mathbf{a}^\* \exp\left(\sum\_{i=1}^n b\_i O\_i(P)\right) \tag{20}$$

where *a* and *bi* are approximation factors, *Oi* . *i* = 1,...,*n*, is some parameter of distribution, *n* defines the number of such parameters. As *Oi* , it is possible to use the distribution mean, median, mode, variance, skewness, and kurtosis. The factors *a* and *bi* , *i* = 1,...,*n* have to be obtained in advance by multidimensional (*n*-dimensional) regression.

The results of using multidimensional regression are presented in Table 2. The abbreviations used are the following: *M* – mean; *Var* – variance; *Med* – median, *Mod* – mode; *K* – kurtosis; *S* – skewness; all calculated for a set of local estimates of probability *P*2σ. The results are given for both considered filters for the metrics *IPSNR* and *IPHVSM.* Only the best sets for *n* from 1 to 5 are presented since the joint use of all considered parameters is less efficient than five input parameters employed together.


**Table 2.** Goodness of the best multiparameter fit for *P*2σ

2

2

(18)

(19)

<sup>2</sup> 1.92 IPSNR= 100 exp - , 0.63

<sup>2</sup> 2.08 IPHVSM= 100\* exp - . 0.67

prediction of *IPHVSM* is worth improving.

384 Environmental Applications of Remote Sensing

**Table 1.** Goodness of fit (*R*<sup>2</sup>

formula:

where *a* and *bi*

*P* s

æ ö æ ö - \* ç ÷ ç ÷ ç ÷ è ø è ø

*P* s

The values of *R*<sup>2</sup> are presented in Table 1. The analysis confirms that it is better to use *P*2σ than *P*2.7σ. Prediction of *κ* is slightly more accurate than the prediction of *IPSNR.* However, the

> **Metric** *P2σ P2.7σ Κ* 0.978 0.955 *IPSNR* 0.963 0.935 *IPHVSM* 0.82 0.78

It has been discovered that not only the mean of local (block) estimates of probability *P*2σ is connected with predicted metrics [51], but the other statistical parameters of the distribution of local estimates can also be exploited to improve prediction. The general framework to obtain an estimate of a predicted metric by multiparameter fitting is described by the following

1

=

The results of using multidimensional regression are presented in Table 2. The abbreviations used are the following: *M* – mean; *Var* – variance; *Med* – median, *Mod* – mode; *K* – kurtosis; *S* – skewness; all calculated for a set of local estimates of probability *P*2σ. The results are given for both considered filters for the metrics *IPSNR* and *IPHVSM.* Only the best sets for *n* from 1

*i i i*

æ ö ç ÷ è ø

*bO P*

å (20)

. *i* = 1,...,*n*, is some parameter of distribution, *n*

, it is possible to use the distribution mean,

, *i* = 1,...,*n* have to be

*n*

Metric = a\* exp ( ) ,

) of the obtained approximations

est

median, mode, variance, skewness, and kurtosis. The factors *a* and *bi*

obtained in advance by multidimensional (*n*-dimensional) regression.

are approximation factors, *Oi*

defines the number of such parameters. As *Oi*

æ ö æ ö - ç ÷ ç ÷ ç ÷ è ø è ø

The conclusions are the following. The use of more input parameters leads to larger (better) *R*2 for both filters and both metrics. The benefit of using several input parameters instead of one is quite small for *IPSNR,* where *R*<sup>2</sup> for one-parameter prediction is already quite high. Meanwhile, for the visual quality metric *IPHVSM,* the improvement is quite large. Interest‐ ingly, the use of median of local estimates instead of the mean considerably improves predic‐ tion (compare the data in Tables 2 and 1) for *IPHVSM* for the DCT-based filter and *P*2σ.


**Table 3.** Coefficient values of the obtained approximations for *P*2σ

More input parameters provide better prediction. At the same time, more time is needed for calculation of input parameters (although their calculation is not difficult). Then, a compromise solution could be the use of the dependence of the type

$$\text{Metric}\_{\text{est}} = \mathbf{a}^\* \exp\{b\_1 \text{mean}(\hat{P}\_{2\sigma \log}) + b\_2 \text{var}(\hat{P}\_{2\sigma \log})\},\tag{21}$$

where *P* ^ <sup>2</sup>*<sup>σ</sup>* loc denotes the local estimates of probabilities obtained in blocks. The approximation coefficients for all cases are presented in Table 3.

The expression (20) is not the only way to combine several input parameters into a joint output. Neural networks (NN) are known to perform this task rather well and to be good approxima‐ tors [52]. This property has been used by us in [53] to make the neural network predict the considered metrics based on multiple input parameters. The obtained results are practically the same as in Table 3. Therefore, there is no need to use a more complex NN approximator instead of expression (20).

A more reasonable solution is to look for better input parameters. Such a study has been conducted in [51]. It has been shown that the probability *P*0.5σ is more informative than *P*2σ, that *P*0.5σ is the mean probability where the magnitudes of DCT coefficients in blocks are smaller than 0.5*σ*. Theoretically, for Gaussian distribution, this probability does not exceed 0.38. Gaussian distribution takes place for DCT coefficients of AWGN. Thus, the mean *P*0.5σ approaches to 0.38 only if a considered image is "very homogeneous" and noise is intensive. This is postulated in further studies.

The obtained results for multiparameter fitting are presented in Table 4. The abbreviations are the same as in Table 2. The first observation is that even for one parameter (mean of local probabilities), the values *R*<sup>2</sup> are sufficiently better than the corresponding values for *P*2σ. Again the results for the BM3D filter are slightly worse than for the DCT-based filter and the results of predicting *IPHVSM* are worse than for predicting *IPSNR.* Again the use of only two input parameters, mean and variance of local estimates, seems to be a good practical choice. Thus, the best parameters of the function (21) are presented for this case in Table 5. Besides, we give an example of scatterplot fitting by 2D surface (function) for two-parameter case of using mean and variance of local estimates of the considered probability for predicting *IPHVSM* (see Fig. 4). Processing of Multichannel Remote-Sensing Images with Prediction of Performance Parameters http://dx.doi.org/10.5772/61853 387


**Table 4.** Goodness of the best multiparameter fit for *P*0.5σ

More input parameters provide better prediction. At the same time, more time is needed for calculation of input parameters (although their calculation is not difficult). Then, a compromise

**Filter Metric** *a b1 b2*

<sup>2</sup>*<sup>σ</sup>* loc denotes the local estimates of probabilities obtained in blocks. The approximation

The expression (20) is not the only way to combine several input parameters into a joint output. Neural networks (NN) are known to perform this task rather well and to be good approxima‐ tors [52]. This property has been used by us in [53] to make the neural network predict the considered metrics based on multiple input parameters. The obtained results are practically the same as in Table 3. Therefore, there is no need to use a more complex NN approximator

A more reasonable solution is to look for better input parameters. Such a study has been conducted in [51]. It has been shown that the probability *P*0.5σ is more informative than *P*2σ, that *P*0.5σ is the mean probability where the magnitudes of DCT coefficients in blocks are smaller than 0.5*σ*. Theoretically, for Gaussian distribution, this probability does not exceed 0.38. Gaussian distribution takes place for DCT coefficients of AWGN. Thus, the mean *P*0.5σ approaches to 0.38 only if a considered image is "very homogeneous" and noise is intensive.

The obtained results for multiparameter fitting are presented in Table 4. The abbreviations are the same as in Table 2. The first observation is that even for one parameter (mean of local

the results for the BM3D filter are slightly worse than for the DCT-based filter and the results of predicting *IPHVSM* are worse than for predicting *IPSNR.* Again the use of only two input parameters, mean and variance of local estimates, seems to be a good practical choice. Thus, the best parameters of the function (21) are presented for this case in Table 5. Besides, we give an example of scatterplot fitting by 2D surface (function) for two-parameter case of using mean and variance of local estimates of the considered probability for predicting *IPHVSM* (see Fig. 4).

 s

*IPSNR* 0.023 6.338 7.459 *IPHVSM* 2.225\*10−4 10.81 37.14

*IPSNR* 0.019 6.591 6.849 *IPHVSM* 5.324\*10−5 12.42 41.36

are sufficiently better than the corresponding values for *P*2σ. Again

+ (21)

est ( <sup>1</sup> 2 loc 2 2 loc ) Metric = a\* exp mean( ) var( ) , ˆ ˆ *b P bP* s

solution could be the use of the dependence of the type

**Table 3.** Coefficient values of the obtained approximations for *P*2σ

coefficients for all cases are presented in Table 3.

where *P* ^

DCT filter

386 Environmental Applications of Remote Sensing

BM3D

instead of expression (20).

This is postulated in further studies.

probabilities), the values *R*<sup>2</sup>


**Table 5.** Approximation coefficients values of obtained approximations for *P*0.5σ

**Figure 4.** Scatterplot of *IPHVSM* for the DCT-based filter efficiency on statistics of *P*0.5σ and the fitted surface

#### **3.2. Analysis for signal-dependent and spatially correlated types of noise**

Let us define the models of signal-dependent noise used. According to a first model [7], [11], the expression (1) transforms to

$$I\_{\rm kj}^{\rm noise} = I\_{\rm kj}^{\rm true} + \mathcal{N}\_{\rm kj}^{\rm SI} + \mathcal{N}\_{\rm kj}^{\rm SD} \, \prime \, \tag{22}$$

where *Nkij SI* , *Nkij SD* denote signal-independent (SI) and signal-dependent (SD) noise components. Both the noise components in expression (22) are assumed zero mean, spatially uncorrelated and Gaussian. Then, the model for the noise variance is *σkij* <sup>2</sup> <sup>=</sup>*σ<sup>k</sup>* <sup>2</sup> <sup>+</sup> *<sup>γ</sup>Ikij* true, where *σ<sup>k</sup>* <sup>2</sup> is the SI noise variance and *γ* is the SD noise parameter (which is usually between zero and unity). A second model [2] presumes purely multiplicative noise with *Ikij* noise= *Ik ij* true*µkij* , where *µkij* denotes unity mean random factor with variance *σµk* <sup>2</sup> that is within the limits from 0 to 1. It is supposed for both the models that the noise is spatially uncorrelated.

As mentioned in Section 2 (expression no. 15), the local threshold is set as *T*bl=*β σ*<sup>0</sup> <sup>2</sup> + *γI* ¯ bl for signal-dependent noise (expression no. 22) and as *T*bl=*βσµI* ¯ bl for pure multiplicative noise. In addition to modifying the filtering algorithm, we need to modify the algorithm of input parameter calculation. Then, the local probability estimate has to consider the local variation of noise standard deviation. For instance, the local estimate of probability *P*2σ is obtained as

$$\hat{P}\_{2\sigma}^{\text{bl}} = \sum\_{q=0}^{7} \sum\_{s=0}^{7} \delta\_{qs} / \text{63}\_{\text{\textdegree}} \tag{23}$$

where *δqs* =1, *if* | *Dqs* | ≤2*σ*bl and 0 otherwise (*σbl* is equal to *σ*<sup>0</sup> <sup>2</sup> + *γI* ¯ bl or to *σµI* ¯ bl depending upon a model used). DC component of DCT coefficients in blocks is not taken into account as it always exceeds the local threshold.

Some of the results of studies in our papers [54], [55] are presented next. One aspect that was specially addressed in these studies was to check the influence of an image set used in forming a scatterplot. In fact, two scatterplots have been formed separately: for the set of standard images used in optical image processing as Baboon, Barbara, Lena, etc., and for the set of images called "Remote Sensing" as Frisco, Diego, etc. The reason for such study was the following fact. Some people from RS community are categorically against using standard grayscale test images in their studies although there are no commonly accepted sets of test RS images.

The methodology of obtaining scatterplot was modified a little. For the noise expression model (22), three different cases were modeled: prevailing influence of SI noise, dominant influence of SD noise, and comparable contribution of both components. As a result, a wide range of mean *P*2σ has been provided. Scatterplot points that belong to different image sets are indicated by different signs (and different colors). There are also two fitted curves. We believe there is no essential difference between the scatterplots and fitted curves. Thus, it can be concluded that the prediction is quite universal and suitable for conventional gray-scale optical images and component-wise (single-channel) RS images. Moreover, it has been shown in a study [55] that prediction is valid for single-look SAR images corrupted by fully developed spatially uncorrelated speckle. It is also possible to compare the results in Fig. 5 with the data in Fig. 3(b). They are very similar. Fig. 4 shows that *IPSNR* is approximately 1 dB or less for *P*2σ approximately 0.5 and then denoising is practically useless. Meanwhile, if *IPSNR* is approxi‐ mately 4 dB for *P*2σ approximately 0.8, then the use of filtering is expedient. The parameter *R*<sup>2</sup> for both fitting curves in Fig. 5 is approximately 0.96, that is, the prediction is approximately as good as for AWGN case. Again, the results for *P*2σ are better than for *P*2.7σ; fitting for *IPSNR* is more accurate than for *IPHVSM*. Improved fitting by means of using multiple input parameters has not been investigated yet.

**Figure 4.** Scatterplot of *IPHVSM* for the DCT-based filter efficiency on statistics of *P*0.5σ and the fitted surface

Let us define the models of signal-dependent noise used. According to a first model [7], [11],

Both the noise components in expression (22) are assumed zero mean, spatially uncorrelated

variance and *γ* is the SD noise parameter (which is usually between zero and unity). A second

addition to modifying the filtering algorithm, we need to modify the algorithm of input parameter calculation. Then, the local probability estimate has to consider the local variation of noise standard deviation. For instance, the local estimate of probability *P*2σ is obtained as

As mentioned in Section 2 (expression no. 15), the local threshold is set as *T*bl=*β σ*<sup>0</sup>

7 7 bl 2

*P* s

0 0 <sup>ˆ</sup> / 63, *qs q s*

d = =

a model used). DC component of DCT coefficients in blocks is not taken into account as it

is equal to *σ*<sup>0</sup>

*kij kij kij kij I I NN* =++ (22)

<sup>2</sup> <sup>+</sup> *<sup>γ</sup>Ikij*

true*µkij*

<sup>2</sup> that is within the limits from 0 to 1. It is supposed for

<sup>=</sup> åå (23)

bl or to *σµI* ¯

<sup>2</sup> + *γI* ¯ true, where *σ<sup>k</sup>*

, where *µkij*

bl for pure multiplicative noise. In

<sup>2</sup> is the SI noise

denotes unity

<sup>2</sup> + *γI* ¯ bl for

bl depending upon

<sup>2</sup> <sup>=</sup>*σ<sup>k</sup>*

noise= *Ik ij*

¯

*SD* denote signal-independent (SI) and signal-dependent (SD) noise components.

noise true , *SI SD*

**3.2. Analysis for signal-dependent and spatially correlated types of noise**

and Gaussian. Then, the model for the noise variance is *σkij*

model [2] presumes purely multiplicative noise with *Ikij*

both the models that the noise is spatially uncorrelated.

signal-dependent noise (expression no. 22) and as *T*bl=*βσµI*

where *δqs* =1, *if* | *Dqs* | ≤2*σ*bl and 0 otherwise (*σbl*

always exceeds the local threshold.

mean random factor with variance *σµk*

the expression (1) transforms to

388 Environmental Applications of Remote Sensing

where *Nkij SI* , *Nkij*

**Figure 5.** Scatterplots of *IPSNR* for the DCT-based filter efficiency on statistics of *P*2σ

Two examples of image processing are presented here. Fig. 6(a) represents the noisy image Frisco, where noise parameters are σ<sup>0</sup> 2 =100; *σ* <sup>2</sup> =100, and *γ* =0.2. The output image for the DCTbased filter is presented in Fig. 6(b). The effect of denoising is obvious. Actual provided improvement of PSNR is equal to 9.77 dB. The predicted value for mean P2σ=0.92 is approxi‐ mately 9.5 dB (see the blue fitted curve in Fig. 5), that is, there is good agreement of attained and predicted values. Prediction shows that it is worth applying denoising in this case.

For a real-life data, it is impossible to determine true values of the considered metrics charac‐ terizing filtering efficiency. However, it is possible to analyze the predicted values and denoising results visually. For fragments of sub-band images of hyperspectral sensor, Hyper‐ ion, such analysis was done. For example, noise parameters of the expression model (22) have been blindly estimated [11]. The noisy image for the 13th sub-band of the set EO1H1800252002116110KZ is depicted in Fig. 7(a). Noise is clearly seen. The prediction of *IPSNR* is approximately 8.5 dB and *IPHVSM* is approximately 5.7 dB. Thus, it is expedient to perform denoising. The denoised image is presented in Fig. 7(b). As can be seen, its quality has very much improved due to filtering.

**Figure 6.** Noisy (a) and output (b) images Frisco

Fig. 6. Noisy (a) and output (b) images Frisco

The sub-bands 13...22 are considered for two sets of Hyperion data. The values *IPSNR* are always larger than *IPHVSM*. This means that it is harder to provide an improvement of image visual quality than to gain improvement according to standard metrics (MSE, PSNR). For the sub-bands with indices *k* = 13...16, *IPSNR* is always larger than 1.6 dB and *IPHVSM* exceeds 0.6 dB, that is, filtering is desirable. For other sub-bands, as the predicted improvements are small, it is doubtful whether it is worth carrying out filtering. Visual inspection of images in sub-bands with *k* = 17...22 has shown that noise is either hardly noticeable or practically invisible. Positive effect of its removal is partly or fully compensated by edge/detail/texture smearing performed by any filter, even the most sophisticated one [56]. The texture filtering is always problematic and the prediction approach is able to reliably predict this [56].

(a) (b)

Fig. 7. Noisy (a) and output (b) images of 13th sub-band images of Hyperion sensor

Fig. 7. Noisy (a) and output (b) images of 13th sub-band images of Hyperion sensor **Figure 7.** Noisy (a) and output (b) images of 13th sub-band images of Hyperion sensor

Fig. 6. Noisy (a) and output (b) images Frisco

Two examples of image processing are presented here. Fig. 6(a) represents the noisy image

based filter is presented in Fig. 6(b). The effect of denoising is obvious. Actual provided improvement of PSNR is equal to 9.77 dB. The predicted value for mean P2σ=0.92 is approxi‐ mately 9.5 dB (see the blue fitted curve in Fig. 5), that is, there is good agreement of attained and predicted values. Prediction shows that it is worth applying denoising in this case.

For a real-life data, it is impossible to determine true values of the considered metrics charac‐ terizing filtering efficiency. However, it is possible to analyze the predicted values and denoising results visually. For fragments of sub-band images of hyperspectral sensor, Hyper‐ ion, such analysis was done. For example, noise parameters of the expression model (22) have been blindly estimated [11]. The noisy image for the 13th sub-band of the set EO1H1800252002116110KZ is depicted in Fig. 7(a). Noise is clearly seen. The prediction of *IPSNR* is approximately 8.5 dB and *IPHVSM* is approximately 5.7 dB. Thus, it is expedient to perform denoising. The denoised image is presented in Fig. 7(b). As can be seen, its quality

(a) (b)

The sub-bands 13...22 are considered for two sets of Hyperion data. The values *IPSNR* are always larger than *IPHVSM*. This means that it is harder to provide an improvement of image visual quality than to gain improvement according to standard metrics (MSE, PSNR). For the sub-bands with indices *k* = 13...16, *IPSNR* is always larger than 1.6 dB and *IPHVSM* exceeds 0.6 dB, that is, filtering is desirable. For other sub-bands, as the predicted improvements are small, it is doubtful whether it is worth carrying out filtering. Visual inspection of images in sub-bands with *k* = 17...22 has shown that noise is either hardly noticeable or practically invisible. Positive effect of its removal is partly or fully compensated by edge/detail/texture smearing performed by any filter, even the most sophisticated one [56]. The texture filtering

(a) (b)

is always problematic and the prediction approach is able to reliably predict this [56].

Fig. 7. Noisy (a) and output (b) images of 13th sub-band images of Hyperion sensor

=100; *σ* <sup>2</sup> =100, and *γ* =0.2. The output image for the DCT-

2

Frisco, where noise parameters are σ<sup>0</sup>

390 Environmental Applications of Remote Sensing

has very much improved due to filtering.

Fig. 6. Noisy (a) and output (b) images Frisco

**Figure 6.** Noisy (a) and output (b) images Frisco

Considering certain benefits achieved due to using *P*0.5*<sup>σ</sup>* as input parameter, the analysis similar to the one presented in Fig. 5 has been performed. The results are presented in Fig. 8. The noise is signal-dependent and most scatterplot points correspond to the expression model (22). The curve is fitted employing all points (although they relate to optical and RS subsets). Obviously, fitting is very good and, according to quantitative criteria, it is better than for the parameter *P*2*<sup>σ</sup>* (Fig. 5). Four black points at the scatterplot in Fig. 8 correspond to one-look SAR images. They fit the curve well and have the arguments close to the maximal potential limit (0.38), where *IPSNR* attains very large values (approximately 10 dB and more).

**Figure 8.** Scatterplots of *IPSNR* for the DCT-based filter efficiency on the statistics of *P*0.5σ (two sets of images and two fitted curves)

Additional studies concentrated on the multi-look SAR images that were corrupted by pure multiplicative noise [57]. Analysis has been done for speckle variance *σ<sup>µ</sup>* <sup>2</sup> =0.273 / *L* , where *L* denotes the number of looks. Scatterplot points are presented in Fig. 9 for different number of looks. An obvious tendency is that mean *P*0.5*<sup>σ</sup>* becomes larger and *IPSNR* increases for smaller number of looks. Other conclusions that can be drawn from analysis in a study in [57] are the following. Prediction is possible for filtering techniques with and without VST, where the prediction quality is better in the latter case. Prediction using different types of functions (polynomial, power, exponential) produce fitting of approximately equal accuracy. Mean‐ while, accuracy of prediction is worth improving (RMSE is approximately 1 dB) since it is sufficiently worse than for the case of AWGN.

Understanding that, in practice, noise can be spatially correlated [33], the case of spatially correlated noise – additive in [45] and multiplicative in [57] – are also studied. A difficulty of dealing with spatially correlated noise is that there are numerous shapes (and parameter sets) of 2D auto-correlation function or spatial spectrum of such a noise. Thus, studying a particular case of spatially correlated noise gives only limited information on general dependences. Hence, two models of spatially correlated noise (called middle correlation and strong corre‐ lation) have been considered [45]. A peculiarity of prediction is that the local estimate of probability *P*2σ is obtained according to expression (23), where, in the general case, *δqs* =1, *if* | *Dqs* | ≤2*σ*bl(*W* (*q*, *s*)) 1/2 and 0 otherwise (*σbl* is the local standard deviation in a considered block; expressions for its derivation depending upon noise model are given above). If the probability *P*0.5σ is used, the condition is *δqs* =1, *if* | *Dqs* | ≤0.5*σ*bl(*W* (*q*, *s*)) 1/2 and 0 otherwise.

**Figure 9.** Scatterplot for *IPSNR* vs. mean *P*0.5*<sup>σ</sup>* for a part of test images corrupted by spatially uncorrelated speckle

The scatterplots and fitted curves are presented in Fig. 10. The fitted curves are similar and they clearly show that there is no reason to filter images if *P*0.5*<sup>σ</sup>* is smaller than 0.15. The difference in the scatterplots for *IPHVSM* and *IPSNR* is that the latter one is more compact and, thus, *IPSNR* can be predicted more accurately. An additional distinctive feature of the plot for

Processing of Multichannel Remote-Sensing Images with Prediction of Performance Parameters http://dx.doi.org/10.5772/61853 393

The scatterplots and fitted curves are presented in Fig. 10. The fitted curves are similar and they clearly show that there is no reason to filter images if *P*0.5*<sup>σ</sup>* is smaller than 0.15. The

Fig. 10. The scatterplots for middle-correlation noise and the fitted curves for *IPHVSM* (a)

and *IPSNR* (b) **Figure 10.** The scatterplots for middle-correlation noise and the fitted curves for *IPHVSM* (a) and *IPSNR* (b) filtering similar to expression (8) have been used. In addition, instead of BM3D, its color version called C-BM3D has been analyzed (Dabov et al., 2007b).

differences in prediction. First, all DCT coefficients in 3D block are subject to analysis for estimating the local probabilities. Second, the metric PSNR-HMA (Ponomarenko et al., The scatterplots have been obtained and curves were fitted to them (see examples in Fig. 11). **Figure 11.** The scatterplots and the fitted curves for *IPSNR* (a) and *IPHVSMA* (b)

Additional studies concentrated on the multi-look SAR images that were corrupted by pure

denotes the number of looks. Scatterplot points are presented in Fig. 9 for different number of looks. An obvious tendency is that mean *P*0.5*<sup>σ</sup>* becomes larger and *IPSNR* increases for smaller number of looks. Other conclusions that can be drawn from analysis in a study in [57] are the following. Prediction is possible for filtering techniques with and without VST, where the prediction quality is better in the latter case. Prediction using different types of functions (polynomial, power, exponential) produce fitting of approximately equal accuracy. Mean‐ while, accuracy of prediction is worth improving (RMSE is approximately 1 dB) since it is

Understanding that, in practice, noise can be spatially correlated [33], the case of spatially correlated noise – additive in [45] and multiplicative in [57] – are also studied. A difficulty of dealing with spatially correlated noise is that there are numerous shapes (and parameter sets) of 2D auto-correlation function or spatial spectrum of such a noise. Thus, studying a particular case of spatially correlated noise gives only limited information on general dependences. Hence, two models of spatially correlated noise (called middle correlation and strong corre‐ lation) have been considered [45]. A peculiarity of prediction is that the local estimate of probability *P*2σ is obtained according to expression (23), where, in the general case,

and 0 otherwise (*σbl*

If the probability *P*0.5σ is used, the condition is *δqs* =1, *if* | *Dqs* | ≤0.5*σ*bl(*W* (*q*, *s*))

considered block; expressions for its derivation depending upon noise model are given above).

**Figure 9.** Scatterplot for *IPSNR* vs. mean *P*0.5*<sup>σ</sup>* for a part of test images corrupted by spatially uncorrelated speckle

The scatterplots and fitted curves are presented in Fig. 10. The fitted curves are similar and they clearly show that there is no reason to filter images if *P*0.5*<sup>σ</sup>* is smaller than 0.15. The difference in the scatterplots for *IPHVSM* and *IPSNR* is that the latter one is more compact and, thus, *IPSNR* can be predicted more accurately. An additional distinctive feature of the plot for

<sup>2</sup> =0.273 / *L* , where *L*

is the local standard deviation in a

1/2

and 0 otherwise.

multiplicative noise [57]. Analysis has been done for speckle variance *σ<sup>µ</sup>*

sufficiently worse than for the case of AWGN.

392 Environmental Applications of Remote Sensing

1/2

*δqs* =1, *if* | *Dqs* | ≤2*σ*bl(*W* (*q*, *s*))

*IPSNR* is that its maximal values are smaller than for AWGN case (data in Fig. 3(b)). The scatterplots for a strong correlation of the noise and the conclusions derived from them are similar. filtering similar to expression (8) have been used. In addition, instead of BM3D, its color version called C-BM3D has been analyzed (Dabov et al., 2007b). for highly textured images when noise standard deviation is small). Another observation is the same as earlier – visual quality can be predicted worse than *IPSNR.* The prediction accuracy for C-BM3D is worse than for 3D DCT filter.

Fig. 11. The scatterplots and the fitted curves for *IPSNR* (a) and *IPHVSMA* (b)

2011), which is a color extension of PSNR-HVS-M, and improvement of this metric due to

As mentioned earlier, filtering is useless for *P*0.5*<sup>σ</sup>* < 0.15. However, this happens rarely (only

Taking into account our previous experience, the multiparameter input was analyzed with

We have also studied the case of spatially correlated speckle [57]. It has been shown that the prediction seems possible for a spatially correlated noise. However, more research is needed to understand how to select a parameter or several parameters to characterize spatial correla‐ tion and how it can be involved in prediction. exponential function expressed in (20). Considerable improvement has been reached, especially for *IPHVSMA,* for the 3D DCT filter. For the C-BM3D filter, the positive effect is less. One has *IPHVSMA* equal to 0.8481 for one input parameter and 0.8555 for four parameters. Again, a reasonable practical solution is to use the mean and variance of local

Finally, a preliminary research has been carried out for denoising color images corrupted by AWGN with equal variance values in channels [58]. There are two differences in prediction. First, all DCT coefficients in 3D block are subject to analysis for estimating the local probabil‐ ities. Second, the metric PSNR-HMA [59], which is a color extension of PSNR-HVS-M, and improvement of this metric due to filtering similar to expression (8) have been used. In addition, instead of BM3D, its color version called C-BM3D has been analyzed [46].

The scatterplots have been obtained and curves were fitted to them (see examples in Fig. 11). As mentioned earlier, filtering is useless for *P*0.5*<sup>σ</sup>* < 0.15. However, this happens rarely (only for highly textured images when noise standard deviation is small). Another observation is the same as earlier – visual quality can be predicted worse than *IPSNR.* The prediction accuracy for C-BM3D is worse than for 3D DCT filter.

Taking into account our previous experience, the multiparameter input was analyzed with exponential function expressed in (20). Considerable improvement has been reached, espe‐ cially for *IPHVSMA,* for the 3D DCT filter. For the C-BM3D filter, the positive effect is less. One has R2 equal to 0.8481 for one input parameter and 0.8555 for four parameters. Again, a reasonable practical solution is to use the mean and variance of local estimates of probability. One more important observation for color image filtering is that *P*0.5*<sup>σ</sup>* for 3D filter is larger than for DCT filter applied to components of a processed color image. This again proves that 3D processing of color and multichannel images iiis are potentially more efficient compared to their component-wise denoising.

#### **4. Prediction in lossy compression of noisy images**

In this section, the compression of images corrupted by AWGN is considered. Lossy com‐ pression is carried out by the aforementioned coder AGU with *QS* =4*σ*. In this case, OOP may exist or be absent. The task is to predict *IPSNR* and *IPHVSM* and to decide whether OOP exists as well as to predict what CR is.

#### **4.1. Prediction of OOP existence and metrics' values in it**

This section shortly describes how the scatterplots were obtained. As in the filtering case, a set of gray-scale test images of different content and complexity was used. AWGN of different intensity has been added and then the obtained images have been compressed by AGU. After this, the parameters (12) and (13) have been calculated as well as *P*2σ for each compressed image. Clearly, all these actions are done off-line before applying the prediction approach in practice.

The obtained scatterplot is presented in Fig. 12. A specific feature of this scatterplot is that it has negative values and they seem to be approximately −3.5 dB for *P*2σ approaching to zero. Therefore, not all fitting functions can be used. The study carried out by Zemliachenko et al. in [44] has shown that the polynomials of the fourth and fifth order usually allow approxi‐ mating the dependence very well (with *R*<sup>2</sup> almost equal to unity and RMSE approximately 0.25 for *IPSNR*). As can be seen from the analysis of the scatterplot in Fig. 12, there are quite many images and/or noise variances when OOP does not exist (*IPSNR* is negative). OOP exists with high probability if *P*2σ exceeds 0.82. This can be used as a basis for predicting OOP existence.

The scatterplot for the metric *IPHVSM* is presented in Fig. 13. In some sense, behavior of the fitted polynomial is similar to the one in Fig. 12. There are many values about −4 dB showing that due to lossy compression the visual quality becomes worse. However, this mainly happens for small *P*2σ that corresponds to high-complexity images and/or low level of the noise. The visual quality improves for *P*2σ exceeding 0.9 and this takes place for low-complexity images and rather intensive noise.

mainly happens for small *P*2σ that correspond to high-complexity images and/or low level

estimates of probability. One more important observation for color image filtering is that *P*0.5*<sup>σ</sup>* for 3D filter is larger than for DCT filter applied to components of a processed color image. This again proves that 3D processing of color and multichannel images iiis are

In this section, the compression of images corrupted by AWGN is considered. Lossy

OOP may exist or be absent. The task is to predict *IPSNR* and *IPHVSM* and to decide

This section shortly describes how the scatterplots were obtained. As in the filtering case, a set of gray-scale test images of different content and complexity was used. AWGN of different intensity has been added and then the obtained images have been compressed by AGU. After this, the parameters (12) and (13) have been calculated as well as *P*2<sup>σ</sup> for each compressed image. Clearly, all these actions are done off-line before applying the prediction

The obtained scatterplot is presented in Fig. 12. A specific feature of this scatterplot is that it has negative values and they seem to be approximately −3.5 dB for *P*2σ approaching to zero. Therefore, not all fitting functions can be used. The study carried out by Zemliachenko et al. in [44] has shown that the polynomials of the fourth and fifth order usually allow approximating the dependence very well (with *R*2 almost equal to unity and RMSE approximately 0.25 for *IPSNR*). As can be seen from the analysis of the scatterplot in Fig. 12, there are quite many images and/or noise variances when OOP does not exist (*IPSNR* is

. In this case,

potentially more efficient compared to their component-wise denoising.

compression is carried out by the aforementioned coder AGU with *QS* 4

**4. Prediction in lossy compression of noisy images**

**4.1 Prediction of OOP existence and metrics' values in it**

whether OOP exists as well as to predict what CR is.

approach in practice.

for predicting OOP existence.

Fig. 12. The scatterplot and the fitted curves for **Figure 12.** The scatterplot and the fitted curves for *IPSNR* and the coder AGU *IPSNR* and the coder AGU of the noise. The visual quality improves for *P*2σ exceeding 0.9 and this takes place for low-

complexity images and rather intensive noise.

same as earlier – visual quality can be predicted worse than *IPSNR.* The prediction accuracy

Taking into account our previous experience, the multiparameter input was analyzed with exponential function expressed in (20). Considerable improvement has been reached, espe‐ cially for *IPHVSMA,* for the 3D DCT filter. For the C-BM3D filter, the positive effect is less. One has R2 equal to 0.8481 for one input parameter and 0.8555 for four parameters. Again, a reasonable practical solution is to use the mean and variance of local estimates of probability. One more important observation for color image filtering is that *P*0.5*<sup>σ</sup>* for 3D filter is larger than for DCT filter applied to components of a processed color image. This again proves that 3D processing of color and multichannel images iiis are potentially more efficient compared to

In this section, the compression of images corrupted by AWGN is considered. Lossy com‐ pression is carried out by the aforementioned coder AGU with *QS* =4*σ*. In this case, OOP may exist or be absent. The task is to predict *IPSNR* and *IPHVSM* and to decide whether OOP exists

This section shortly describes how the scatterplots were obtained. As in the filtering case, a set of gray-scale test images of different content and complexity was used. AWGN of different intensity has been added and then the obtained images have been compressed by AGU. After this, the parameters (12) and (13) have been calculated as well as *P*2σ for each compressed image. Clearly, all these actions are done off-line before applying the prediction approach in practice.

The obtained scatterplot is presented in Fig. 12. A specific feature of this scatterplot is that it has negative values and they seem to be approximately −3.5 dB for *P*2σ approaching to zero. Therefore, not all fitting functions can be used. The study carried out by Zemliachenko et al. in [44] has shown that the polynomials of the fourth and fifth order usually allow approxi‐

for *IPSNR*). As can be seen from the analysis of the scatterplot in Fig. 12, there are quite many images and/or noise variances when OOP does not exist (*IPSNR* is negative). OOP exists with high probability if *P*2σ exceeds 0.82. This can be used as a basis for predicting OOP existence.

The scatterplot for the metric *IPHVSM* is presented in Fig. 13. In some sense, behavior of the fitted polynomial is similar to the one in Fig. 12. There are many values about −4 dB showing that due to lossy compression the visual quality becomes worse. However, this mainly happens for small *P*2σ that corresponds to high-complexity images and/or low level of the noise. The visual quality improves for *P*2σ exceeding 0.9 and this takes place for low-complexity images

almost equal to unity and RMSE approximately 0.25

for C-BM3D is worse than for 3D DCT filter.

394 Environmental Applications of Remote Sensing

their component-wise denoising.

as well as to predict what CR is.

mating the dependence very well (with *R*<sup>2</sup>

and rather intensive noise.

**4. Prediction in lossy compression of noisy images**

**4.1. Prediction of OOP existence and metrics' values in it**

Although prediction has been studied by simulations only for images corrupted by AWGN, it can also be applied to images corrupted by a signal-dependent spatially uncorrelated noise under condition that a proper VST is applied to them before compressing. Such VST (a generalized Anscombe transform in this case) provides approximately constant noise variance that usually equals to unity. Thus, QS = 4 is used. This approach has been used for Hyperion data and the results are presented in Fig. 14. There are two groups of sub-bands that are usually not analyzed in Hyperion data since they are too noisy. Thus, the prediction values are not given for all sub-bands. Analysis of the presented values shows that there are only a few sub-bands where it is worth expecting OOP. For most other sub-bands, *IPSNR* is about −3 dB and the ways of dealing with them are considered in a study [44]. One

Fig. 15 shows the original and the decompressed images in 110-th sub-band, where decrease of visual quality according to quantitative criteria is predicted. Noise is not seen in the original image and the compression practically does not influence the image quality (in our

Fig. 13. The scatterplot and the fitted curves for *IPHVSM* and the coder AGU **Figure 13.** The scatterplot and the fitted curves for *IPHVSM* and the coder AGU

proposition is to set less QS but this leads to smaller CR.

opinion, both images look the same).

Although prediction has been studied by simulations only for images corrupted by AWGN, it can also be applied to images corrupted by a signal-dependent spatially uncorrelated noise under condition that a proper VST is applied to them before compressing. Such VST (a generalized Anscombe transform in this case) provides approximately constant noise variance that usually equals to unity. Thus, QS = 4 is used. This approach has been used for Hyperion data and the results are presented in Fig. 14. There are two groups of sub-bands that are usually not analyzed in Hyperion data since they are too noisy. Thus, the prediction values are not given for all sub-bands. Analysis of the presented values shows that there are only a few subbands where it is worth expecting OOP. For most other sub-bands, *IPSNR* is about −3 dB and the ways of dealing with them are considered in a study [44]. One proposition is to set less QS but this leads to smaller CR.

Fig. 15 shows the original and the decompressed images in 110-th sub-band, where decrease of visual quality according to quantitative criteria is predicted. Noise is not seen in the original image and the compression practically does not influence the image quality (in our opinion, both images look the same).

**Figure 14.** Predicted *IPSNR* for Hyperion hyperspectral image compressed by the coder AGU

Fig. 15. The 110-th sub-band images before (a) and after (b) compression

moreover, are even less efficient in many practical situations.

**4.2 Prediction of compression ratio in OOP**

A study [44] also presents data for three other DCT-based coders, where two of them are specially suited for providing better visual quality. It is demonstrated that the coder adaptive DCT (ADCT), which exploits the optimized partition schemes [60], provides certain improve‐ ments compared to AGU. Meanwhile, DCT coders oriented on improving the visual quality being applied to noisy images do not offer substantial benefits and, moreover, are even less efficient in many practical situations.

(a) (b)

A study [44] also presents data for three other DCT-based coders, where two of them are specially suited for providing better visual quality. It is demonstrated that the coder adaptive DCT (ADCT), which exploits the optimized partition schemes [60], provides certain improvements compared to AGU. Meanwhile, DCT coders oriented on improving the visual quality being applied to noisy images do not offer substantial benefits and,

The methodology of predicting CR in OOP is the same as that for filtering. It is based on the scatterplot obtaining and curve fitting. The only difference is that the vertical axis relates to CR, while the horizontal axis, as earlier, corresponds to mean probability. Two mean probabilities *P2*σ and *P*2.7σ have been considered where the latter occurred to be worse again.

Therefore, the obtained results for the mean probability *P*2σ only are presented below.

Processing of Multichannel Remote-Sensing Images with Prediction of Performance Parameters http://dx.doi.org/10.5772/61853 397

**Figure 15.** The 110-th sub-band images before (a) and after (b) compression

#### **4.2. Prediction of compression ratio in OOP** A study by Zemliachenko et al. (2015b) also presents data for three other DCT-based coders,

noise.

additive with unity variance.

Although prediction has been studied by simulations only for images corrupted by AWGN, it can also be applied to images corrupted by a signal-dependent spatially uncorrelated noise under condition that a proper VST is applied to them before compressing. Such VST (a generalized Anscombe transform in this case) provides approximately constant noise variance that usually equals to unity. Thus, QS = 4 is used. This approach has been used for Hyperion data and the results are presented in Fig. 14. There are two groups of sub-bands that are usually not analyzed in Hyperion data since they are too noisy. Thus, the prediction values are not given for all sub-bands. Analysis of the presented values shows that there are only a few subbands where it is worth expecting OOP. For most other sub-bands, *IPSNR* is about −3 dB and the ways of dealing with them are considered in a study [44]. One proposition is to set less QS

Fig. 15 shows the original and the decompressed images in 110-th sub-band, where decrease of visual quality according to quantitative criteria is predicted. Noise is not seen in the original image and the compression practically does not influence the image quality (in our opinion,

Fig. 14. Predicted *IPSNR* for Hyperion hyperspectral image compressed by the coder AGU

A study [44] also presents data for three other DCT-based coders, where two of them are specially suited for providing better visual quality. It is demonstrated that the coder adaptive DCT (ADCT), which exploits the optimized partition schemes [60], provides certain improve‐ ments compared to AGU. Meanwhile, DCT coders oriented on improving the visual quality being applied to noisy images do not offer substantial benefits and, moreover, are even less

**Figure 14.** Predicted *IPSNR* for Hyperion hyperspectral image compressed by the coder AGU

(a) (b)

A study [44] also presents data for three other DCT-based coders, where two of them are specially suited for providing better visual quality. It is demonstrated that the coder adaptive DCT (ADCT), which exploits the optimized partition schemes [60], provides certain improvements compared to AGU. Meanwhile, DCT coders oriented on improving the visual quality being applied to noisy images do not offer substantial benefits and,

The methodology of predicting CR in OOP is the same as that for filtering. It is based on the scatterplot obtaining and curve fitting. The only difference is that the vertical axis relates to CR, while the horizontal axis, as earlier, corresponds to mean probability. Two mean probabilities *P2*σ and *P*2.7σ have been considered where the latter occurred to be worse again.

Therefore, the obtained results for the mean probability *P*2σ only are presented below.

Fig. 15. The 110-th sub-band images before (a) and after (b) compression

moreover, are even less efficient in many practical situations.

**4.2 Prediction of compression ratio in OOP**

efficient in many practical situations.

but this leads to smaller CR.

396 Environmental Applications of Remote Sensing

both images look the same).

The methodology of predicting CR in OOP is the same as that for filtering. It is based on the scatterplot obtaining and curve fitting. The only difference is that the vertical axis relates to CR, while the horizontal axis, as earlier, corresponds to mean probability. Two mean proba‐ bilities *P2*σ and *P*2.7σ have been considered where the latter occurred to be worse again. Therefore, the obtained results for the mean probability *P*2σ only are presented below. where two of them are specially suited for providing better visual quality. It is demonstrated that the coder adaptive DCT (ADCT), which exploits the optimized partition schemes (Ponomarenko et al., 2008), provides certain improvements compared to AGU. Meanwhile, DCT coders oriented on improving the visual quality being applied to noisy images do not offer substantial benefits and, moreover, are even less efficient in many practical situations.

Fig. 15. The 110-th sub-band images before (a) and after (b) compression

Two lossy compression methods, namely, the coders AGU and ADCT, have been studied. Their scatterplots are presented in Fig. 16. Contrary to other cases considered above, fitting is performed using a sum of two weighted exponential functions. As can be seen, fitting in both cases is very good with *R*<sup>2</sup> exceeding 0.99. Slightly larger values of CR are provided by the more sophisticated coder ADCT [60]. Very large (over 20) values of CR are provided for *P*2σ > 0.93, that is, for simple structure images corrupted by intensive noise. **4.2 Prediction of compression ratio in OOP** The methodology of predicting CR in OOP is the same as that for filtering. It is based on the scatterplot obtaining and curve fitting. The only difference is that the vertical axis relates to

We did not have real-life multichannel images corrupted by AWGN. But the hyperspectral data for the sensors Hyperion and airborne visible/infrared imaging spectrometer (AVIRIS) were available. Noise in them is signal dependent [14] with prevailing SD component for the model (22). The parameters of this noise were estimated in an automatic manner [11] and, thus, it became possible to apply VST (a generalized Anscombe transform with properly adjusted parameters) with converting noise into pure additive with unity variance. CR, while the horizontal axis, as earlier, corresponds to mean probability. Two mean probabilities *P2*σ and *P*2.7σ have been considered where the latter occurred to be worse again. Therefore, the obtained results for the mean probability *P*2σ only are presented below. Two lossy compression methods, namely, the coders AGU and ADCT, have been studied. Their scatterplots are presented in Fig. 16. Contrary to other cases considered above, fitting

Lossy compression in OOP neighborhood has been applied after VST. After decompression, inverse transform has to be applied, respectively. The obtained and predicted values of CR for Hyperion data are depicted in Fig. 17(a). As can be seen, the curves are in good agreement. There are some channels where predicted CRs are slightly larger than attained ones. This is is performed using a sum of two weighted exponential functions. As can be seen, fitting in both cases is very good with *R*2 exceeding 0.99. Slightly larger values of CR are provided by the more sophisticated coder ADCT (Ponomarenko et al., 2008). Very large (over 20) values of CR are provided for *P*2<sup>σ</sup> > 0.93, that is, for simple structure images corrupted by intensive

We did not have real-life multichannel images corrupted by AWGN. But the hyperspectral data for the sensors Hyperion and airborne visible/infrared imaging spectrometer (AVIRIS) were available. Noise in them is signal dependent (Abramov et al., 2015) with prevailing SD component for the model (22). The parameters of this noise were estimated in an automatic manner (Uss et al., 2011) and, thus, it became possible to apply VST (a generalized Anscombe transform with properly adjusted parameters) with converting noise into pure

Fig. 16. Scatterplots and fitted curves of dependences of CR vs. *P*2σ for the coders AGU (a)

and ADCT (b) **Figure 16.** Scatterplots and fitted curves of dependences of CR vs. *P*2σ for the coders AGU (a) and ADCT (b) parameters for channels with high signal-to-noise ratio. The largest CRs take place for subbands with low SNR (these are the sub-bands with indices 13–20, 125–130, and 175–180).

EO1H1800252002116110KZ) (a) and AVIRIS Lunar Lake image (b) by the coder AGU after the generalized Anscombe transform **Figure 17.** Results of a component-wise compression of Hyperion data (the analyzed set is EO1H1800252002116110KZ) (a) and AVIRIS Lunar Lake image (b) by the coder AGU after the generalized Anscombe transform

Fig. 17. Results of a component-wise compression of Hyperion data (the analyzed set is

explained by the imperfectness of VST and blind estimation of noise parameters for channels with high signal-to-noise ratio. The largest CRs take place for sub-bands with low SNR (these are the sub-bands with indices 13–20, 125–130, and 175–180). (a) (b) The results for the AVIRIS test image Lunar Lake are given in Fig. 17(b). Here, the agreement between the predicted and the attained values is even better than for the Hyperion data. Again, the largest CR is observed for sub-bands with low SNR. There are

The results for the AVIRIS test image Lunar Lake are given in Fig. 17(b). Here, the agreement between the predicted and the attained values is even better than for the Hyperion data. Again, the largest CR is observed for sub-bands with low SNR. There are considerable differences in maximal and minimal values of CR. The main reason is the different SNR and different dynamic range in sub-band images. Certainly, CR also depends upon the image content. Fig. 17. Results of a component-wise compression of Hyperion data (the analyzed set is EO1H1800252002116110KZ) (a) and AVIRIS Lunar Lake image (b) by the coder AGU after the generalized Anscombe transform The results for the AVIRIS test image Lunar Lake are given in Fig. 17(b). Here, the agreement between the predicted and the attained values is even better than for the considerable differences in maximal and minimal values of CR. The main reason is the different SNR and different dynamic range in sub-band images. Certainly, CR also depends upon the image content. **5. Conclusions and future work**

Hyperion data. Again, the largest CR is observed for sub-bands with low SNR. There are

It is demonstrated that it is possible to predict the efficiency of image filtering as well as the parameters of lossy compression of a noisy image in OOP neighborhood. As opposed to the

approach predicts practically a reachable performance and makes this very rapidly, by one

#### **5. Conclusions and future work** considerable differences in maximal and minimal values of CR. The main reason is the different SNR and different dynamic range in sub-band images. Certainly, CR also depends earlier known approaches that allow predicting potential efficiency of filtering, the present

upon the image content.

It is demonstrated that it is possible to predict the efficiency of image filtering as well as the parameters of lossy compression of a noisy image in OOP neighborhood. As opposed to the earlier known approaches that allow predicting potential efficiency of filtering, the present approach predicts practically a reachable performance and makes this very rapidly, by one or more orders faster than filtering or compression itself. **5. Conclusions and future work** It is demonstrated that it is possible to predict the efficiency of image filtering as well as the parameters of lossy compression of a noisy image in OOP neighborhood. As opposed to the or more orders faster than filtering or compression itself. Certainly, a limited number of quality metrics, filtering, and compression techniques have been considered. However, it is important that a general methodology of prediction is proposed, and it is shown there are somewhat strict connections between simple input parameters (that can be easily and quickly calculated) and output parameters that are able to

adequately characterize the efficiency of filtering or lossy compression techniques. In favor of this methodology, there are certain facts. First, there are many modern filters that have filtering efficiency of the same order as the DCT-based filter and BM3D. Thus, predicting denoising efficiency for the filters mentioned above, it is possible to approximately predict performance for other modern filters (although such prediction would be less accurate). Second, the same holds for lossy compression methods. For example, AGU and JPEG2000 Certainly, a limited number of quality metrics, filtering, and compression techniques have been considered. However, it is important that a general methodology of prediction is proposed, and it is shown there are somewhat strict connections between simple input parameters (that can be easily and quickly calculated) and output parameters that are able to adequately characterize the efficiency of filtering or lossy compression techniques. In favor of this methodology, there are certain facts. First, there are many modern filters that have filtering efficiency of the same order as the DCT-based filter and BM3D. Thus, predicting denoising efficiency for the filters mentioned above, it is possible to approximately predict performance for other modern filters (although such prediction would be less accurate). Second, the same holds for lossy compression methods. For example, AGU and JPEG2000 provide similar performance characteristics. Then, by predicting compression parameters for AGU, they are, in fact, estimated for JPEG2000 as well.

Concerning the decision making, whether to perform filtering or not, strict recommendations have been given for probabilities *P*2σ and *P*0.5σ. Filtering can be expedient if *P*2σ exceeds 0.5 or *P*0.5σ exceeds 0.15. Similarly, OOP is quite possible if *P*2σ is approximately 0.85 or larger. A very important fact is that these rules for filtering are valid for different types of noise (pure additive and signal-dependent, additive white Gaussian and spatially correlated). This generalization can be considered as one of the main contributions of this chapter. Meanwhile, the case of spatially correlated noise requires more attention in future. In prediction of filtering efficiency, general prediction approximations for spatially correlated noise with *a priori* known or preestimated properties (e.g., 2D spectrum) have not been obtained yet. It can only be expected that the scatterplots for spatially correlated noise with other (not analyzed yet) shapes and parameters of spatial power spectrum behave similarly. The studies for lossy compression of images corrupted by spatially correlated noise are yet to be started. This opens a very wide field for future research.

The results of this research show that although sometimes the prediction of performance characteristics based on one input parameter is appropriately accurate, there are several means to improve the prediction accuracy. One way that deals with multiparameter input has been already used for particular cases. The use of mean *P*0.5σ has shown itself a good solution, although it has not yet been tried for all possible applications. In particular, mean *P*0.5σ has not been tested for lossy compression. It is hoped that performance can be improved due to this reason. Neural networks or other approximators of multidimensional functions (surfaces) can be useful.

explained by the imperfectness of VST and blind estimation of noise parameters for channels with high signal-to-noise ratio. The largest CRs take place for sub-bands with low SNR (these

The results for the AVIRIS test image Lunar Lake are given in Fig. 17(b). Here, the agreement between the predicted and the attained values is even better than for the Hyperion data. Again, the largest CR is observed for sub-bands with low SNR. There are considerable differences in maximal and minimal values of CR. The main reason is the different SNR and different dynamic range in sub-band images. Certainly, CR also depends

(a) (b) Fig. 17. Results of a component-wise compression of Hyperion data (the analyzed set is EO1H1800252002116110KZ) (a) and AVIRIS Lunar Lake image (b) by the coder AGU after

**Figure 17.** Results of a component-wise compression of Hyperion data (the analyzed set is EO1H1800252002116110KZ)

(a) and AVIRIS Lunar Lake image (b) by the coder AGU after the generalized Anscombe transform

(a) (b) Fig. 16. Scatterplots and fitted curves of dependences of CR vs. *P*2σ for the coders AGU (a)

for Hyperion data are depicted in Fig. 17(a). As can be seen, the curves are in good agreement. There are some channels where predicted CRs are slightly larger than attained ones. This is explained by the imperfectness of VST and blind estimation of noise parameters for channels with high signal-to-noise ratio. The largest CRs take place for subbands with low SNR (these are the sub-bands with indices 13–20, 125–130, and 175–180).

Lossy compression in OOP neighborhood has been applied after VST. After decompression, inverse transform has to be applied, respectively. The obtained and predicted values of CR for Hyperion data are depicted in Fig. 17(a). As can be seen, the curves are in good agreement. There are some channels where predicted CRs are slightly larger than attained ones. This is explained by the imperfectness of VST and blind estimation of noise parameters for channels with high signal-to-noise ratio. The largest CRs take place for subbands with low SNR (these are the sub-bands with indices 13–20, 125–130, and 175–180).

**Figure 16.** Scatterplots and fitted curves of dependences of CR vs. *P*2σ for the coders AGU (a) and ADCT (b)

The results for the AVIRIS test image Lunar Lake are given in Fig. 17(b). Here, the agreement between the predicted and the attained values is even better than for the Hyperion data. Again, the largest CR is observed for sub-bands with low SNR. There are considerable differences in maximal and minimal values of CR. The main reason is the different SNR and different dynamic range in sub-band images. Certainly, CR also depends upon the image content.

The results for the AVIRIS test image Lunar Lake are given in Fig. 17(b). Here, the agreement between the predicted and the attained values is even better than for the Hyperion data. Again, the largest CR is observed for sub-bands with low SNR. There are considerable differences in maximal and minimal values of CR. The main reason is the different SNR and different dynamic range in sub-band images. Certainly, CR also depends

It is demonstrated that it is possible to predict the efficiency of image filtering as well as the parameters of lossy compression of a noisy image in OOP neighborhood. As opposed to the earlier known approaches that allow predicting potential efficiency of filtering, the present approach predicts practically a reachable performance and makes this very rapidly, by one

(a) (b) Fig. 17. Results of a component-wise compression of Hyperion data (the analyzed set is EO1H1800252002116110KZ) (a) and AVIRIS Lunar Lake image (b) by the coder AGU after

It is demonstrated that it is possible to predict the efficiency of image filtering as well as the parameters of lossy compression of a noisy image in OOP neighborhood. As opposed to the earlier known approaches that allow predicting potential efficiency of filtering, the present approach predicts practically a reachable performance and makes this very rapidly, by one or

Certainly, a limited number of quality metrics, filtering, and compression techniques have been considered. However, it is important that a general methodology of prediction is proposed, and it is shown there are somewhat strict connections between simple input parameters (that can be easily and quickly calculated) and output parameters that are able to adequately characterize the efficiency of filtering or lossy compression techniques. In favor of this methodology, there are certain facts. First, there are many modern filters that have filtering efficiency of the same order as the DCT-based filter and BM3D. Thus, predicting denoising efficiency for the filters mentioned above, it is possible to approximately predict performance for other modern filters (although such prediction would be less accurate). Second, the same holds for lossy compression methods. For example, AGU and JPEG2000

It is demonstrated that it is possible to predict the efficiency of image filtering as well as the parameters of lossy compression of a noisy image in OOP neighborhood. As opposed to the

are the sub-bands with indices 13–20, 125–130, and 175–180).

**5. Conclusions and future work**

**5. Conclusions and future work**

upon the image content.

upon the image content.

the generalized Anscombe transform

**5. Conclusions and future work**

the generalized Anscombe transform

and ADCT (b)

398 Environmental Applications of Remote Sensing

more orders faster than filtering or compression itself.

or more orders faster than filtering or compression itself.

There are also other possible directions for future research. 3D filtering warrants a more thorough study, at least, for the case of more than three channels. The same relates to 3D lossy compression performance, which has not been tried to predict yet. Compression parameters for QS other than the one recommended for OOP is also of sufficient interest in DCT-based lossy compression. Influence of errors in *a priori* information on noise parameters or their blind estimates on prediction accuracy has to be studied as well.

#### **Author details**

Benoit Vozel2\*, Oleksiy Rubel1 , Alexander Zemliachenko1 , Sergey Abramov1 , Sergey Krivenko1 , Ruslan Kozhemiakin1 , Vladimir Lukin1 and Kacem Chehdi2

\*Address all correspondence to: benoit.vozel@univ-rennes1.fr

1 National Aerospace University, Ukraine

2 University of Rennes 1, France

#### **References**


[10] Meola J., Eismann M. T., Moses R. L., & Ash J. N. (2011). Modeling and estimation of signal-dependent noise in hyperspectral imagery. *Applied Optics*, Vol. 50, No. 21, pp. 3829-3846.

**Author details**

Sergey Krivenko1

**References**

Benoit Vozel2\*, Oleksiy Rubel1

400 Environmental Applications of Remote Sensing

2 University of Rennes 1, France

1 National Aerospace University, Ukraine

Publishing, Herndon, VA.

13, pp. 1125-1139.

ley-Interscience, Hoboken, NJ.

*Processing,* pp. 263-295. France.

Doi: 10.1007/s11760-013-0446-1, 16 p.

, Ruslan Kozhemiakin1

*Processing*, Academic Press, Orlando, FL.

M., and Chanussot J., pp. 9-29. Springer.

Vol. 5, No. 1, 053502. Doi: 10.1117/1.3539768.

eters. *Annals of Geophysics*, Vol. 49, No. 1, pp. 1-9.

\*Address all correspondence to: benoit.vozel@univ-rennes1.fr

, Alexander Zemliachenko1

, Vladimir Lukin1

[1] Schowengerdt R. (2006). Remote Sensing, Third Edition. *Models and Methods for Image*

[2] Oliver C. & Quegan S. (2004). *Understanding Synthetic Aperture Radar Images.* SciTech

[3] Christophe, E. (2011). Hyperspectral Data Compression Tradeoff. In: *Optical Remote Sensing in Advances, Signal Processing and Exploitation Techniques*, Eds. Prasad S., Bruce L.

[4] Lukin V., Abramov S., Ponomarenko N., Uss M., Zriakhov M., Vozel B., Chehdi K., & AstolaJ.(2011).Methodsandautomaticproceduresforprocessingimagesbasedonblind evaluation of noise type and characteristics. *SPIE Journal on Advances in Remote Sensing*,

[5] Lukin V., Abramov S., Ponomarenko N., Krivenko S., Uss M., Vozel B., Chehdi K., Egiazarian K., & Astola J. (2014). Approaches to Automatic Data Processing in Hy‐ perspectral Remote Sensing. *Telecommunications and Radio Engineering*, Vol. 73, No.

[6] Chang C. I. (Ed.). (2007). *Hyperspectral Data Exploitation: Theory and Applications.* Wi‐

[7] Aiazzi B., Alparone L., Barducci A., Baronti S., Marcoinni P., Pippi I., & Selva M. (2006). Noisemodelling andestimationofhyperspectraldata fromairborne imaging spectrom‐

[8] Vozel B., Abramov S., Chehdi K., Lukin V., Ponomarenko N., Uss M., & Astola J. (2009). Blind methods for noise evaluation in multi-component images, In: *Multivariate Image*

[9] Ponomarenko N., Lukin V., Egiazarian K., & Lepisto L. (2013). Adaptive Visually Lossless JPEG-Based Color Image Compression. Signal*, Image and Video Processing*,

, Sergey Abramov1

and Kacem Chehdi2

,


[35] Bekhtin Yu. S. (2011). Adaptive Wavelet Codec for Noisy Image Compression. *Pro‐ ceedings of the 9-th East-West Design and Test Symposium*, Sevastopol, Ukraine, Sept. 2011, pp. 184-188.

[22] Sendur L. & Selesnick I. W. (2002). Bivariate shrinkage with local variance estimation.

[23] Ponomarenko N. N., Lukin V. V., Egiazarian K. O., & Astola J. T. (2010). A method for blind estimation of spatially correlated noise characteristics. *Proceedings of SPIE 7532 of Image Processing: Algorithms and Systems VIII,* 753208, San Jose, USA, January

[24] Lebrun M., Colom M., Buades A., & Morel J. M. (2012). Secrets of image denoising

[25] Van Zyl Marais I., Steyn W.H., & du Preez J.A. (2009). On-Board Image Quality As‐ sessment for a Small Low Earth Orbit Satellite. *Proceedings of the 7th IAA Symp. on*

[26] Anfinsen S. N., Doulgeris A. P., & Eltoft T. (2009). Estimation of the equivalent num‐ ber of looks in polarimetric synthetic aperture radar imagery. *IEEE Transactions on*

[27] Liu C., Szeliski R., Kang S. B., Zitnick C. L., & Freeman W. T. (2008). Automatic esti‐ mation and removal of noise from a single image. *IEEE Transactions on Pattern Analy‐*

[28] Colom M., Lebrun M., Buades A., & Morel J. M. (2014). A Non-Parametric Approach for the Estimation of Intensity-Frequency Dependent Noise. IEEE International Con‐

[30] Öktem R., Yaroslavsky L., Egiazarian K., & Astola J. (2002). Transform domain ap‐ proaches for image denoising. *Journal of Electronic Imaging*, Vol. 11, No. 2, pp. 149 –

[31] Solbo S. & Eltoft T. (2004). Homomorphic wavelet-based statistical despeckling of SAR images. *IEEE Transactions on Geoscience and Remote Sensing*, Vol. GRS-42, No. 4,

[32] Öktem R., Egiazarian K., Lukin V.V., Ponomarenko N.N., & Tsymbal O.V. (2007). Lo‐ cally adaptive DCT filtering for signal-dependent noise removal. *EURASIP Journal on*

[33] Lukin V., Ponomarenko N., Egiazarian K., & Astola J. (2008). Adaptive DCT-based filtering of images corrupted by spatially correlated noise. *Proceedings of SPIE 6812 of Image Processing: Algorithms and Systems VI*, 68120W, San Jose, USA. Doi:

[34] Dabov K., Foi A., Katkovnik V., & Egiazarian K. (2007). Image denoising by sparse 3- D transform-domain collaborative filtering. *IEEE Transactions on Image Processing*,

ference on Image Processing (ICIP). Doi: 10.1109/ICIP.2014.7025865.

[29] Mallat S. (1998). *A Wavelet Tour of Signal Processing*. Academic Press, San Diego.

*IEEE Signal Processing Letters*, Vol. 9, No. 12, pp. 438-441.

2010. Doi: 10.1117/12.847986.

402 Environmental Applications of Remote Sensing

156.

pp. 711-721.

10.1117/12.764893.

Vol. 16, No. 8, pp. 2080-2095.

cuisine. *Acta Numerica*, Vol. 21, pp. 475-576.

*Small Satellites for Earth Observation*, Berlin, Germany.

*sis and Machine Intelligence*, Vol. 30, No. 2, pp. 299-314.

*Advances in Signal Processing*, Vol. 2007, 10 p.

*Geoscience and Remote Sensing*, Vol. 47, No. 11, pp. 3795-3809.


nance Space. *IEEE International Conference on Image Processing, ICIP,* Vol. 1, pp. 313-316.


[59] Ponomarenko N., Ieremeiev O., Lukin V., Egiazarian K., & Carli M. (2011). Modified Image Visual Quality Metrics for Contrast Change and Mean Shift Accounting. *Pro‐ ceedings of CADSM*, Ukraine, pp. 305 - 311.

nance Space. *IEEE International Conference on Image Processing, ICIP,* Vol. 1, pp.

[47] Lee J.S. (1983). Digital image smoothing and the sigma filter. *Computer Vision, Graph‐*

[48] Chatterjee P. & Milanfar P. (2010). Is denoising dead? *IEEE Transactions on Image*

[49] Levin A. and Nadler B. (2011). Natural image denoising: Optimality and inherent bounds. *IEEE Conference on Computer Vision and Pattern Recognition (CVPR)*, pp.

[50] [50 Cameron C., Windmeijer A., Frank A.G., Gramajo H., Cane D.E., & Khosla C. (1997). An R-squared measure of goodness of fit for some common nonlinear regres‐

[51] Rubel O. & Lukin V. (2014) An Improved Prediction of DCT-Based Filters Efficiency Using Regression Analysis. *Information and Telecommunication Sciences*, Kiev, Ukraine,

[52] Badiru A. & Cheung J. (2002). Fuzzy Engineering Expert Systems with Neural Net‐

[53] Rubel A., Naumenko A., & Lukin V. (2014). A Neural Network Based Predictor of

[54] Krivenko S., Lukin V., Vozel B., & Chehdi K. (2014). Prediction of DCT-based Denois‐ ing Eficiency for Images Corrupted by Signal-Dependent Noise. *Proceedings of IEEE 34*th *International Scientific Conference Electronics and Nanotechnology*, Kiev, Ukraine,

[55] Lukin V., Abramov S., Rubel A., Naumenko A., Krivenko S., Vozel B., Chehdi K., Egiazarian K., & Astola J. (2014). An approach to prediction of signal-dependent noise removal efficiency by DCT-based filter. *Telecommunications and Radio Engineer‐*

[56] Rubel A., Lukin V., & Pogrebnyak O. (2014). Efficiency of DCT-based denoising tech‐ niques applied to texture images, *Proceedings of Mexican Conference of Pattern Recogni‐*

[57] Rubel O., Lukin V., & de Medeiros F.S. (2015). Prediction of Despeckling Efficiency of DCT-based Filters Applied to SAR Images, *Proceedings of 2015 International Conference*

[58] Rubel O. S., Kozhemiakin R. O., Krivenko S. S., & Lukin V. V. (2015). A Method for Predicting Denoising Efficiency for Color Images. *Proceedings of 2015 IEEE 35th Inter‐ national Conference on Electronics and Nanotechnology (ELNANO)*, Kiev, Ukraine, pp.

*on Distributed Computing in Sensor Systems*, Fortaleza, Brazil, pp. 159-168.

sion models. *Journal of Econometrics,* Vol. 77, No. 2, pp. 1790–1792.

*ics, and Image Processing*, Vol. 24, No. 2, pp. 255-269.

work Applications*.* Wiley-Interscience, New York.

Filtering. *Proceedings of MRRS*, Kiev, Ukraine, pp. 14-17.

*Processing*, Vol. 19, No. 4, pp. 895-911.

313-316.

404 Environmental Applications of Remote Sensing

2833-2840.

pp. 254-258.

304-309.

*ing*, Vol. 73, No. 18, pp. 1645-1659.

*tion*, Cancun, Mexico, pp. 261-270.

Vol. 5, No. 1, pp. 30-41.

[60] Ponomarenko N., Lukin V., Egiazarian K., & Astola J. (2008). ADCT: A New High Quality DCT Based Coder for Lossy Image Compression. CD-ROM Proceedings of LNLA, Switzerland, 6 p.

## *Edited by Maged Marghany*

Nowadays, the innovation in space technologies creates a new trend for the Earth observation and monitoring from space. This book contains high quality and compressive work on both microwave and optical remote sensing applications. This book is divided into five sections: (i) remote sensing for biomass estimation, (ii) remote sensing-based glacier studies, (iii) remote sensing for coastal and ocean applications, (iv) sewage leaks and environment disasters, and (v) remote sensing image processing.

Each chapter offers an opportunity to expand the knowledge about various remote sensing techniques and persuade researchers to deliver new research novelty for environment studies.

Photo by Anastasiya\_Voronova / iStock

Environmental Applications of Remote Sensing

Environmental Applications

of Remote Sensing

*Edited by Maged Marghany*