Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Enhancing sparse representation of color images by cross channel transformation

Abstract

Transformations for enhancing sparsity in the approximation of color images by 2D atomic decomposition are discussed. The sparsity is firstly considered with respect to the most significant coefficients in the wavelet decomposition of the color image. The discrete cosine transform is singled out as an effective 3 point transformation for this purpose. The enhanced feature is further exploited by approximating the transformed arrays using an effective greedy strategy with a separable highly redundant dictionary. The relevance of the achieved sparsity is illustrated by a simple encoding procedure. On typical test images the compression at high quality recovery is shown to significantly improve upon JPEG and WebP formats.

1 Introduction

In the signal processing field sparse representation usually refers to the approximation of a signal as superposition of elementary components, called atoms, which are members of a large redundant set, called a dictionary [1]. The superposition, termed atomic decomposition, aims at approximating the signal involving as few atoms as possible [14]. Sparsity is also relevant to data collection. Within the emerging theory of sampling known as compressive sensing (CS) this property is key for reconstructing signals from a reduced number of measures [57]. In particular, distributed compressive sensing (DCS) algorithms exploit inter signal correlation structure for multiple signal recovery [8].

Sparse image representation using redundant dictionaries has been considered in numerous works e.g. [911] and in the context of applications such as image restoration [1214], feature extraction [1518] and super resolution [1922].

The sparsity property of some types of 3D images benefits from 3D processing [2326]. In particular most RGB (Red-Green-Blue) images admit a sparser atomic decomposition if approximated by 3D atoms [27]. Within the 3D framework the gain in sparsity comes at expenses of computational cost though.

The purpose of this work is to show that the application of a transformation across the direction of the colors improves sparsity in the wavelet domain representation of the 2D color channels. The relevance of this feature is demonstrated by a simple encoding procedure rendering good compression results in comparison to commonly used image compression standards.

The work is organized as follows: Sec. 2 introduces the mathematical notation. Sec. 3 compares several cross color transformations enhancing sparsity. A numerical example on a large data set is used to illustrate the suitability of the dct for that purpose. Sec. 4 demonstrates the gain in sparsity obtained by the atomic decomposition of color images when the dct is applied across the RGB channels. Sec. 5 illustrates the relevance of the approach to image compression with high quality recovery. The conclusions are summarized in Sec. 6.

2 Mathematical notation

Throughout the paper we use the following notational convention. represents the set of real numbers. Boldface letters indicate Euclidean vectors, 2D and 3D arrays. Standard mathematical fonts indicate components, e.g., is a vector of components . The elements of a 2D array are indicated as I(i, j), i = 1, …, Lx, j = 1, …, Ly and the color channels of as I(:, :, z), z = 1, 2, 3. The transpose of a matrix, G say, is indicated as G. A set of, say M, color images is represented by the arrays

The inner product between arrays in is given by the Frobenius inner product 〈⋅, ⋅〉F as

Consequently, the Frobenius norm ‖⋅‖F is calculated as

The inner produce between arrays in is given by the Euclidean inner product 〈⋅, ⋅〉 as

3 Cross color transformations

Given a color image I(i, j, z), i = 1, …, Lx, j = 1, …, Ly, z = 1, 2, 3 the processing of the 3 channels can be realized either in the pixel/intensity or in the wavelet domain. Since the representation of most images is sparser in the wavelet domain [2730] we approximate in that domain and reconstruct the approximated image by the inverse wavelet transform. Thus, using a 3 × 3 matrix T, we construct the transformed arrays and as follows (1) (2) where dwt indicates the 2D wavelet transform. For the transformation T we consider the following cases

  1. (i). The dct.
  2. (ii). The reversible YCbCr color space transform.
  3. (iii). The principal components (PC) transform.
  4. (iv). A transformation learned from an independent set of images.

The dct is given by the matrix whereas the YCbCr color space transform is given by the matrix below [31]

The columns of the principal components transform are the normalized eigenvectors of covariance matrix of the RGB pixels. Thus, the transformation is image dependent and optimal with respect to decorrelating the color channels.

As a first test, the approximation of the transformed channels is realized by keeping a fixed number of the largest absolute value entries, and setting the others equal to zero. In relation to this, for an image of size Lx × Ly × 3 we define the Sparsity Ratio (SR) as follows (3)

The quality of a reconstructed image , with respect to the original 8-bit image I, is compared using the Peak Signal-to-Noise Ratio (PSNR)

For the numerical examples below the transformation corresponding to case (iv) is learned from a set of images I{m}, m = 1, …, M all of the same size. Starting from an invertible 3 × 3 matrix Tk, with k = 1, the learning algorithm proceeds through the following instructions.

  1. Use Tk to transform the arrays I{m}→Uk{m}→Wk{m} as in (1) and (2).
  2. Approximate each transformed array Wk{m} to obtain by keeping the largest K absolute value entries.
  3. Apply the inverse 2D wavelet transform idwt to reconstruct the approximated arrays as
  4. Use the original images I{m}, m = 1, …, M to find G = T−1 by least squares fitting, i.e where with
  5. While decreases, or the maximum number of allowed iterations has not been reached, set kk + 1, Tk = (G*)−1 and repeat steps 1)–5).

Given the arrays the least squares problem for determining the transformation Tk has a unique solution. However, the joint optimization with respect to the arrays and the transformation Tk is not convex. Hence, the algorithm’s outcome depends on the initial value.

The transformation (iv) used in the numerical examples of Secs. 3.1 and 4.1 has been learned from M = 100 images, all of size 384 × 512 × 3, from the UCID data set [32], which contains images of buildings, places and cars. The learning curves for two random orthonormal initializations is shown in Fig 1.

thumbnail
Fig 1. Learning curves for the transformation (iv) corresponding to two different random orthogonal transforms initializing the process.

The mean value PSNR with respect to the 100 images in the training set corresponds to SR = 15 for all the images.

https://doi.org/10.1371/journal.pone.0279917.g001

It is worth mentioning that, as shown in Fig 1, learning is richer when starting comparatively distant from a local minimum (left graph in Fig 1). However, since the convergence is to a local minimum other random initializations, even if generating less learning, may produce better results (right graph in Fig 1).

The aim of the next numerical test is to demonstrate the effect on the SR (c.f. (3)) produced by the above (i)–(iv) transformations across the color channels.

3.1 Numerical test I

Using the whole Berkeley data set [33], consisting of 300 images all of size 321 × 481 × 3 we proceed with each image as in (1) and (2). The dwt corresponds to the Cohen-Daubechies-Feauveau 9/7 (CDF97) wavelet family. Fig 2 shows the transformed channels of the image in Fig 3, including the dct transformation across channels (right graph) and without T transformation (left graph). As noticeable in the figure, the effect of the dct is to re-distribute the intensity in the color channels by transferring values between channels.

thumbnail
Fig 2. Magnitude of the entries in the array W constructed as in (2) from the image of Fig 3, with T the dct (right graph) and without T (left graph).

https://doi.org/10.1371/journal.pone.0279917.g002

thumbnail
Fig 3. One of the RGB images in the Berkeley’s data set.

https://doi.org/10.1371/journal.pone.0279917.g003

For the numerical test the approximations are realized fixing SR = 20 and SR = 10. The reconstructed images are obtained for the approximated arrays as (4) (5) where T−1 is the inverse of T. When no transformation T is considered the image is reconstructed directly from (4).

The 2nd and 4th columns of Table 1 show the mean value PSNR (), with respect to the whole data set, for SR = 20 and SR = 10 respectively, corresponding to the transformations T listed in the 1st column of the table. The 3rd and 5th columns give the standard deviations (std).

thumbnail
Table 1. Mean value PSNR, with respect to the 300 images in the Berkeley data set.

The approximation of each image is realized by setting the least significant entries in the arrays W{m} = 1, …, 300 equal zero, in order to obtain SR = 20 for all the images (2nd column) and SR = 10 for all the images (4th column).

https://doi.org/10.1371/journal.pone.0279917.t001

While Table 1 shows that all (i)–(iv) transformations render equivalent superior results, with respect to not applying a transformation (case (v)), the dct slightly exceeds the others. Case (iv) refers to the best result achieved when initializing the learning algorithm with 500 different random orthonormal transformations. When initializing the algorithm with transformations (i) and (ii) it appears that each of these transformations is close to a local minimizer of the method. This stems from the fact that such initializations do not generate significant learning.

The common feature of most of the 300 images in the data set used in this numerical example is the correlation property of the three color channels. This property is assessed by the correlation coefficients where (8) and indicates the mean value of channel z.

The range of the correlation coefficient r(z), z = 1, 2, 3 for the images in the data set can be better estimated using the Fisher transform [34, 35] which is defined as follows

As seen in the graphs of Fig 4, the histograms of the transformed coefficients ζ(z), z = 1, 2, 3 resemble normal distributions. Hence, the confidence intervals can be well estimated in this domain. Once that is done, the intervals for the correlation coefficients are retrieved through the inverse transformation

thumbnail
Fig 4. Histograms of the transformed correlation coefficients between the RGB channels: ζ(1) (top graph), ζ(2) (middle graph), and ζ(3) (bottom graph).

https://doi.org/10.1371/journal.pone.0279917.g004

Table 2 gives the range of the correlation coefficients concerning approximately 68% and 95% of the images in the data set.

thumbnail
Table 2. Intervals for the correlation coefficients between R and G channels, r(1), G and B channels, r(2), and R and B channels, r(3), involving approximately 68% and 95% of the images in the data set.

https://doi.org/10.1371/journal.pone.0279917.t002

In view of the high correlation between the color channels for most images in the data set it is surprising than the PC transformation (iii), which completely decorrelates the channels, does not overperform the other transformations, on the contrary. This feature has also been noticed in the context of bit allocation for subband color image compression [36].

4 Approximations by atomic decomposition

We have seen (c.f. Table 1) that by transformation of channels it is possible to gain quality when reducing nonzero entries in the channels. Now we discuss how to improve quality further by approximating the 2D arrays (2) by an atomic decomposition, other than just by neglecting their less significant entries. For the approximation to be successful it is important to use an appropriate dictionary. To this end, one possibility could be to learn the dictionary from training data [3740]. However as demonstrated in previous works [27, 29, 30] a separable dictionary, which is easy to construct, is well suited for the purposes of achieving sparsity and delivers a fast implementation of the approach. Since we use that dictionary in the numerical examples, below we describe the method for constructing the atomic decomposition of the array W considering specifically a separable dictionary.

Firstly we concatenate the 3 planes W(:, :, z), z = 1, 2, 3 into an extended 2D array and divide this array in small blocks . Without loss of generality the blocks are assumed to be square of size Nb × Nb say, and are approximated using separable dictionaries and .

For q = 1, …, Q every element is approximated by an atomic decomposition as below: (6) where is the index in the set {1, 2, …, Mb} corresponding to the label of the atom in the dictionary contributing to the n-th term in the approximation of the q-th block. The index has the equivalent description. The assembling of the approximated blocks gives rise to the approximated array , where represents the assembling operation, i.e. the operation that retrieves the approximation of the array from the approximation of the blocks in the partition. The approximated array is reshaped back into 3 planes, , to be converted back to the approximated RGB intensity image as in (4) and (5).

The approximation of the partition is carried out iteratively as a two step process which selects i) the atoms in the atomic decomposition (6) and ii) the sequence in which the blocks in the partition are approximated. The procedure is called Hierarchized Block Wise (HBW) implementation of greedy pursuit strategies [28, 41]. For the selection of the atoms we apply the Orthogonal Matching Pursuit (OMP) approach [42] dedicated to 2D with separable dictionaries (OMP2D) [43]. Thus, the whole algorithm is termed HBW-OMP2D [28]. The method iterates as described below.

On setting kq = 0 and at iteration kq + 1 the algorithm selects the indices and , as follows: (7) where . The calculation of is realized in order to minimize , which is equivalent to finding the orthogonal projection onto the subspace spanned by the selected atoms . In our implementation, the calculation of the coefficients cq(n), n = 1, …, kq in (6) is realized as (8) where the set is biorthogonal to the set and needs to be upgraded and updated to account for each newly selected atom. Starting from the updating and upgrading is realized through the recursive equations [43, 44]: (9) with the additional re-orthogonalization step (10)

As discussed in [28, 41], for and the indices resulting from (7), the block to be approximated in the next iteration corresponds to the value q such that

The algorithm stops when the required total number of atoms has been selected. This number can be fixed using the SR, which is now calculated as (11)

Remark 1. The above described implementation of HBW-OMP2D is very effective in terms of speed, but demanding in terms of memory (the partial outputs corresponding to all the blocks in the partition need to be stored at every iteration). An alternative implementation, termed HBW Self Projected Matching Pursuit (HBW-SPMP) [29, 45], would enable the application of the identical strategy to much larger images than the ones considered in this work.

4.1 Numerical example II

For this and the next numerical example, we use a mixed dictionary consisting of two classes of sub-dictionaries of different nature:

  1. I). The trigonometric dictionaries and , defined below, for i = 1…, Nb wc(n) and ws(n), n = 1, …, Mx are normalization factors.
  2. II). The dictionary , which is constructed by translation of the prototype atoms in Fig 5.
thumbnail
Fig 5. Prototypes (each in different color) that generate the dictionaries by sequential translations of one point.

https://doi.org/10.1371/journal.pone.0279917.g005

The mixed dictionary is built as and .

Table 3 shows the improvement in achieved by atomic decompositions using the mixed dictionary for SR = 20 and SR = 10.

thumbnail
Table 3. Mean value PSNR, with respect to 300 images in the Berkeley data set, produced by 2D atomic decomposition of the arrays W{m} = 1, …, 300 in order to obtain SR = 20 (2nd and 6th column) and SR = 10 (4th and 8th column).

https://doi.org/10.1371/journal.pone.0279917.t003

Notice that while case (v), which does not include any T transformation, gives superior results than by disregarding entries (c.f. Table 1) when applying any of the transformations (i)–(iv) the results improve further. The PC transform, however, appears significantly less effective than the others. In addition to rendering the best results, the dct brings along the additional advantage of being orthonormal. Consequently, it does not magnify errors at the inversion step. Because of this, we single out the dct as the most convenient cross color transformation out of the four considered here.

5 Application to image compression

In order achieve compression by filing an atomic decomposition we need to address two issues. Namely, the quantization of the coefficients cq(n)n = 1, …, kq, q = 1, …, Q in (6) and the storage of the indices . We tackle the matters by simple but effective procedures [30].

For q = 1, …, Q the absolute value coefficients |cq(n)|, n = 1, …, kq are converted to integers through uniform quantization as follows (12)

The signs of the coefficient are encoded separately as a vector, sq, using a binary alphabet. Each pair of indices corresponding to the atoms in the decompositions of the block is mapped into a single index oq(n). The set oq(1), …, oq(kq) is sorted in ascending order to take the differences and construct the string of non-negative numbers . The order of the set induces order in the unsigned coefficients, , and in the corresponding signs .

For each q the number 0 is added at the end of the indices before concatenation, to be able to separate strings corresponding to different blocks. Each sequence of strings corresponding to q = 1, …, Q is concatenated and encoded using the off-the-shelf MATLAB function Huff06 [46], which implements Huffman coding.

The compression rate is given in bits-per-pixel (bpp) which is defined as

At the reconstruction stage the indices are recovered from the string of differences δq(n), n = 2, …, kq. The signs of the coefficients are read from the binary string. The quantized unsigned coefficients are read and transformed into real numbers as:

The codec for reproducing the examples in the next sections has been made available on [47].

5.1 Numerical example III

The relevance to image compression of the achieved sparsity by dct cross color transformation is illustrated in this section by comparison with results yielded by the compression standards JPEG, and WebP, on the 15 images in Table 4. These are typical images, used for compression tests, available in ppm or png format. The first 9 images are classic test images taken from [48]. The last six images are portions of 1024 × 1024 × 3 pixels shown in Fig 6 from very large high resolution images available on [49].

thumbnail
Table 4. Test images.

The last column gives the approximation times to produce the results in Table 5.

https://doi.org/10.1371/journal.pone.0279917.t004

thumbnail
Fig 6. Illustration of the 1024 × 1024 × 3 panels from the six high resolution test images (No 10–15) listed in Table 4.

https://doi.org/10.1371/journal.pone.0279917.g006

All the results have been obtained in the MATLAB environment (version R2019a), using a machine with CPU Intel(R) Core(TM) i7–3520M RAM 8GB CPU @ 2.90GHz. For the image approximation the HBW-OMP2D method was implemented with a C++ MEX file. All the channels and all the images were partitioned into blocks of size 16 × 16. The approximation times to produce the results in Table 5 are displayed in the last column of Table 4.

thumbnail
Table 5. Compression rate (bpp) corresponding to JPEG (bjp), WebP (bwb) and the proposed sparse representation (bsr), for the values of PSNR given in the 2nd column.

The corresponding values of MSSIM are given in the 3rd to 5th columns. ssjp, sswb, and sssr indicate the MSSIM produced by JPEG, WebP and the sparse representation codec respectively.

https://doi.org/10.1371/journal.pone.0279917.t005

For realizing the comparison we proceed as follows: we set the required value of PSNR as that produced by JPEG at quality = 95 and tune compression with the other methods to produce the same PSNR. In our codec the tuning is realized by approximating the image up to PSNRo = 1.025 ⋅ PSNR (where PSNR is the targeted quality) and setting the quantization parameter Δ so as to reproduce the targeted value of PSNR.

For compression with JPEG we use the MATLAB imwrite command. The compression with WebP was realized using the software for Ubuntu distributed on [50]. All the approaches were tuned for producing the same value of PSNR as JPEG for quality 95. The Mean Structural SIMilarity (MSSIM) index [51] was then calculated with the approximation corresponding to those values of PSNR.

6 Conclusions

The application of a cross color transformation for enhancing sparsity in the atomic decomposition of RGB images has been proposed. It was demonstrated that the effect of the transformation is to re-distribute the most significant values in the dwt of the 2D channels. As a result, when approximating the arrays by disregarding the less significant entries, the quality of the reconstructed image improves with respect to disabling the cross color transformation. Four transformations have been considered: (i) a 3 point dct, (ii) the reversible YCbCr color space transform, (iii) the PC transform, (iv) a transformation learned from an independent set of images.

The quality of the image approximation was improved further by approximating the transformed arrays by an atomic decomposition using a separable dictionary and the greedy pursuit strategy HBW-OMP2D. The dct was singled out as the most convenient cross color transformation for approximating RGB color images in the wavelet domain.

The approximation approach has been shown to be relevant for image compression. By means of a simple coding strategy the achieved compression for typical test images considerably improves upon the most commonly used compression standards, namely JPEG and WebP.

References

  1. 1. Mallat S. and Zhang Z., “Matching pursuit with time-frequency dictionaries,” IEEE Trans. Signal Process., Vol (41,12) 3397–3415 (1993).
  2. 2. Chen S., Donoho D. and Saunders M. “Atomic Decomposition by Basis Pursuit”, SIAM Review, 129–159 (2001).
  3. 3. Mallat S., A Wavelet Tour of Signal Processing: The Sparse Way, Academic Press (2009).
  4. 4. Eldar Y., Kuppinger P. and Biölcskei H., “Block-Sparse Signals: Uncertainty Relations and Efficient Recovery”, IEEE Trans. Signal Process., 58, 3042–3054 (2010).
  5. 5. Candès E. and Wakin M., “An introduction to compressive sampling”, IEEE Signal Processing Magazine, 25, 21–30 (2008).
  6. 6. Romberg J., “Imaging via compressive sampling”, IEEE Signal Processing Magazine, 25, 14–20 (2008).
  7. 7. Baraniuk R., “More Is less: Signal processing and the data deluge”, Science 331, 717–719 (2011). pmid:21311012
  8. 8. Torkamani R., Zayyani H., and Sadeghzadeh R., “Model-based decentralized Bayesian algorithm for distributed compressed sensing”, Signal Processing: Image Communication, 95, (2021), 116212.
  9. 9. Wright J., Ma Yi, Mairal J., Sapiro G., Huang T.S., and Yan S., “Sparse Representation for Computer Vision and Pattern Recognition”, Proc. of the IEEE, 98, 1031–1044 (2010).
  10. 10. Elad M., Sparse and Redundant Representations: From Theory to Applications in Signal and Image Processing, Springer (2010).
  11. 11. Zhang Z., Xu Y., Yang J., Li X., and Zhang D., “A survey of sparse representation: algorithms and applications”, IEEE access, (2015).
  12. 12. Mairal J., Eldar M., and Sapiro G., “Sparse Representation for Color Image Restoration”, IEEE Trans. Image Proces., 17, 53–69 (2008). pmid:18229804
  13. 13. Dong W., Zhang L., Shi G., and Li X., “Nonlocally Centralized Sparse Representation for Image Restoration”, IEEE Trans. Image Proces., 22, 1620–1630 (2013). pmid:23269751
  14. 14. Su Zhenming, Zhu Simiao, Lv Xin, and Wan Yi, “Image restoration using structured sparse representation with a novel parametric data-adaptive transformation matrix”, Signal Processing: Image Communication, 52, 151–172 (2017).
  15. 15. Wright J., Yang A. Y., and Ganesh A., “Robust Face Recognition via Sparse Representation”, IEEE Trans. Pattern Analysis and Machine Intelligence, 31, 210–227 (2009). pmid:19110489
  16. 16. Yuan XT., Liu X., and Yan S., “Visual Classification With Multitask Joint Sparse Representation”, IEEE Trans. Image Proces., 21, 4349–4360 (2012). pmid:22736645
  17. 17. Heinsohn D., Villalobos E., Prieto L., and Mery D. “Face recognition in low-quality images using adaptive sparse representations”, Image and Vision Computing, 85, 46–58 (2019).
  18. 18. Raja K., Ramachandra R., and Busch C. “Collaborative representation of blur invariant deep sparse features for periocular recognition from smartphones”. Image and Vision Computing, 101, 46–58 (2020).
  19. 19. Yang J., Wright J., and Huang T., “Image Super-Resolution via Sparse Representation”, IEEE Trans. Image Proces., 19, 2861–2873 (2010). pmid:20483687
  20. 20. Zhang Y., Liu J., Yang W., and Guo Z., “Image Super-Resolution Based on Structure-Modulated Sparse Representation”, IEEE Trans. Image Proces., 9, 2797–2810 (2015). pmid:25966473
  21. 21. Zhang Chaopeng, Liu Weirong, Lui Jie, Liu Chaorong, and Shi Changhong, “Sparse representation and adaptive mixed samples regression for single image super-resolution”, Signal Processing: Image Communication, 67, 79–89 (2018).
  22. 22. Li Xuesong, Cao Guo, Zhang Youqiang, Shafique Ayesha, and Fu Peng, “Combining synthesis sparse with analysis sparse for single image super-resolution”, Signal Processing: Image Communication, 83 115805 (2020).
  23. 23. C. Caiafa F. and Cichocki A., “Computing sparse representations of multidimensional signals Using Kronecker Bases”, Neural computation, 25, 186–220 (2013). pmid:23020110
  24. 24. Cichocki A., Mandic D., De Lathauweri L., Zhou G., Zhao Q., Caiafai C., et al, “Tensor decompositions for signal processing applications: From two-way to multiway component analysis”, IEEE Signal Processing Magazine, 32, 145–163 (2015).
  25. 25. Dai Q., Yoo S., and Kappeler A. “Sparse Representation-Based Multiple Frame Video Super-Resolution”, IEEE Trans. Image Proces., 26, 2080–2095 (2017).
  26. 26. Mousavi H. S. and Monga V., “Sparsity-based Color Image Super Resolution via Exploiting Cross Channel Constraints”, IEEE Trans. Image Proces., 26, 5094–5106 (2017). pmid:28534773
  27. 27. Rebollo-Neira L. and Whitehouse D., “Sparse representation of 3D images for piecewise dimensionality reduction with high quality reconstruction”, Array, 1 (2019).
  28. 28. Rebollo-Neira L., Matiol R., and Bibi S., “Hierarchized block wise image approximation by greedy pursuit strategies”, IEEE Signal Process. Letters, 20, 1175–1178 (2013).
  29. 29. Rebollo-Neira L., “Effective sparse representation of X-Ray medical images”, International Journal for Numerical Methods in Biomedical Engineering, 33 e2886 (2017). pmid:28388837
  30. 30. Rebollo-Neira L., “A competitive scheme for storing sparse representation of X-Ray medical images”, PLoS ONE, 13(8): e0201455. (2018). pmid:30114219
  31. 31. Taubman D. S. and Marcellin M. W., JPEG2000 Image Compression, Fundamentals, Standards and Practice, Kluwer Academic Publishers (2002).
  32. 32. G. Schaefer and M. Stich, “UCID: an uncompressed color image database”, Proceedings Volume 5307, Storage and Retrieval Methods and Applications for Multimedia 2004; (2003).
  33. 33. https://www2.eecs.berkeley.edu/Research/Projects/CS/vision/bsds/ (Last access December 2022).
  34. 34. Fisher R.A., “Frequency distribution of the values of the correlation coefficient in samples from an indefinitely large population”, Biometrika, 10, 507–521 (1915).
  35. 35. Fisher R.A., “On the probable error of a coefficient of correlation deduced from small sample”, Metron, 1, 3–32 (1921).
  36. 36. Gershikov E. and Porat M., “On color transforms and bit allocation for optimal subband image compression”, Signal Processing: Image Communication, 22, 1–18 (2007).
  37. 37. Tošić I. and Frossard P., “Dictionary Learning: What is the right representation for my signal?”, IEEE Signal Processesing Magazine, 28, 27–38 (2011).
  38. 38. Zepeda J., Guillemot C., and Kijak E., “Image Compression Using Sparse Representations and the Iteration-Tuned and Aligned Dictionary”, IEEE Journal of Selected Topics in Signal Processing, 5,1061–1073 (2011).
  39. 39. Srinivas M., Naidu R. R., Sastry C.S., and Krishna Mohana C., “Content based medical image retrieval using dictionary learning”, Neurocomputing, 168, 880–895 (2015).
  40. 40. Wen B., Ravishankar S., and Bresler Y., “‘Structured Overcomplete Sparsifying Transform Learning with Convergence Guarantees and Applications”, International Journal of Computer Vision, 114, 137–167 (2015).
  41. 41. Rebollo-Neira L., “Cooperative greedy pursuit strategies for sparse signal representation by partitioning”, Signal Processing, 125, 365–375 (2016).
  42. 42. Pati Y.C., Rezaiifar R., and Krishnaprasad P.S., “Orthogonal matching pursuit: recursive function approximation with applications to wavelet decomposition”, Proc. of the 27th ACSSC,1,40–44 (1993).
  43. 43. Rebollo-Neira L., Bowley J., Constantinides A. and Plastino A., “Self contained encrypted image folding”, Physica A, 391, 5858–5870 (2012).
  44. 44. Rebollo-Neira L. and Lowe D., “Optimized orthogonal matching pursuit approach”, IEEE Signal Process. Letters, 9, 137–140 (2002).
  45. 45. Rebollo-Neira L., Rozložník M., and Sasmal P., “Analysis of the Self Projected Matching Pursuit Algorithm”, Journal of The Franklin Institute 8980–8994 (2020).
  46. 46. K. Skretting, https://www.mathworks.com/matlabcentral/fileexchange/2818-huffman-coding-and-arithmetic-codin (Last access December 2022).
  47. 47. http://www.nonlinear-approx.info/examples/node015.html (Last access December 2022).
  48. 48. https://www.hlevkin.com/hlevkin/06testimages.htm (Last access December 2022).
  49. 49. https://imagecompression.info/test_images (Last access December 2022).
  50. 50. https://developers.google.com/speed/webp (Last access December 2022).
  51. 51. Wang Z., Bovik A. C., Sheikh H. R., and Simoncelli E. P., “Image quality assessment: From error visibility to structural similarity”, IEEE Trans. Image Process., 13, 600–612 (2004). pmid:15376593