and limits.
Once again, image indexes are presented to the
look-up table created with MSCL
PICT
(l) (according
to the block size) that returns the block shape. Fi-
nal image is regenerated by adding means of block-
saliency, masking each block and positioning it in the
image (adding it to the regenerated image as we had
done before with the saliency map).
3.6 Extension to Color Images
Figure 3 defines the flowchart to use MSCL in the
case of color images. The process is similar to the
used in the case of grayscale images, but applied to
each of the color components of the image.
First, we calculate the saliency map from the color
image. With this saliency map we extract and quan-
tify blocks as described in subsection 3.1, blocks
which were restored at transmitter as mentioned in
3.2. As a result of this step we get the map block-
centers, block-means and indexes. Encoding is made
with the previously trained MSCL
PICT
(l) picture li-
brary.
Then, original RGB image is transformed to the
L-a-b color space. The reason of selecting this color
codification is that it has been demonstrated its suit-
ability for interpreting the real world (Cheung, 2005).
Now with these L-a-b color components of the im-
age, we follow the process indicated in 3.3. Each
of them will be trained with a MSCL neural net-
work (MSCL
IC−L
, MSCL
IC−a
, MSCL
IC−b
,) and it will
return the block sizes and indexes for each compo-
nent. The indexes of the blocks are also encoded with
MSCL
PICT
(l).
Once at receiver saliency map is restored (see
3.4). Then, we follow the image restoration
step, applied to each L-a-b component. Its cen-
ters are calculated training three MSCL networks
(MSCL
IC2−L
, MSCL
IC2−a
, MSCL
IC2−b
,), with the co-
ordinates of each pixel, and the regenerated saliency
map. These neural networks becomes identical to
those at the transmitter.
To get the final image, we transform the restored
L-a-b image to RGB.
4 EXPERIMENTAL RESULTS
4.1 Grayscale Images
Simulations were conducted on four 256x256 gray
scaled images (65536 bytes), all of them are typical
in image compression benchmarking tasks.
We applied the MSIC algorithm, with the follow-
ing MSCL training parameters: 15 cycles and learn-
ing factor varying along the training process from
0.9 to 0.05. We used Graph-Based Visual Saliency
GBV S(x) ( (Harel et al., 2006) ) as the pixel saliency
of the corresponding sample. However, it is possible
to use other kind of magnitudes to define which areas
of the image are compressed more or less deeply.
JPEG was done with the standard Matlab imple-
mentation and a compression Quality of Q = 3 or Q=5
(i.e., with a high compression ratio).
We also compare with the algorithm described in
(Liou, 2007), whose main steps are followed for all
the mentioned SOM based algorithms: The original
image is divided into small blocks (we select a size
of 8x8 to achieve a similar compression ratio to JPEG
or MSCL). The 2-D DCT is first performed on each
block. The DC term is directly send for reconstruc-
tion. The AC terms after low-pass filtering (we only
consider 8 AC coefficients) is fed to a SOM network
for training or testing. All experiments were carried
out with the following parameters: 256 units, 5 train-
ing cycles and learning factor decreasing from 0.9 to
0.05.
The number of bytes used to compress each image
was the same for MSCL and JPEG (see table 1) and
fixed to 2048 for SOM.
For evaluation purpose, we use the mean squared
error (MSE) as an objective measurement for the per-
formance. Table 1 shows the resulting mean of the
MSE in 10 tests using our algorithm compared to
JPEG and SOM applied to 4 test images. We present
a second column showing the value of MSE but only
calculated in those pixels which saliency is over 50%.
Standard deviation is also shown (in brackets).
To obtain the generic pictorial library
MSCL
PICT
(l) we used three additional images
from the (Computer Vision Group, 2002) with the
same training parameters. This number is quite low,
but enough to show the good performance of our
proposal. However, in a real scenario it would be
necessary to use a higher number of images to get a
suitable pictorial library. Moreover, we have not used
any entropic coding applied to indexes which would
have result in a further compression.
As expected, the MSE value calculated for the
whole image area given by JPEG is lower than the one
provided by MSIC, because prototypes tend to focus
on zones with high saliency while other areas in the
image are under-represented.
However, when MSE was calculated taking into
account only those pixels with high saliency, MSIC
obtained better results than JPEG or SOM. This ef-
fect can be clearly appreciated by visual inspection of
MagnitudeSensitiveImageCompression
375