cryption. Empirically, it seems that the DC coefficient
is a good candidate for encryption and authentication,
and generally leads to a good compromise between
computation overhead and security. However, to the
best of our knowledge, there are neither theoretical
nor experimental results to quantify how much infor-
mation can be guaranteed or hindered by authenticat-
ing or encrypting the DC coefficient. Without such
results, it is difficult for designers to decide whether
to adopt similar approaches. Motivated by this fact,
we would like to investigate the above problem in a
quantitative way and measure the actual performance
of DC coefficient encryption and authentication.
In this work, the evaluation of encryption and au-
thentication performance is considered as an image
quality evaluation problem. An image quality met-
ric called “structural similarity” (Wang et al., 2004)
is adopted to measure the amount of information hin-
dered by partial encryption or guaranteed by partial
authentication. From various experiments applied to
an image database of about 420 JPEG images, we
show that by authenticating DC coefficients, about
60% of the information can be guaranteed; by en-
crypting DC coefficients, about 80% of the informa-
tion can be hindered. Although these are not absolute
results, they might give insights into the problem.
The rest of the work is organized as follows: Sec-
tion 2 first introduces our approach to evaluate en-
cryption and authentication performance by structural
similarity; Subsection 2.1 gives the background of
partial image authentication and results of various ex-
periments in terms of structural similarity; Subsec-
tion 2.2 is a similar approach to partial image encryp-
tion. Section 3 concludes our work.
2 BENCHMARKING DC
COEFFICIENT ENCRYPTION
AND AUTHENTICATION
Although encryption and authentication are quite dif-
ferent, a uniform approach can be used to evaluate
their performance for visual data. Because they both
consider how much information is preserved after the
operation, one could think of them as an image qual-
ity problem. Assuming that the original image has full
quality, for partial encryption, we measure the quality
of the encrypted version compared to the original one;
for partial authentication, we measure the quality of
the authenticated part compared to the original one.
Therefore, the problem is converted into the choice of
a proper image quality metric.
The most widely used image quality metrics are
the peak signal-to-noise ratio (PSNR) and the mean
square error (MSE). They are simple and efficient for
general purposes. However, “they are not very well
matched to perceived visual quality” (Wang et al.,
2004), because they only concentrate on the amount
of errors, but not the perceived information. For these
metrics, image quality measure is quite different from
the amount of information expressed through the im-
age. For example, Figure 1 shows a gray-scale Lena
image (a) and several distorted versions: (b) subtract-
ing a constant from all pixel values and setting neg-
ative results to zero; (c) applying an averaging filter;
(d) JPEG compression. The PSNRs of the distorted
versions, compared to the original one, are given be-
low the images. Although the distorted versions look
rather similar to the original one, the PSNRs are quite
low. One can also note that the PSNR of (b) is lower
than the one of (c), while (b) actually has better per-
ceptual quality. Therefore, it might not be appropriate
to use simple metrics that measure the error visibility,
such as PSNR and MSE; instead, we need a metric
which indeed measures image similarity.
We find that the image quality metric structural
similarity (SSIM) (Wang et al., 2004) fulfills the re-
quirement. This metric compares luminance, con-
trast, and structure information between two gray-
scale images of the same size and returns an average
score between zero and one, with one meaning ex-
actly the same and zero meaning completely different.
It is defined as:
SSIM(x, y) = [l(x, y)]
α
· [c(x, y)]
β
· [s(x, y)]
γ
, (1)
where x and y represent two test images; functions
l(), c(), and s() correspond to luminance, contrast,
and structure similarity, respectively; α, β, and γ are
weighting factors. In our experiments, we use its sim-
plified form:
SSIM(x, y) =
(2µ
x
µ
y
+C
1
)(2σ
xy
+C
2
)
(µ
2
x
+ µ
2
y
+C
1
)(σ
2
x
+ σ
2
y
+C
2
)
, (2)
where µ represents mean; σ represents (co)variance;
C
1
and C
2
are numerical constants for stability (we
use 6.5025 and 58.5225 as suggested). This metric
satisfies the following conditions:
• Symmetry: SSIM(x, y) = SSIM(y, x);
• Boundedness: SSIM(x, y) ≤ 1;
• Unique maximum: SSIM(x, y) = 1 iff x = y.
These properties make it easy to interpret the meaning
of an SSIM score. Due to space constraints, we skip
elaboration on the details of this metric. For more
information, one can refer to the paper (Wang et al.,
2004). Although this metric does not cover all aspects
of image similarity measure, it gives more reasonable
SIGMAP 2007 - International Conference on Signal Processing and Multimedia Applications
376