• 沒有找到結果。

Watermarking approach to embedded signature-based authentication by channel statistics

N/A
N/A
Protected

Academic year: 2021

Share "Watermarking approach to embedded signature-based authentication by channel statistics"

Copied!
9
0
0

加載中.... (立即查看全文)

全文

(1)

Watermarking approach to embedded

signature-based authentication by channel statistics

Zhi-Fang Yang Wen-Hsiang Tsai

National Chiao Tung University

Department of Computer and Information Science

Hsinchu, Taiwan 300 Republic of China

E-mail: gis82504@cis.nctu.edu.tw whtsai@cis.nctu.edu.tw

Abstract. Embedded signature-based authentication is a highly promis-ing watermarkpromis-ing method for solvpromis-ing high-level authentication problems. This work proposes such a watermarking approach using channel sta-tistics evaluated from a reference watermark. The basic concept is to consider the problem to be an example of communications with side information, which is the reference watermark embedded in the host with the signature at the sender site. At the receiver site, the reference wa-termark is first extracted to determine channel statistics, which are used to generate information about signature reliability. Three kinds of infor-mation are derived, including a reliability measure, an entropy measure, and an authenticated signature. Experimental results establish that the measures of reliability and uncertainty are meaningful, and that the em-bedded signature can survive high-quality JPEG compression and ma-nipulations such as negation, cropping, replacement, and modification. ©2003 Society of Photo-Optical Instrumentation Engineers.

[DOI: 10.1117/1.1556394]

Subject terms: watermarking; authentication; embedded signature; channel sta-tistics; reliability; uncertainty.

Paper 020226 received Jun. 5, 2002; revised manuscript received Oct. 9, 2002; accepted for publication Oct. 10, 2002.

1 Introduction

Currently, a mass of multimedia data can be produced and rapidly transmitted using modern digital technologies. Naturally, solutions to digital copyright protection are re-quired urgently to tackle the problem of unauthorized copy-ing and distribution. The research field of copyright mark-ing, concerned with inserting copyright information into host documents, has thus attracted extensive interest in the last decade.1,2

In the early 1990s, the methods proposed for copyright marking involved the property of robustness, and robust watermarking algorithms were extensively studied.1–3 The aim was to make the destruction of embedded watermarks impractical without severely degrading visual quality. Re-cently, attention has been switched from robust watermark-ing to fragile watermarkwatermark-ing,3 in which the main aim is to authenticate the host using fragile watermarks.

With respect to watermarking-based authentication, ap-proaches developed so far fall into two main categories, namely fragile watermarking methods and robust watermarking-based methods.3 Fragile watermarking is a more mature field, since it was studied earlier in the history of visual authentication.4Whether or not data is modified, a ‘‘yes-or-no’’ question can be answered by fragile water-marking. However, to distinguish different types of distor-tions, the high-level authentication problem is more desir-able for authenticating multimedia data. In Ref. 5, observing modifications of a fragile watermark in different localized spatial and frequency regions make solving the problem possible. But sometimes it is not easy to distin-guish new kinds of distortions based on preselected fea-tures, especially when only semantic meanings of

distor-tions are described. A similar viewpoint can also be found in Ref. 3, in which robust watermarking is recommended and the watermark is suggested to be a low-resolution im-age of the host. Generally, the watermark in the second category is a signature of the host embedded in the host at the sender site and can be extracted to authenticate the host at the receiver site.

Research into this topic is still in its infancy.3 Related work6 – 8has revealed two problems particularly worthy of investigation.3 One concerns the accuracy of the informa-tion used by the signature in summarizing the host. The other concerns the accuracy of the extracted signature at the receiver site. The former is an issue of signature design, and the latter is one of robustness. Generally, more data contained in the signature yield greater accuracy. However, more data hidden in the host yield lower robustness. That is, a tradeoff exists between these two kinds of accuracy. Based on previous work, a compromise involves finding how many data can be reliably embedded in the host; this problem is called the capacity problem.3However, the ca-pacity problem remains unsolved3,9 because of a lack of ability to model arbitrary manipulations that might occur during transmission or other applications.

This work uses a reference watermark to derive channel statistics to address the issue of the accuracy of the embed-ded signature. The basic concept is to view the problem as one of communications with side information,14 which is designed to be the reference watermark embedded in the host with the signature. The underlying motivation of this approach is as follows. Comparing accurate original data with received data is the key to successful authentication. However, the extracted signature just cannot be treated as

(2)

an accurate one because the capacity problem, namely, how many bits can be reliably embedded, is still open. In this study, the original reference watermark is available at the receiver site. That is, authentication of the reference water-mark is guaranteed. The authentication results of the refer-ence watermark are thus used to generate information about signature reliability. In this study, three kinds of informa-tion are provided, including a reliability measure, an en-tropy measure, and an authenticated signature. The former two measures are used to inform users at the receiver site about some quantitative measures regarding the reliability and uncertainty of the transmitted image.

Research on using reference or pilot watermarks to de-rive channel statistics can also be found in Refs. 10–12. In Voloshynovskiy et al.,10 a key-dependent reference pilot watermark is used to estimate channel state, the noise dis-tribution, after attack. Their goal is to design an optimal decoder for images under attacks, including geometrical at-tacks and additive non-Gaussian noise. In Kundur and Hatzinakos,11 a reference sequence is utilized for channel estimation that is used to adjust the receiver filter to maxi-mize detection reliability of the watermark. In addition, a new watermarking channel, a nonstationary parallel binary symmetric channel共BSC兲 model, is introduced. In Ba¨uml, Eggers, and Huber12 a securely embedded pilot signal is exploited to estimate the transformation of the sampling grid. The channel estimate is then used to invert the desyn-chronization attack before applying the Scalar Costa Scheme.

The rest of this work is organized as follows. Section 2 formulates the problem to explain how robust watermarking-based authentication can be viewed as com-munications with side information. The computation of channel statistics is also discussed. Section 3 describes the proposed approach. Section 4 gives some experimental re-sults and related discussion. Conclusions and recommenda-tions for future work are finally made in Sec. 5.

2 Authentication as Communications with Side Information

According to Fig. 1, a communications system encodes a message at the sender site, sends it through a channel, and decodes it at the receiver site. The side information is used by both the encoder and the decoder to improve the trans-mission rate or enhance the detectability of the signature embedded in the message.

Figure 2 illustrates the proposed approach. The signature

is embedded in the host data with a reference watermark. After transmission through the channel, the embedded ref-erence watermark is extracted and used to determinethe channel statistics. The signature is then detected and used to authenticate the transmitted host data.

In this work, the side information, which is a reference watermark, is known by both the watermark embedder at the sender site and the watermark extractor at the receiver site. The channel statistics are provided to the signature extractor using the side information. This mapping natu-rally leads to an example of communications models with side information, as shown in Fig. 3.

The concept of the proposed approach is thus clarified: a reference watermark is used as side information to derive channel statistics that are then used to reliably extract the signature to enable authentication. Two assumptions are made without a loss of generality. The first is that manipu-lations in the channel cannot distinguish the embedded sig-nature and the embedded reference watermark in the host data. The second is that the manipulations on the host are locally consistent. Thus, local channel statistics can be ana-lyzed by both a designed reference watermark at the sender site and the corresponding extracted reference watermark at the receiver site.

3 Proposed Approach

3.1 Watermark Design and Embedding

The watermark is composed of a signature and a reference watermark. The total size cannot be so large as to destroy the transparency of the image. Moreover, the number of reference watermark bits cannot be too few with respect to the number of signature bits; otherwise, the computed channel statistics are unreliable in the local area. A binary bit stream is used as the reference watermark. It is used to detect channel statistics based on the two assumptions con-cerning manipulations stated before. The signature is gen-erated by thresholding the host image according to a thresh-old preselected manually for each individual host image. That is, a set of threshold candidates is tested for a specific host image. The threshold candidate with the clearest sig-nature is selected according to the subjective judgment of the user. Doing so is an attempt to produce a distinct sig-nature for human visibility. Such a design avoids the diffi-culty of summarizing information contained in the host data. Other kinds of signatures are also suitable.

After being created, the signature and the reference wa-termark are randomly permuted using a permutation key to increase security and distribute the reference watermark Fig. 1 A communications model with side information.

Fig. 2 The model of the proposed approach.

Fig. 3 The proposed approach viewed as communications with side information.

(3)

bits as uniformly as possible. The permutation includes two stages. In the first stage, the signature is randomly per-muted alone. After the permutation, two equally long bit streams of zeros and ones, the reference watermark, are appended to the permuted signature bit stream. Then, the three kinds of bit streams, the permuted signature, the ze-ros, and the ones, are divided into equal numbers of sub-streams. Substreams with the same order of corresponding bit streams are treated as a subgroup. In the second stage, all the subgroups are permuted, respectively. The final com-plete watermark is created when all permutations are per-formed.

The embedding method is basically the quantization-based method developed by Kundur and Hatzinakos11: the host image is transformed using a wavelet transform, and the watermark is embedded bit by bit in the randomly cho-sen detail coefficients. Two modifications are made to im-prove robustness: one is not to embed watermark bits at the highest resolution; the other is to set the quantization pa-rameter as a constant. This study suggests that if many wavelet levels are applied, the constant quantization param-eter at a middle resolution, as computed in Ref. 5, is se-lected. However, if fewer levels are applied, the selected constant quantization parameter is the one calculated in Ref. 5 at the lowest resolution. Experiments performed for this study reveal that if the number of levels is five, the selected quantization parameter should be 23; if the num-ber of levels is three, the selected quantization parameter should also be 23. Briefly, the embedding method is de-signed to be insensitive to smaller modifications and to pre-vent the host from severe visual distortion.

The following rules summarize the proposed embedding method. Notation, statements, and rules are consistent with those developed in Ref. 5. For each watermark bit w(i), a wavelet coefficient fk,l(m,n) is randomly selected, where

k⫽h, v, or d, which stand for horizontal, vertical, and

diagonal detail coefficients, respectively; l⫽1,2,...,L and is a specific resolution, and (m,n) specifies a spatial location. Notably, each wavelet coefficient can be selected only once to prevent the production of errors by repeatedly embed-ding one coefficient. The embedembed-ding rules are as follows:

fk,l共m,n兲 ª

fk,l共m,n兲⫹⌬ if fk,l共m,n兲⭐0 and Q关 fk,l共m,n兲兴⫽w共i兲 fk,l共m,n兲⫺⌬ if fk,l共m,n兲⭓0 and Q关 fk,l共m,n兲兴⫽w共i兲 fk,l共m,n兲 if Q关 fk,l共m,n兲兴⫽w共i兲 , 共1兲 where ª is the assignment operator; ⌬ represents the unique quantization parameter discussed in this section; and Q(.) is a quantization function which is defined as follows: Q关 fk,l共m,n兲兴⫽

0 if

b

fk,l共m,n兲

c

is an even number 1 otherwise . 共2兲

3.2 Watermark Extraction and Authentication At the receiver site, localizing the watermark is rather simple by the coefficient selection and permutation keys. The transmitted host image is first transformed into the pre-defined wavelet domain. Then the coefficient selection key is used to locate the wavelet coefficient positions in which the watermark bits are embedded. Later, the permutation key is used to separate the embedded signature bits from the embedded reference watermark bits.

The channel statistics are determined by analyzing the reference watermark bits in a predefined neighborhood of each signature bit. In this study, the neighborhood is an N ⫻N square centered at the point in which the signature bit is embedded in the host image. The size of the neighbor-hood should contain sufficient reference watermark bits to make the statistics meaningful. Furthermore, the neighbor-hood should not be so large that it violates the local statis-tics of the channel condition. Therefore, the size of the neighborhood should be chosen appropriately. In this work, experiments show that a 21⫻21 neighborhood is suited to analyze a 512⫻512 host image. Conditional probabilities

pi兩 j 共where i is the value of the extracted reference

water-mark bit and j is the value of the corresponding original reference watermark bit兲 are evaluated according to the fol-lowing equation after the reference watermark bits are lo-cated in the neighborhood ofa specific signature bit b:

pi兩 j

兺1⭐p,q⭐N共1⫺ep,qi兲共1⫺rp,qj

兺1⭐p,q⭐N共1⫺rp,qj

for i, j苸兵0,1其, 共3兲 where丣 is the XOR operator; ep,q represents an extracted

reference watermark bit at position ( p,q) in the N⫻N neighborhood, and rp,q represents the original reference

watermark bit embedded at the same position at the sender site. Four kinds of conditional probabilities, p0兩0, p0兩1,

p1兩0, and p1兩1, are computed, and correspond to pairs of the extracted and original bits共0, 0兲, 共0, 1兲, 共1, 0兲, and 共1, 1兲, respectively. Notably, Eq. 共3兲 can be easily extended to the case of more signature bits being considered simulta-neously.

Since the a priori probabilities Pj of the classes Cj, which denote the classes of the signature bits with the value

j苸兵0,1其, are unknown at the receiver site in this study, they are set to be identical, without a loss of generality. That is, P0 and P1 are both 0.5. Then, the estimated bit value b

of the extracted signature bit b is determined us-ing the Bayes theorem13with an additional constraint:

b⬘⫽

arg max j苸兵0,1其 pb兩 jPj 兺0⭐k⭐1pb兩kPk for max j苸兵0,1其 pb兩 jPj 兺0⭐k⭐1pb兩kPk⬎T unreliable otherwise , 共4兲 where T is the threshold of the additional constraint:

pb兩 jPj

兺0⭐k⭐1pb兩kPk

(4)

This constraint makes the maximum a posteriori prob-ability much larger than the other a posteriori probprob-ability. The extracted signature bit b

is classified as ‘‘reliable’’ if one of the a posteriori probabilities passes the test; other-wise, b

is classified as ‘‘unreliable.’’ The concept is that the computed value may not be sufficiently representative if the difference between the two a posteriori probabilities is insignificant. In this work, T is suggested to be 0.8, follow-ing the experimental results. Now, every bit of the signature can be estimated from the channel statistics. A reliability measure, denoted as R, of the extracted signature is then defined as the following statistical value:

R

兺1⭐i⭐Ls共1⫺bi

bi

Ls

for bi

is in the ‘‘reliable’’ class, 共5兲 where Ls is the length of the signature stream. In Eq.共5兲,

the number of the extracted signature bits classified as ‘‘re-liable’’ is divided by the signature length Ls. This percent-age measure R is then used to represent the reliability of the extracted signature. The uncertainty of the corresponding channel statistics of the extracted signature is defined as the entropy H. The entropy H of an extracted signature is de-termined as follows:

H

兺1⭐i⭐Ls兺0⭐i, j⭐1共⫺pi, jlog2pi, j

Ls

, 共6兲

where pi, j are the joint probabilities determined from the

following equation:

pi, j⫽ 兺1⭐p,q⭐N共1⫺ep,qi兲共1⫺rp,qj

兺1⭐p,q⭐N关共1⫺rp,q丣0兲⫹共1⫺rp,q丣1兲兴

for i, j苸兵0,1其, 共7兲

where ep,qand rp,qrepresent an extracted reference

water-mark bit and the corresponding correct bit at position ( p,q) in the tested N⫻N neighborhood, respectively. This en-tropy measure H reflects the uncertainty of the channel in the local area. Finally, the watermarked image is compared to the extracted signature by visual judgments, by typical matching techniques, or by ad hoc approaches, to classify various manipulations. In this work, for simplicity, visual judgments are adopted in experiments about authentication and classification of various manipulations.

4 Experimental Results and Discussion

The experiments involve evaluating the reliability of the extracted signature and demonstrating the authentication ef-fectiveness of the proposed approach. First, we evaluate the reliability measure R, computed from Eq. 共5兲. More spe-cifically, the most interesting question is, ‘‘Is the measure of reliability based on channel statistics R itself reliable?’’ Two experiments were conducted to answer the question. One compared the reliability measure R to the hit ratio of the corresponding extracted signature bits. The hit ratio A of an extracted signature is determined as follows:

A

兺1⭐i⭐Ls共1⫺bi

bi o 兺1⭐i⭐Lsbi

for bi

is classified into the ‘‘reliable’’ class, 共8兲

where bi

is a reliable bit and the i’th bit of the extracted signature, and bio is the i’th bit of the original signature stream. The hit ratio A is the percentage of the reliable signature bits that arecorrect. The signature bits detected as reliable can be confidently held to contain a large percent-age of correct bits if a reliability measure R appears with a large hit ratio. That is, the reliable signature bits can be trusted to perform authentication under such circumstances. The other experiment concerning the evaluation of reli-ability compared the relireli-ability measure R to the uncer-tainty of the corresponding channel statistics H. As is well known, entropy represents uncertainty in information; larger entropy values imply greater uncertainty of informa-tion. Accordingly, when a reliability measure R appears with a low entropy value, little uncertainty surrounds the detected signature bits that are recognized as reliable. That is, the values of the reliable signature bits can be believed. The experiments concerning authentication were per-formed with the aim of demonstrating some authentication abilities of the proposed method. In this work, four kinds of manipulations are considered—negation, cropping, replace-ment, and modification. The desired aim is to extract rect signatures from transmitted images. Therefore, the cor-rect signatures can be used to authenticate visually whether the transmitted images are attacked by the four manipula-tions. The following paragraph introduces parameters used in the experiments.

The experiments were performed on monochrome im-ages with a size of 512⫻512. Three images were used— Airplane, Lena, and Baboon—and were selected to repre-sent three kinds of images: those containing large smooth areas, those containing both smooth and detailed areas, and those containing huge amounts of details. The signatures are generated by thresholding the monochrome images to black and white ones and then reducing them to a size of 128⫻128 by selecting the first pixel of every four pixels in a row from top to bottom, left to right. In the experiments, a gray-level threshold of 128 was used for the images Lena and Baboon. A gray-level threshold of 185 was selected for the image Airplane. The Haar wavelet transform is used and the resolution levels are chosen as three. The length of the reference watermark was set to a quarter of the signa-ture size in these experiments. After the permutations speci-fied in Sec. 3.1 are applied, the watermark is embedded into the transformed images using a unique quantization param-eter, 23, according to the method presented in Ref. 5. The PSNRs, which are calculated from the formula given in Ref. 5, of the embedded images in the experiments are all between 40 and 50. At the receiver site, the watermark is extracted from the transmitted images using the permuta-tion keys, following what is described in Sec. 3.2. And the channel statistics are analyzed in a 21⫻21 neighborhood around each signature bit position. The threshold T is set at 0.8 in the experiments. Then, the value of each signature bit can be computed from the corresponding channel statistics.

(5)

Finally, the computed signatures are visually compared with the transmitted images to perform the authentication.

With respect to the properties of the reliability measure

R, a series of JPEG operations with various compression

ratios are used. These JPEG images are generated by ad-justing parameters pertaining to functions of saving images provided by the software, MATLAB. In the experiments, the images are compressed with lost data percentages be-tween 3 and 98%, which is the ratio of the size difference between an original image and its corresponding JPEG im-age to the size of the original imim-age. These imim-ages exhibit various degrees of distortion, as therefore do the embedded signatures. Figure 4 shows some examples of extracted sig-natures from these JPEG images. In the figure, the reliable signature bits are represented by black and white pixels, which represent the bit value zero and the bit value one, respectively. The unreliable signature bits are represented by gray pixels.

Figure 5 compares the measures of reliability with hit ratios. Larger reliability measures are associated with larger hit ratios for all three images. Therefore, such a reliability

measure based on channel statistics is meaningful. In par-ticular, the hit ratios exceed 90% when the reliability ex-ceeds 0.5, and under such circumstances, the lost data per-centages can be 60% for JPEG images of Airplane and Lena and 38% for the Baboon image. The reliability mea-sures approach zero when the lost data percentages increase to a high value. That is, the distortion caused by JPEG compression so severely damaged the images that very few or no extracted signature bits can be considered to be reli-able. Hit ratios also reflect the same situation: the hit ratios are either smaller than 0.5 or are 0 under such circum-stances. Notably, the values of the reliable signature bits are random if the corresponding hit ratio is one half. The num-bers of zeros in the hit ratios of the three images are quite different. The largest number of zeros appears in the Ba-boon image. This result is because the image BaBa-boon in-cludes many details, which are more easily distorted by JPEG compression, than those in the other two images. Under such circumstances, channel statistics tend to have equal shares and then the extracted signature bits are judged as unreliable. Therefore, the corresponding hit ratio Fig. 4 Examples of signatures extracted from JPEG images used in experiments on hit ratios and

entropies. The reliable extracted signature bits are represented by black and white pixels, which indicate the bit values 0 and 1, respectively. The unreliable signature bits are represented by gray pixels. Three images, Airplane, Lena, and Baboon, are shown for different reliability measuresRand entropy measuresH. For the Airplane image, (a) original; (b)R⫽0.52 andH⫽1.24; (c)R⫽0.43 and

H1.72; (d)R0.24 andH1.78; and (e)R0.03 andH1.87. For the Lena image, (f) original; (g)

R0.54 andH1.44; (h)R0.5 and H1.69; (i)R0.33 andH1.75; and (j)R0.03 andH

⫽1.86. For the Baboon image, (k) original; (l)R⫽0.55 andH⫽1.27; (m)R⫽0.5 andH⫽1.67; (n)R

(6)

and its reliability measure are both zero. With respect to the image of Lena, which has fewer details than the Baboon image, the hit ratio has fewer zeros than the Baboon image. The image Airplane, which is smooth in most areas, con-tained no zero hit ratios in the experiments.

Under sever image distortion, the values of the extracted reference bits and the signature bits are unpredictable 共we may consider in this case that p0兩0⬃p0兩1⬃p1兩0⬃p1兩1) and the conclusion about the reliability of these bits will be in most of the cases unreliable. Then users will know that this transmission is not unreliable. That is, the unreliable mation is also a kind of information. Note that such infor-mation cannot be provided by using an embedded signature alone.

Figure 6 compares the measures of reliability with the entropy of the signatures. Notably, larger reliability mea-sures accompany lower entropy in the experimental results;

that is, the informational uncertainty is quite low for those computed reliable signature bits. When the reliability mea-sures are near zero, the corresponding entropy values are rather high共the maximum entropy value is 2 in the experi-ments兲, which implies that the uncertainty of the informa-tion is high. Therefore, the channel statistics-based reliabil-ity measure is meaningful. Moreover, the curve shapes of the entropy values are rather symmetrical; that is, entropy is low at both smaller and larger degrees of distortion. There-fore, the channel statistics reveal high certainty in these two situations.

Figures 7–12 show experimental results concerning au-thentication. For instance, Fig. 7共a兲 is the original host im-age. After the watermark is embedded, the host image is manipulated by the negation operation. As shown in Fig. 7共b兲, the airplane seems to be imaged at night. Now, the power of the approach based on channel statistics can be Fig. 5 Experimental comparisons between reliability measureRand hit ratioAobtained by testing a

(7)

illustrated as follows. In Fig. 7共c兲, the extracted signature is obtained by directly retrieving the bit values stored in the signature positions, and the airplane still looks as if it is flying at night. However, after the proposed channel-statistics-based approach is used to detect the embedded signature, the result correctly reveals the airplane actually flies by day. Figures 8 and 9 show similar situations. All three images give successful results.

The proposed approach should now be compared to the conventional approach of embedding the signature

repeat-edly without any side information such as channel statis-tics. Two points are worthy of note. The first arises from experiments on negation attack. The traditional repeating method clearly cannot survive such an attack, since no knowledge is available concerning what happens in the transmission. Sometimes, a conventional voting mechanism can be used to guess the transmission conditions. However, voting-based methods fail if the negation manipulation is performed. Under such circumstances, the approach based on channel statistics is superior, because the channel statis-Fig. 6 Experimental comparisons between reliability measureRand entropyH, obtained by testing a

(8)

tics can be used to collect information about transmission quality. The second point is that the repeating method uses more space than the proposed approach. As described in this section, the reference watermark size is only one quar-ter of the signature size. However, the conventional repeat-ing method uses double the signature size even if the sig-nature is embedded only twice.

Other kinds of attacks are also examined, including cropping, replacement, and modification. In Fig. 10, the top wing on the rear of the airplane is cropped after the water-mark is embedded. Figure 10共b兲 shows the cropped image. However, the extracted signature presents the original pic-ture of the airplane, which is different from the transmitted one. Thus, the authentication can be performed success-fully. In Fig. 11共b兲, the original white hat is replaced by a black one. The replacement manipulation can be detected by comparing the attacked image with the black hat to and the extracted signature with a white hat. The areas near the nose of the baboon in Fig. 12共a兲 are modified into inter-laced bands of black and white, as shown in Fig. 12共b兲. By comparing the attacked image in Fig. 12共b兲 with the ex-tracted signature in Fig. 12共c兲, the original baboon is shown to differ from the transmitted image in Fig. 12共c兲. There-fore, the authentication has also been performed success-fully.

5 Conclusions

Authentication based on embedded signatures constitutes a growing field of research. In previous work, robust watermarking-based methods were designed to tackle this problem by focusing on the robustness of embedding sig-natures and generating efficient sigsig-natures. This study seeks to develop an approach based on channel statistics, which gives useful information for authentication, including a measure of reliability of the extracted signature and a mea-Fig. 7 Example of negation attack: (a) the original host image

Air-plane; (b) negation of the watermarked image; (c) signature directly extracted without applying the proposed mechanism; and (d) signa-ture extracted by the proposed approach.

Fig. 8 Example of negation attack: (a) the original host image Lena; (b) negation of the watermarked image; (c) signature directly ex-tracted without applying the proposed mechanism; and (d) signature extracted by the proposed approach.

Fig. 9 Example of negation attack: (a) the original host image Ba-boon; (b) negation of the watermarked image; (c) signature directly extracted without applying the proposed mechanism; and (d) signa-ture extracted by the proposed approach.

Fig. 10 Example of cropping attack: (a) original host image; (b) at-tacked image (the top rear wing is cropped from the watermarked image); and (c) extracted signature showing the top rear wing.

Fig. 11 Example of replacement attack: (a) original host image; (b) attacked image (the hat is replaced by a black one in the water-marked image); and (c) extracted signature showing the original hat color as white.

(9)

sure of the uncertainty of the channel status. The goal is to extract reliable information based on the transmission qual-ity at the receiver site rather than to consider the problem of robustness at the sender site. The watermarking-based au-thentication problem is thus viewed as an example of com-munications with side information. The important point is to use a reference watermark as side information to derive channel statistics.

The experimental results concerning hit ratios and un-certainties have shown that such a method based on chan-nel statistics can be used to describe reasonably the chanchan-nel status. The experimental results reveal that the proposed approach can be used to authenticate transmitted images manipulated by high-quality JPEG compression, negation, cropping, replacement, or modification. Both the compari-son of the watermarked image with the extracted signature and the classification of various manipulations were achieved by visual judgments. Future research is suggested to develop more objective methods, such as methods based on typical matching techniques, to do the comparison and the classification. Other future research may be directed to developing more robust methods of embedding, deriving the optimal size of the neighborhood used to authenticate the signature, and generating more representative signa-tures.

Acknowledgment

This work was supported by the MOE Program of Excel-lence under Grant No. 89-E-FA04-1-4.

References

1. F. Hartung and M. Kutter, ‘‘Multimedia watermarking techniques,’’ Proc. IEEE 87共7兲, 1079–1107 共July 1999兲.

2. F. A. P. Petitcolas, R. J. Anderson, and M. G. Kuhn, ‘‘Information hiding—a survey,’’ Proc. IEEE 87共7兲, 1062–1078 共July 1999兲. 3. F. Bartolini, A. Tefas, M. Barni, and I. Pitas, ‘‘Image authentication

techniques for surveillance applications,’’ Proc. IEEE 89共10兲, 1403– 1418共Oct. 2001兲.

4. G. L. Friedman, ‘‘The trustworthy digital camera: restoring credibility to the photographic image,’’ IEEE Trans. Consumer Electron. 39, 905–910共Nov. 1993兲.

5. D. Kundur and D. Hatzinakos, ‘‘Digital watermarking for telltale tamper proofing and authentication,’’ Proc. IEEE 87共7兲, 1167–1180 共July 1999兲.

6. J. Fridrich, ‘‘Image watermarking for tamper detection,’’ Proc. ICIP98, IEEE Int. Conf. Image Process. 2, 404 – 408共1998兲. 7. C. Y. Lin and S. F. Chang, ‘‘Issues and solutions for authenticating

MPEG video,’’ Proc. SPIE 3657, 54 – 65共1999兲.

8. M. Wu and B. Liu, ‘‘Watermarking for mage authentication,’’ Proc. ICIP98, IEEE Int. Conf. Image Process. pp. 437– 441共1998兲. 9. P. Moulin, ‘‘The role of information theory in watermarking and its

application to image,’’ Signal Process. 81共6兲, 1121–1139 共June 2001兲. 10. S. Voloshynovskiy, F. Deguillaume, S. Pereira, and T. Pun, ‘‘Optimal adaptive diversity watermarking with channel state estimation,’’ Proc. SPIE 4314, 673– 685共2001兲.

11. D. Kundur and D. Hatzinakos, ‘‘Diversity and attack characterization for improved robust watermarking,’’ Signal Process. 49共10兲, 2383– 2396共Oct. 2001兲.

12. R. Ba¨uml, J. J. Eggers, and J. Huber, ‘‘A channel model for desyn-chronization attacks on watermarks,’’ Proc. SPIE 4675, 281–292 共2002兲.

13. K. Fukunaga, Introduction to Statistical Pattern Recognition, 2nd ed., Academic Press, Boston共1990兲.

14. I. J. Cox, M. L. Miller, and A. L. McKellips, ‘‘Watermarking as com-munications with side information,’’ Proc. IEEE 87共7兲, 1127–1141 共July 1999兲.

Zhi-Fang Yang received her PhD degree in computer and information science from Chiao Tung University, Hsinchu, Taiwan in 1999. She was a postdoctoral fellow at the Institute of Information Science, Academia Sinica, Taiwan from December 1999 to De-cember 2000. Since January 2001, she has been a contracted assistant professor in the Department of Computer and Infor-mation Science at National Chiao Tung University. Her current research interests include multimedia security and communication, multimedia infor-mation processing, and watermarking theory.

Wen-Hsiang Tsai received his PhD de-gree in electrical engineering from Purdue University in 1979. Dr. Tsai joined the fac-ulty of National Chiao Tung University, Hsinchu, Taiwan in November 1979. He has also been the Head of the Computer and Information Science Department, the Associate Director of the Microelectronics and Information System Research Center, the Dean of General Affairs, and the Dean of Academic Affairs there. He is currently the Vice President of the University. He has been the editor of sev-eral academic journals, includingComputer Quarterly(nowJournal of Computers),Proceedings of the National Science Council, Inter-national Journal of Pattern Recognition, and theJournal of Chinese Engineers; and the Editor-in-Chief of theJournal of Information Sci-ence and Engineering. His major research interests include image processing, pattern recognition, computer vision, document image analysis, and information hiding. He has published more than 268 academic papers, including 109 journal papers. He has received many awards from the National Science Council and from the Chi-nese Image Processing and Pattern Recognition Society. He is a senior member of IEEE and a member and the Director of the Chi-nese Image Processing and Pattern Recognition Society. He is also a member of the Medical Engineering Society of the Republic of China, the Information Processing Society, and the International Chinese Computer Society.

Fig. 12 Example of modification attack: (a) original host image; (b) attacked image (the face of the baboon is modified in the water-marked image); and (c) extracted signature showing the original ba-boon face.

數據

Figure 2 illustrates the proposed approach. The signature
Figure 5 compares the measures of reliability with hit ratios. Larger reliability measures are associated with larger hit ratios for all three images
Fig. 5 Experimental comparisons between reliability measure R and hit ratio A obtained by testing a series of JPEG images: (a) Airplane; (b) Lena; and (c) Baboon.
Fig. 6 Experimental comparisons between reliability measure R and entropy H , obtained by testing a series of JPEG images: (a) Airplane; (b) Lena; and (c) Baboon.
+3

參考文獻

相關文件

Students are able to use different learning strategies such as inquiry, reasoning, and problem solving skills in various learning activities. Teachers will employ a variety

- promoting discussion before writing to equip students with more ideas and vocabulary to use in their writing and to enable students to learn how to work in discussion groups and

好了既然 Z[x] 中的 ideal 不一定是 principle ideal 那麼我們就不能學 Proposition 7.2.11 的方法得到 Z[x] 中的 irreducible element 就是 prime element 了..

Wang, Solving pseudomonotone variational inequalities and pseudocon- vex optimization problems using the projection neural network, IEEE Transactions on Neural Networks 17

volume suppressed mass: (TeV) 2 /M P ∼ 10 −4 eV → mm range can be experimentally tested for any number of extra dimensions - Light U(1) gauge bosons: no derivative couplings. =>

Define instead the imaginary.. potential, magnetic field, lattice…) Dirac-BdG Hamiltonian:. with small, and matrix

incapable to extract any quantities from QCD, nor to tackle the most interesting physics, namely, the spontaneously chiral symmetry breaking and the color confinement.. 

• Formation of massive primordial stars as origin of objects in the early universe. • Supernova explosions might be visible to the most