• 沒有找到結果。

Scene-effect detection and insertion MPEG encoding scheme for video browsing and error concealment

N/A
N/A
Protected

Academic year: 2021

Share "Scene-effect detection and insertion MPEG encoding scheme for video browsing and error concealment"

Copied!
9
0
0

加載中.... (立即查看全文)

全文

(1)

Abstract—For an MPEG coding scheme, the encoder can do

more than video compression. In this paper, a novel MPEG codec embedded with scene-effect information detection and insertion is proposed to provide more functionality at the decoder end. Based on the macroblock (MB) type of information that is generated simultaneously in the encoding process, a single-pass automatic scene-effect insertion MPEG coding scheme can be achieved. Using the USER_DATA of picture header, the video output bitstreams by our method still conform to the conventional MPEG decoding system. The proposed method provides a solution toward upgrading the existing MPEG codec with low complexity to accomplish at least two major advantages. Precise and effective video browsing resulting from the scene-effect extraction can significantly reduce the user’s time to look up what they are interested in. For video transmission, the bitstreams containing scene-effect information can obtain better error concealment performance when scene changes are involved. Compared with the gain it achieves, the payout of our algorithm is very worthy with comparatively small efforts.

Index Terms—Error concealment, MPEG, scene change, scene

cut, video analysis, video transmission.

I. INTRODUCTION

A

S THE Internet becomes more and more popular and mul-timedia can be widely spread and applied, the video data booms to a tremendous amount and keeps on increasing rapidly. In addition to the problem of large memory storage, to browse and retrieve among this huge amount of video databases is also very difficult. MPEG codec nowadays is widely used to save the storage space and proven to be very effective. VCD and DVD players, which are based on the MPEG standard, have also be-come the common appliance in families. Since most video se-quences are stored in MPEG compressed form, many browsing and retrieving algorithms based on MPEG compressed video have been proposed recently. Hence, to achieve this purpose, the scene change detection method, acting as the initial step to extract the video editing points also arouses a lot of interests. A number of researches have focused on scene-change detec-tion methods [1]–[28] and many effective methods are devel-oped. Although the scene-change detection mechanism at the

Manuscript received July 3, 2003; revised February 17, 2004. This work was supported by the National Science Council of Taiwan, R.O.C., under contract NSC 93-2219-E-002-004 and by the Ministry of Education under Contract 89-E-FA06-2-4. The associate editor coordinating the review of this manuscript and approving it for publication was Dr. Wen-Tsung Chang.

The authors are with the Department of Electrical Engineering National Taiwan University Taipei, Taiwan, R.O.C. (e-mail: pei@cc.ee.ntu.edu.tw).

Digital Object Identifier 10.1109/TMM.2005.850961

decoder end works effectively, it still puts great efforts on the decoder structures. It violates the rule of MPEG coding scheme to keep the decoder as simple as possible and to concentrate the complex part at the encoder end. As a result, it is a preferred way to detect the interesting frame, such as scene changes, at the encoder side and insert the indications of the findings in the video bitstreams. Thus, video browsing can work by means of extracting the specific indications instead of complex analysis and collection of huge amount of overhead data. This idea is simple and intuitive but very attractive. However, it is not as easy as it seems. The difficulty lies in how to provide a novel MPEG encoding scheme with effective scene-effect detection, and how to design the format of overhead scene-effect indication to make the video bitstreams still conform to the conventional existing MPEG decoder. It is crucial to provide a novel solution with low complexity on the conventional operative MPEG encoder.

In addition to video storage, MPEG also makes video trans-mission possible because it significantly reduces large amount of transmission data. However, due to the property of predic-tive coding, the penalty of propagation error to video quality is quite huge. Even a single bit error can produce the corruption of a series of macroblocks (MBs) and these erroneous blocks will propagate to all GOP. To solve this problem, many error-re-silience and error-concealment methods have been proposed [29]–[54]. Error concealment method uses the spatial or tem-poral redundancy to conceal the corrupted MBs. Although the temporal error concealment is proven effective, scene changes often occur and can do much harm for sudden loss of tem-poral redundancy. As a result, if the scene change is found, the error concealment should be adjusted correspondingly. For ex-ample, the bidirectional (B) frames should choose the reference P frame of the same video clip to conceal the erroneous block because the temporal redundancy is still valid. In other words, the scene-change information is also very helpful for error con-cealment algorithm. Therefore, besides easy video browsing and description, a more precise error concealment mechanism can be also obtained if the scene-effect indication is added in the headers of MPEG bitstreams.

To conclude from the above discussions, the benefit of inserting a scene-effect indication is clear. Nevertheless, how to minimize the efforts to MPEG encoding when the scene-effect detection and indication insertion is involved also becomes an important issue. In this paper, a novel MPEG encoding scheme with low-complexity scene-effect detection method and simultaneous insertion is proposed. The scene-effect

(2)

Fig. 1. Illustration of Picture-Header insertion method.

tection method is based on MB-type information generated in the encoding processes. In addition, only simple analysis on MB-type information can achieve effective scene-effect detec-tion [23]. It is worthwhile to note that only one-pass processing is needed, that is to say, our method creates very little efforts on conventional MPEG encoding scheme. To switch the existing encoder to our proposed scheme, the cost is acceptably low.

In order to evaluate our method, the overheads for computa-tions, memory and encoding data are taken into considerations and experimented. It is shown that the cost is so insignificant that this method is very practical to be embedded into conventional MPEG codec. Moreover, the corresponding decoding scheme is also proposed to provide easy browsing, effective video descrip-tion and smarter error concealment mechanism. Experimental results show that our method works very well and effectively.

II. PROPOSEDMETHOD

It is obvious that to insert the scene-effect information into the bitstream can do a lot of good to the applications of the decoder side. Very fast and precise video browsing can be ob-tained and more effective error concealment to deal with scene effects can provide better quality under the noisy transmission environments. The benefits is evident, and the only question lies in how many efforts is needed to detect and insert the scene-effect information into the compressed video sequences. The most straightfoward and intuitive way is a two-phase method, in which the first pass is for scene-change detection for raw data and the second for the video compression. In the first pass, a scene-change detection method based on image difference or histogram comparison is used to decide which frame header needs to be inserted scene-change information and then the video sequence and the detection information are put into the conventional MPEG encoding scheme. This method works but the computation overhead is enormous, and moreover, the above encoding scheme brings out great efforts on the conventional encoding mechanism. In other words, the cost is tremendous, either from software design or hardware structure, to integrate the first phase detection algorithm into widely-used MPEG coding scheme.

Accordingly, in order to reduce the efforts to be integrated into conventional MPEG coding scheme, the embedded scene-change detection algorithm, which achieves based on the in-formation collected during the conventional encoding process, is adopted. During this single phase process, video compres-sion and scene-change detection perform simultaneously, and as a result, there is very low impact and integration cost for

present MPEG coding systems. In this paper, an integration in-formation insertion coding mechanism with a scene-change de-tection method based on the MB-type information is proposed. Our proposed method can be easily integrated into the present MPEG system with only MB-type information collection and little computation overhead.

There exists one another problem: how to insert the scene-ef-fect information conforming to the existing MPEG format. It is mandatory that the conventional existing decoding tool should be capable of handling the video sequence with the picture header (PH) containing scene effects. The MPEG standard reserves the USER_DATA field to insert useful data for some specific purpose and this field can be used to in-dicate the scene effects. In MPEG, USER_DATA is defined by the users for their specific applications, which follows the USER_DATA_START_CODE that is the bit string 000 001B2 in hexadecimal. It identifies the beginning of user data and the user data continues until receipt of another start code. In this paper, the format of USER_DATA for scene effects will be discussed in the later paragraph.

In order to lower the storage and compuation load, we leverage our past research which addressed a scene-effect detection method based on MB-type information. This method exploit the computation of motion estimation at the encoder side and only simple analysis can detect the scene effect such as abrupt and gradual scene change, flashlight or caption, etc. This method benefits from low computation load and small storage space, and also provides a good solution toward a single phase scene-effect PH insertion coding scheme.

Single phase algorithm, although it reduces the integration complexity, faces one another difficulty. The encoding process has been finished befor the scene-effect detection completes. As a result, the USER_DATA in the PH has to be inserted in ad-vance no matter whether this frame is a scene-change frame or not. In order to lower the overhead data for scene-effect identifi-cation, not all frames’ PHs are inserted USER_DATA. Because the P frame is decoded in advance of the following B frames, it is practicable that the scene-effect indication is only appended to P and I frames. This concept is illustrated in Fig. 1 and the pseudocode of the proposed algorithm is shown in Fig. 2.

As to the format of USER_DATA to represent the scene ef-fects, two kinds of information should be carried. One is for showing which frame in this SGOP (PBB ) is a scene-change frame, and the other is for indicating what kind of the scene effect is. For the above purpose, the USER_DATA for scene effects are shown in Tables I–III. The 2 bytes following the user_data_code is defined as the scene-effect message (SEM).

(3)

Fig. 2. Pseudocodes of scene-effect detection algorithm.

In order to discriminate SEM from other user defined data, the first four bits of the first byte is set 1111 to stand for SEM. The least significant four bits of the first bit is used for representing the scene-effect location in this SGOP. The last byte of SEM is designed to indicate the scene-effect types which are classified and listed in Table I.

In our proposed scene-effect detection method, three types of scene effects, including abrupt scene change, gradual scene change, and fast panning sequence, can be detected. Based on our past research [23], the abrupt scene change can be found from the prediction pattern of MB-type information of B frames. As to fast panning and gradual scene change, significant number of Intra-coded MBs of P frame is used for indication and MB-type information of B frames can used to distinguish these two apart. With these three scene-effect indications, it is essential for fundamental video browsing. However, it is worthwhile to note that the scene-effect detection method can be extended to other effects certainly, such as caption or flashlight detection. It depends on what functionality and what key frame the system needs to provide. Compared with conven-tional MPEG coded, our proposed method with scene-effect indication requires three kinds of overheads: computational

TABLE II

FORMAT OFLOSE FIELD OFSEMUDI

TABLE III

FORMAT OFSETYPE FIELD OFSEMUDI

overhead, memory overhead and data overhead. Computational overhead means that the operations for scene-effect detection, PH position recording and insertion, and all other extra process compared with the conventional coding scheme. Memory overhead represents the extra memory space requirement to store MB-type information and the information relative to PH insertion. As to data overhead, it comes from the USER_DATA of PHs of P and I frames. Because these three parameters will determine the efficiency of our proposed method, we design a series of experiments to verify our coding scheme.

A. Computational Overhead

A 30-min video (i.e. frames) is encoded by the conventional MPEG TM5 codec and our proposed method, respectively; and the computational time is recorded. The test

(4)

TABLE IV

COMPUTATION COMPARISONS OFCONVENTIONAL ANDPROPOSEDENCODINGMETHOD

TABLE V

OVERHEADTIME OFCONVENTIONAL ANDPROPOSEDMETHODS(CIF, secs)

TABLE VI

MEMORYOVERHEADEVALUATION

platform is based on a PIII 866 PC with 256 M RAM. The ex-perimental results are listed in Table IV and Table V. From these tables, only 0.1% computational overhead is required to handle the extra process of our scene-effect detection and PH inser-tion. It also implies that our proposed method is very practical in real-world video applications.

B. Memory Overhead

The extra requirement of memory space is shown in Table VI. For each MB, only one type information is required in our de-tection method. Thus, for a SGOP containing 1P2B frames (352 240), only 330 3 byte of memory space is required to be recorded. Compared with the memory space used to store the

decoded reference image , the overhead

memory is quite small and acceptable.

C. Data Overhead

Owing to the insertion of PH header, the video data stream will be elonged. Equation (1) shows the ratio of extra bits to the original data stream

where is the number of pictures in a GOP and is the dis-tance from a P frame to its previous P frame in a GOP.

is in the unit of bit per second and represents the number of GOP in a second. For the case of a 4-Mbps bitstream and

30 frames per second ( , , )

It is clear that the ratio is very insignificant.

D. Easy Embedding

Besides the advantage of lower overhead, the proposed method also benefits the architectural advantage. The scene-ef-fect detection based on MB-type information is very easy to be integrated into conventional MPEG encoding scheme. Fig. 3 illustrates the simpleness of the integration. The scene-effect detection engine can operate totally separately without inter-fering the other functional blocks of the conventional MPEG encoder. The result of scene-effect detection will be used to modify the pre-pending user data field in the PH, as Fig. 1 shows. The change of MPEG encoding architecture is in very small scale and the video sequence will carry the scene-effect information for the decoder’s further processing. The advantage of this kind of sequences will be discussed in the next section.

III. ADVANTAGES OFPROPOSEDMETHOD

As what we have mentioned, if the PH contains the scene-ef-fect information, a corresponding MPEG decoder can provide more precise and efficient video browsing, an effective video description, and more accurate error concealment. We will de-scribe these three advantages, respectively.

A. Video Browsing

A traditional way to browse video sequences is forwarding-and-searching. For MPEG sequences, the I-frame will be de-coded as the key frame for browsing. Unfortunately, the number of I frames is large and makes the video browsing imprecise and time-consuming. Taking a 90-min movie video, for example,

there exist I-frames to be browsed. It

is a tremendous large number. Actually, a 90-min movie may contain only 100 to 300 video clips, that is, 100 to 300 scene changes. If the scene changes can be found as the key frames for browsing, the user can save a lot of time and more precise video browsing is obtained. The browing decoder is quite simple and only the user_data of GOP header are required to be checked. A simulation to compare the efficiency of proposed algorithm and conventional browsing method have been tested and the results are shown in Table VII.

It is obvious from Table VII that our method provides a more effective and efficient video browsing, which significantly re-duce the computation load and relieve the browser from the te-dious searching task. However, it is noted that the actual video browsing is an interactive action and Table VII simply shows the improvement of browsing by the reduction of key frames.

B. Accurate Error Concealment

Recently, video transmission gets more and more important because the internet become common and widely spread. MPEG makes video transmission practical owing to the significant re-duction of data amount but it also brings great impact. The characteristic of temporal prediction and variable length coding makes MPEG very sensitive to transmission error. Error prop-agation can do great harm to decoded video sequences. As a

(5)

Fig. 3. How to embed our scene-effect detection engine into conventional MPEG encoder: (a) two-pass solution and (b) proposed single-pass solution.

TABLE VII

COMPARISON OFDECODINGTIME FORBROWSING

result, error-resilient coding to resist error and error conceal-ment to recover the damaged video stream nowadays arouse more and more attention. Regarding error concealment, tem-poral error concealment is widely used and proven superior to spatial error concealment in most of cases. However, the scene change may do great harm to temporal error concealment be-cause it be-cause the temporal corelation entirely ineffective. As a result, a special treatment should be adopted to handle the error

concealment under the circumstance of scene changes. Because the scene effect is known after the PH decoding, the choice of error concealment strategy can be made more precisely. If an abrupt scene change is indicated in the PH of this P frame, it implies that the temporal correlation is invalid for error conceal-ment. As a result, once the error is encountered in the decoding process of the P frame, the spatial error concealment is adopted. For the B frames in this SGOP, the error concealment direction is toward the reference frame in the same shot. Figs. 4 and 7 il-lustrates this idea.

As to the dissolve sequences, the temporal error concealment is also ineffective because the temporal correlation is almost insignificant. A more precise error concealment based on the linear property can be applied to obtain better quality of decoded frames. When the SGOP is decoded as a dissolve SGOP, which is indicated by PH of the P frame, the corrupted MBs of B frames can be concealed by the interpolation of two reference frames.

(6)

Fig. 4. Solution to scene-change error concealment.

The weighting of the interpolation can be evaluated by the error-free decoded MB subgroup, MBSGOP [23]. For P frames, the linear property also can be applied to recover the error MBs. If there exists three consecutive reference frames in the dissolve sequence found by the PHs, the successful error-free co-located decoded MBs in the two of these three frames can be used to predict the corrupted MB of the other one. Interpolation method is applied for the error of the middle one and the extrapolation is for the latest reference frame. This method is shown in Figs. 5, 6, and 8and it is noted again the parameters of interpolation and extrapolation is calculated by the successful error-free decoded MBs.

A series of experiments are tested to verify our proposed scene-change dependent error concealment algorithm. Fig. 9 shows the comparison between conventional temporal error concealment and proposed method. From Fig. 9, the effective-ness of error concealment of our method is obviously better than the conventional one because the more accurate prediction reference frame is chosen. Tables VIII– X show the statistics of the PSNR performance comparison. It is found that about 3–5 dB is gained by adopting our method. Fig. 10 shows similar performance for the proposed algorithm, and about 3–5 dB in PSNR is gained, corresponding to different bit error rates (BERs). In all the experiments, it is worthwhile to emphasize that only the GOP with a scene change is taken into consider-ation and PSNR calculconsider-ation. It is because the error resulting from the poorly concealed block due to the scene change will not propagate to the next GOP.

Fig. 5. Interpolated temporal error concealment method to handle B frames in the dissolve sequence.

Fig. 6. Illustration of interpolated and extrapolated temporal error concealment to handle P frames in the dissolve sequence.

IV. CONCLUSIONS

In this paper, a novel idea and the corresponding imple-mentation to insert the scene-effect information into the video bitstream is proposed. In order to lower the impact on the existing MPEG coding scheme, we presented a single-pass MPEG codec with low-complexity scene-effect detection. The method based on MB-type information from our past studies benefits from the simultaneous data generation and simple analysis of low computational complexity. Moreover, USER_DATA of a PH is applied to represent the scene effect

(7)

Fig. 7. Pseudocodes of abrupt scene-change error concealment.

and keeps the consequent video stream conformable to con-ventional MPEG codec. Integrated with the above two ideas, a single-pass automatic scene-effect insertion MPEG codec is proposed and experimented. Based on our evaluation, less than 1% overhead on the computational complexity, memory space, and extra data is required to achieve our proposed codec. This is very attractive for conventional MPEG codecs to adopt. At the decoder end, the video bitstream with the scene-effect information can provide some useful functions. First, because the scene effects are known in advance of picture decoding, the video browsing can precisely focus on those containing specific scene effects, such as scene changes. A lot of browsing time and computation efforts can be saved by this design. Besides, the scene-effect information can be applied to help to describe video clips. For a higher level of video retrieval, scene-effect information can cooperate with some other video descriptors, such as color histogram and edge contour, to achieve the pur-pose of video description. For example, a TV commercial clip would show frequent scene changes and short video clip span and in the basketball sequences, frequent and steady appear-ance of fast panning effects could be found. In other words,

Fig. 8. Pseudocodes of dissolve error concealment.

Fig. 9. Plot of comparisons of four methods (abrupt scene change).

to embed the scene-effect information into video headers may also provide a useful video descriptor in the decoder end. For video transmission, we also examine a scene-effect dependent

(8)

TABLE VIII

SIMULATION OFDIFFERENTLENGTH OFDISSOLVESEQUENCE(4 Mbps,BER = 0:5%, Total Block error rate = 5%).

Where M0: Error free decoding, M1: BREC, M2: MCTEC, M3: WISEC, M4: Proposed TABLE IX

SIMULATION OFDIFFERENTLENGTH OFDISSOLVESEQUENCE(BER = 1%, Total Block error rate = 10%)

Where M0: Error free decoding, M1: BREC, M2: MCTEC, M3: WISEC, M4: Proposed TABLE X

SIMULATION OFDIFFERENTLENGTH OFDISSOLVESEQUENCE(BER = 2%, Total Block error rate = 20%)

Where M0: Error free decoding, M1: BREC, M2: MCTEC, M3: WISEC, M4: Proposed

Fig. 10. Plot of comparisons of four methods (gradual scene change).

error concealment to handle the inaccurate frame prediction when scene change is involved. Based on the experiments, it gains more than 3 dB in PSNR by this algorithm.

REFERENCES

[1] A. Nagasaka and Y. Tanaka, “Automatic video indexing and full mo-tion search for object appearances,” Visual Database Systems, II, pp. 119–133, 1991.

[2] F. Arman, R. Depommier, A. Hsu, and M.-Y. Chiu, “Content-based browsing of video sequences,” in Proc. ACM Multimedia’94, Aug. 1994, pp. 97–103.

[3] F. Arman, A. Hsu, and M. Y. Chiu, “Feature management for large video databases,” Storage and Retrieval for Image and Video Databases, vol. SPIE-1908, pp. 2–12, 1993.

[4] Y. Nakajuma, “A video browsing using fast scene cut detection for an efficient networked video database access,” IEICE Trans. Inform. Syst., vol. E77-D, no. 12, pp. 1355–1364, Dec. 1994.

[5] F. Arman, A. Hsu, and M. Y. Chiu, “Image processing on compressed data for large video databases,” in Proc. 1st ACM Int. Conf. Multimedia, Aug. 1993, pp. 267–272.

[6] H. J. Zhang, A. Kankanhalli, and S. W. Smoliar, “Automatic partitioning of full-motion video,” Multimedia Syst., vol. 1, pp. 10–28, Jul. 1993. [7] H. J. Zhang, C. Y. Low, and S. W. Smoliar, “Video parsing and browsing

using compressed data,” Multimedia Tools Applicat., vol. 1, no. 1, pp. 89–111, Mar. 1995.

[8] B.-L. Yeo and B. Liu, “Rapid scene analysis on compressed video,”

IEEE Trans. Circuits Syst. Video Technol, vol. 5, no. 8, pp. 533–544,

Dec. 1995.

[9] K. Otsuji and Y. Tonomura, “Projection detecting filter for video cut detection,” in Proc. 1st ACM Int. Conf. Multimedia, Aug. 1993, pp. 251–257.

[10] S. F. Chang and D. G. Messerschmitt, “Manipulation and compositing of MC-DCT compressed video,” IEEE J. Select. Areas Commun., vol. 13, no. 1, pp. 1–11, Jan. 1995.

[11] N. V. Patel and I. K. Sethi, “Compressed video processing for cut detec-tion,” Proc. Inst. Elect. Eng.—Vis., Image, Signal Process., vol. 143, no. 5, pp. 315–323, Oct. 1996.

[12] , “Video shot detection and characterization for video databases,”

Pattern Recognit., vol. 30, no. 4, pp. 583–592, Mar. 1997.

[13] B.-L. Yeo and B. Liu, “Visual content highlighting via automatic extrac-tion of embedded capextrac-tions on MPEG compressed video,” in Proc. SPIE,

Digital Video Compression: Algorithms Technology., vol. 2668, 1996,

pp. 47–58.

[14] J. Meng, Y. Juan, and S.-F. Chang, “Scene change detection in a MPEG compressed video sequence,” in Proc. SPIE, Digital Video Compression:

Algorithms Technol., vol. 2419, 1995, pp. 14–25.

[15] Q. Wei, H. Zhang, and Y. Zhong, “A robust approach to video segmen-tation using compressed data,” in Proc. SPIE, Storage and Retrieval for

Still Image and Video Database, vol. 3022, 1997, pp. 448–456.

[16] L. A. Rowe, S. Smoot, K. Patel, B. Smith, K. Gong, E. Hung, D. Banks, S. T.-S. Fung, D. Brown, and D. Wallach. (1995, Aug.) Berkeley MPEG Tools. Version. 1.0, Rel. 2. [Online] Available: ftp://mm-ftp.cs.berkeley.edu/pub/multimedia/mpeg/ bmt1r1.tar.gz [17] A. M. Alattar, “Wipe scene change detector for use with video

compres-sion algorithms and MPEG-7,” IEEE Trans. Consumer Electron., vol. 44, no. 1, pp. 43–51, Feb. 1998.

[18] W. A. C. Fernando, C. N. Canagarajah, and D. R. Bull, “Wipe scene change detection in video sequence,” Proc. ICASP, pp. 294–298, 1999. [19] M. Wu, W. Wolf, and B. Liu, “An algorithm for wipe detection,” Proc.

ICIP, pp. 893–897, 1998.

[20] S. J. Golin, “New metric to detect wipes and other gradual transitions in video,” SPIE, vol. 3653, pp. 1464–1474, 1998.

(9)

[24] N. Gamaz, X. Huang, and S. Panchanathan, “Scene change detection in MPEG domain,” in Proc. IEEE Southwest Symp. Image Analysis and

Interpretation, 1998, pp. 12–17.

[25] U. Gargi, R. Kasturi, and S. H. Strayer, “Performance chararterization of video-shot-change detection methods,” IEEE Trans. Circuits Syst. for

Video Technol., vol. 10, no. 1, pp. 1–13, Feb. 2000.

[26] K. Tse, J. Wei, and S. Pancanathan, “A scene change detection algorithm for MPEG compressed video sequences,” in Proc. Canadian Conf.

Elec-trical and Computer Engineering, vol. 2, 1995, pp. 827–830.

[27] J. Meng, Y. Juan, and S.-F. Chang, “Scene change detection in a MPEG compressed video sequence,” in IS&T/SPIE Symp. Proc., vol. 2419, Feb. 1995.

[28] T. Shin, J.-G. Kim, H. Lee, and J. Kim, “Hierarchical scene change de-tection in an Mpeg-2 compressed video sequence,” in Proc. 1998 IEEE

Int. Symp. Circuits and Systems, vol. 4, pp. 253–256.

[29] G.-S. Yu, M. M.-K. Liu, and M. W. Marcellin, “POCS-based error con-cealment for packet video using multiframe overlap information,” IEEE

Trans. Circuits Syst. Video Technol., vol. 8, no. 4, pp. 422–434, Aug.

1998.

[30] Y. J. Chung, J. W. Kim, and C.-C. J. Kuo, “Real-time streaming video with adaptive bandwidth control and DCT-based error concealment,”

IEEE Trans. Circuits Syst. II, vol. 46, no. 7, pp. 951–956, Jul. 1999.

[31] H. Sun and W. Kwok, “Concealment of damaged block transform coded images using projections onto convex sets,” IEEE Trans. Image Process., vol. 4, no. 4, pp. 470–477, Apr. 1995.

[32] V. Parthasarathy, J. W. Modestino, and K. S. Vastola, “Design of a trans-port coding scheme for high-quality video over ATM networks,” IEEE

Trans. Circuits Syst. Video Technol., vol. 7, no. 2, pp. 358–376, Apr.

1997.

[33] W. Keck, “A method for robust decoding of erroneous MPEG-2 video bitstreams,” IEEE Trans. Consumer Electron., vol. 42, no. 3, pp. 411–421, Aug. 1996.

[34] J. Feng, K.-T. Lo, and H. Mehrpour, “Error concealment for MPEG video transmissions,” IEEE Trans. Consumer Electron., vol. 43, no. 2, pp. 183–187, May 1997.

[35] W. Zhu, Y. Wang, and Q.-F. Zhu, “Second-order derivative-based smoothness measure for error concealment in DCT-based codecs,”

IEEE Trans. Circuits Syst. Video Technol., vol. 8, no. 6, pp. 713–718,

Oct. 1998.

[36] J. F. Arnold, M. R. Frater, and J. Zhang, “Error resilience in the MPEG-2 video coding standard for cell based networks—a review,” Signal

Pro-cessing: Image Commun., vol. 14, pp. 607–633, 1999.

[37] S. Kaiser and K. Fazel, “Comparison of error concealment techniques for an MPEG-2 video decoder in terrestrial TV-broadcasting,” Signal

Processing: Image Commun., vol. 14, pp. 655–676, 1999.

[38] H.-C. Shyu and J.-J. Leou, “Detection and concealment of transmission errors in MPEG-2 images—a genetic algorithm approach,” IEEE Trans.

Circuits Syst. Video Technol., vol. 9, no. 6, pp. 937–948, Sep. 1999.

[39] J. M. Boyce, “Packet loss resilient transmission of MPEG video over the internet,” Signal Processing: Image Commun., vol. 15, pp. 7–24, 1999. [40] D. W. Redmill, D. R. Bull, J. T. Chung-How, and N. G. Kingsbury, “Error-resilient image and video coding for wireless communication systems,” Electron. Commun. Eng. J., Aug. 1998.

[41] M. Bystrom, V. Parthasarathy, and J. W. Modestino, “Hybrid error concealment schemes for broadcast video transmission over ATM networks,” IEEE Trans. Circuits Syst. Video Technol., vol. 9, no. 6, pp. 868–881, Sep. 1999.

[42] P.-J. Lee and M.-J. Chen, “Robust error concealment algorithm for video decoder,” IEEE Trans. Consumer Electron., vol. 45, no. 3, pp. 851–859, Aug. 1999.

[43] M.-C. Hong, H. Schwab, L. P. Kondi, and A. K. Katsaggelos, “Error con-cealment algorithms for compressed video,” Signal Processing: Image

Commun., vol. 14, pp. 473–492, 1999.

[44] J.-W. Suh and Y.-S. Ho, “Error concealment based on directional inter-polation,” IEEE Trans. Consumer Electron., vol. 43, no. 3, pp. 295–302, Aug. 1997.

novel applications in block-based low-bit-rate coding,” IEEE Trans.

Cir-cuits Syst. Video Technol., vol. 9, no. 4, pp. 648–665, Jun. 1999.

[49] J.-T. Wang and P.-C. Chang, “Error-propagation prevention technique for real-time video transmission over ATM networks,” IEEE Trans.

Cir-cuits Syst. Video Technol., vol. 9, no. 3, pp. 513–523, Apr. 1999.

[50] M.-J. Chen, L.-G. Chen, and R.-M. Weng, “Error concealment of lost motion vectors with overlapped motion compensation,” IEEE Trans.

Circuits Syst. Video Technol., vol. 7, no. 3, pp. 560–563, Jun. 1997.

[51] V. Parthasarathy, J. W. Modestino, and K. S. Vastola, “Reliable trans-mission of high-quality video over ATM networks,” IEEE Trans. Image

Process., vol. 8, no. 3, pp. 361–374, Mar. 1999.

[52] Z. Wang, Y. Yu, and D. Zhang, “Best neighborhood matching: An in-formation loss restoration technique for block-based image coding sys-tems,” IEEE Trans. Image Process., vol. 7, no. 7, pp. 1056–1061, Jul. 1998.

[53] X. Lee, Y.-Q. Zhang, and A. Leon-Garcia, “Information loss recovery for block-based image coding techniques-a fussy logic approach,” IEEE

Trans. Image Proces., vol. 4, no. 3, pp. 259–273, Mar. 1995.

[54] J. W. Park and S. U. Lee, “Recovery of corrupted image data based on the NURBS interpolation,” IEEE Trans. Circuits Syst. Video Technol., vol. 9, no. 7, pp. 1003–1008, Oct 1999.

Soo-Chang Pei (SM’89–F’00) was born in Soo-Auo, Taiwan, R.O.C., in 1949. He received the B.S.E.E. degree from National Taiwan University (NTU), Taipei, in 1970 and the M.S.E.E. and Ph.D. degrees from the University of California, Santa Barbara (UCSB), in 1972 and 1975, respectively.

He was an Engineering Officer in the Chinese Navy Shipyard from 1970 to 1971. From 1971 to 1975, he was a Research Assistant at UCSB. He was the Professor and Chairman in the Electrical Engineering Department, Tatung Institute of Tech-nology and NTU from 1981 to 1983 and 1995 to 1998, respectively. Presently, he is the Professor in the Electrical Engineering Department at NTU. His research interests include digital signal processing, image processing, optical information processing, and laser holography.

Dr. Pei received the National Sun Yet-Sen Academic Achievement Award in Engineering in 1984, the Distinguished Research Award from the National Sci-ence Council, R.O.C., from the Chinese Institute of Electrical Engineering in 1998, the Academic Achievement Award in Engineering from the Ministry of Education in 1998, the Pan Wen-Yuan Distinguished Research Award in 2002; and the National Chair Professor Award from Ministry of Education in 2002. He has been President of the Chinese Image Processing and Pattern Recogni-tion Society in Taiwan from 1996-1998, and is a member of Eta Kappa Nu and the Optical Society of America. He became an IEEE Fellow in 2000 for con-tributions to the development of digital eigenfilter design, color image coding, and signal compression, and to electrical engineering education in Taiwan.

Yu-Zuong Chou was born in Tainan, Taiwan, R.O.C. He received the B.S. degree from the National Tsing Hua University, Hsinchu, Taiwan, in 1994, and the M.S. and Ph.D. degrees from the National Taiwan University (NTU), Taipei, Taiwan, in 1996 and 2002, respectively, both in electrical engineering.

He is currently an Engineer with Realtek Semicon-ductor Corporation, Hsinchu. His research interests include video compression, image processing, multi-media application, and SOC design.

數據

Fig. 1. Illustration of Picture-Header insertion method.
Fig. 2. Pseudocodes of scene-effect detection algorithm.
TABLE IV
Fig. 3. How to embed our scene-effect detection engine into conventional MPEG encoder: (a) two-pass solution and (b) proposed single-pass solution.
+4

參考文獻

相關文件

• The scene with depth variations and the camera has movement... Planar scene (or a

• The scene with depth variations and the camera has movement... Planar scene (or a

– Change Window Type to Video Scene Editor – Select Add → Images and select all images – Drag the strip to the “1st Frame” in Layer

Research method is to use the Mirror method or as light reflection principle, which commonly used in geometry, and classified into odd and even side polygon various situations

The hashCode method for a given class can be used to test for object equality and object inequality for that class. The hashCode method is used by the java.util.SortedSet

• When a system undergoes any chemical or physical change, the accompanying change in internal energy, ΔE, is the sum of the heat added to or liberated from the system, q, and the

request even if the header is absent), O (optional), T (the header should be included in the request if a stream-based transport is used), C (the presence of the header depends on

n The information contained in the Record-Route: header is used in the subsequent requests related to the same call. n The Route: header is used to record the path that the request