University of Bristol - Explore Bristol Research. Peer reviewed version Link to published version (if available): /ISCAS.1999.

Similar documents
Dumanli, S., Paul, D. L., & Railton, C. J. (2010). LTCC or LCP, a comparison using cavity backed slot antennas with pin curtains at 60 GHz. 1-5.

University of Bristol - Explore Bristol Research. Link to publication record in Explore Bristol Research PDF-document.

University of Bristol - Explore Bristol Research. Link to published version (if available): /VTCF

University of Bristol - Explore Bristol Research. Peer reviewed version. Link to publication record in Explore Bristol Research PDF-document

Personalized Karaoke

Segmentation using Saturation Thresholding and its Application in Content-Based Retrieval of Images

University of Bristol - Explore Bristol Research. Link to published version (if available): /LAWP

Linear Gaussian Method to Detect Blurry Digital Images using SIFT

Content Based Image Retrieval Using Color Histogram

A Novel Method for Enhancing Satellite & Land Survey Images Using Color Filter Array Interpolation Technique (CFA)

University of Bristol - Explore Bristol Research. Peer reviewed version Link to published version (if available): /LSP.2004.

Stamp detection in scanned documents

Finding Text Regions Using Localised Measures

Zero-Based Code Modulation Technique for Digital Video Fingerprinting

Bogdan Smolka. Polish-Japanese Institute of Information Technology Koszykowa 86, , Warsaw

Practical Content-Adaptive Subsampling for Image and Video Compression

An Efficient Color Image Segmentation using Edge Detection and Thresholding Methods

REIHE INFORMATIK TR Automatic Generation of Video Summaries for Historical Films

Types of Angles. Low Angle: High Angle: Dutch Angle:

Solution for Image & Video Processing

CSSE463: Image Recognition Day 2

Live Hand Gesture Recognition using an Android Device

Background Pixel Classification for Motion Detection in Video Image Sequences

An Efficient Method for Landscape Image Classification and Matching Based on MPEG-7 Descriptors

OBJECTIVE OF THE BOOK ORGANIZATION OF THE BOOK

A JOINT MODULATION IDENTIFICATION AND FREQUENCY OFFSET CORRECTION ALGORITHM FOR QAM SYSTEMS

Background Subtraction Fusing Colour, Intensity and Edge Cues

COMPARATIVE PERFORMANCE ANALYSIS OF HAND GESTURE RECOGNITION TECHNIQUES

Monophony/Polyphony Classification System using Fourier of Fourier Transform

Memory-Efficient Algorithms for Raster Document Image Compression*

Recognition Of Vehicle Number Plate Using MATLAB

Image Segmentation of Historical Handwriting from Palm Leaf Manuscripts

Fig Color spectrum seen by passing white light through a prism.

License Plate Localisation based on Morphological Operations

Video Encoder Optimization for Efficient Video Analysis in Resource-limited Systems

MICROCHIP PATTERN RECOGNITION BASED ON OPTICAL CORRELATOR

McCarthy, M. R., Muller, H. L., Calway, A. D., & Wilson, R. E. (2006). Position and velocity recovery from independent ultrasonic beacons.

Digital Image Processing. Lecture # 6 Corner Detection & Color Processing

(i) Understanding the basic concepts of signal modeling, correlation, maximum likelihood estimation, least squares and iterative numerical methods

Development of an Automatic Camera Control System for Videoing a Normal Classroom to Realize a Distant Lecture

Terms and Techniques

Detection of Out-Of-Focus Digital Photographs

Nonuniform multi level crossing for signal reconstruction

Autocomplete Sketch Tool

Experimental Images Analysis with Linear Change Positive and Negative Degree of Brightness

Detecting Intra-Room Mobility with Signal Strength Descriptors

Image Forgery Detection Using Svm Classifier

Main Subject Detection of Image by Cropping Specific Sharp Area

Distributed Algorithms. Image and Video Processing

Response Curve Programming of HDR Image Sensors based on Discretized Information Transfer and Scene Information

Performance Analysis of Color Components in Histogram-Based Image Retrieval

Retrieval of Large Scale Images and Camera Identification via Random Projections

Computer Vision. Howie Choset Introduction to Robotics

No-Reference Image Quality Assessment using Blur and Noise

3D and Sequential Representations of Spatial Relationships among Photos

HIGH DYNAMIC RANGE VERSUS STANDARD DYNAMIC RANGE COMPRESSION EFFICIENCY

Chapter 8. Representing Multimedia Digitally

DESIGN & DEVELOPMENT OF COLOR MATCHING ALGORITHM FOR IMAGE RETRIEVAL USING HISTOGRAM AND SEGMENTATION TECHNIQUES

APPLYING EDGE INFORMATION IN YCbCr COLOR SPACE ON THE IMAGE WATERMARKING

Blur Detection for Historical Document Images

University of Bristol - Explore Bristol Research. Peer reviewed version. Link to published version (if available): /VETEC.1997.

Impulse noise features for automatic selection of noise cleaning filter

Automated Detection of Early Lung Cancer and Tuberculosis Based on X- Ray Image Analysis

REVERSIBLE MEDICAL IMAGE WATERMARKING TECHNIQUE USING HISTOGRAM SHIFTING

An Algorithm and Implementation for Image Segmentation

EFFICIENT COLOR IMAGE INDEXING AND RETRIEVAL USING A VECTOR-BASED SCHEME

A Method of Multi-License Plate Location in Road Bayonet Image

Real Time Video Analysis using Smart Phone Camera for Stroboscopic Image

CONNECTICUT Statewide Career & Technical Education ASSESSMENT

Adaptive Feature Analysis Based SAR Image Classification

RECOMMENDATION ITU-R BT SUBJECTIVE ASSESSMENT OF STANDARD DEFINITION DIGITAL TELEVISION (SDTV) SYSTEMS. (Question ITU-R 211/11)

Wadehra Kartik, Kathpalia Mukul, Bahl Vasudha, International Journal of Advance Research, Ideas and Innovations in Technology

Comparison of Static Background Segmentation Methods

In literary texts, we speak of the contributing parts as words, phrases, sentences, paragraphs, and chapters. In film, there are:

Travel Photo Album Summarization based on Aesthetic quality, Interestingness, and Memorableness

Dynamic thresholding for automated analysis of bobbin probe eddy current data

Color Image Processing

Filtering in the spatial domain (Spatial Filtering)

Movie 10 (Chapter 17 extract) Photomerge

Non Linear Image Enhancement

SONG RETRIEVAL SYSTEM USING HIDDEN MARKOV MODELS

INDIAN VEHICLE LICENSE PLATE EXTRACTION AND SEGMENTATION

Fuzzy Statistics Based Multi-HE for Image Enhancement with Brightness Preserving Behaviour

A TWO-PART PREDICTIVE CODER FOR MULTITASK SIGNAL COMPRESSION. Scott Deeann Chen and Pierre Moulin

Feature Analysis for Audio Classification

Evaluation of image quality of the compression schemes JPEG & JPEG 2000 using a Modular Colour Image Difference Model.

A Proficient Roi Segmentation with Denoising and Resolution Enhancement

Analysis on Color Filter Array Image Compression Methods

Effective and Efficient Fingerprint Image Postprocessing

Performance of Combined Error Correction and Error Detection for very Short Block Length Codes

NON UNIFORM BACKGROUND REMOVAL FOR PARTICLE ANALYSIS BASED ON MORPHOLOGICAL STRUCTURING ELEMENT:

Matching Words and Pictures

Robust Voice Activity Detection Based on Discrete Wavelet. Transform

A Comparison of the Multiscale Retinex With Other Image Enhancement Techniques

CAMERA BASICS. Stops of light

Introduction. Article 50 million: an estimate of the number of scholarly articles in existence RESEARCH ARTICLE

Extraction and Recognition of Text From Digital English Comic Image Using Median Filter

A Model of Color Appearance of Printed Textile Materials

Real-Time Face Detection and Tracking for High Resolution Smart Camera System

Lecture Notes 11 Introduction to Color Imaging

Transcription:

Fernando, W. A. C., Canagarajah, C. N., & Bull, D. R. (1999). Automatic detection of fade-in and fade-out in video sequences. In Proceddings of ISACAS, Image and Video Processing, Multimedia and Communications, Florida. (Vol. 4, pp. 255-258). Institute of Electrical and Electronics Engineers (IEEE). 10.1109/ISCAS.1999.779990 Peer reviewed version Link to published version (if available): 10.1109/ISCAS.1999.779990 Link to publication record in Explore Bristol Research PDF-document University of Bristol - Explore Bristol Research General rights This document is made available in accordance with publisher policies. Please cite only the published version using the reference above. Full terms of use are available: http://www.bristol.ac.uk/pure/about/ebr-terms.html Take down policy Explore Bristol Research is a digital archive and the intention is that deposited content should not be removed. However, if you believe that this version of the work breaches copyright law please contact open-access@bristol.ac.uk and include the following information in your message: Your contact details Bibliographic details for the item, including a URL An outline of the nature of the complaint On receipt of your message the Open Access Team will immediately investigate your claim, make an initial judgement of the validity of the claim and, where appropriate, withdraw the item in question from public view.

AUTOMATIC DETECTION OF FADE-IN AND FADE-OUT IN VIDEO SEQUENCES W.A.C. Fernando, C.N. Canagarajah, D.R. Bull Centre for Communications Research, Department of Electrical and Electronic Engineering, University of Bristol, Merchant Ventures Building, Woodland Road, Bristol BS8 IUB, United Kingdom. Voice - +44-117-954-5198, Fax - +44-117-954-5206 Email - W. A. C. Fernando @ bris t ol. ac. uk ABSTRACT A common video indexing technique is to segment video shots by identifying scene changes and then to extract features. This paper discusses a novel algorithm for detecting fade-in and fade-out using statistical features of both luminance and chrominance signals. The ratio between incremental change in the mean of the luminance signal to the chrominance (average sum of C, and C,) is considered to identify fade-in and fade-out. Results show that the algorithm is capable of detecting all fade regions accurately even the video sequence contains other special effects. 1. INTRODUCTION A major feature required in a visual information system is an efficient indexing technique to enable fast access to the stored data. A video index serves as a descriptor of the video, thus enabling rapid access to the video clips stored in a multimedia databases. Another consideration is that a user who is interested in searching, browsing or retrieving video clips needs a way to interface with the database by formulating appropriate queries. These queries need to be appropriately translated into a form that can be used to search index and retrieve the matching clips. A typical approach to indexing and archiving video for retrieval requires parsing the video, extracting key information from each clip, indexing the information and providing a representation which allows accurate and efficient retrieval based on the user s request. A common technique is to index the video sequences first into video shots by identifying scene changes and then to extract features. Scene changes (transitions) can be divided into two categories: abrupt transitions and gradual transitions. Gradual transitions include camera movements: panning, tilting, zooming and video editing special effects: fade-in, fade-out, dissolving, wiping. Considerable work has been done for automatic sudden change detection and camera movements [ 1-41. However, automatic special effect detection is still in a very early stage. Zabih et a1 [5] proposed a feature-based algorithm for detecting fade-in and fade-out detection. This algorithm needs to detect edges in every frame and hence it is very costly. Another limitation of this scheme is that the edge detection method does not handle rapid changes in overall scene brightness, or scenes, which are very dark or very bright. Furthermore, automatic segmentation and classification is not possible with this scheme. Alattar proposed an algorithm for detecting fade-in and fade-out by exploiting the semi-parabolic behaviour of the variance curve [6]. This algorithm can only detect fade-in and fade-out when the sequence is freezed. When the sequence has considerable motion, this algorithm fails to identify correct fade-in and fadeout regions. In this paper a real time algorithm is proposed for fade-in and fade-out detection using the statistical features of both luminance and chrominance signals. The paper is organised as follows. Section 2 gives a brief introduction to fade-in and fade-out and their mathematical analysis. Section 3 describes the proposed algorithm. Results are presented in section 4. Section 5 includes the conclusions. 2. FADE-IN AND FADE-OUT In video editing and production, proportions of two or more picture signals are simply added together so that the two pictures appear to merge on the output screen. Very often this process is used to move on from picture A to picture B. In this case, the proportions of the two signals are such that as the contribution of picture A changes from 100% to zero and the contribution of picture B changes from zero to 100%. This is called dissolving. When picture A is a solid colour, it is called as fade-in and when picture B is a solid colour, it is known as fade-out. Consider a sequence, which is 0-7803-5471-0/99/$10.0001999 IEEE IV-255

subjected to fade-in. Mathematically, fade-in can be expressed as shown in Equation (I). Where, S,(x,y) - Resultant video signal, f,(x,y) - Picture A, g, (x, y) - Picture B, C - Video signal level (solid value) at the start of the fade-in sequence, L, - Length of sequence A, F- Length of fading sequence, L, - Length of the total sequence. Assume that video sequence f, (x, y ) and g, (x, y) are ergodic with mean mf and mg, and variance 0; and 0:. Let, ms,, is the mean of the resultant video sequence S, and o:,, is the variance of the resultant video sequence. transition between the two scenes. However, still the process is not an ergodic process. Therefore, fading cannot be detected using only the linear behaviour of mean and the quadratic behaviour of variance as these parameters are very sensitive to even for a small motion. Since video signal consists of a linear combination of luminance (Y) and two chrominance (C, and cb) signals, similar set of equations can be derived for both luminance and chrominance signals. Mean of the chrominance signal is not sensitive as mean of the luminance signal for a video sequence. However, for a fading sequence both signals are equally sensitive. We combine this behaviour of the chrominance signal together with the luminance signal to identify fading. Since mean of the two signals have a linear characteristic during fading, the ratio ( R(n) ) of incremental change in the mean of the luminance signal to the chrominance signal (average sum of C, and cb) should be a constant during a fading sequence (Equation - 5 and 6). During a non-fading sequence incremental change in the mean of the chrominance signal is small compared to the luminance signal. Therefore, naturally ratio R(n) should have a larger variation. In this proposed scheme, we exploit the linear behaviour of the mean and zero variance at the beginning of fade-in process(variance is zero at the end of the fade-out process). It can be shown that the change in mean of the video signal can be expressed as in Equation (4). (L,+F)iniL2 (3) where, A,, = Imn+l -m,l. Taking the mean shows that during fade-in the mean has a linear characteristic as given by Equation (2). Furthermore, the variance has a quadratic behaviour as shown in Equation (3). It should be noted that all these mathematical derivations are valid under the assumption of the video sequences are an ergodic process. However in practice this is not true and alternative strategies are needed in order to identify these special effects. 3. PROPOSED SCHEME During special effects like fade-in and fade-out large movements between frames are not allowed as it causes inconvenience to the viewer and also it leaves a rough IV-256

where, S, - Chrominance Signal s, - Luminance Signal En - Incrernental change in mean of the chrominance signal Y in mean of the A,, - Incremental change luminance signal Simulation results confirmed that the ratio of has a large variance and hence the differentiation of R(n) should also have a larger variance. During a fade-in sequence, however R(n) is almost a constant or varying very slowly with reference to the non-fade-in case. As a consequence, differentiation is close to zero and enables to detect fade-in regions. Fade-out : Detect a sequence of continuous region, which is identified by the above algorithm followed by detecting a frame with zero variance. 4. SIMULATION RESULTS Consider a test sequence to describe the performance of the above algorithm. Figure 1 shows mean of the luminance signal for a test sequence. The same information for chrominance signal is presented in Figure 2. Figure 3 shows the function DR for the sequence. Figure 3 reveals that there is a region, which satisfies the condition (D, < Tfade ), right after the 21 frame. Thus, the fading region is identified from 2lSt frame to 70 frame. However, variance of the 2lSt frame is zero. Therefore this identifies the fade-in sequence accurately. D, = ~bs(r(n) - R(n -1)) (7) Therefore, considering the absolute value of the differentiation as defined in Equation (7), it is possible to detect the fade regions by setting a very low threshold ( Tfilde). It may be possible to satisfy the condition (DR < Trade ) for a small number of consecutive frames in a non fade-in sequences. But, small size of fade-in sequences are not common in practice and also there should be a frame with zero variance either at the beginning or at the end of the detected region. Using this argument false regions are eliminated from the proposed algorithm. Furthermore, if there are two consecutive fade regions separated by very small gap, they are bridged to form a longer fade region. Similar set of equations and arguments can be derived easily for fade-out with the initiative function in Equation (8). Finally, algorithm identifies the fade-in and fade-out as follows. FrameNrmber Figure 1: Mean of the luminance signal of the test sequence 160 120 590 B 40 0 0 40 83 120 160 m 240 FrmNmber Figure 2: Mean of the chrominance signal of the test sequence Fade-in: Detect a frame with zero variance followed by a sequence of continuous region, which is identified by DR. IV-257

900 3 800 700 L600 g 500.= 400 ;. 300 f 200 E 100 n JlWl Sequence 0 50 100 150 200 250 Frame Number Figure 3: Differentiation ofr(n) ( D, ) for the test sequence ( Tfade = 5) I 2 Actual fade Detected Nature of region fade the region region 21-70 I 21-70 I fade-in 299-337 I 299-337 fade-in 348-403 1 348-403 I fade-out 444-546 I 444-546 I fade-out 595-667 I 595-668 I fade-out 44-86 44-87 fade-in 302-467 302-467 fade-out 565-620 565-620 fade-in Table 1: Summarised results for two different sequences Sequence 1: Length of 700 frames and contains two fade-in and three fade-out regions. This sequence doesn t contain either any other special effects or sudden scene changes. 5. CONCLUSIONS The ratio between incremental change in the mean of the luminance signal to the chrominance signal (average sum of C, and C,) is considered as the criteria of identifying fading transitions. Results show that the algorithm is capable of detecting all fade regions accurately even the video sequence contains other special effects. Therefore, the proposed algorithm can be used in uncompressed video to detect fade-regions with a very higher reliability rate. Further work is required to extend this algorithm for compressed video. ACKNOWLEDGEMENTS Authors would like to express their gratitude and sincere appreciation to the university of Bristol and CVCP for providing financial support for this work. 3. REFERENCES Zhang, H.J., Kankanhalli, A., and Smoliar, S.W., Automatic Partitioning of Full-Motion Video, ACWSpinger Multimedia Systems, Vol. 1, No. 1,pp. 10-28, 1993. Otsuji, K., Tonomura, Y., Projection Detecting Filter for Video Cut Detection, Proc. 1 ACM International Conference on Multimedia, Anaheim CA, pp.251-257, August, 1993. Shahraryay, B., Scene Change Detection and Content Based Sampling of Video Sequences, Digital Video Compression: Algorithms and Technologies, Vol. SPIE-24 19, pp. 2-13, February 1995. Sequence 2: Length of 700 frames and contains one fade-out regions, two fade-in regions, two sudden scene changes, and several other special effects like zoom-in, zoom-out, panning and tilting. Table 1 shows the summarised results of the proposed algorithm with the above two sequences. Results show that the algorithm is capable of detecting all fade regions accurately even the video sequence contains other special effects. However, this algorithm may fail to identify fading regions when the solid colour is very close to the mean of the original sequence (before fading is applied). 4-5- 6. Fernando, W.A.C., Canagarajah, C.N., Bull, D. R., Video Segmentation and Classification for Content Based Storage and Retrieval Using Motion Vectors, Paper Number-3656-69, SPIE 99, San Jose, California, USA. Zabih, R., Miller J, Mai, K., Feature-Based Algorithms for Detecting and Classifying Scene Breaks, Proc. 4 h ACM International Conference on Multimedia, San Francisco, California, November 1995. Alattar, A, Detecting Fade Regions in Uncompressed Video Sequences, pp.3025-3028, ICASSP 1997. IV-258