American Journal of Signal Processing
p-ISSN: 2165-9354 e-ISSN: 2165-9362
2012; 2(2): 35-40
doi: 10.5923/j.ajsp.20120202.06
G. G. Lakshmi Priya , S. Domnic
Department of Computer Applications, National Institute of Technology, Tiruchirappalli, Tamilnadu, India
Correspondence to: G. G. Lakshmi Priya , Department of Computer Applications, National Institute of Technology, Tiruchirappalli, Tamilnadu, India.
| Email: | ![]() |
Copyright © 2012 Scientific & Academic Publishing. All Rights Reserved.
In this paper, we propose a new method for detecting shot boundaries in video sequences by performing Hilbert transform and extracting feature vectors from Gray Level Co-occurrence Matrix (GLCM). The proposed method is capable of detecting both abrupt and gradual transitions such as dissolves, fades and wipes in the video sequences. The derived features on processing through Kernel k-means clustering procedure results in efficient detection of abrupt and gradual transitions, as tested on TRECVid video test set containing various types of shot transition with illumination effects, object and camera movement in the shots. The results show that the proposed method yields better result compared to that of the existing transition detection methods.
Keywords: Transition detection, Hilbert Transform, GLCM, Feature Vectors, Kernel k-means clustering, Performance Evaluation
![]() | (1) |
,![]() | (2) |
is the Fourier transform of the function
,
is the odd signum function.
is +1 for
> 0, 0 for
= 0 and -1 for
< 0.The Hilbert transform is considered as a filter, which simply shifts phases of all frequency components of its input, by±π/2 radians. In general, video is a sequential collection of frames whose intensity values are represented as a set of real numbers R. More features like color, edge, texture etc., information can be extracted from each frame. Instead of using the whole frames’ intensity values as features, transformed information can also be considered, is the condensed representation of the original. In our work the transformed information is obtained by applying above discussed Hilbert transformation over the frames. To perform Hilbert transformation, the input video data is to be converted into 1-D signal f (t). The RGB frame shown in figure 2a is converted into gray scale image as shown in figure 2b and the Hilbert transformed image is shown in figure 2c. Based on this transformed information, GLCM features are extracted.![]() | Figure 1. Block diagram of the proposed method |
![]() | Figure 2. (a) Original frame (b) Gray scale frame (c) Hilbert transformed frame |
. ‘d’ is measured in pixel number. Normally
is quantized in four angles (0°, 45°, 90°, 135°). Let
represents the GLCM for an image
for distance d and direction
can be defined as ![]() | (3) |
. For each
value, it’s
and
values are (0, 1) for 0°, (-1,-1) for 45°, (1, 0) for 90°, (1,-1) for 135°.![]() | (4) |
![]() | (5) |
![]() | (6) |
![]() | (7) |
is mean,
is the standard deviation of GLCM
. However, for a constant image the texture features are
. The four features
are the functions of distance and angle. For a chosen distance d, four angular GLCM are measured and hence a set of four values for each of the four features are obtained. On the whole 16 feature measures are generated. To reduce the number of features, the average of four angular GLCM is taken using (8) and then the four features are extracted as the Average GLCM (AGLCM). ![]() | (8) |
from AGLCM is constructed which is feature representing each frame. In order to find the relationship between the consecutive frames, dissimilarity / similarity between these frames are to be carried out. The simplest method for shot detection is to compute absolute difference
between the F of k and k+1th frame. Thus the relation between the consecutive frames is calculated using:![]() | (9) |
, the k-means algorithm finds clusters
that minimizes the objective function![]() | (10) |
[16] before clustering points. Then kernel k-means partitions the points by linear separators in the new space. The main objective of the kernel k-means can be written as a minimization of ![]() | (11) |
The higher these ratios are, the better the performance. Our proposed algorithm is implemented in Matlab 7.6, Table.2 shows the experimental results of the proposed method for detection of abrupt and gradual transition.
| ||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
|
![]() | Figure 3. Results obtained for Anni001 video sequence |
| [1] | Tudor Barbu, “A novel automatic video cut detection techniques using Gabor filtering,” Computer and Electrical Engineering, vol.35, pp. 712-721, Sep. 2009. |
| [2] | U.Gargi, Kasturi, and S.H.Strayer, “Performance characterization of video shot change detection methods,” IEEE Trans. on Circuits and Systems for Video Technology, CSVT-10(1), pp.1-13, 2000. |
| [3] | Rainer Lienhart, “Comparison of automatic shot boundary detection algorithm,” Image and video processing VII, in Proc. SPIE, 3656-3659, 1999. |
| [4] | J. S. Boreczky and L. Rowe, “Comparison of video shot boundary detection techniques,” in Proc. IS&T/SPIE Storage and Retrieval for Still Image and Video Databases IV, vol. 2670, pp. 170–179,1996. |
| [5] | R. Zabih, J. Miller, and K. Mai, “A feature-based algorithm for detecting cuts and classifying scene breaks,” in Proc. ACM Multimedia ’95, San Francisco, CA, pp. 189–200, 1995. |
| [6] | Hun-Woo Yoo & Han-Jin Ryoo & Dong-Sik Jang, Gradual shot boundary detection using localized edge blocks, Multimedia tools appls. 28: 283-300, 2006. |
| [7] | Y. Kawai, H. Sumiyoshi, and N. Yagi. “Shot Boundary Detection at TRECVID 2007,” In TRECVID 2007 Workshop, Gaithersburg, 2007. |
| [8] | Shiguo Lian. Automatic video temporal segmentation based on multiple features. Soft Computing, Vol. 15, 469-482, 2011. |
| [9] | Weigang Zhang, et. al., “Video Shot Detection Using Hidden Markov Models with Complementary Features,” In Proceedings of the First International Conference on Innovative Computing, Information and Control.Vol.3.http://doi.ieeecomputersociety.org/10.1109/ICICIC.2006.549, 2006. |
| [10] | Damian Borth, Adrian Ulges, Christian Schulze, Thomas M. Breuel: Keyframe Extraction for Video Tagging & Summarization. In: Informatiktage pp. 45-48, 2003. |
| [11] | Ahmed O. Abdul Salam, “Hilbert transform in image processing” in Proc. ISIE, pp. 111-113, 1999. |
| [12] | Lakshmi Priya G.G., Domnic S., “ Video cut detection using Hilbert transform and GLCM”, in proc. of IEEE-International Conference on Recent Trends in Information Technology, 749-753, 2011. |
| [13] | Inderjit S. Dhillon, Yuqiang Guan, and Brian Kulis, Weighted Graph Cuts without Eigenvectors: A Multilevel Approach, IEEE Transactions On Pattern Analysis And Machine Intelligence, VOL. 29, NO. 11,1944- 1957, 2007. |
| [14] | Panchamkumar D Shukla, Complex wavelet transforms and their applications, Thesis of Master of Philosophy(M.Phil.), University of Strathclyde , United Kingdom, 2003. |
| [15] | Haralick, R.M., Shanmugam, K., Dinstein, I.: “Textural Features for Image Classification,” IEEE Trans. Systems Man Cybernet. SMC-3, 610 – 621, 1973. |
| [16] | B. Scholkopf, A. Smola, and K.-R. Muller, “Nonlinear Component Analysis as a Kernel Eigenvalue Problem,” Neural Computation, vol. 10, pp. 1299-1319, 1998. |
| [17] | TRECVID Dataset website: http://trecvid.nist.gov/. |
| [18] | Video Dataset: http://www.open-video.org. |