Upload
rashmi-shankar
View
213
Download
0
Embed Size (px)
Citation preview
7/29/2019 Object Extraction in Data Mining Framework for Video Sequenc
1/6
P21/1
Abstract -We are aware that the emerging video coding standard such as MPEG-4 enables various content-
based functionalities for multimedia applications. To support such functionalities and also to improve the
coding efficiency, each frame of an image sequence is decomposed into video object planes (VOPs). Each VOP
corresponds to a single moving object in the scene. This paper proposes an algorithm of key object(s)
extraction from video database. Here we describe a framework of data mining by applying segmentation and
enhancement based technique to the image and video data set, for the purpose of knowledge discovery. The
objective of this particular segmentation algorithm is to segment the huge image and video data sets in order to
extract important objects from them. This proposed method covers both spatial and temporal aspects of data
mining. The video data set is parsed to create the image data set by preserving temporal information, then the
background subtraction method is applied to get the moving object, and then double differencing operator is
applied to get better result. The gamma enhancement function is applied with histogram equalization with a
number of different parameters to improve the contrast of the resultant images. This method is thus able to
detect moving object(s) with surprisingly good location capability, and eliminate noise, which occurs due to
camera movements. The proposed approach is applied to a traffic video data set to demonstrate the
segmentation for data mining of video database.
KEYWORDS: Video mining, Video object segmentation, background subtraction, enhancement technique.
1 Introduction
There is an enormous need for visual information management in the growing field of medical imaging and
telemedicine, traffic monitoring etc. In particular, image understanding form image database queries and
retrieval are major topics of research [1,2]. This newly emerging field combines aspects from databases, imageprocessing and understanding, knowledge-based systems and context-based compression. Thus this growing
amount of video information requires efficient management, and to support more efficient video database
management, a video segmentation method has become necessary.
Segmentation is one of the most challenging tasks in the field of video mining, since it lies at the base of
virtually any scene analysis problem. The general approach that motivates the need for segmentation is that the
problem of analyzing a complete scene can be subdivided into a set of simpler problems, of analyzing its
primitives. Then, each primitive object in the scene can be processed independently [3][4]. This approach
greatly simplifies many high-level vision problems, such as recognition, classification, scene analysis, etc.
However, the ultimate goal of the segmentation process is to divide an image into the physical objects
composing it, so that each region constitutes a semantically meaningful entity.
In order to identify and track the temporal and relative spatial positions of objects in video sequences, it is
necessary to have object-based representation of video data. For this purpose, attention has been devoted tosegmenting video frames into regions such that each region, or a group of regions, corresponds to an object that
is meaningful to human viewers [7, 8]. While most of the previous works [5] [6] are based on low-level global
features, such as color histogram, shape and texture feature, our method focuses on obtaining object level
segmentation; obtaining objects in each frame and their traces across the frames.
There has been a large amount of researches [18, 19] to find object from video sequence when the segmentation
problem is considered for a fixed camera domain, a usual technique is required to resolve the foreground
Ms. Stuti Dave
Samrat Ashok Technological Institute,
Vidisha, M.P., India
Prof. Manish Manoria
Samrat Ashok Technological Institute,
Vidisha,M.P.,India
Object Extraction in Data Mining Framework for Video Sequence
7/29/2019 Object Extraction in Data Mining Framework for Video Sequenc
2/6
P21/2
objects i.e. background subtraction [9]. This involves the creation of a background model that is subtracted
from the input image to create a difference image. The new difference image only contains objects or new
features that have not yet appeared in the background. This method is especially suitable for video conferencing
[10] and surveillance [11] applications, where the backgrounds remain still. Neri et al. [12] suggested a
different segmentation technique based on change detection. The method assumes that the objects are moving
over a static background, thus, if a moving camera is detected, then global motion compensation be applied
first. Potential foreground regions are detected in a preliminary stage by applying a higher-order statistics
(HOS) test to a group of inter-frame differences. The non-zero values in these difference frames are due to
moving objects.Few references [13, 14] have addressed the problem of object segmentation and tracking using an active
camera, they cannot able to give spatial information of moving object, which is very inefficient and non-flexible
This method is thus able to detect moving object(s) with surprisingly good location capability.
2 Video Objects Extraction
Figure 1 depicts the diagram of the proposed object extraction scheme. Prior to performing segmentation and
extraction, a video database is created. It is a temporal data because it contains before-after sequence
relationship that is first frame comes before second frame and second frame comes after first frame. Video
parsing divides large video sequence in to frames or images.
Figure 1 diagram of the proposed object segmentation and extraction method
The number of images/frames per second gives motion in video. Then reference frame of stationary component
is created. The proposed background substraction method is applied which is able to detect moving object from
each frame.
2.1 Background subtraction
Background subtraction is a technique to remove nonmoving components from a video sequence. The
main assumption for its application is that the camera remains stationary. The basic principle is to create a
reference frame of the stationary components in the image.
Step1: Once reference frame of stationary component is created, the reference frame is subtracted from any
subsequent images. Those pixels resulting from new (moving) objects will generate a difference not equal to
zero.
Step2: In the video sequences the nonmoving objects were manually selected from the video data and then
averaged together. Here accumulating and averaging images of the target area in some time interval construct a
reference frame. The image sequence used consists of about 16 minutes of video from a traffic intersectionwith approximately constant lighting conditions. (The difference image is created by subtracting the reference
frame from the current image) .The results are scaled by eq. 1.
s = clog (1+|dij|) (1)
Where dij is the value for the difference at pixel ij
Step3: The double-difference operator (also called three-frame difference) has been applied to improve the
performance of background subtraction. It is able to detect moving points with surprisingly good location
Video file
ImageDatabase
BackgroundSubtraction
ObjectDetection
Perform gammacorrection
Object
extracted
Create negative
image
Improve
contrast
7/29/2019 Object Extraction in Data Mining Framework for Video Sequenc
3/6
P21/3
capability, and eliminate noise, which occurs due to camera movements and able to filter very small moving
objects in the scene. Hereafter, we sketch the double-difference operator. Eq.2 defines the difference-image.
Dri (i,j) = [ Iri(i,j) - I r,i-1(i,j) ] (2)
Step 4: The double-difference image is obtained by eq.3 which perform a logical AND between pixels
belonging to two subsequent difference-images, threshold by T
1, if (Dri+1 (i,j) >T ) ( Dri (i,j) > T )D` Dri (i,j) =
0 otherwise (3)
The non-zero values in these difference frames are either due to noise or moving objects. The
intensity of the resultant image is improved by gamma correction; here different perimeters are tried to obtain
best result.
2(a) Reference frame 2(b) Frame 4
2(c) key object extracted from frame 4
Figure 2: Example result of background subtraction
Figure 2(c) is the resultant image having key object occurring after applying the background substraction
algorithm.
Some approaches [17] or [12], which used background subtraction method, constructed the reference by
accumulating and averaging images of the target area for some time interval. As mentioned above, this is not a
robust technique as it is sensitive to intensity variations [15],[16]. The segmentation algorithms usually start
with the gray value difference image between two consecutive frames. In this paper the computation of images
has performed on gray level images because the gray level of the images or region is having uniform intensity.
The intensity change in data flow can very well be observed in gray level. Then the gamma correction function
is applied to improve the contrast of the resultant image as shown in figure 3.
3 Experimental Setup
We have applied the proposed scheme to the traffic video sequence in our experiments. Their resolution
is 195 x 255 pixels. The color video sequence consists of about 10 minutes of video from a traffic intersection
with approximately constant lighting conditions. The color video clips used for this study were originally
digitized in AVI format at 16 frames/second. The original video frames were of size 390 rows 510 columns, 24-
bit color. To reduce computation time, we made our test video clips by extracting frames from these originals at
the rate of 4 frames per second. For simplicity and real-time processing purpose, we transform the color video
frames to gray scale images and resize them to half of the original size (195 rows 255 columns). A small portion
7/29/2019 Object Extraction in Data Mining Framework for Video Sequenc
4/6
P21/4
of the traffic video is used to illustrate how the proposed framework can be applied in various kinds (i.e., TV
commercials, movie, and documentary, TV dramas, and animation). This process can be done in real-time or
off-line.
3.1 Experiment Results
The enhanced background subtraction method to the video sequences is able to remove nonmoving
components from a video sequence. The main principle is to create a reference frame of the stationary
components in the image. The reference frame is subtracted from any subsequent images. Those pixels resulting
from new (moving) objects will generate a difference not equal to zero .The double-difference operator is used,which is able to detect moving points with amazingly good location capability, and is proved robust to noise
due to camera movements and able to filter very small moving objects in the scene. By using this background
subtraction method the accuracy of segmentation is increased and this thus provides support for real-time
processing. The segmentation results for a few frames are shown in Figure 2 along with the original frames
adjacent to them. Related work has been done based on image processing and rule-based reasoning. Here we
applied the gamma correction and tried different perimeters to increase the contrast of the resultant image.
This proposed framework, however can deal with more complex situations with more accuracy. Some of the
objects, which are closely located to each other, are extracted in single segment. In Figure 2, the frame in the
left most column (Figure 2(a)) is the background reference frame. The second column (Figure 2(b)) shows the
original frame. The figure 2(c) shows the images after background subtraction. The extracted key object is
shown in the third column (Figure 2(d)). Figure 3 shows the result gamma corrected images with parameters
0.3, 1.5, 3, 2.1. The result shown in figure 5 shows the improved intensity of the resultant frame compared with
original one.
Figure 3 gamma corrected frame with parameters 0.3, 1.5, 3, 2.1
Figure 4 Histogram before Equalization (a) and After Equalization (b)
7/29/2019 Object Extraction in Data Mining Framework for Video Sequenc
5/6
P21/5
4. Conclusion
This paper introduces an algorithm for key object extraction for segmentation in video sequences. The
main assumption underlying by proposed algorithm is that a motion characterizes physical objects that is
separate from background. The algorithm is based on background substraction method. Experimental results
demonstrate that proposed technique can successfully extract moving objects from video sequence. The
boundaries of the extracted objects are accurate enough to place them in different frames. If there is insufficient
motion, the algorithm is able to perform the correct segmentation and show the position of object in each frame.
In order to extract the key object(s) from the video sequence, background subtraction techniques areemployed. Using the background subtraction technique, both the efficiency of the segmentation process and the
accuracy of the segmentation results are improved achieving more accurate video indexing and annotation. The
existing schemes try to extract key frames from the video shot, but in this proposed scheme is able to extract
key object(s) instead of key frame(s). Here we proposed a simple and efficient technique for object
segmentation in video sequence. It is able to detect moving points with surprisingly good location capability,
and eliminate noise, which occurs due to camera movements. It is easy to implement and fast to compute since
it uses the segmentation result of the previous video frame to speed up the segmentation process of the current
video frame. The proposed work extract small moving component(s) from image sequence and therefore this
opens scope for possible future work in the area of doing all operation in image processing, and the global
motion estimation and compensation. By considering more than one reference frame in the segmentation
process, the motion can be estimated by comparing each frame with more than one reference frame so as to
detect all the information of video scene. This work opens scope for futher study on data mining work,clusttering and association mining on video data set.
5. Acknowledgement
This work was supported by all India council of technical education AICTE),New Delhi,India under RPS
grant F.No.-8022/RID/NPROJ/RPS-14/2003-04.
References
1. Jiawei Han, Data Mining Techniques, DBLP, SIGMOD, 1996, pp: 545 [Conf]
2. Michael C. Burl, Mining Patterns of Activity from Video Data, Department of Computer Science University
of Colorado, Boulder, First SIAM International Conference on Data Mining, 2002.
3 R. Mech, M. Wollborn, "A Noise Robust Method for Segmentation of Moving Objects in Video Sequences,"
p. 2657, 1997 IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP'97) -
Volume 4, 1997.
4 Xiaodong CAI, F.H.Ali and E.Stipidis An Integrated framework for Robust and Fast Automatic Video
Segmentation Communication Research Group School of Engineering and Information Technology University
of Sussex, Publication Date: Feb,2002,Volume,12,Issue:2,On page(s): [122-129]P21-5
5 P.Stanchev, IASTED, Using image mining for image retrieval, Cancum, Maxico, 214-218, conf., CSE, May
19-21,2003
6 QBIC project: querying images by content, using color, texture, and shape Carlton W. Niblack, Ron Barber,
Will Equitz, Myron D. Flickner, Eduardo H. Glasman, Dragutin Petkovic, Peter Yanker, Christos Faloutsos,
and Gabriel Taubin , April 1993, pp. [173-187]
Paper No. P21
7 Teuner, A.Pichler, Hosticka, B.J. Fraunhofer Inst. of Microelectron. Circuits & Syst., Duisburg;
Unsupervised texture segmentation of images using tuned matched Gabor filters, Image Processing, IEEE
Transactions on, Jun 1995, Volume: 4, Issue: 6 On page(s): 863-870
8 Jia-Yu Pan and Christos Faloutsos, "Geoplot": Spatial Data Mining on Video Libraries. In Proceedings of the
Eleventh International conference on information and,Knowledge, Management, CIKM,2002, Mclean, VA
(near Washington DC), November ,pp [4-9], 2002
7/29/2019 Object Extraction in Data Mining Framework for Video Sequenc
6/6
P21/6
9 YuriA. Ivanov, Aaron, F.Bobick, JohnLiu,Fast Lighting Independent Background Subtraction,DBLP,
International Journal of Computer Vision, 2000, v:37, n:2, pp:199-207 [Journal]
10 C.W. Lin, Y.-J. Chang, Y.-C. Chen, and M.-T. Sun, Implementation of a real time object-based virtual
meeting system, in Proc. IEEE Int. Conf. Multimedia and Expo, pp. [565-568], August 2001, Tokyo, Japan.
11 I. Haritaoglu, D. Harwood, and L.S. Davis, W4: Who? When? Where? What? A real-time system for
detecting and tracking people, in Proc. IEEE Int. Conf. Automatic Face and Gesture Recognition, pp. [222-
227],1998, Nara, Japan.
12 A.Neri, S. Colonnese, G. Russo, P. Talone, Automatic moving object and background separation, Signal
Processing 66 (1998), pp. [219-232].13 Y. Ye, J. K. Tsotsos, K. Bennet, and E. Harley, Tracking a person with pre-recorded image database and a
pan, tilt, and zoom camera, in Proc. IEEE Workshop on Visual Surveillance, pp.[10-17], 1998.
14 Hat, M. Saptharishi, and P. K. Khosla, Motion detection and segmentation using image mosaics, in Proc.
IEEE Int. Conf. Multimedia and Expo, pp. 1577-1580, July 2000, NY, USA.
15 Cheung, C. Kamath, Robust Techniques for Back-ground Subtraction in Urban Traffic Video", In Video
Comm and Image Proc., SPIE Electronic Imaging, San Jose 2004
16. Ismail Haritaoglu, David Harwood, Larry S. Davis: A Fast Background Scene Modeling and Maintenance
for Outdoor Surveillance. ICPR 2000: 4179-4183
17. T. Meier and K.N. Ngan, Automatic segmentation based on Hausdorff object tracking,.Doc. ISO/IEC
JTC1/SC29/WG11 MPEG97/m2238, July 1997.
18.T. Meier, and K.N. Ngan, Automatic segmentation of moving objects for video object plane generation.
IEEE Trans. on Circuits and Systems for Video Technology, Vol. 8, No.5, September 1998.