Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
0% found this document useful (0 votes)
22 views3 pages

Moving Object Detection in The Real World Video: Dr. G. Shobha, Vartika Mudgal

Download as pdf or txt
Download as pdf or txt
Download as pdf or txt
You are on page 1/ 3

IOSR Journal of Computer Engineering (IOSRJCE) ISSN: 2278-0661 Volume 1, Issue 2 (May-June 2012), PP 31-33 www.iosrjournals.

org

Moving object detection in the real world video


Dr. G. Shobha, Vartika Mudgal
2

(Dean PG CSE, CSE Department, RVCE, Bangalore) (2nd year M.tech(CSE), CSE Department, RVCE, Bangalore)

Abstract: Detecting moving objects using stationary cameras has been responsible for reducing the reliability
of many computer vision algorithms, including segmentation, object detection, scene analysis, stereo, tracking, etc. Therefore, detecting moving object is an important pre-processing for improving performance of such vision tasks. In this paper we proposed a method to extract the key frames, detect the moving objects . Firstly the video is segmented into shots after that the key frames are extracted. The moving objects is detected using background subtraction.

Keywordsmoving object detection, background subtraction I. Introduction

In recent years, there has been a dramatic proliferation of research works concerning with content based video retrieval. Video segmentation is first step towards the content based video search aiming to segment moving objects in video sequences. Video segmentation initially segments the first image frame as the image frame into some moving objects and then it tracks the evolution of the moving objects in the subsequent image frames. After segmenting objects in each image frame, these segmented objects have many applications, such as surveillance, object manipulation, scene composition, and video retrieval [1]. Video is created by taking a set of shots and composing them together using specified composition operators. A shot is defined as an image sequence that presents continuous action which is captured from a single operation of single camera. Shots are joined together in the editing stage of video production to form the complete sequence. Shots can be effectively considered as the smallest indexing unit where no changes in scene content can be perceived and higher level concepts are often constructed by combining and analysing the inter and intra shot relationships. Key-frames are still images extracted from original video data that best represent the content of shots in an abstract manner. Key-frames have been frequently used to supplement the text of a video log, though they were selected manually in the past. Key-frames, if extracted properly, are a very effective visual abstract of video contents and are very useful for fast video browsing. A video summary, such as a movie preview, is a set of selected segments from a long video program that highlight the video content, and it is best suited for sequential browsing of long video programs. Colour component, or a colour channel, is one of the dimensions.

II. METHODOLOGY In the following section we will explain the methods used in moving object detection: Shot Segmentation
The video has to be split into chunks or video shots prior of conducting any video object analysis. Scene change detection, either abrupt scene changes or transitional (e.g. dissolve, fade in/out, wipe) is employed to achieve the video shot separation [2]. A shot can be defined as a sequence of frames taken by a single camera without any significant change in the colour content of consecutive images

III.

Key frame extraction

Key frames may be one frame or some frames reflecting the main information content in the shot. The extraction of key frames is the selection of a frame image in the shot of represents the video contents of the shot. By the calculation of key frame extraction, the redundancy would decrease in the scene effectively [3]. For the selection of the key frames in the videos, the frame must can reflect the primary affairs in the shots, which exactly and fully describes the scene as much as possible.

www.iosrjournals.org

31 | Page

IV.

Moving object detection in the real world video Motion detection

We assume that the camera for recording video data is stationary and the background changes are relatively slow to the targets motion in the scene. We use some image frames without moving objects to compute statistical quantities for the background scene. Then, we detect the foreground pixels and extract features [4]. We will explain how to extract moving objects in this sub-section.

V.

Background image

The initial background image modelling is carried out over first few seconds image frames (n) according to the video shots.. With the assumption of no moving objects in those few image frames (n), build the reference colour background image with a normal distribution. We model the background by computing the sample mean (x,y) and variance 2(x,y) in the colour images over a sequence of those(n) frames. These statistics are calculated separately for each one of the RGB components by using the following iterative formula [5]. For image frame f=2,,n , we have [( )]f =[( )]f-1 + ([ C( )]f - [( )]f-1), (1)
f

f-1

+ ([(

)]f-1 -([C(

)2

Where [.] f , denotes the corresponding value at frames f, and [( )] = [C( )] and = 0. Equations (1) can be shown by straightforward calculations that they yield the sample mean and variance over the first (n) image frames [4]. The sample mean is the background image. The background image of the (x,y) h pixel s RGB values over the first (n) frames is given by [5] : ( ) = [ R ( ) , G ( ), B ( ,] (2)

and the variance of the ( =[


R(

)h pixel's RGB values over the first 30 image frames is given by[5] :
G

),

(3)

Which the image brightness changes sharply or, more formally, has discontinuities. The purpose of detecting sharp changes in image brightness is to capture important events and changes in properties of the world.

Figure1. Block diagram of moving object detection

i. Background subtraction We defined the term moving pixels as the foreground. In each new image frame C(x,y), the foreground can be obtained by comparing their RGB values to the corresponding mean values. We first create a binary image D(x,y), with the same dimension as the image C(x,y), and set all its pixel values to 0. The output of the background subtraction method is defined as follows:
www.ijmer.com 32 | Page

Moving object detection in the real world video


D(x,y) = (foreground) , if [ C(x,y)- (x,y) ] > . . (x,y) 0(background), otherwise (4)

A pixel (x,y) is extracted as a foreground if it RGB values C(x,y), satisfies the absolute difference value with (x,y), where the parameter can be adjusted to yield more or less foreground. In all experiments of this paper, we use 3 as the threshold in background subtraction. ii. Background updating The background cannot be expected to be stationary for a long period of time. We need to update the background parameters for the lighting changes. An adaptive scheme makes a constant updating of background as linear combination of previous background image and current image frame. The recursive estimation of mean and variance can be performed using equations (5.1) and (5.2). Equations (5.1) and (5.2) update the background image and sample variance, respectively [6,7]: ( )f = .C(x,y)f + (1-) .[ ( )]f-1 = .( C(x,y)f - ( )]f )2 + (1-) . f (5.1)
f-1

(5.2) Where , C( ) is the current image frame, ( )f and f are mean and variance values which updates the current image frame values, determines the speed rate at which distribution parameters change. (0 < < 1).

VI.

Conclusion

In this paper, we apply the technology of moving-object tracking to content base video retrieval. We use background subtraction algorithm to detect moving pixels. Apart from moving object detection, the work can be done in eliminating shadows from the video.

REFERENCES
[1] [2] [3] [4] [5] [6] B. V. Patel, B. B. Meshram (2007), Retrieving and Summarizing Images from PDF Documents, International Conference on Soft computing and Intelligent Systems (ICSCSI-07),Jabalpur, India. AN ENHANCED CONTENT-BASED VIDEO RETRIEVAL SYSTEM BASED ON QUERY CLIP T.N.SHANMUGAMand PRIYA RAJENDRAN International Journal of Research and Reviews in Applied Sciences (December 2009) 3rd International congress on Image and Signal Processing (CSIP) 2010: Battlefield video target mining based image retrieval at the end of the early years. Content Based Video retrieval with Motion vectors and the RGB Colour Model. Forensic Science Journal 2007. H. T. Chen, H. H. Lin, T. L. Liu. Multi-object tracking using dynamical graph matching, IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR01) - Vol. 2, p.210, 2001. S. J. McKenna, S. Jabri, Z. Duric, A. Rosenfeld, and. H. Wechsler, Tracking groups of people, Computer Vision and Image Understanding 80, 4256, 2000. [7] J. R. Smithand S.-F. Chang, Single color extraction and image query, in Proc. IEEE Int. Conf. on Image Proc., 1995. S. Jabri, Z. Duric, H. Wechsler, A. Rosenfeld. Detection and location of people in video images using adaptive fusion of color and edge information, 15th International Conference on Pattern Recognition ,Vol. 4, p.4627, 2000.

[7]

www.ijmer.com

33 | Page

You might also like