Top Banner

of 8

Important Application of Composite

Jun 03, 2018

Download

Documents

chethugowda7
Welcome message from author
This document is posted to help you gain knowledge. Please leave a comment to let me know what you think about it! Share it to your friends and learn new things together.
Transcript
  • 8/12/2019 Important Application of Composite

    1/8

    IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 11, NO. 2, FEBRUARY 2002 105

    Integrating Intensity and Texture Differences forRobust Change Detection

    Liyuan Li, Student Member, IEEE, and Maylor K. H. Leung, Member, IEEE

    AbstractWe propose a novel technique for robust changedetection based upon the integration of intensity and texturedifferences between two frames. A new accurate texture differ-ence measure based on the relations between gradient vectorsis proposed. The mathematical analysis shows that the measureis robust with respect to noise and illumination changes. Twoways to integrate the intensity and texture differences havebeen developed. The first combines the two measures adaptivelyaccording to the weightage of texture evidence, while the seconddoes it optimally with additional constraint of smoothness. Theparameters of the algorithm are selected automatically based ona statistic analysis. An algorithm is developed for fast implemen-tation. The computational complexity analysis indicates that theproposed technique can run in real-time. The experiment resultsare evaluated both visually and quantitatively. They show thatby exploiting both intensity and texture differences for changedetection, one can obtain much better segmentation results thanusing the intensity or structure difference alone.

    Index TermsChange detection, difference pictures, integra-tion, motion detection, segmentation, subtraction, video analysis.

    I. INTRODUCTION

    FINDING moving objects in image sequences is one of

    the most important tasks in computer vision. Common

    applications for motion analysis include object tracking [1],

    security surveillance [2], [3], intelligent user interfaces [4],

    and traffic flow analysis [5]. In all these applications, the first

    fundamental problem encountered is motion segmentation

    which extracts moving objects from the scene. Most motion

    segmentation techniques for dynamic-scene analysis are based

    on the detection of changes in a frame sequence [6]. Change

    detection reduces the amount of data for further processing.

    Because of its importance in the preprocessing of dynamic

    scene analysis, many techniques have been developed for

    robust change detection.

    The existing change detection techniques can be classified

    into two categories: pixel-based and region-based methods. The

    most intuitive technique to detect change is simple differencingfollowed by thresholding. Changeat a pixel is detected if the dif-

    ference in gray levels of the corresponding pixels in two images

    Manuscript received July 25, 2000; revised October 10, 2001. The associateeditor coordinating the review of this manuscript and approving it for publica-tion was Prof. Jezekiel Ben-Arie.

    L. Li is with Kent Ridge Digital Labs, Singapore, 119613 (e-mail:[email protected]).

    M. K. H. Leung is with the School of Computer Engineering, NanyangTechnological University, Singapore, 639798 (e-mail: [email protected];http://www.ntu.edu.sg/home/asmkleung).

    Publisher Item Identifier S 1057-7149(02)00461-X.

    exceeds a preset threshold. More robust methods adaptively se-

    lect the threshold based on the noise estimation of difference

    picture [7] or thresholds at each pixel based on the gray level

    distributions of the background points [4], [8]. Change detec-

    tion at pixel level requires less computational cost since only

    one pixel is considered at a time. But it is very sensitive to noise

    and illumination changes since it does not exploit local struc-

    tural information. Hence, it is not suitable for applications in

    complex environments. To make change detection more robust,

    intensity characteristics of regions at the same location in two

    images are compared using statistical approaches. A straight-

    forward solution is hypothesis testing to check if statistics ofthe two corresponding regions come from the same intensity

    distribution. The likelihood ratio test taken by Nagel [9] is one

    example of such approach. The performance of the likelihood

    ratio test can be improved by using quadratic surfaces to ap-

    proximate the intensity values of pixels belonging to the re-

    gions. Hsu et al. [10] proposed a second-order bivariate poly-

    nomial in the pixel coordinates to model the gray value varia-

    tion in the regions. A difference metric is then defined to deter-

    mine if changes occur in corresponding regions of two images.

    Approaches based on the comparison of intensity distributions

    at region level for change detection work well in noisy envi-

    ronments, but they are sensitive to illumination changes since

    they still only employ intensity difference measures. Recently,region-based illumination independent statistics have been pro-

    posed. Skifstad and Jain [11] presented a Shading Model (SM)

    which uses the ratio of the intensities recorded in the corre-

    sponding regions of two images to detect changes. The basic

    idea behind the SM method is that a change in the physical sur-

    face in the region will make the intensity ratios vary inconsis-

    tently in the corresponding regions of two images. Liu et al. [12]

    defined circular shift moments (CSM) to describe the illumina-

    tion-independent characteristics of a region. Change detection

    is performed as a hypothesis testing based on the characteris-

    tics of the defined moments. These methods can detect changes

    accurately under time-varying illumination since they are devel-

    oped based on shading or reflection models [13]. However, theyall employ ratio of intensities or sums of intensities to describe

    structural characteristics of a region. This means they may per-

    form poorly over the dark regions in images due to the denom-

    inator of the ratio becoming insignificant for these regions. In

    addition, the illumination independency of the measures is ob-

    tained only on the simple diffuse reflection model.

    The robustness of region-based difference measure is based

    on its dissimilar representations of different local structural fea-

    tures, or texture. Background texture remains relatively stable

    with respect to noise and illumination changes unless it is cov-

    10577149/02$17.00 2002 IEEE

    http://-/?-http://-/?-http://-/?-http://-/?-http://-/?-http://-/?-http://-/?-http://-/?-http://-/?-http://-/?-http://-/?-http://-/?-http://-/?-http://-/?-http://-/?-http://-/?-http://-/?-http://-/?-http://-/?-http://-/?-http://-/?-http://-/?-http://-/?-http://-/?-http://-/?-http://-/?-http://-/?-http://-/?-
  • 8/12/2019 Important Application of Composite

    2/8

    106 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 11, NO. 2, FEBRUARY 2002

    ered by moving objects or abrupt lighting change occurs. On the

    other hand, for a background region covered by a moving object,

    even though the gray level distributions of the foreground and

    background regions may be similar, the texture features of the

    two regions are usually different. From this point of view, a re-

    gion-based difference measure based on texture representation

    will be more accurate for change detection than those based on

    local statistics. However, for the homogeneous regions of fore-ground and background, the texture difference measure will be-

    come less valid than the simple intensity difference. There are

    usually many such regions in man-made objects. Therefore, it is

    desirable to integrate the intensity and texture differences adap-

    tively for robust change detection.

    In this paper, a novel texture difference measure is proposed.

    It is derived from the relations of local gradient vectors to

    compute the difference of texture patterns. Mathematical

    analysis indicates that the measure is robust with respect

    to noise and illumination changes based on a more general

    illumination model. Two methods to integrate intensity and

    texture differences have been developed. One is the adaptive

    weighted sum of two difference measures, and the other is anoptimal integration by minimizing an energy function with

    additional constraint of smoothness. Real-time implementation

    of the new technique is also proposed. Experimental results

    indicate that the proposed methods can detect changes much

    more accurately than those based on intensity or structure

    difference only.

    The rest of this paper is organized as follows. In Section II,

    a novel texture difference measure is proposed. Section III de-

    scribes two methods to integrate the differences of intensity and

    texture. In Section IV, the experiment results and comparisons

    of the proposed technique with the pixel-based and region-based

    methods are presented. Finally, this paper is concluded in Sec-

    tion V with discussions.

    II. TEXTUREDIFFERENCEMEASURE

    Texture is an important feature for image representation. It

    represents the spatial arrangement of pixel gray levels in a re-

    gion [14]. There are many approaches developed to describe the

    texture feature [15], such as the co-occurrence matrix, Fourier

    power spectrum, Markov random field, and Gabor filters. Being

    developed for image segmentation and classification, they ag-

    gregate pixels based on the similarities of local structures within

    a neighborhood. These similarities are extracted from statis-

    tics of gray level arrangement or spectrum feature over a greatnumber of samples. This goes against the requirement for real

    time change detection to compare two small regions. For this

    purpose, a simple and efficient texture difference measure is de-

    rived here.

    A good texture difference measure should be able to repre-

    sent the difference between two local spatial gray level arrange-

    ments accurately. Since the gradient value is a good measure to

    describe how the gray level changes within a neighborhood and

    it is less sensitive to light changes, it can be used to derive an

    accurate local texture difference measure. Let be

    a point in an image plane, then the th frame and its gradient

    vector can be represented as and

    with and . Here the partial

    derivatives are generated using the Sobel operator. At a point ,

    the cross-correlationof gradient vectors of two frames can be

    defined as

    (1)

    where is the angle between the vectors. Similarly, theauto-

    correlation of a gradient vector can be defined as

    (2)

    Obviously, we have

    (3)

    If there is no change within the neighborhood of a point in

    the two frames, the local texture features of this region in the

    two frames would be similar, i.e., there are no great differences

    in length and direction of and . Hence, one has

    . On the other hand, if there is a

    change in the corresponding regions of two frames, there wouldusually be large differences between the local textures of the

    two frames since they are from surfaces of different objects.

    In this case, the gradient vectors and would be

    different in both length and direction. Then would

    become much smaller than . Lets define a

    measure of the neighborhood gradient difference as

    (4)

    where denotes the 5 5 neighborhood centered at . Since90 would indicate much difference between the vectors

    with , we normalize by setting as the

    maximum value. The robustness of with respect to illumina-

    tion changes and noise can be illustrated as follows.

    Illumination Factor: According to the general illumination

    model [17], the light from a point is composed of three parts,

    which are the ambient light, diffuse reflection, and specular re-

    flection. Theintensity of a pixel in an imageframeis represented

    as

    (5)

    where and are the reflection coefficients, the

    ambient-light intensity, the intensity of light source, the

    angle of incidence between the light-source direction and the

    normal of the surface, and the viewing angle relative to the

    specular reflection direction. To visualize the effect of illumi-

    nation change on a particular background pixel , one can treat

    and as constants from one frame to the other [11], [12] and

    rewrite the intensity equation as

    (6)

    Consequently, one gets

    (7)

    http://-/?-http://-/?-http://-/?-http://-/?-http://-/?-http://-/?-http://-/?-http://-/?-http://-/?-http://-/?-
  • 8/12/2019 Important Application of Composite

    3/8

    LI AND LEUNG: INTEGRATING INTENSITY AND TEXTURE DIFFERENCES 107

    Fig. 1. The values with respect to .

    with

    (8)

    where and can be assumed constants within a small image

    region between two frames under the assumption of the smooth-ness of local illumination [11], [12], [16]. Equation (7) becomes

    the same as the general linear brightness model proposed by Ne-

    gahdaripour [16] and it is more general than the models used

    in [11], [12]. Since the gradient values are calculated by using

    Sobel operator, it can be shown that

    (9)

    This equation reveals that the gradient is invariant to the ambient

    illumination changes. Substituting (9) into (1), (2), and (4), one

    obtains

    (10)

    When the intensity of light-source does not change too much,

    one has , and . In practice, even though the

    light-source intensity decreases to 26.8% ( ) or in-

    creases to 373.2% ( ) of the original level, is

    still below 0.5 as shown in Fig. 1. Hence, it can be concluded

    that the gradient difference measure is robust to illumination

    changes.

    Noisy Environment: Image noise is usually modeled as an

    additive white Gaussian with zero mean and variance of .

    Moreover, the additive noise is assumed to be independent of

    brightness. Hence, the noise in the partial derivatives ofand which are obtained by applying the Sobel operator

    would follow a Gaussian distribution of . The par-

    tial derivative images can be represented as

    (11)

    where is the partial derivative of the original noise-free

    image and the additivewhite Gaussian noise. Substituting

    (1), (2), and (11) to (4) with statistic analysis, one can get

    SNR

    SNR SNR(12)

    Fig. 2. versus with respect to different SNR values of 1, 1.5, 2, 2.5, and3 from top to bottom.

    where is the measure obtained with noise-free data, SNR

    is the signal-noise-ratio which is defined as ( ) with

    (13)

    where is the average of texture signal power. The relations of

    and withrespect todifferentSNRvaluesare plotted

    in Fig. 2. The ideal relation is a straight line from (0, 0) to (1,

    1). It can be observed that as long as there are significant tex-

    ture features in the region , one can obtain .

    Furthermore, if there is significant structure difference between

    frames in the region , one has

    even when SNR is small. These features indicate that the gra-

    dient difference measure is robust to noise for texture regions.

    On the other hand, if the neighborhoods of a point in both

    frames are homogeneous with low SNR, becomes invalid.This can also be concluded from (4) since the denominator of

    the second term would be small. In this situation, one can notclassify a point with certainty based only on the local gradient

    difference. To tackle this, a validity weight, , for gradient

    difference at each point is computed. Let

    (14)

    then

    if ,

    otherwise.(15)

    where is a parameter based on image noise distribution that

    will be determined later. Consequently, one can define

    (16)

    as the texture difference between two frames.

    III. INTEGRATION OFTEXTURE ANDINTENSITYDIFFERENCES

    The texture and intensity differences can complement each

    other. They are two different views to the difference between

    two frames. A better change detection can therefore be achieved

    by integrating the information from these two sources.

    http://-/?-http://-/?-http://-/?-http://-/?-http://-/?-http://-/?-http://-/?-http://-/?-http://-/?-http://-/?-http://-/?-http://-/?-
  • 8/12/2019 Important Application of Composite

    4/8

    http://-/?-http://-/?-http://-/?-http://-/?-
  • 8/12/2019 Important Application of Composite

    5/8

    http://-/?-http://-/?-http://-/?-http://-/?-http://-/?-
  • 8/12/2019 Important Application of Composite

    6/8

  • 8/12/2019 Important Application of Composite

    7/8

    LI AND LEUNG: INTEGRATING INTENSITY AND TEXTURE DIFFERENCES 111

    Fig. 6. Example three.

    TABLE IQUANTITATIVEEVALUATION OFEXAMPLEONE

    TABLE IIQUANTITATIVEEVALUATION OFEXAMPLETWO

    TABLE IIIQUANTITATIVEEVALUATION OFEXAMPLETHREE

    front of a dark region of the background. Much of it did not

    show up in the result of AID. In the result of SM, the interior

    parts of the human bodies were not detected. But in the results of

    WI and MEI, both the regions of human heads and bodies were

    segmented quite well, and the shadows on the frame of lift door

    (about 55 gray level decreases) were reduced and separated.

    The behaviors of the proposed techniques with respect to the

    parameters were also investigated. Different values of and

    were also tested. had been selected as with

    being 2, 3, 4, and 5, and as with being 4, 5, and

    6. In general, the performance of the proposed techniques are

    not too sensitive to and . Obviously, the lesser the

    and are, the more the points are marked. Different normal-

    ization functions used in (15) and (17), such as the step func-

    tion and the sine function, were also experimented with. The

    performances of the sine and slope functions were similar but

    better than other functions such as the step function. However,

    the slope function is simpler in calculation.

    V. CONCLUSION

    In this work, the integration of intensity and texture differ-

    ences for robust change detection has been investigated. A newaccurate texture difference measure based upon the relation of

    two gradient vectors was derived. The mathematical analysis

    shows that the measure is robust with respect to noise and illu-

    mination changes based on a more general illumination model.

    Two methods to integrate the differences of intensity and tex-

    ture, i.e., the adaptive weighting combination and the optimal

    integration by minimizing an error energy function, have been

    proposed. An analysis for automatic selection of parameters was

    also presented. The computational complexity of the proposed

    techniques was analyzed and real-time implementation was de-

    veloped. Both the visual and quantitative evaluations of exper-

    iment results showed that impressive improvement of change

    detection has been achieved by integrating intensity and texturedifferences. Hence, the proposed technique is more accurate and

    robust for change detection in complex environments.

    To get a meaningful segmentation, the future task would be

    toward recognizing relevant and irrelevant changes in the scene.

    This can be achieved by exploiting not only the low-level differ-

    ence measure but also the high-level knowledge about 3-D ob-

    jects, 3-D perspective projections, and usual events in a scene.

    ACKNOWLEDGMENT

    The authors would like to thank CGIT from NTU for pro-

    viding the experiment facilities.

  • 8/12/2019 Important Application of Composite

    8/8

    112 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 11, NO. 2, FEBRUARY 2002

    REFERENCES

    [1] L. H. Chen and S. Chang, A video tracking system with adaptive pre-dictors, Pattern Recognit., vol. 25, pp. 11711180, 1992.

    [2] I. Haritaoglu, D. Harwood, and L. Davis, W : who? when? where?what? A real time system for detecting and tracking people, in Proc.

    Int. Conf. Face Gesture Recognition, 1998, pp. 222227.[3] E. Stringa and C. S. Regazzoni, Real-time video-shot detection for

    scene surveillance applications,IEEE Trans. Image Processing, vol. 9,pp. 6979, Jan. 2000.

    [4] C. Wren, A. Azarbaygaui, T. Darrell, and A. Pentland, Pfinder:Real-time tracking of the human body, IEEE Trans. Pattern Anal.

    Machine Intell., vol. 19, pp. 780785, July 1997.[5] D. Koller, J. Weber, T. Huang, J. Malik, G. Ogasawara, B. Rao, and S.

    Russell, Toward robust automatic traffic scene analysis in real-time,inProc. Int. Conf. Pattern Recognition, vol. 1, 1994, pp. 126131.

    [6] R. Jain, R. Kasturi, and B. G. Schunck, Machine Vision. New York:McGraw-Hill, 1995.

    [7] P. Rosin, Thresholding for change detection, inProc. Int. Conf. Com-puter Vision, 1998, pp. 274279.

    [8] W. Grimson, C. Stauffer, R. Romano, and L. Lee, Using adaptivetracking to classify and monitor activities in a site, in Proc. ComputerVision Pattern Recognition, 1998, pp. 2231.

    [9] H.Nagel, Formation ofan objectconcept byanalysisof systematictimevariations in the optically perceptible environment, Comput. Graph.

    Image Process., vol. 7, pp. 149194, 1978.[10] Y. Hsu, H. Nagel, and G. Rekers, New likelihood test methods for

    change detection in image sequences, Comput. Vis., Graph., ImageProcess., vol. 26, pp. 73106, 1984.

    [11] K. Skifstad and R. Jain, Illumination independent change detectionfrom real world image sequences, Comput. Vis., Graph., ImageProcess., vol. 46, pp. 387399, 1989.

    [12] S.C. Liu, C. W. Fu,and S. Chang,Statisticalchange detection with mo-mentsunder time-varying illumination,IEEE Trans. Image Processing,vol. 7, pp. 12581268, Sept. 1998.

    [13] B. T. Phong, Illumination for computer generated pictures,Commun.ACM, vol. 18, pp. 311317, 1975.

    [14] IEEE Standard Glossary of Image Processing and Pattern RecognitionTerminology, IEEE Std. 610.4-1990, 1990.

    [15] T. Reed and J. du Buf, A review of recent texture segmentation andfeature extraction techniques,CVGIP: Image Understand., vol. 57, pp.359372, 1993.

    [16] S. Negahdaripour, Revised definition of optical flow: Integration of ra-diometric and geometric cues for dynamic scene analysis, IEEE Trans.

    Pattern Anal. Machine Intell., vol. 20, pp. 961979, 1998.

    [17] J. Vince, Virtual Reality Systems. Reading, MA: Addison-Wesley,1995.

    [18] J. Clark and A. Yuille,Data Fusion for Sensory Information ProcessingSystem. Norwell, MA: Kluwer, 1990.

    [19] D. Murray and B. Buxton, Scene segmentation from visual motionusing global optimization, IEEE Trans. Pattern Anal. Machine Intell. ,vol. PAMI-9, pp. 220228, 1987.

    [20] A. Tekalp, Digital Video Processing. Englewood Cliffs, NJ: Prentice-Hall, 1995.

    Liyuan Li(S99) received the B.Sc. degree in 1985and the M.Sc. degree in 1988 in automatic controlfrom Southeast University of China. Since 1999, hehas been pursuing the Ph.D. degree in the School ofComputer Engineering of Nanyang TechnologicalUniversity, Singapore.

    From 1988 to 1999, he was on the faculty ofSoutheast University, where he was an AssistantLecturer (19881990), Lecturer (19901994), and anAssociate Professor (19951999). Since Septemberof 2001, he has been a Research Staff Member

    with the Kent Ridge Digital Labs, Singapore. His research interests includecomputer vision, image processing, pattern recognition, machine learning, andartificial intelligence.

    Mr. Li is a member of the IEEE Signal Processing Society and IEEE Com-puter Society.

    Maylor K. H. Leung (M92) received the B.Sc.degree in physics from the National Taiwan Univer-sity, Taipei, Taiwan, R.O.C., in 1979, and the B.Sc.,M.Sc., and Ph.D. degrees in computer science fromthe University of Saskatchewan, Saskatoon, SK,Canada, in 1983, 1985, and 1992, respectively.

    Currently he is an Associate Professor withNanyang Technological University, Singapore. Hisresearch interest is in the area of computer vision,pattern recognition, and image processing. Hisparticular interest is on improving security using

    visual information.