CA2200732A1 - Method and system for estimating motion within a video sequence - Google Patents

Method and system for estimating motion within a video sequence

Info

Publication number
CA2200732A1
CA2200732A1 CA002200732A CA2200732A CA2200732A1 CA 2200732 A1 CA2200732 A1 CA 2200732A1 CA 002200732 A CA002200732 A CA 002200732A CA 2200732 A CA2200732 A CA 2200732A CA 2200732 A1 CA2200732 A1 CA 2200732A1
Authority
CA
Canada
Prior art keywords
estimate
dvf
motion
moving object
object boundary
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
CA002200732A
Other languages
French (fr)
Inventor
Taner Ozcelik
James Charles Brailean
Aggelos K. Katsaggelos
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Motorola Solutions Inc
Northwestern University
Original Assignee
Individual
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Individual filed Critical Individual
Publication of CA2200732A1 publication Critical patent/CA2200732A1/en
Abandoned legal-status Critical Current

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • H04N19/51Motion estimation or motion compensation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • H04N19/51Motion estimation or motion compensation
    • H04N19/537Motion estimation other than block-based
    • H04N19/543Motion estimation other than block-based using regions
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Image Analysis (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)
  • Color Television Systems (AREA)

Abstract

The present invention provides a method and system for estimating the motion within a video sequence. The invention provides very accurate estimates of both the displacement vector field, as well as, the boundaries of moving objects. The system comprises a preprocessor (102), a spatially adaptive pixel motion estimator (104), a motion boundary estimator (106), and a motion analyzer (108). The preprocessor (102) provides a first estimate of the displacement vector field, and the spatially adaptive pixel motion estimator (104) provides a first estimate of object boundaries. The motion boundary estimator (106) and the motion analyzer (108) improve the accuracy of the first estimates.

Description

~ 7 ~ ~
wo 97/04600 pcTtuss6/lol4l METHOD AND SYSTEM FOR ESTIMATING ~IOTION Wl l ~ll~ A VIDEO
SEQUENCE

Field of the Invention The present invention relates generally to video coding, and more particularly to using motion estimation in video coding.

Background of the Invention A video sequence consists of temporally sampled projections of a three f1imen~ional, 3-D, scene onto the two ~iim~n~ional, 2-D, image plane. The 3-D
motion that occurs within this scene is captured as 2-D displacements of these 15 projections. The displacement of a particular picture element, pixel, in the current 2-D image plane, may be represented by a vector which points to the location of the particular pixel in a previous image plane. The displacement vector field, DVF, describes the motion of all pixels between a given set of image planes, and therefore represents the 3-D motion of objects projected onto the image plane.
Accurate estim~tion of the DVF within a video sequence is crucial in many applications of image sequence processing. Video coding. frame interpolation. object tracking, and spatio-temporal motion compensated filtering are all applications that require accurate estimation of the DVF to utilize the interframe correlation that exists 25 in a video sequence.

Compression of digital video to a very low bit rate. VLBR, is a very important problem in the field of communications. In general, a VLBR is considered not to exceed 64 kilo-bits per second (Kbps) and is associated with existing personal 30 commlmic~tion systems, such as the public switch telephone network and cellular systems. To provide services like video on demand and video conferencing on these systems~ would require the information contained in a digital video sequence to be ~ ~ ~ Q ~
WO s7/o46oa~ pcT/uss compressed by a factor of 300 to 1. To achieve such large compression ratios, requires that all redundancy present in a video sequence be removed.

Current standards, such as H.261, MPEGl, and MPEG2 provide compression 5 of a digital video sequence by utilizing a block motion-compensated Discrete Cosine Transform, DCT, approach. This video encoding technique removes the redundancy present in a video sequence by utili7in~ a two-step process. In the first step, a block-m~t~hing, BM, motion estim~tion and compensation algorithm estimates the motion that occurs between two temporally adjacent frames. The frames are then 10 compensated for the estimated motion and compared to form a difference image. By taking the difference between the two temporally adjacent frames, all existing temporally redundancy is removed. The only information that remains is new information that could not be comp~n~t~d for in the motion estimation and comp~ns~tiQn algorithm.
In the second step, this new information is transformed into the frequency domain using the DCT. The DCT has the property of compacting the energy of this new information into a few low frequency components. Further compression of the video sequence is obtained by limiting the amount of high frequency information 20 encoded.

The majority of the compression provided by this approach to video encoding is obtained by the motion estimation and compensation algorithm. That is, it is much more efficient to transmit information regarding the motion that exists in a video 25 sequence. as opposed to information about the intensity and color. The motioninformation is represented using vectors which point from a particular location in the current intensity frame to where that same location originated in the previous intensity frame. For BM, the locations are predetermined non-overlapping blocks of equal size. All pixels contained in these blocks are assumed to have the same motion 30 The motion vector associated with a particular block in the present frame of a video sequence is found by searching over a predetermined search area, in the previoustemporally adjacent frame for a best match. This best match is generally determined -wo 97/04600 PCT/US96/10141 using the mean-squared-error (MSE) or mean-absolute-difference (MAD) between ~ the two blocks. The motion vector points from the center of the block in the current frame, to ~e center of the block which provides the best match in the previous frame.

Utilizing the çstim~ted motion vectors, a copy of the previous frame is altered by each vector to produce a prediction of the current frame. This operation is referred to as motion compensation. As described above. the predicted frame is subtractedfrom the current frame to produce a difference frame which is transformed into the spatial frequency domain by the DCT. These spatial frequency coefficients are quantized and entropy encoded providing further compression of the original video sequence. Both the motion vectors and the DCT coefficients are tr~n~mitt~d to the decoder, where the inverse operations are performed to produce the decode video sequence.

The estim~tion of the DVF within a video sequence is an extremely difficult problem. The two main sources of the difficulties in obtaining accurate estimates are the nonst~tion~rity of the DVF and the ill-posed nature of the problem. The ill-posedness results from the violation of ~a~m~rd's definition of a well-posed problem. which is characterized by the existencc of a solution, the uniqueness of a solution, and the continuity of a solution. The problem of estim~ting the displacement field violates all three of these properties. Objects moving in an image sequence will occlude certain areas and uncover others: at these locations the DVF is undefined and no solution exists. Object deformation and changes in the camera'sdepth of field can also result in regions where the DVF i~ undefined. For a given image sequence, many displacement fields may satisfy thc data and thus the solutions are not unique. The continuity property is also violated. since in some image sequences even slight modifications in the local intensity values can cause large changes in the m~gnit~lde and/or direction of the displacement estimate. The fact that the DVF estimation problem is ill-posed. must be taken into account if any useful results are to be obtained.

wo 97/04C00 2 2 ~ ~ 7 ~ ~ PCT/USg6/10141 The non.st~tinn~rity of the DVF results from objects moving along dirrelent tr~ject-~ri~s within the scene that causes disc~ntinllitiPs to occur at the object boundaries, as well as from regions where the motion is undefined due to covered or uncovered parts of the moving scene. The nonstationarity of the DVF implies that5 any approach used in its estim~tion~ must be adaptable. That is, approaches which assume stationarity will result in estimates of the DVF that distort the boundaries between dirren~ ly moving objects. These distortions in the motion boundaries result directly in distorted objects, and therefore an increase in the entropy of the difference image.
The assumption used by BM that the motion within a blocl~ is constant, effectively constrains the problem of estim~ting the DVF so that the problem is no longer ill-posed. However, this same assumption also requires that if the motionboundaries are to be preserved they must coincide with the block boundaries. Since 15 in real sequences this rarely occurs, significant errors in the estimation of the DVF
result at the boundaries. Such errors can be described as a blurring of the boundaries of the DVF. For example, if a block contains a moving object and stationary background, then depending on the size of the object either part of the stationary background will be displaced or the moving object will be treated as stationary 20 background. In either case, the motion compensated block of the previous frame will be a poor prediction of the block in the present frame, resulting in a DFD image with increased entropy.

Another approach for estim~ting the DVF which has recently gained in 25 popularity is the spatio-temporal gradient approach. ~patio-temporal gradient-based motion estim~tion algorithms are derived by minimi7ing the displaced frame difference, DFD, at each pixel using a temporal gradient and spatial gradient in the previous frame, based on an initial estimate of the DVF. One common approach to carry out this minimi7~tion is that of first linearizing the DFD using a prediction of 30 the DVF. By ~snming that all the pixels in a neighborhood of the working point undergo the same motion, a set of linear equations results, instead of a single equation. If the neighborhood allows recursive computability, then algorithms 2 ~ ~ ~ 7 3 ~ PCT/US96/10141 resulting from this approach are generally referred to as pixel recursive, PR, motion estim~tors. PR algorithms are generally faster than BM algorithms, and do not require spatial interpolation for subpixel accuracy, although PR algorithms may require spatial interpolation to compute the spatial gradients in the previous frame 5 along the direction of motion. Furthermore, PR algorithms can be extended to handle more complicated types of motion than the purely translational type.

Similar to block m~t~hin~, the performance of PR algorithms also suffers at the boundaries separating ~ nlly moving objects. That is, at these motion 10 boundaries, pixels in the local neighborhood do not undergo similar motion. By ~sllming that they do, PR algorithms produce inaccurate estim~t~s near these boundaries. As in the case of block matching, the poor estimates of the DVF nearthese bolln~l~ri~s will result in a marked increase in the entropy of the DFD.

There are several major problems with both block m~t~hing and spatio-temporal gradient approaches. For instance, the assumption that the motion within a block or local neighborhood is homogeneous results in severe degradation of the boundaries within the DVF. Block matching algorithms can not resolve within a single block complicated types of motion, such as rotations. Also, they require 20 spatial interpolation for sub-pixel accuracy. Spatio-temporal gradient approaches suffer from the fact that the linearization of the DFD is extremely dependent on the initial prediction of the DVF. These difficulties result in an increase in the DFD's entropy, which is prohibitive for very low bitrate coding applications. A very low bitrate is defined as a rate less than 64 kbits/sec.
Therefore, a need exists for a method and system for accurately estim~tin~
motion within a video sequence. The method and system are required to regularizethe estimation of the displacement vector field, DVF, to combat ill-posedness and estimate the boundaries or discontinuities that exist within the DVF.

W097/04600 22~73 ~ PCT/US96/1014!

Brief Description of the Drawings FIG. 1 is a ~ gr~m of a preferred embodiment of a system for estim~ting 5 motion in accordance with the present invention.

FIG. 2 is a diagram of a pl~rell~d embodiment of a preprocessor in accordance with the present invention.

10FIG. 3 is a diagram of a plerell~d embodiment of a spatially adaptive motion estim~tor in accordance with the present invention.

FIG. 4 is a diagram of a preferred embodiment of a first DVF update circuit in accordance with the present invention.
FIG. S is a diagram of a preferred embodiment of a motion boundary estim:~tor in accordance with the present invention.

FIG. 6 is a diagram of a preferred embodiment of a motion analyzer in 20 accordance with the present invention.

FIG. 7 is a flow diagram of steps for implementing a plefeli~d embodiment of a method for estim~ting motion in accordance with the present invention.

25FIG. 8 is a flow diagram of steps for implementing a LJl~rell~d embodiment of a method for se~m~nting in accordance with the present invention.

FIG. 9 is a flow diagram of steps for implementing a ~ fel,~d embodiment of a method for adaptively estim:~ting motion in accordance with the present invention.

wo 97/04600 ~ 7 3 ~ PCTIUS96/10141 FIG. 10 is a flow diagram of steps for implementing a ~ d embodiment of a method for estim~ting motion boundaries in accordance with the present invention.

FIG. 11 is a flow diagram of steps for implementing a preferred embodiment of a method for analyzing motion in accordance with the present invention.

Detailed Description of the Preferred Embodiments The present invention provides a method for obtaining very accurate estimates of both the displacement, as well as, the boundaries ot the moving objects within a video sequence. Therefore, the problems of ill-posedness and nonstationarity of the boundaries are solved simult~neously. An estimate of the object boundaries within a video sequence is obtained through the segmentation of the present image frame into regions of similar intensity. The boundaries separatin~ these regions are considered to be the boundaries of the objects within the sequencc. This object boundary information is used to determine a first estimate of thc DVF.

Utilizing the first estimate of the DVF. a first cstimate of the object boundaries is obtained with the accuracy of the location of each boundary improved, as well as, the removal of any boundary found in thc cstimate to be non-moving.
Based on this estimate of the moving object boundaries. a second and improved estimate of the DVF is also determined . As a final stcp the second estimate of the DVF and the estimated moving object boundaries arc turther refined based on a predetermined model of the DVF. Such models typically tal;e into account complicated types of motion, such as object rotations in the field of view and apparent object motion due to changes in the camcra focal length or position. The third estimate of the DVF and the second estimate of the moving object boundaries are the result of fitting such a model to the second estimate of the DVF and themoving object boundary estimate.

W097/04600 ~ 2 ~ ~ 7 ~ ~ PCT/US96/10141 FIG. 1, numeral 100, is a diagram of a p,efel,ed embodiment of a system for estim:~ting motion in accordance with the present invention. The system includes a preprocessor (102), a spatially adaptive motion estimator (104), a motion boundary estimator (106), and a motion analyzer (108). The preprocessing section (102) S provides a object boundary çstim~tç (112) of a current intensity frame (110), f,,, at time instant k . Based on this object boundary çstim~tç (112), the current intensity frame (110), fk . and a previous intensity frame (114), fk-l . the spatially adaptive motion estim~tor (104) provides a first DVF estimate (116), dk . The DVF
characterizes the motion within the scene that occurs during the time separating10 frames fk and fk-l . Both the first DVF estimate (116) and the object boundary estimate (112) are further refined in the motion boundary estimator (106) and motion analyzer (108). Specifically, using the first DVF estimate (116), dk . the motion boundary estimator (106) refines the object boundary estimate (112), elimin~ting the intensity boundaries which do not correspond to motion boundaries. The output of15 the motion boundary estim~tor (106) is the first moving object boundary estimate (1 18), lk . and represents only the boundaries of objects that are moving within the video sequence. Also, inaccuracies in the first DVF estimate (116) are removed in the motion boundary estim~tor (106) resulting in the second DVF estimate (120),dk .
These inaccuracies are generally due to the corruption of the input video sequence 20 with noise. The motion analyzer (108) further refines both estimates, dk and Ik, using modeling approaches based on the principles of object motion. The third DVF
estimate (122), dk . and the second moving object boundary estimate (124), Ik, represent extremely accurate estimates of the DVF and the moving object boundaries, respec~ively.
FIG. 2, numeral 200, s a diagram of a preferred embodiment of a preprocessor in accordance with the present invention. The system includes an order statistics filter (202) and an object boundary estimator which is comprised of a object grower (206), and a object merger (206). The major function of the preprocessing30 step is to provide the spatially adaptive motion estimator (104) with an accurate WO 97/04600 ~ 7 3 ~ PCT/US96/10141 segm~nt~tion of the current inten~ity frame (210) into regions or objects. To achieve this goal, a copy of the current intensity framefk (210) is ~lrst ~lltered separately using an order statistics filter (202). This operation is performed to remove any small objects from the sequence prior to segmçnt~tion. These smaller objects, which are in S general the result of illl~min~tion changes in the video se~uence and not a true object.
can degrade the spatially adaptive motion estimator's (104) ability to accurately ~stim~te the motion in that object.

A filter which is typically used to remove smaller objects from an image 10 frame is a 7x7 median filter. This filter is described by the following operation y~j(i,j) = median~fk(i--3,j--3),fk(i--3,j--2),.. ,fk(i+3,j+2),fk(j+3,j+3)}

where yk(i, j) is the filtered version of fk(i, j). The filtered output (212). yk(i, j), is the median value of f,,.(i, j) (210) and its 49 nearest spatial neighbors. Any objects with a spatial support, or size, that is less than 7x7 are removed by this filter. As discussed in the paragraph above, these smaller objects must be removed prior to the 20 segmentation of the image frame to avoid degradation to the accuracy of the resulting segmentation.

The object boundaries contained in the filtered output (212) are estimated using a two-step process (204). The first step (206) is called object growing, wherein 25 each pixel is compared to its neighbors. Based on this comparison, the pixel is classified as either belonging to the same object as its neighbors, or to a new object.
The test used to deter~nine the classification of pixel (i,j) is given by ¦Yk (i- j) Yk (i--m, j--n)¦ S T ( 1 ) where m and n can take on the values of ~-1,0,1~. The threshold T is used to determine whether Yk (i, j) iS a member of the same object as Yk (i - m, j - n) . If the WO 97/04600 ~ 2 ~ ~ 7 3 PCT/uss6tlol41 absolute value of the difference between two neighboring pixels, ¦Yk (i, j) - Yk(i - m, j - n)¦, is less than the threshold T, then pixel (i,j) is cl~csi~l~d as a membler of the same object as pixel (i-m,j-n). This object is denoted as Obj(i - m, i - n). If the difference is greater than T, then pixel (i,j) is considered not 5 to be a member of Obj(i - m,i- n). For the case when the neighbors of (i,j) have not all been cl~sifi~cl or yk(i, j) is determined to not be a member of the neighboring objects, than pixel (i,j) is cl~ified as the initial pixel of a new object.

The threshold T is typically fixed to a predetermined value. A drawback of 10 this approach is that this value is dependent upon the particular image to besegmented. An approach which removes this dependency is described by the following expression Yk (i, j) S MIN + Tl yk(i, j) ~ Obj(i - m,i - n) ~ ~ and ' (2) ~yk(i, j) 2 MAX--T, where MIN and MAX are the maximum and minimum intensity values contained within Obj(i - m,i - n) and Tl is again a predetermined parameter. After each pixel iS Cl~s~cifi~l as belonging to a specific object, the maximum and minimnm values of that object are tested to determine if an adjustment is needed. The tests and 20 adjustrnents are the following if (yk (i, j) < MIN), ~ MIN = Yk (i~
and if (yk(i" j) 2 MAX'),~ MAX = yk (i, j) .

This operation has the effect of providing an adjustable window for each object ~hat can adapt to any particular intensity frame. Typically, the predetermined value for the threshold T~ is 25. This method for adjusting the window, i.e.
thresholds, is extremely effective at proving a consistent segmentation over a wide 30 variety of video sequences.

~ 3 ~

The second step of the object boundary estimation process, each object is tested to determine whether it should be merged with a neighboring object or not.
The pixels residing along the border of an object are compared with the pixels on the 5 borders of the neighboring objects. If the difference between these border pixels is small, then the objects are merged. Specifically. thc following test is conducted on the border pixels of two neighboring objects if (¦yk(i, j)--yk(i--m, j--n)¦ S T2) ~ mer~e(O~7j(i~ j)& Obj(i--n, j--m)) 10 (4) where T2 is a predetermined parameter typically set to a value of 30.

A second comparison is also performed in an attempt to again guard against 15 the creation of small objects. These smaller objects ~rc generally the result of gradually varying illllmin~tion changes which the objcct growing algorithm interprets as object boundaries. This test compares the size of an object, i.e. the number of pixels contained in an object, to a third predetermincd threshold. If the number of pixels is less than this predetermined threshold. than the object is merged with a 20 neighboring object. This merging operation is performed using the neighboringobject which is the most similar to the current object. Thc amount of similaritybetween two neighboring objects is measured using thc border differences described in Equation (4). Typically. the size threshold is set at 256 pixels. The object merging operation (208) is effective at removing smaller objccts which may have been formed 25 during object growing operation.

An advantage of this two-step object boundary estimation approach (204), as opposed to other segm~nt~tion approaches. is that it is guaranteed to provide boundaries that are continuous and closed. For the application of motion estimation 30 and analysis this is a very important result, since we know that in general the objects within a video sequence are defined by a continuous and closed boundary.
Furthermore, the boundaries contained in an intensity frame are a super-set of the Wo 97/04600 Pcr/uSs6/10141 ~ Q7 ~ ~

corresponding DVF boundaries. Therefore, the boundaries or discontinnities present in the DVF are also continuous, closed, and well defined.

The output of the preprocessing unit (102) is the object boundary estimate (216) Sk . The object boundary estim~fe (216) S,~ assigns each pixel in yk(i, j) (212) a value corresponding to the object it is a member of. The boundary is retrieved from (216) Sk by determining where the object numbering changes. The object boundary S,~ (216) combined with the present intensity frame fk (110) and the previous intensity frame fk-l (114), are used by the spatially adaptive motion estimation unit (104) to determine the first estimate of the DVF dk (116). As ~ cu~se~l above, the object boundary estimate Sk (112) contains the discontinuities of the DVF dk . and therefore, is considered as the first estimate of the boundaries between differently moving objects. The first or initial estim~t~ of the line process, Sk, is used as the mechanism by which the spatially adaptive motion estimation algorithm adapts to these discontinuities.

FIG. 3, numeral 300, is a diagram of a preferred embodiment of a spatially adaptive motion estimator in accordance with the present invention. The spatially adaptive motion estimation unit (104) comprises a causal look-up table (312), a DVF
predicl:or device (318), and an update circuit (320). The inputs to the spatially adaptive motion estimation unit are the estimate object boundaries Sk (304), a predetermined causal local neighborhood of past displacement estimates d~,. (306), the previous intensity frame f"_, (308), and a current intensity frame f~ (310). The causal look-up table device (312) stores a set of predetermined autoregressive, AR, prediction coefficients (314), and corresponding prediction uncertainties (316).
The predetermined set of causal AR prediction coefficients (314) and corresponding uncertainty terms (316) are typically found empirically. Generally, a least squares estimation approach is used on either a prototype or previously 30 estimated DVF. Based on the object boundary estimation Sk (304) a subset of the predetermined AR prediction coefficients (314), a(m, nl Sk ), and a corresponding uncertainty term, w(i, jl Sk~ (316), are chosen from the look-up table (312) for use in WO 97/04600 ~ ) 7 3 2 PCT/US96/10141 the DVF predictor (318) and update circuit (320). The DVF predictor circuit ~ determines a prediction of the DVF based on the subset of AR prediction coefficients (314) and a local neighborhood of displacement vectors (306). The prediction operation is described by the following equation dA(i,j)= ~,a(m,nlSI;)d~ m,j-n). (5) I,j~R

where dk(i, j) is the prediction of the motion occurring at pixel location r -- (i, j) in the current image frame, and a(m,nlSk) are the AR prediction coefficients with a10 local support R.

The local neighborhood R consists of the following pixel locations: the pixel in the column directly to the left (i,j-1), the pixel in the row above and the column to the left (i-l,j-l), the pixel in the row above (i-l,j). and the pixel in the row above and 15 the column to the right (i-l,j+l). It should be noted that the choice of R is made at the time of implementation and is dependent on the method used to navigate through the two ~im~ncional data sets used to represent the image and displacement information at a particular time instant. For this particular R, it is assunned that the data is access from left to right across each row, starting with the top row. Other 20 methods for navigating through the image and displacement data can also be used.
This would require a sli~ht modification to the local neighborhood R. however the operations would remain the same.

As described by Equation (5) and discussed above, if the object boundary 25 estimate (304) S~ indicates that an object boundary is present in a predetermined local neighborhood R, the AR coefficients a(m,nlSk) (314) are adapted so as not to include in the prediction any displacement vectors belonging to another object. The - non-stationary assumption on which the above equation is based, is valid throughout the DVF and results in consistently accurate predictions. The inaccuracies incurred 30 by a stationary model due to the mixing or blurring of displacement vectors located near the object boundaries, is alleviated.

Wo 97/04600 ~ PCT/USg6/1014 The predicted displacement vector dk(i,j) (324), the ~.~soci~f~ uncertainty term W(i,flSk) (316), the previous intensity frame fk-l (308), and current intensity frame fk (310) are inputs to the first DVF update circuit (320). The first DVF update 5 circuit updates the predicted displacement vector dk (i, j) (324) resulting in the first DVF estim~t~ (322).

FIG. 4, numeral 400, is a diagram of a preferred embodiment of a first DVF
update determiner in accordance with the present invention. The first update circuit comprises a motion compen~tion unit (402) and a gain calculator unit (404). The motion compensation operation performed by the motion compensation unit (402) isa nonlinear operation which is described by the following equation f k(r) = fk-l (r--dk(r)), (2) where fk~r) (422) is the motion compensated previous frame. The value fk_l (r - dk(r)) (422) is found by taking the intensity value of the pixel located spatially at r - dk(r) in the previous frame fk-l (408). The motion compensated value f k (r) (422) is subtracted from the value of the pixel at location (r) in the 20 current frame fk(r)(412) resulting in the displaced frame difference? DFD, signal (424), ek (i? j) . The DFD (424) signal is an error signal which is related to the accuracy of the predicted displacement vector d~ (i. j) . The DFD signal e . (i? j) (424) is multiplied by a gain term (426) and added to the DVF prediction (406) resultimg in the first estimate DVF estimate (414) .
The gain determined by the gain calculator unit (404) is used to scale the DFD
signal ek(r) (424) prior to updating the DVF prediction (406) dk(r). The expression used to calculate the gain is found by linearizing the motion compensated frame about l:he spatial location (r - dk(r) - u(r))~ where u(r) iS the error in the prediction WO 97/04600 ~ ) 7 3 ~ PCT/US96/10141 dk(r), (406) and solving for the best linear estimate of u(r). The resulting expression for ~e gain, K(i, j), (426) is given by K(i, j~ = [G J(i, j)G(i, j) + W(i, j)] G(i, j) where G(i, j) is a nx2 matrix of spatial gradients evaluated in the motion compensated previous frame (422) for each pixel in thc local neighborhood R .
Specifically, for the local neighborhood R described above, G(i, j) is given by V f ,~; (i, j) VJf,;(i,j-l) G(i, j) = VT fk(i--1, j--1) V f,~,(i--1.j) V f, (i--I~j+l) where df V f 1: (i, i) = df dy is the two ~iimen~ional spatial gradient evaluated at pixcl location (i,j) in the motion compensated previous frame (422) f k_l (r - d k (r ) ) .

As mentioned above, the gain K(i, j) (426) determined by the gain calculator 20 is used to the scale the DFD signal (424). This scaled DFD signal is added to the predicted displacement vector (406), d~ (i, j), resulting in the first estimate of the DVF (414),d~- (i, j) . The operation used to update d~ (i, j) is described in more detail in the following equation wo 97/04600 2 2 ~ ~ 7 ~ ~ PCT/US96/10141 dk(i, j)=dk(i, j)+K(i, j)~E(i, j), where E(i, j) is the DFD evaluated for each pixel in the local neighborhood R .

FIG.5, numeral 500, is a diagram of a preferred embodiment of a motion boundary estimator in accordance with the present invention. The motion boundaryestim~tion unit (106) comprises a noncausal look-up table (510), a moving objectboundary estimator (506), a DVF estim~tor (508), a noncausal local neighborhood of the DVF (502), and a moving object boundary estimator (506). The inputs to the moving boundary estimator are the object boundary estimate Sk (526), the first DVF
çstim:~tç dk (524), the previous intensity frame fk-l (528), and a current intensity framefk (530). The moving object boundary unit (504) stores the current estimate of the moving object boundary for possible use in an iterative solution. Specifically, the solution provided by the moving object boundary estimator (506) can be used as an initial condition for an improved solution. A method for carrying out this iteration is discussed below. The moving object boundary estimator (506) is initi~li7ed with the object boundary estimate (526).

The moving object boundary estimator (506) provides an estimate of the moving object boundaries Ik (546) based on the object boundary Sk (526) and the corresponding first estimate of the DVF (524). The moving object boundary estimalor (506) is comprised of a moving edge determiner unit (514), edge continuity unit (516), and a moving object update unit (518). The estimate of the moving object boundaries, lk (546), is determined by adjusting the first estimate S~ . These adjustment~, which are based on known statistical properties of these boundaries and characterized through the use of confidence measures, result in the removal of boundaries that do not belong to a moving object. The particular confidence measures used are described in detail below. As mentioned above, the moving object boundary estimator (506) is initialized with the object boundary estim~te (526) .

~ ~2~ ~7~ ~

The moving edge determiner unit (514) evaluates whether or not pixel (i,j) is ~ part of a moving object boundary and assigns a confidence measure to tlle current estim~te Ik(i, j). This evaluation is performed by comparing ~ cent displacementvectors. More specifically, the evaluation and corresponding confic~nce measure 5 (540) are given by D(i~ j) = (dk (i, j)--dk (i, j -1)) + (dk (i, j) - dk (i -1, j)) If confidence measure D(i, j) (540) is large, then a moving object boundary is likely to exist. Conversely. if D(i, j) (540) is small, than the likelihood is small that a 10 moving object boundary exists.

Similar to the moving edge determiner unit (514), the edge continuity unit (516) also assigns a confidence measure (542) to the current estimate of the moving object boundary at pixel (i,j). This confidence measurement (542) however, is a 15 function of the values of the neighboring boundary elements. Based on the assumption that objects within a video scene have boundaries that lie on a closed continuous curve. a pixel (i,j) is a boundary element, if and only if, it lies on such a curve. In other words, if pixel (i,j) is determined to be a moving boundary element, than the ~dj~çent pixels in a particular direction most also be boundary elements. The 20 above charac~ ~alion is captured in the following expression C(i, j) = 0.5(1(i, j) + I(i - 1, j) + I(i, j - 1)) .

The moving object update determiner adjusts the moving object boundary estimate ll.(i, j), at pixel (i,j), based on the current estimate of the moving object boundary, which is initially the object boundary estimate (526) Sk(i, j), and the confidence measure D(i, j) (5 4)and C(i, j) (542). The expression which characterizes the moving object boundary update determiner is given by W097/04C00 2~ ~ O ~J PCT/US96/10141 _ i 1 I ( i) 1 + exp(L(i, j)) where L(i~ (a[(l - ~ C(i, j))+ (1- Sh (i~ j))]_ i~2D(i, The coefficients a and ~ are predetermined parameters used to control the amountthat each conficlence measure can influence the estimate Ik The parameter ~ (556) is used to control the response of the update deterrniner. If ~ is small, then the effect of the confidence measures on the adjusted 1,~ (i, j) is also small. If on the other hand 10 ~ is large. then the adjusted Ik(i, j) iS domin~ted by the confidence measure.
Typically, the estim~tion of Ik and dk is done iteratively, with ,~ being initi~li7ed to a small value and increased slowly at the beginning of each new iteration. The decision to increase ,~ and to refine the estimates of Ik (i, j) and dk (i, j) is made in the es~imation terminator (516). The estimation termination circuit (514) is discussed 15 in more detail below.

The DVF estimation determiner (508) is comprised of an advanced prediction determiner (520) and a second DVF update determiner (522). Similar to the DVF
predictor (318) described above, the advanced predictor (520) provides a prediction 20 of the DVF based on the moving object boundary estimate (536). Utilizing thismoving object boundary estimate (546) l/, (i, j), a subset of the pre ~eterminednoncausal AR prediction coefficients, b(m,nllk) (548). and a corresponding uncertainty term, v(i, jllk) (552), are chosen from the non-causal look-up table (510) for use in the advanced predictor (520) and second DVF update determiner (522).
25 The advanced predictor (520) determines a prediction of the DVF based on the subset of AR prediction coefficients (548) and a local neighborhood of displacement vectors 9~. It should be noted that the local neighborhood 9~ is different from the R used by the DVF predictor (318). This diffeience comes from the fact that a complete estimate of the DVF already exists. Therefore, the dependence on the scan that exists 30 for the DVF predictor is not an issue for the advanced predictor. In other words, the ~ 7 3 ~
WO 97/04600 PCT/ltJS96/10141 restriction to use a neighborhood that contains only previously estim:~ted ~ displ~çe.ment vectors is no longer n~ce~ry. The local neighborhood 9~ includes the nearest neighbor pixels surrounding (i,j). Speci~lcally, 9~ includes the following pixels, (i,j-1), (i,j+l), (i~ i), (i+l,j). In addition to the model described by Equation 5 (5) the advanced predictor also utilizes a rigid body assumption to further constrain the prediction. The rigid body assumption specifies that all pixels within an object undergo similar motion.

The advantage of combining this additional constraint with the AR model 10 described in Equation. (5), is that a probabilistic characterization of the DVF can be derived in the forrn of an A Posteriori probability density function. The advanced prediction unit det~rmin~s a prediction of the displacement vector dk(i, j) that maximizes the A Posteriori function. Typically, this ma~imi7~tio~ process is pelrolllled using the following iterative algorithm -n dkn+l(i~ j) = dkn(i~ dkn(i~ j)-- [d~ (i, j)]
i~2(1- ~ a(m.tlll~ (i. j))) m~n ~R
where [D(i, j)(l--I(i, j))+D(i+1. j)(l -I(i+1, j))], ~ is the correction step size, i.e., it is equivalent to a steepest descent minimi7~tion step, and n is the iteration number. The initial condition used to begin the iteration is 20 d~~ = d ~ . Since the first DVF estimate d,~. is of hi~h quality. the improvement resulting from iterating is small. Generally. the numbcr of iterations performed is fixed to 1. However, a criterion that monitors thc pcrcent change between iterations can also be used to terminate the iteration. More spccifically, if ~ dkn+~ d~ (i, j)¦¦
~¦¦dn+l (i j~

22~ ~7 ~ ~ ~

the iteration is termin~t~i Typically, the threshold T is set to 10-3 . Again due to the qualily of the first DVF estimate the correction step size is fixed to a value of 0.1.

The prediction dk (i, j) provided by the advanced prediction unit (520) is updated by the second DVF update determiner (522) to produce the second DVF
estimate dk (534). The second DVF update determiner (522) the same update method as described for the first DVF update circuit (320).

As discussed above, the coupled solution for dk (534) and IL (532) is typically solved using an iterative process. That is, steps used to obtain both IL and dL are repeated using the previous results as the initial conditions. Furthermore, ,8 is increased after each iteration providing more weight to the confidence measures. By iterating the overall accuracy of both lk and dk is improved. The decision whether to perform another iteration or not, is determined in the iteration terminator (516). The iteration terminator utilizes a termination criteria similar the one used by the second DVF estimator (510). That is~ the number of iterations set to a f~ed number or it can determined based on the percent change between iterations. Typically, a maximum number of 3 iterations is set. However, if the percent change between iterations is below a threshold the iteration is terminated. The specific, if the following is true ~¦¦dL (i~ dk (i,i)¦¦

¦¦dk (i~

and ~¦¦ILn+I (i j)--ILn (i i)¦¦
~¦¦ln+'(i i)¦¦~ xlOO C T,, ~ 7 ~ 2 Wo 97/04600 PCT/US96/10141 the iteration is termin~t~l FIG. 6, numeral 600, is a diagram of a ~r~rel~d embodiment of a motion analyzer in accordance with the present invention. The motion analyzer which 5 provides a parameterized representation of the DVF using d~ and 1~ as inputs. The motion analyzer comprises a memory device (614), translational motion estimator (604), zoom estimator (606), rotation estim~tor (608), and a object labeler and center determiner (612). The second estimates of the DVF dk and line process 1,~ are the inputs to the motion analyzer unit. Based on these inputs and a four parameter 10 motion model, a third and final set of estim~t~s of the DVF and line process is determined.

The four parameter motion model characterizes the DVF by utili7inp, as rigid bodies, the objects defined by the line process. Specifically. the displacement of each 15 pixel contained within a particular object is characterized by the following the expression d,~. (i, j) = tk (i, jl Obj(i, j)) + Z(Obj(i, j) - cp(i. jl Obj(i, j)) + ~(Obj(i, j)) - cp(i, jl Obj(i, j)) 20 where tk(i, jl Obj(i, j)) = [t,~ ,; (i, jl Obj(i, j)), tV ,, (i, jl Obj(i, j))~is a vector representing the translational motion component of d~ (i, j), while Z(Obj(i, j) and ~(Obj(i, j)) represent the zoom and rotation components. It is important to note that each of these four parameters is dependent upon the object. Obj(i, j), that a particular pixel (i,j) is assigned. Furthermore, the term cp(i, jl Obj(i, j)) represents the distance ~rom the 25 center of Obj(i, j) to the pixel (i,j). The position of each object's center is located by the object center determiner (612). Based on the estimate 7~ . the center of each object is determined to be the intersection of the two lines which contain the maximum and minimum horizontal and vertical pixel locations in Obj(i, j).

The translational motion estim~tor determines the translational motion component for each object. This is accomplished by averaging the horizontal and WO 97/04600 ~a 2 ~) Q PCT/US96/10141 vertical displacement components over each object. Speci~lcally, the tr~nsl~ti~ n~l components tX k and ty k are calculated using the following expressions t" ,~ (i, jl Obj(i, j)) = N ~,d,~ ,.(i, j) i, j~Re~ion(i, j) 5 and t~,k (i, jl Obj(i, i)) = N ~ dy k (i~ j) i, j~gion(i, j) where N is the total number of pixels contained within Obj(i, j).

The zoom estimator (606) estimates the zoom parameter Z(Obj(i, j)). This 10 parameter is used to characterized any changes that may occur in an object's focal length. These changes may be caused by the camera or by the motion of the objectcloser to or away from the camera location. Utilizing the dense DVF, dk (i, j~, the zoom parameter Z(Obj(i, j)) is estim~ted for object Obj(i, j) using the following expression Z(i, jl Obj(i, j)) = N ~ r-' (i, j) - { (i - cx(i, j)) - (d ~ k (i, j)--tl ~. (i, jl Obj(i, j))) +( j - cy(i, j)) ~ (dy ,~. (i, j) - ty k (i, jl Obj(i, j)))~

20 where cx(i. j) and cy(i, j)represent the horizontal and vertical indices of the center for the object which includes pixel (i,j) and r(i, j) = (i - cx(i, j))2 + ( j - Cy(i~ j))2 .

The rotation estimator (608) estimates the rotation parameter ~(Obj(i, j)) for each object. The rotation parameter is used to characterize any rotation that may 25 occur in an object. These rotations are generally the result of object motion and are estim~ l in a manner similar to that used to determine Z(Obj(i, j)). Specifically, the rotation parameter f~(Obj(i, j)) is estimated using the following expression WO 97/04600 ~ 7 ~ ~ PCT/US96/10141 r jl Obj(i, i)) = N ~, r~ --cy(i, ;)) ~ (dX ~ (i, jl Obj(i, j))) Obj(i, j) ~(i--cx(i, j)) ~ (dy k (i, j)--~y k (i~ jl Obj(i, j)))~

S As mentioned above, ~is four parameter representation of the DVF is very useful in applications that rcquire object manipulation.

As a final step, the parameterized representation of the DVF is used to refine estim~t~ of the line process Ik. Based on the four motion parameters, each object is compared with its neighboring objects. If an object is determined to be similar to a neighboring region they are merged to form a single region. Two regions are determine to be similar if the following four condition~ are met:

Condition 1.
(tX~,(i, jlRegion(i,j))--tXk(i+rl. j+ mlRegiorl(i+n~j+m)))- < Pl, Condition 2.
(t,,k(i, jlRegion(i,j))--tyk(i+~l.j+mlRegio~l(i+n,j+m))) < P2, Condition 3.
(Z(i, jl Re gion(i, j))--Z(i + n, j + ml Re gion(i + n, j + m))) S P3 Condition 4.
(~1(i,jlRegion(i,j))-~(i+n,j+mlRe~ 7rl(i+tl.j+m)))<P4 where (i+n,j+m) indicates the location of the neighboring region Obj(i + n, j + m) .

FIG. 7, numeral 700 is a flow diagram of steps for implementing a preferred embodiment of a method for e,stim~ting motion in accordance with the present invention. The current intensity frame is segmented (702) using a preprocessor to determine the boundaries of the objects contained in the scene captured by the video sequence. These boundaries are used by the spatially adaptive motion estimator wo 97/04600 2 ~ Q ~ PCT/USg6/10141 (704~ to adapt to objects that are moving differently in the video sequence. A first estimate of the DVF is provided by the spatially adaptive motion çstim~tor (704).
Based on the estim~tor of the object boundaries and the first estimate of the DVF, a first estimate of the moving object boundaries and a second çstim~te of the DVF are 5 obtained (706). Analyzing and modelling the first estimate of the moving object boundaries and the second eStim:~t(e of the DVF provides a third estimate of the DVF
and a second estimate of the moving object boundaries.

FIG. 8, numeral 800, is a flow diagram of steps for implementing a preferred 10 embodiment of a method for segm~nting in accordance with the present invention.
The first step is to remove noise from the image frame by filtering using an order St~lti!StiC'S filter (802). Next pixels are grouped into objects using a region grower (804). The final step includes merging small objects into larger ones using a region merger and a predetermined comparison test (806).
FIG. 9, numeral 900, is a flow diagram of steps for implementing a preferred embodiment of a method for adaptively estimating motion in accordance with the present invention. First, based on the object boundary estimate, a causal look-up table is accessed to provide at least one causal prediction coefficient and a causal 20 uncertainty parameter (902). Next, from a memory device containing previouslyestimated displacement vectors, a causal local neighborhood is set to a predeterrnined initi~li7~tion (904). A prediction of the current displacement vector is determined next, based on the previously estim~ted displacement vectors and at least one causal prediction coefficient (906). The predicted displacement vector is updated to become 25 the first displacement estimate and is stored in the causal local neighborhood (908).
The process is repeated for the entire current intensity frame (910) resulting in the first estimate of the DVF.

FIG. 10, numeral 1000, is a flow diagram of steps for implementing a 30 preferred embodiment of a method for estimating motion boundaries in accordance with the present invention. The first estimate of the DVF and the previously estimated displ~cement vectors of the second estimated DVF are stored in the Wo 97/04600 ~ 7 3 ~ PCT/US96/10141 nonc~us~l local neighborhood memory (1002). The previously estim~ted object boundaries are stored in the moving object boundary d~telll..ller (1004). Utilizin~ the object boundaries and the previously estimated displacement vectors a ~lr;st estim~t~
of ~e moving object boundaries is determined (1006). Based on the first estimate of 5 the moving object boundaries, a noncausal look-up table is accessed to provide at least one noncausal prediction coefficient and a noncausal uncertainty parameter(1008). A second estim~te of the current displacement vector is determined next,based on the previously estim~ted displacement vectors, at least one noncausal prediction coefficient and the DVF update determiner (lOlO). This estimation 10 process is repeated until all pixels in the current intensity frame have a displacement vector associated with them tlO12). The process is terminated when the percent total change in the second estimate of the DVF and the moving boundary estimate are below a set of thresholds (1014).

FIG. 11, numeral 1 lOO, is a flow diagram of steps for implementing a pl~rell~d embodiment of a method for analyzing motion in accordance with the present invention. First a center for each object is determined (1102). Next, anestimate of the translational motion is determined for each object (1104). This is followed by the estimation of a zoom parameter for each object (1106). An estimate 20 of a rotation parameter is also determined for each object (1108). Finally, the translation, rotation, and zoom parameters are used in fitting a model to the estimated DVF (lllO).

The system described in FIGs. 1 through 6 may be implemented in various 25 embodiments, such as Application Specific Integrated Circuit, ASIC, Digital Signal Processor, DSP, Gate Array. GA, and any tangible medium of/for a computer.

Claims (10)

26We claim:
1. A system for estimating motion, comprising:

A) a preprocessor, operably coupled to receive a current intensity frame, for segmenting the current intensity frame to provide a first object boundary estimate;

B) a spatially adaptive motion estimator, operably coupled to the preprocessor and operably coupled to receive the current intensity frame and a previous intensity frame, for providing a first displacement vector field estimate, a first DVF estimate; and C) a motion boundary estimator operably coupled to the spatially adaptive motion estimator and preprocessor and operably coupled to receive the current intensity frame and the previous intensity frame, for providing a seconddisplacement vector field estimate, a second DVF estimate, and a first moving object boundary estimate.
2. The system for estimating motion according to claim 1, wherein:

the preprocessor includes:
an order statistics filter, coupled to receive the current intensity frame, for removing noise and small objects from the current intensity frame; and an object boundary estimator, coupled to the order statistics filter, for providing the object boundary estimate;

the spatially adaptive motion estimator includes:
a causal look-up table, coupled to receive the object boundary estimate, for providing at least one causal prediction coefficient and a causal uncertainty parameter based on the object boundary estimate;
a causal local neighborhood for storing previously estimated DVFs, the causal local neighborhood is set to a predetermined initialization before DVF
estimation;
a DVF predictor, coupled to the causal look-up table and the causal local neighborhood, for providing a DVF prediction based on the previously estimated DVFs and at least one causal prediction coefficient: and a first DVF update determiner, coupled to receive the DVF prediction, the causal uncertainty parameter, the current intensity frame and the previous intensity frame, for providing the first DVF estimate, the first DVF estimate is an input to the causal local neighborhood where it is stored as a previously estimated DVF;

the motion boundary estimator includes:
a noncausal local neighborhood for storing previously estimated DVFs, the noncausal local neighborhood is initialized by the first DVF estimate;a moving object boundary unit for storing previously estimated moving object boundaries, the moving object boundary unit is initialized by the object boundary estimate;

a moving object estimator, coupled to the noncausal local neighborhood and the moving object boundary unit, for providing a first moving object boundary estimate;
a noncausal look-up table, coupled to the moving object estimator, for providing at least one noncausal prediction coefficient and a noncausal uncertainty parameter based on the first moving object boundary estimate;
a DVF estimator, coupled to the moving object estimator and the noncausal look-up table, for providing the second DVF estimate; and an estimation terminator coupled to the moving object estimator and the DVF estimator, for passing the first moving object boundary estimate and thesecond DVF estimate upon termination;
wherein the first moving object boundary estimate is an input to the moving object boundary unit and the second DVF estimate is an input to the noncausal local neighborhood;

and where further selected, wherein:

the object boundary estimator in the preprocessor includes a region grower that groups pixels into regions based on an adjustable window and a region merger that merges small regions based on a predetermined comparison test.
3. The system for estimating motion according to claim 1, further comprising:

D) a motion analyzer, operably coupled to the motion boundary estimator.
for providing a third DVF estimate and a second moving object boundary estimate based on the second DVF estimate and the first moving object boundary estimate;

and where further selected, wherein:

the motion analyzer incudes:
a center determiner, coupled to receive the first moving object boundary estimate, for providing a region location and a center point;

a translational motion estimator, coupled to receive the second DVF
estimate and the center point, for providing a translational motion estimate;
a zoom estimator, coupled to receive the center point, the region location, and the translational motion estimate, for providing a zoom estimate;
a rotation estimator, coupled to receive the center point, the region location, and the translational motion estimate, for providing a rotation estimate; and a model fitting unit, coupled to the translational motion estimator, the zoom estimator, and the rotation estimator, for providing the third DVF estimate and the second moving object boundary estimate based on the translational motion estimate, the zoom estimate, and the rotation estimate.
4. The system of claim 1 wherein the system is embodied in a tangible medium of/for a computer, and where selected, one of 4A-4B:

4A) wherein the tangible medium is a computer diskette; and 4B) wherein the tangible medium is a memory unit of the computer.
5. The system of claim 1 wherein one of 5A-5C:

5A) the system is embodied in a Digital Signal Processor, DSP;

5B) the system is embodied in an Application Specific Integrated Circuit, ASIC; and 5C) the system is embodied in a gate array.
6. A method for estimating motion, comprising:

segmenting, using a preprocessor, a current intensity frame to provide an object boundary estimate;

adaptively estimating the motion from a previous intensity frame to the current intensity frame to provide a first displacement vector field estimate, a first DVF estimate; and estimating motion boundaries based on the object boundary estimate and the first DVF estimate to provide a second displacement vector field estimate, a second DVF estimate, and a first moving object boundary estimate.
7. The method for estimating motion according to claim 6, wherein:

segmenting includes:
filtering, using an order statistics filter, the current intensity frame to remove noise and small objects from the current intensity frame;
region growing to group pixels into regions based on an adjustable window; and region merging to merge small regions based on a predetermined comparison test;

adaptively estimating the motion includes:
accessing a causal look-up table to provide at least one causal prediction coefficient and a causal uncertainty parameter based on the object boundary estimate;
storing previously estimated DVFs in a causal local neighborhood, the causal local neighborhood is set to a predetermined initialization before DVF
estimation;
predicting a DVF based on the previously estimated DVFs and at least one causal prediction coefficient; and updating the predicted DVF to provide the first DVF estimate, the first DVF estimate is an input to the causal local neighborhood where it is stored as a previously estimated DVF; and estimating motion boundaries includes:

storing previously estimated DVFs in a noncausal local neighborhood, the noncausal local neighborhood is initialized by the first DVF estimate;
storing previously estimated moving object boundaries in a moving object boundary unit, the moving object boundary unit is initialized by the object boundary estimate;
estimating moving objects to provide a first moving object boundary estimate;
accessing a noncausal look-up table to provide at least one noncausal prediction coefficient and a noncausal uncertainty parameter based on the first moving object boundary estimate; and estimating a DVF to provide the second DVF estimate;
terminating the estimating step when an entire frame has been estimated, and passing the first moving object boundary estimate and the second DVF
estimate upon termination.
8. The method for estimating motion according to claim 6, further comprising:

analyzing motion to provide a third DVF estimate and a second moving object boundary estimate based on the second DVF estimate and the first moving object boundary estimate;

and where further selected, wherein:

analyzing motion incudes:
determining a center to provide a region location and a center point;
estimating translational motion to provide a translational motion estimate;
estimating zoom to provide a zoom estimate;
estimating rotation to provide a rotation estimate; and fitting the translational motion estimate, the zoom estimate, and the rotation estimate to a predetermined model.
9. The method of claim 6 wherein the steps of the method are embodied in a tangible medium of/for a computer; and where further selected one of 9A-9B:

9A) wherein the tangible medium is a computer diskette; and 9B) wherein the tangible medium is a memory unit of the computer.
10. The method of claim 6 wherein one of 10A-10C:

10A) the steps of the method are embodied in a tangible medium of/for a Digital Signal Processor, DSP;

10B) the steps of the method are embodied in a tangible medium of/for an Application Specific Integrated Circuit, ASIC; and 10C) the steps of the method are embodied in a tangible medium of/for a gate array.
CA002200732A 1995-07-24 1996-06-12 Method and system for estimating motion within a video sequence Abandoned CA2200732A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US08/505,992 1995-07-24
US08/505,992 US5717463A (en) 1995-07-24 1995-07-24 Method and system for estimating motion within a video sequence

Publications (1)

Publication Number Publication Date
CA2200732A1 true CA2200732A1 (en) 1997-02-06

Family

ID=24012729

Family Applications (1)

Application Number Title Priority Date Filing Date
CA002200732A Abandoned CA2200732A1 (en) 1995-07-24 1996-06-12 Method and system for estimating motion within a video sequence

Country Status (7)

Country Link
US (1) US5717463A (en)
EP (1) EP0783821A4 (en)
KR (1) KR100256194B1 (en)
CN (1) CN1159276A (en)
CA (1) CA2200732A1 (en)
TW (1) TW299553B (en)
WO (1) WO1997004600A1 (en)

Families Citing this family (39)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
IL114838A0 (en) * 1995-08-04 1996-11-14 Spiegel Ehud Apparatus and method for object tracking
BR9805971A (en) * 1997-02-06 1999-08-31 Koninkl Philips Electronics Nv Process and system for object segmentation and tracking
US6137544A (en) * 1997-06-02 2000-10-24 Philips Electronics North America Corporation Significant scene detection and frame filtering for a visual indexing system
EP0936575B1 (en) * 1998-02-17 2004-07-28 Koninklijke Philips Electronics N.V. Image processing method for motion estimation in a sequence of images
US6400831B2 (en) 1998-04-02 2002-06-04 Microsoft Corporation Semantic video object segmentation and tracking
US6192079B1 (en) * 1998-05-07 2001-02-20 Intel Corporation Method and apparatus for increasing video frame rate
US6750900B1 (en) * 1998-07-23 2004-06-15 Eastman Kodak Company Method and apparatus for computing motion tracking parameters
EP1048170A1 (en) 1998-08-21 2000-11-02 Koninklijke Philips Electronics N.V. Problem area location in an image signal
US6711278B1 (en) * 1998-09-10 2004-03-23 Microsoft Corporation Tracking semantic objects in vector image sequences
US6307964B1 (en) * 1999-06-04 2001-10-23 Mitsubishi Electric Research Laboratories, Inc. Method for ordering image spaces to represent object shapes
US6400846B1 (en) * 1999-06-04 2002-06-04 Mitsubishi Electric Research Laboratories, Inc. Method for ordering image spaces to search for object surfaces
US7123745B1 (en) 1999-11-24 2006-10-17 Koninklijke Philips Electronics N.V. Method and apparatus for detecting moving objects in video conferencing and other applications
IL134182A (en) 2000-01-23 2006-08-01 Vls Com Ltd Method and apparatus for visual lossless pre-processing
US6731799B1 (en) * 2000-06-01 2004-05-04 University Of Washington Object segmentation with background extraction and moving boundary techniques
US6753929B1 (en) 2000-06-28 2004-06-22 Vls Com Ltd. Method and system for real time motion picture segmentation and superposition
EP1294194B8 (en) * 2001-09-10 2010-08-04 Texas Instruments Incorporated Apparatus and method for motion vector estimation
US20050104964A1 (en) * 2001-10-22 2005-05-19 Bovyrin Alexandr V. Method and apparatus for background segmentation based on motion localization
IL147370A (en) * 2001-12-27 2007-07-24 Itzhak Florentin Method and system for guiding a remote vehicle via lagged communication channel
US20040131117A1 (en) * 2003-01-07 2004-07-08 Sheraizin Vitaly S. Method and apparatus for improving MPEG picture compression
US20040146108A1 (en) * 2003-01-23 2004-07-29 Shih-Chang Hsia MPEG-II video encoder chip design
KR100598093B1 (en) * 2003-01-29 2006-07-07 삼성전자주식회사 Apparatus and method with low memory bandwidth for video data compression
US7349583B2 (en) * 2003-09-05 2008-03-25 The Regents Of The University Of California Global motion estimation image coding and processing
US7903902B2 (en) 2004-07-26 2011-03-08 Sheraizin Semion M Adaptive image improvement
US7639892B2 (en) * 2004-07-26 2009-12-29 Sheraizin Semion M Adaptive image improvement
JP3880985B2 (en) * 2004-08-05 2007-02-14 松下電器産業株式会社 Motion vector detection apparatus and motion vector detection method
US7526142B2 (en) * 2005-02-22 2009-04-28 Sheraizin Vitaly S Enhancement of decompressed video
AU2006201210A1 (en) * 2006-03-23 2007-10-11 Canon Information Systems Research Australia Pty Ltd Motion characterisation
WO2008005007A1 (en) 2006-06-29 2008-01-10 Thomson Licensing Adaptive pixel-based filtering
WO2008042029A2 (en) * 2006-09-29 2008-04-10 Thomson Licensing Automatic parameter estimation for adaptive pixel-based filtering
US20090141802A1 (en) * 2007-11-29 2009-06-04 Sony Corporation Motion vector detecting apparatus, motion vector detecting method, and program
US8605942B2 (en) * 2009-02-26 2013-12-10 Nikon Corporation Subject tracking apparatus, imaging apparatus and subject tracking method
TWI471521B (en) 2010-07-23 2015-02-01 Pixart Imaging Inc Displacement estimation method and displacement estimation device using the same
CN102346571B (en) * 2010-07-29 2014-03-19 原相科技股份有限公司 Displacement estimation method and displacement estimation device applying same
CN102075757B (en) * 2011-02-10 2013-08-28 北京航空航天大学 Video foreground object coding method by taking boundary detection as motion estimation reference
JP2013048717A (en) 2011-08-31 2013-03-14 Sony Corp Image processing device and method, recording medium, and program
US9046627B2 (en) 2012-05-30 2015-06-02 Massachusetts Institute Of Technology Generating a forecast by field coalescence
RU2534827C2 (en) * 2013-04-12 2014-12-10 Закрытое акционерное общество "НТК" Method for video surveillance of open space with fire hazard monitoring
CN104680504B (en) * 2013-11-26 2018-06-08 杭州海康威视数字技术股份有限公司 Scene-change detecting method and its device
KR102156410B1 (en) * 2014-04-14 2020-09-15 삼성전자주식회사 Apparatus and method for processing image considering motion of object

Family Cites Families (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4661849A (en) * 1985-06-03 1987-04-28 Pictel Corporation Method and apparatus for providing motion estimation signals for communicating image sequences
US4727422A (en) * 1985-06-03 1988-02-23 Picturetel Corporation Method and apparatus for efficiently communicating image sequence having improved motion compensation
EP0236519B1 (en) * 1986-03-08 1989-06-07 ANT Nachrichtentechnik GmbH Motion compensating field interpolation method using a hierarchically structured displacement estimator
DE3869475D1 (en) * 1987-06-02 1992-04-30 Siemens Ag METHOD FOR DETERMINING MOTION VECTOR FIELDS FROM DIGITAL IMAGE SEQUENCES.
FR2624997B1 (en) * 1987-12-16 1990-06-01 France Etat MULTI-PREDICTIVE ESTIMATION OF POINT MOVEMENT OF AN ELECTRONIC IMAGE
FR2651399B1 (en) * 1989-08-29 1996-05-15 Thomson Consumer Electronics METHOD AND DEVICE FOR ESTIMATING AND HIERARCHIZED CODING OF THE MOTION OF IMAGE SEQUENCES.
GB9321372D0 (en) * 1993-10-15 1993-12-08 Avt Communications Ltd Video signal processing

Also Published As

Publication number Publication date
AU6275596A (en) 1997-02-18
KR100256194B1 (en) 2000-05-15
US5717463A (en) 1998-02-10
AU682135B2 (en) 1997-09-18
CN1159276A (en) 1997-09-10
WO1997004600A1 (en) 1997-02-06
TW299553B (en) 1997-03-01
KR970706697A (en) 1997-11-03
EP0783821A4 (en) 1998-05-20
EP0783821A1 (en) 1997-07-16

Similar Documents

Publication Publication Date Title
CA2200732A1 (en) Method and system for estimating motion within a video sequence
US6625216B1 (en) Motion estimation using orthogonal transform-domain block matching
JP4906864B2 (en) Scalable video coding method
Chan et al. An efficient search strategy for block motion estimation using image features
Hong et al. Error concealment algorithms for compressed video
JP2009510869A5 (en)
Ribas-Corbera et al. Optimizing block size in motion-compensated video coding
Chan et al. Edge oriented block motion estimation for video coding
Sorwar et al. Fast global motion estimation using iterative least-square estimation technique
US20080112487A1 (en) Image search methods for reducing computational complexity of motion estimation
Yang et al. A low-complexity region-based video coder using backward morphological motion field segmentation
Dekeyser et al. Spatio-temporal Wiener filtering of image sequences using a parametric motion model
AU682135C (en) Method and system for estimating motion within a video sequence
Chan et al. On block motion estimation using a novel search strategy for an improved adaptive pixel decimation
Wang et al. Comparison of block-based and mesh-based motion estimation algorithms
Strobach Quadtree-structured interframe coding of hdtv sequences
Chen et al. A joint motion-image inpainting method for error concealment in video coding
Hamid et al. A new orthogonal-diamond search algorithm for motion estimation
Csillag et al. Frame-rate conversion based on acceleration and motion-based segmentation
Xu et al. An improved three-step search block-matching algorithm for low bit-rate video coding applications
EP0914738A2 (en) Method of estimation of motion between images
LeQuang et al. Object-oriented coding using successive motion field segmentation and estimation
Arora et al. A new fast motion estimation algorithm using adaptive size diamond pattern search with early search termination
Rehan et al. A new fast block matching algorithm using the simplex technique
Rajani et al. Implementation and analysis of Video Error Concealment using Moment Invariants

Legal Events

Date Code Title Description
EEER Examination request
FZDE Dead