A method and an apparatus for detecting an object in the image pickup field by template matching and tracking the detected object. An intruding object is detected from an image acquired by an image pickup device which takes the image of an image pickup field. A template image is formed from the detected object and is stored in a storing unit. A template matching is performed between a present input image of the image pickup field and the template image stored in the storing unit to detect the location of that part of the object which has a maximum degree of coincidence with the template image. An edge detection of the object is performed over a predetermined search area in the present input image which area includes the detected part of the object having a maximum degree of coincidence. Based on results of the edge detection, the detected part of the object is corrected and is determined as the present location of the object. A part of the present input image having the corrected location is a new template image with which the template image stored in the storing unit is updated. Instead of a single template image, a plurality of template images may be used for template matching to assure the tracking of the object with even higher accuracy and stability.
|
1. A tracking method for detecting an object in an image pickup field of an image pickup device, comprising the steps of:
forming from an image of said image pickup field a template image of a predetermined size including at least a part of said object; storing said template image in a memory; performing a template matching between an input image of said image pickup field from said image pickup device and said template image from said memory to detect a location of a part of said object in said input image which has a maximum degree of coincidence with said template image; performing an edge density detection of said object over an expanded part of said input image which has a larger size than said predetermined size and includes said part of said object of said template image; detecting a location of said object having a maximum edge density in said performing an edge density detection step; compensating a location of said template image based on said location of said object having a maximum edge density so that a greater part of said template image contains said detected object; and updating said template image, a location of which is compensated, into a new template image for a next input image.
5. A tracking method for detecting an object in an image pickup field of an image pickup device, comprising the steps of:
forming from images of said image pickup field a predetermined number of template images, each of which has a predetermined size and includes at least a part of said object; storing said template images in a memory; performing a template matching between an input image of said image pickup field from said image pickup device and each of said template images from said memory to detect a location of a part of said object in said input images; selecting a template image which has a maximum degree of coincidence in said performing a template matching step; performing an edge density detection of said object over an expanded part of said input image which has a larger size than said predetermined size and includes said at least a part of said object at a location of said selected template image; detecting a location of said object having a maximum edge density in said performing an edge density detection step; compensating the location of said template image having a maximum degree of coincidence based on said location of said object having a maximum edge density so that a greater part of said template image contains said detected object; and updating said template image, a location of which is compensated, as a new template image for a next input image.
14. A computer program product comprising:
a computer usable medium having computer readable program code means embodied in said medium for detecting and tracking an object in an image pickup field, said program code means comprising: means for cutting out that part of an input image signal from an image pickup device which includes at least a part of said object to produce a template image of a predetermined size and storing said template image on a template image storing unit; means for performing a template matching between an input image signal from said image pickup device and said template image from said memory to detect a location of a part of said object having a maximum degree of coincidence with said template image from said input image signal; means for performing an edge density detection of said object over an expanded part of said input image which has a larger size than said predetermined size and includes said part of said object at a location of said template image; means for detecting a location of said object having a maximum edge density in said means for performing an edge density detection; means for compensating a location of said template image based on said location of said object having a maximum edge density so that a greater part of said template image contains said detected object; means for updating said template image, the location of which is compensated, into a new template image for a next input image signal; and means for controlling said pan and tilt head based on said compensated location of said object so that said image pickup device is directed toward said object. 15. A computer program product comprising:
a computer usable medium having computer readable program code means embodied in said medium for detecting and tracking an object in an image pickup field, said program code means comprising: means for cutting out from input image signals each sequentially supplied from an image pickup device a plurality of template images each having a predetermined size and including at least a part of said object and storing said template images on a template image storing unit; means for performing a template matching between an input image signal from said image pickup device and each of said template images from said memory to detect a location of a part of said object in said input image; means for selecting a template image which has a maximum degree of coincidence in said template matching means; means for performing an edge detection of said object over an expanded part of said input image which has a larger size than said predetermined size and includes said part of said object at a said location of said selected template image; means for detecting a location of said object having a maximum edge density in said means for performing said edge density detection; means for compensating a location of said template image having a maximum degree of coincidence based on said location of said object having a maximum edge density so that a greater part of said template image contains said detected object; means for updating said template image, a location of which is compensated into, a new template image for a next input image; and means for controlling said pan and tilt head based on the compensated location of said object so that said image pickup device is directed toward said object. 13. An object tracking apparatus for detecting and tracking an object in an image pickup field, comprising:
an image pickup device which takes an image in a range to be monitored; a pan and tilt head with said image pickup device mounted thereon; an image input interface for sequentially converting a video signal of an object acquired by said image pickup device in said monitor range into an input image signal; an image processing unit connected to said image input interface which processes said input image signal; a pan and tilt head control interface connected to said image processing unit and said pan and tilt head; and a template image storing unit connected to said image input interface and said image processing unit, wherein said image processing unit cuts out that part of an input image signal in advance to produce a template image of a predetermined size which includes at least a part of said object and store said template image in said template image storing unit, performs a template matching between said input image signal and said template image to detect a location of a part of said object in said input image having a maximum degree of coincidence with said template image, performs an edge detection of said object over an expanded part of said input image which has a larger size than said predetermined size and which includes said part of said object at location of said template image, detects a location of said object having a maximum edge density in said edge density detection, compensates the location of said template image based on said location of said object having a maximum edge density so that a greater part of aid template image contains said detected object, updates said template image, a location of which is compensated, into a new template image for a next input image signal, and controls said pan and tilt head based on said compensated location of said template image to direct said image pickup device toward said object. 16. An object tracking apparatus for detecting and tracking an object in an image pickup field, comprising:
an image pickup device which takes an image in a range to be monitored; a pan and tilt head with said image pickup device mounted thereon; an image input interface for sequentially converting a video signal of an object acquired by said image pickup device in said monitor range into an input image signal; an image processing unit connected to said image input interface which processes said input image signal; a template image storing unit connected to said image input interface and said image processing unit; and a pan and tilt head control interface connected to said image processing unit and said pan and tilt head, wherein said image processing unit cuts out from sequentially input image signals a predetermined number of template images each including at least a part of said object, stores said template images in said template image storing unit, performs a template matching between an input image of said image pickup field from said image pickup device and each of said template images from said memory to detect a location of a part of said object in said input images, selects a template image which has a maximum degree of coincidence in said template matching step, performs an edge density detection of said object over an expanded part of said input image which has a larger size than said predetermined size and includes said part of said object at a location of said selected template image, detects a location of said object having a maximum edge density in said performing an edge density detection step, compensates the location of said template image having a maximum degree of coincidence based on said location of said object having a maximum edge density so that a greater part of said template image contains said detected object, updates said template image, a location of which is compensated, into a new template image for a next input image, and controls said pan and tilt head based on said compensated location of said template image so as to direct said image pickup device toward said object. 2. A method according to
extracting edge density components contained in said expanded part of said present input image; displaying accumulated quantities of x-axis and y-axis edge components on x-axis and y-axis, respectively; and detecting said portion having the maximum edge density from the accumulated quantities of x-axis and y-axis edge components on x-axis and y-axis.
3. A method according to
4. A method according to
controlling pan and tilt of said image pickup device so as to be directed toward said object based on a relation between said compensated location of said template image and a predetermined reference position in said image pickup field.
6. A method according to
7. A method according to
8. A method according to
9. A method according to
controlling said image pickup device to be directed toward said object based on a relation between said detected location of said object and a predetermined reference position in said image pickup field.
10. A method according to
extracting edge components contained in said expanded part of said input image; displaying accumulated quantities of x-axis and y-axis edge components on x-axis and y-axis, respectively; and detecting said part of said object having the maximum edge density from the accumulated quantities of x-axis and y-axis edge components on x-axis and y-axis.
11. A method according to
12. A method according to
|
The present invention relates to a monitor apparatus using an imaging unit, or in particular to an object tracking method for automatically detecting an object intruding into an imaging field or image pickup field from a video signal inputted from the imaging unit and automatically tracking the motion of the detected object and an object tracking apparatus for adjusting the imaging direction in accordance with the detected motion of the object.
A video monitor apparatus using an imaging unit or image pickup unit such as a camera has been widely used. A monitor system using a video monitor apparatus is in demand, in which an intruding object is automatically detected from the image input from an image input means such as a camera, the motion of the object is automatically tracked and a predetermined announcement or alarm action can be taken, apart from the manned monitoring system in which an intruding object such as a man or an automotive vehicle entering the monitor field is detected or tracked by a human monitor while watching the image displayed on the monitor.
For realizing such an unmanned system, the first step is to detect an intruding object in the view field by a so-called subtraction method or the like. In the subtraction method, the input image obtained by an imaging device such as a television camera (hereinafter referred to as the TV camera) is compared with a reference background image prepared in advance, i.e. an image not including the object to be detected to determine the brightness (or intensity) difference for each pixel and detect an area with a large difference value as an object. The part of the input image (hereinafter referred to as partial image) corresponding to the position of the intruding object detected in this way is registered as a template, so that a position associated with the maximum degree of coincidence with the template image is detected in the sequentially input images. This method is widely known as the template matching, and is described in detail, for example, in the book entitled "Introduction to Computer Image Processing" edited by Hideyuki Tamura, published by Soken Publishing Co. pp. 149-153, 1985. Further, this method is disclosed in "Digital Picture Processing" published by ACADEMIC PRESS pp. 296-303, 1976 and U.S. Pat. No. 5,554,983, the disclosures of each is hereby incorporated herein by reference.
The aforementioned method of tracking an intruding object by template matching poses the problem that with the change of the orientation or posture of the target object (when the target object person turns to the right or turns around, for example), the deviation of the target object from the matching position increases to such an extent that the accurate and stable tracking becomes impossible.
Specifically, the template matching has the property that the pattern portion high in contrast in the template images coincides for matching. In the case where a vehicle is a target object, for example, substantially the whole vehicle facing forward first constituting an object of matching (the input image 802 in
This will be explained with reference to FIG. 8.
In
This phenomenon is caused by the fact that the matching is conducted in such a manner as to reduce the deviation of the positions between the input image target for template matching and the image portion high in contrast in the template image. In this case, such a portion is the light of the vehicle. As a result, in the case where the target object turns to the left in
Further, at time t1, only the vehicle image is included in the template image 801a. With the direction change of the target object and the resulting deviation of the template position, however, the image of the background portion other than the object image occupies into the template image 807a. In the case where the tracking is continued using a template image like the template image 807a including many images other than that of the target object, the target object cannot be matched and the background portion that has occupied into the template is matched. Thus, in the case where the target object changes in direction, for example, the object tracking method using the template matching cannot guarantee the tracking of the target object and cannot assure stable tracking because the pattern of the target object apparently moves, which causes the position of the template to deviate.
Further, the aforementioned method of tracking an intruding object by the template matching, which is the process of detecting the portion of the input image associated with the maximum degree of coincidence with the template image, poses another problem that the target object, if temporarily hidden behind some blocking object, cannot be found. Still another problem is that in the case where a plurality of objects (moving objects) exist in the view field and the target object is temporarily hidden behind another moving object, the template image may be updated undesirably with the blocking object in front of the target object as a template image, thereby making it impossible to track the target object.
An example in which the target object is cannot be tracked is explained with reference to FIG. 14.
A similar phenomenon occurs also in the absence of another moving object. The target object being hidden behind a block is a case in point. In such a case, the template image is occupied in a lesser proportion by the target object, and therefore the tracking is terminated at the particular time point or a template image is updated to a different template image than that of the target object, thereby making the tracking at the next time point difficult.
As described above, the conventional method of tracking an intruding object by template matching has the disadvantage that stable tracking is impossible in the case where the target object changes its orientation or posture considerably.
An object of the present invention is to provide a reliable object tracking method and apparatus by which the disadvantages of the prior art are obviated and an object can be accurately detected and tracked even in the case where the orientation or posture of the target object is changed considerably.
In order to achieve this object, according to one aspect of the invention, there is provided an object tracking method used with an object tracking apparatus in which a template image is registered, and an object in the imaging field is automatically detected by template matching based on the registered template image and the detected object is automatically tracked, comprising the steps of registering an image including at least a part of a detected object as a template image, detecting the position of a partial image of the input image signal associated with the maximum degree of coincidence between the registered template image and the input image signal by template matching, correcting the detected template position to the position of the object detected based on the edge image of the input image and updating the template image based on the corrected detected position thereby to track the object in the imaging field.
According to another aspect of the invention, there is provided an object tracking apparatus comprising an imaging unit for picking up an image of the monitor range for detecting and tracking an object in the imaging field, a pan and tilt head for changing the direction of the view field of the imaging unit, an image input interface for converting the video signals acquired by the imaging unit sequentially into image signals, an image processor for processing the image signals converted by the image input interface, and a pan and tilt head control interface for supplying a control signal for controlling the pan and tilt head to enable the image processor to change the direction of the view field of the imaging unit, wherein the image processor matches a template registered in advance with the image signals sequentially input from the imaging unit, the template matching position is corrected to a new template matching position associated with the maximum edge density in a predetermined range (expanded partial image) in the neighborhood of the template matching position associated with the maximum degree of coincidence of the image signal, the image at the newly corrected matching position is updated as a template, the direction of the target object is detected based on the newly corrected matching position, the direction of the view field of the imaging unit is adjusted through the pan and tilt head control interface from the direction thus obtained, and thereby the object intruding into the imaging field of the imaging unit is tracked.
As explained above, the conventional intruding object tracking method using the template matching described above also has another disadvantage that an object cannot be accurately tracked in the case where another object passes the front of a target object in relative motion.
Accordingly, another object of the invention is to provide a method and an apparatus for tracking an object high in reliability, in which the disadvantages of the prior art described above are obviated and an object can be accurately detected and tracked even in the case where another object passes the front of the target object.
In order to achieve this object, according to one aspect of the invention, there is provided an object tracking method used with an object tracking apparatus in which a predetermined number of template images are registered, an object in the imaging field is automatically detected by template matching based on the predetermined number of the template images and the detected object is automatically tracked, the method comprising the steps of registering the images of a detected object as template images, detecting the position of an image associated with the maximum degree of coincidence between an input image signal and any one of the predetermined number of registered template images, determining the detected image position by template matching as the position of the object tracked, and updating the template image based on the detected position, thereby tracking an object in the imaging field.
According to another aspect of the invention, there is provided an object tracking apparatus comprising an imaging unit for imaging the monitor range for detecting and tracking an object in the imaging field, an image input interface for converting the video signals acquired by the imaging unit sequentially into an image signal, and an image processor for processing the image signal converted by the image input interface, wherein the image processor conducts the template matching between a predetermined number of templates and the image signals sequentially input from the imaging unit, the template matching position is corrected by determining the template matching position associated with the maximum degree of coincidence obtained by template matching as a new template matching position, the image at the newly corrected matching position is updated as a template, and the direction of the target object is detected from the newly corrected matching position, thereby tracking an object intruding into the imaging field of the imaging unit.
Further, an object tracking apparatus according to an embodiment of the invention comprises a pan and tilt head for changing the direction of the view field of the imaging unit, and a pan and tilt head control interface connected to the image processor for supplying a control signal for controlling the pan and tilt head, wherein the direction of the view field of the imaging unit is adjusted through the pan and tilt head control interface toward the direction detected by the image processor thereby to track an object intruding into the view field of the imaging unit.
Further objects, features and merits of the present invention appear from the following detailed description of a few embodiments of the invention, and from the appended claims as well as the drawings.
In the case where a target object is tracked using the template matching, a template image or picture template is normally sequentially updated using the image of the position of the target object detected by the matching processing in order to follow the change of the position of the target object. Before explaining the embodiment, these processes will be explained with reference to
In
In
Then, in the object presence determining step 404, the image extraction unit 608 detects a cluster of the pixels of which the pixel value is "255" in the binarized image 604, and in the presence of a cluster of pixels having the pixel value of "255", the object detection processing is terminated, and the partial image of the input image corresponding to the circumscribed rectangle of the existing cluster is registered in the image memory (to be described later) as a new template image 613. In the absence of such pixels, on the other hand, the process branches to the image input step 401.
The flow of the object tracking processing is explained with reference to FIG. 5. First, the explanation is made with reference to
In
Specifically, in
In other words, the maximum degree of coincidence and the position associated with the maximum degree of coincidence are obtained in the template matching step 103.
As a method of calculating this coincidence degree r(Δx,Δy), the index called the normalized correlation obtained from equation (1) below can be used, for example.
where,
In the case where the template matching is conducted with respect to the input image 702, ft(x, y) designates the input image 702, ft(x, y) the template image 701a, (x, y) is the coordinate indicating the pixel portion, x-axis is a horizontal direction, y-axis is a vertical direction, (xo, yo) is the upper left coordinate (in the image with the origin located at the upper left) of the registered template image 701a, Δx is an axis of abscissas of the search range, Δy is an axis of ordinates of the search range, and Dt designates the search range 702c of the template matching process. In the case where an image having exactly the same pixel value as the template image 701a exists in the search range 702c, the coincidence degree r(Δx,Δy) is given as 1∅ In the template matching step 103, the index expressed by equation (1) is calculated for the search range 702c indicated by (Δx,Δy) ∈ D, in which the position (circumscribed rectangle) 702a associated with the maximum coincidence degree r(Δx,Δy) is detected. This search range 702c is determined by the apparent amount of movement of the target object. For example, assume that an object moving at the rate of 40 km/h is monitored by a TV camera (CCD of element size 6.5 mm×4.8 mm, lens of focal length 25 mm, input image size 320×240 pixels, processing intervals 0.1 frame/sec) located 50 m away. The apparent amount of movement of the object is given as 27.4 pix/frame for horizontal direction and 27.8 pix/frame for vertical direction. Thus, D can be set at about a value satisfying the relations -30 pix<Δx<30 pix, -30 pix<Δy<30 pix.
The method of calculating the coincidence degree is not limited to the aforementioned index of normalization correlation. Instead, the difference of the pixel value may be determined for each pixel between the input image and the template image, for example, and the reciprocal of the accumulated value of the absolute values of the particular difference may be determined as the coincidence degree.
After it is determined in the template matching step 103 that the object has moved to the position in the input image 702 associated with the maximum coincidence degree with the template image 701a (from the circumscribed rectangle 702b to the circumscribed rectangle 702a), assume that the maximum coincidence degree has decreased to a predetermined value or less (for example, less than 0.5). Then, the maximum coincidence degree determining step 104 determines that the target object has disappeared from the input image, and the process branches to the object detection processing step 101. In the case where the maximum coincidence degree is not less than the predetermined value (not less than, for example, 0.5), on the other hand, the process branches to the template update step 106.
In the template update step 106, the template image 701a is updated to the template image 703a using the partial image 702a having the maximum degree r(Δx,Δy) of coincidence with the template image 701a in the search range 702c of the input image 702. The template image is updated by reason of the fact that if the posture of the target object changes (for example, the image changes as the man constituting the target object raises his hand, bends himself or raises his leg) and the template image is not updated, the coincidence degree would decrease for a reduced reliability of the tracking result. For this reason, the template image is updated with the partial image 702e of the detected target object as a new template image 703a, so that a stable tracking is secured even in the case where the target object changes his posture.
Then the process proceeds to step 107 for controlling the pan and tilt head of the camera.
In the pan and tilt head control step 107, the pan and tilt motor of the camera pan and tilt head 302 is controlled based on the displacement between the image center and the position of the target object detected by template matching, i.e. the direction of the target object with respect to the optical axis of the camera. Specifically, the center position (x0+Δx+dx/2, y0+Δy+dy/2) ((dx,dy) indicates the size of the template) of the target object detected by template matching is compared with the center position (160, 120) of the image (assuming that the image size is 320×240), and in the case where the center position of the target object detected is located to the left of the center position of the image, the pan motor of the camera pan and tilt head is controlled to move the optical axis of the camera leftward, while in the case where the center position of the target objected is located to the right of the center position of the image, on the other hand, the pan motor of the camera pan and tilt head is controlled to move the optical axis of the camera rightward. Also, in the case where the center position of the target object detected is located above the center position of the image, the tilt motor of the camera pan and tilt head is controlled to move the optical axis of the camera upward, while in the case where the center position of the target object detected is located below the center position of the image, the tilt motor of the camera pan and tilt head is controlled to move the optical axis of the camera downward. The pan motor and the tilt motor can be controlled at the same time. In the case where the center position of the target object detected is located to the left above the center position of the image, for example, the tilt motor of the camera pan and tilt head is controlled to move the optical axis of the camera leftward while at the same time controlling the pan motor to move the optical axis of the camera upward. By doing so, the camera pan and tilt head can be controlled in such a manner as to hold the target object on the optical axis of the camera.
Then, in the alarm/monitor display step 108, an alarm is issued or the image of the target object is displayed on the monitor in the case where the target object is located in such a range that a predetermined alarm is to be issued.
Upon completion of the alarm/monitor display step 108, the process is returned to the image input step 401, where a new input image is acquired and the template matching is conducted again. Specifically, the template matching is conducted using the template image 703a updated by the input image 702 at time point t0 and the input image 704 at time point t0+1. By this time, the search range 704c has been moved to the position centered at the template image 704b updated at time point t0, and the new search range is searched. An object associated with the maximum coincidence degree is detected, and a new template image 705a is generated based on the position 704a of the object thus detected.
As described above, as long as a target object exists, the process of steps 401, 103, 104, 106, 107 and 108 are repeated, so that the template image is updated to the new template images 706a, 708a and so on, thus continuing to track the target object.
In the object tracking method according to the invention, in order to solve the aforementioned problem that the pattern of the target object apparently moves causing the template position to be displaced, the feature that the target object has more edge components than the background is utilized. In other words, the position of the template image updated during the tracking process is corrected based on the density of the edge image of the input image.
Specifically, according to this invention, an object is detected by the subtraction method, the image of the detected object is held as a template image, and the object is tracked while correcting the detected position of the image of the object to a position associated with the maximum density of the edge image searched over an area consisting of the portion detected by the template matching and a peripheral portion extending around the portion detected by the template matching. In this way, a stable tracking is secured even in the case where the orientation of the target object is changed.
In the above described embodiment, a template image formed with respect to an intruding object detected by the subtraction method is formed such that a circumscribed rectangle of a cluster of pixels detected by the subtraction method is formed and the part of the input image or partial image surrounded by the circumscribed rectangle is cut out as a template image. However, the method of deciding the size of the template image to be cut out is not limited to this method. For example, the size may be determined by multiplying the size of the circumscribed rectangle by a predetermined coefficient such as 0.8, 1.1 or the like. Further, as will be described below, when a CCD is used as the image pickup device, the size of the object regarded as an object to be tracked can be calculated from the size of the CCD, the focal length of the lens used and the distance from the CCD to the detected object and the thus calculated size of the object may be the size of the template image.
More specifically, the apparent vertical size (A) and the apparent horizontal size (B) of the object to be monitored are given as
where T is the vertical size in mm of the CCD, S the horizontal size in mm of the CCD, f the focal length in mm of the lens, L the distance in m up to the object, H the minimum height in m of the object, W the minimum width in m of the object, X the vertical image size in the number of pixels in the vertical direction and Y the horizontal image size in the number of pixels in the horizontal direction.
For example, when an object to be tracked 200 m away is to be monitored with an image of 256×192 pixels using a ½-type CCD (6.5 mm (W)×4.8 mm (H)) and a lens having a focal length of 112 mm, the apparent vertical size (A) and the apparent horizontal size (B) are
Alternatively, the A and B thus calculated may further be multiplied by 0.8, 1.1 or other coefficient as described above to set the size of the template image. In this way, by excluding objects in the view field that are smaller than the predetermined size, a high precision object tracking becomes possible.
In
All the flowcharts used for explanation hereinbelow are based on the hardware configuration of the object tracking apparatus described with reference to FIG. 3.
A first embodiment of the invention will be explained with reference to FIG. 1.
In step 104 for determining the maximum coincidence degree, assume that the maximum coincidence degree is not less than a predetermined value. The process proceeds to step 105 for correcting or compensating for the template position. The contents of the process in the template position correction step 105 will be explained with reference to
In
Then, from the edge image 902, the search area 903a (defined by dotted frame in the view 903, i.e. upper left coordinate (x0-d, y0-d), size (dx+2d, dy+2d)) is cut out, which is the result of expanding the range of the detected position 804a obtained in the template matching step 103 by a predetermined pixel amount d (d: tolerable displacement of the matching position with the change in the orientation of the target object) in four directions, thereby producing the projected image 903b of the edge image on the x axis and the projected image 904c of the edge image on the y axis. Therefore, the search area 903a is an expanded partial image including the range of the detected position 804a.
In the graph 904, the abscissa represents the horizontal direction (x axis), and the ordinate the value hx(x) of the projected image 903b of the edge image for each pixel (pix) along the horizontal (x axis) direction. In the graph 905, on the other hand, the abscissa represents the vertical (y axis) direction, and the ordinate the value hy(y) of the projected image 903c of the edge image for each pixel (pix) along the vertical (y axis) direction.
The projection value x(x0) of the projected image 903b along x axis at x=x0 is obtained by changing (x, y) in such a manner that y0-d<y<y0+dy+d at x=x0 in the edge image cut out as the search area 903a and counting the number of pixels corresponding to the pixel value of "255". Also, the projection value y(y0) of the projected image 903c at y=y0 along y axis is obtained by changing (x, y) in such a manner that x0-d<x<x0+dx+d at y=y0 in the edge image cut out as the search area 903a, and counting the number of pixels corresponding to the pixel value of "255". The range 904b is the range (x1<x<x1+dx) associated with the maximum accumulated projection value, i.e. the maximum edge density, and this position is obtained from the following equation (2).
This equation (2) is to determine x1 associated with the maximum accumulated value of hx(x) in the relation x1<x<x1+dx for x changed so that x0-d<x1<x0+d. In similar fashion, the range (y1<y<y1+dy) associated with the maximum accumulated edge value is obtained for the projected image on y axis. Thus, the position (upper left coordinate (x0, y0)) of the target object detected in the template matching step 103 is changed to the position (upper left coordinate (x1, x1)) corrected in the template position correction step 105.
The effect of the aforementioned embodiment will be explained with reference to FIG. 10.
In the method shown in
In the case of
After the template position correction step 105, the template update step 106 is executed by updating the corrected position of the target object as a new template image. A similar process to that of
As described above, according to this embodiment of the invention, the position detected in the template matching step 103 is corrected to the position associated with the maximum edge density by detecting the edges included in the target object. Even in the case where the target object changes the orientation thereof, therefore, the position of the template is not displaced from the target object and the target object can be accurately tracked.
A second embodiment of the invention will be explained with reference to FIG. 2.
In
After that, in the template matching step 103, the stored template image at time point t0-1 is matched with the input image at time point t0. Through the branching step 201 (described later), the process proceeds to the maximum coincidence degree determining step 104'.
In the case where the maximum coincidence degree is not less than a predetermined value in the maximum coincidence degree determining step 104', the process proceeds to the template position correction step 105, while in the case where the maximum coincidence degree is less than a predetermined value, the process returns to the object detection processing step 101.
In the template position correction step 105, the position extracted in the maximum coincidence degree determining step 104' is corrected as a detected position for time point t0. In the next step 202 for storing a plurality of templates, the template at time point t0 is newly stored based on the corrected detected position for time point t0. At the same time, the template image at time point t0-1 already registered in the initial template registration step 102 is stored as it is. Then, the process proceeds to the camera pan and tilt head control step, where the view field of the camera is directed toward the target object based on the corrected detected position for time point t0.
Then, the process proceeds to the alarm/monitor display step 107 for sounding an alarm or displaying the image of the target object on the monitor.
Upon completion of the alarm/monitor display step 107, the process returns to the image input step 401 where a new input image is acquired and the template matching is conducted again.
When the process is returned to the template matching step 103, there are stored two templates including the one at time point t0-2 and the other at time point t0-1 ("-1" is added since the time is advanced by "1"). In the template matching step 103, the input image for time point t0 is matched with the template for time point t0-1, and then the process proceeds to step 201.
In the branch step 201, all the template images in store are checked whether they are subjected to template matching or not. Assume that the template matching for the template at time t0-1 has been completed but the template matching process remains undone for time point t0-2. In that case, the process returns to step 103 for conducting the template matching between the template for time point t0-2 and the template for time point t0. In this way, the remaining templates are matched one by one, and upon complete template matching of all the templates, the process proceeds from the branch step 201 to the maximum coincidence degree determining step 104'.
In the maximum coincidence degree determining step 104', the largest value is selected from the maximum coincidence degrees obtained for a plurality of template images by template matching. In the case where the largest maximum coincidence degree thus selected is not less than a predetermined value (for example, 0.5), the process proceeds to the template position correction step 105, while in the case where the largest value of the maximum coincidence degree is less than a predetermined value, on the other hand, the process returns to the object detection step 103, regarding the target object as nonexistent now in the input image.
In the template position correction step 105, the input image edge processing is performed on the template image associated with the largest value of the maximum coincidence degree selected in the maximum coincidence degree determining step 104', and the position of the target object is corrected based on the edge image obtained.
In the next step 202 for storing a plurality of templates, the template at time point t0 is newly stored based on the corrected detected position for time point t0. At the same time, the template image at time point t0-1 registered in the template registration step 102 is held as it is.
The number of template images stored in the template storage step 202 is predetermined at an arbitrary number (for example, "3"), and when the predetermined number is exceeded, the oldest template acquired is deleted. Then, the process proceeds to the camera pan and tilt head control step for controlling the camera position.
The process further proceeds to the alarm/monitor display step 108 for sounding an alarm or displaying the image of the target object on the monitor, for example.
Upon completion of the alarm/monitor display step 108, the process returns to the image input step 401 to acquire a new input image thereby to continue the template matching process again.
According to this second embodiment, the edges of the target object are detected based on the position detected in the template matching step 103, and the detected position is corrected to the position associated with the maximum edge density. The template images for a predetermined number of frames obtained at different time points are matched independently of each other. Even in the case where the target object changes the orientation (posture) thereof or another object transverses in front of the target object, therefore, the area associated with the maximum coincidence degree is used as the template matching position based on a plurality of the past template images for the purpose of correction. As a result, the template position is not displaced from the target object and the target object can be accurately tracked without tracking another object.
As described above, according to this embodiment, the target object changing the orientation or posture thereof can be stably tracked and thus the monitor used with the imaging unit can find considerably wider applications.
Now, a third embodiment of the invention will be explained with reference to
In the object tracking method according to this embodiment, a plurality of frames of template images used for template matching are held in the process of tracking taking advantage of the feature that the coincidence degree of matching decreases when a plurality of objects pass each other. A plurality of template images obtained at different time points are matched independently of each other, and the target object is tracked based on the template image having the highest degree of coincidence.
Specifically, according to this embodiment, an object is detected by the subtraction method, and the images of the detected object are held in a predetermined number of frames as templates, each of which is matched. The template associated with the maximum coincidence degree and the position thereof are detected, thereby making it possible to stably track the target object even when another object passes in front of the target object.
The example of hardware configuration of the object tracking apparatus shown in
The flowchart below will be described with reference to the hardware configuration of the object tracking and monitor apparatus shown in FIG. 3. Nevertheless, the invention is not limited to the hardware configuration shown in
An embodiment of the invention will be explained with reference to FIG. 12.
Once the process is started in
In the template matching step 103, the template matching is conducted between the template image stored at time point t0-1 and the input image for time point t0. Through the branching step 210 (to be described later), the process proceeds to the maximum coincidence degree determining step 214.
In the case where the maximum coincidence degree is not less than a predetermined value in the maximum coincidence degree determining step 214, the process proceeds to the plural-template storage step 202', while in the case where the maximum coincidence degree is less than the predetermined value, the process returns to the object detection step 101.
In the plural-template storage step 202', the template at time point t0 is newly stored based on the detected position of the object detected in the input image for time point t0. In the process, the template image at time point t0-1 already registered in the template registration step 102 is held as it is.
Then, the process proceeds to the camera pan and tilt head control step 107 for directing the view field of the camera 301 toward the target object.
The process proceeds to the alarm/monitor display step 108 for sounding an alarm or displaying the image of the target object on the monitor.
Upon completion of the alarm/monitor display step 108, the process returns to the image input step 401 for acquiring a new input image and conducting the template matching process again.
By the time the process is returned to the template matching step 103, the two templates including the template at time point t0-2 and the template at time point t0-1 are stored ("-1" is added since the time is advanced by "+1"). In the template matching step 103, the input image at time point t0 is matched with the template at time point t0-1 and the process proceeds to the branching step 210.
In the branching step 210, all the template images in store are checked whether they have been subjected to template matching or not. The template at time point t0-1 is subjected to template matching, but the template at time point t0-2 is not yet subjected to template matching. Therefore, the process is returned to step 103, and the template matching is conducted between the template at time point t0-2 and the input image at time point t0. In this way, the remaining templates are subjected to template matching one by one, and upon complete template matching for all the templates, the process proceeds from the branching step 201 to the maximum coincidence degree determining step 214.
In the maximum coincidence degree determining step 214, the largest value is selected from a plurality of maximum coincidence degrees for a plurality of template images. In the case where the maximum coincidence degree of the selected largest value is not less than a predetermined value, the process proceeds to the template position correction step 105, while in the case where the maximum coincidence degree of the largest value is less than the predetermined value, the process returns to the template matching step 103.
In the plural-template storage step 202', the template at time point t0 is newly stored based on the position associated with the largest value of the maximum coincidence degree determined in the maximum coincidence degree determining step 214 among the objects detected in the input image for time point t0.
At the same time, the template image at time point t0-1 already registered in the initial template registration step 102 is kept stored as it is.
The number of the template images stored in the plural-template storage step 202'is predetermined (at "3", for example), and when the predetermined number is exceeded, the oldest acquired template, for example, is deleted.
The template is updated in step 202' in this way. As an alternative, the template of the lowest coincidence degree in the template matching step 103 may be deleted.
Then, the process proceeds to the camera pan and tilt head control step 107 for directing the view field of the camera 301 toward the target object, followed by proceeding to the alarm/monitor display step 108 for sounding an alarm or displaying the image of the target object on the monitor, as the case may be.
Upon completion of the alarm/monitor display step 108, the process is returned to the template matching step 103 for continuing the template matching process again.
The effects of this embodiment will be explained with reference to FIG. 13.
In the method shown in
As seen from the input images 1904, 1906, in the case where the template is updated at a time point when a plurality of objects pass each other (the time point when two objects are superposed one on the other), the two objects in the template image are picked up undesirably. As a result, the pixels of the object being tracked occupy a lesser proportion of the template image. In the input image 1902 at time point t0, the objects have not yet passed each other, and therefore no tracking problem occurs. In this case, a small target object is being tracked. In the input image 1904 in which another moving object 1904c for time point t0+1 begins to be superposed, the template image produced is not yet imaged with another moving object. Therefore no problem is posed as yet and the target object is still captured. However, another moving object 1904c is imaged in the template image 905a produced based on this matching position. Even in the input image 906 which is substantially superposed by another moving object 1906c, the detected position 1906a used for matching with the template image 1905a having a greater proportion of another moving object 1904c imaged therein is not so displaced from the detected position 1906b used for matching with the template image 1903a in the immediately preceding frame. In the template 1907a stored at this time point, however, the pixels of the large object 1906c occupy a larger proportion of the template image than the pixels of the original target object in the template image. Thus, the template image is stolen by the large object 1906c, so that the large object 1908c is erroneously tracked at and after time point t0+2.
According to this invention, however, the matching is conducted using the template image 1905a previously stored, and therefore the original target object can be captured for the detected position 1908b.
As described above, according to the embodiment shown in
Now, a fourth embodiment of the invention will be explained with reference to FIG. 15.
In
Then, in the template counter reset step 991, the value of a template counter i (not shown) stored in the work memory 306 is initialized to 1. After that, in the template matching step 992, the template matching is performed between the template image for time point t0-1 in store and the input image for time point t0 obtained in the image input step 401. In the maximum coincidence degree determining step 993, the process proceeds to the plural-template storage step 202 in the case where the maximum coincidence degree obtained in the template matching step 992 is not less than a predetermined value, while in the case where the maximum coincidence degree is less than the predetermined value, the process proceeds to the counter increment step 994. In the counter increment step 994, 1 is added to the value of the template counter i stored in the work memory 306 thereby to update the value of the template counter. Then, in the branching step 995, the process returns to the object detection step 101 in the case where the value of the template counter i stored in the work memory 306 reaches not less than a predetermined value (for example, 5 which is the value indicating the preceding number of frames up to which the template images are stored), while in the case where the value of the template counter i is less than the predetermined value, the process returns to the template matching step 992.
According to the embodiment of
It will thus be understood from the foregoing description that according to this embodiment, even in the case where a plurality of moving objects are present within the monitor field or the target object is hidden behind another object temporarily, the target object can be tracked steadily and the application of the monitor using the imaging unit can be widened considerably.
While the present invention has been particularly described and shown with reference to the presently preferred embodiments thereof, it will be understood by those ordinary skilled in the art that various changes in form and detail and omissions may be made therein without departing from the scope of the invention.
For example, a computer program product embodying the computer program code means in a computable usable medium for implementing the object tracking method according to the invention described above is of course included in the scope of the present invention.
Patent | Priority | Assignee | Title |
10133950, | Mar 04 2011 | Qualcomm Incorporated | Dynamic template tracking |
10438360, | May 25 2016 | Canon Kabushiki Kaisha | Video processing apparatus, video processing method, and storage medium |
10489917, | Apr 14 2008 | GRASS VALLEY CANADA | Technique for automatically tracking an object in a defined tracking window by a camera based on identification of an object |
10497133, | Jul 08 2016 | Omron Corporation | Object tracking apparatus and object tracking method |
10521917, | Jul 29 2016 | Omron Corporation | Image processing apparatus and image processing method for object tracking |
10645299, | Jun 23 2017 | GOERTEK INC | Method for tracking and shooting moving target and tracking device |
10872421, | May 11 2017 | BOE TECHNOLOGY GROUP CO , LTD | Object tracking method and object tracking device |
10878677, | Apr 14 2017 | HANWHA VISION CO , LTD | Method of controlling panning and tilting of surveillance camera using edge value |
10932445, | Nov 08 2002 | Biopar, LLC | System for uniquely identifying subjects from a target population |
10949995, | May 22 2017 | Alibaba Group Holding Limited | Image capture direction recognition method and server, surveillance method and system and image capture device |
11020563, | Jul 14 2016 | C R BARD, INC | Automated catheter-to-vessel size comparison tool and related methods |
11026630, | Jun 26 2015 | C. R. Bard, Inc. | Connector interface for ECG-based catheter positioning system |
11074695, | Jan 05 2017 | NEC Corporation | Image processing device, image processing method, and recording medium storing program |
11074696, | Jan 05 2017 | NEC Corporation | Image processing device, image processing method, and recording medium storing program |
11123099, | Nov 26 2007 | C. R. Bard, Inc. | Apparatus for use with needle insertion guidance system |
11134915, | Nov 26 2007 | C. R. Bard, Inc. | System for placement of a catheter including a signal-generating stylet |
11159798, | Aug 21 2018 | International Business Machines Corporation | Video compression using cognitive semantics object analysis |
11413015, | Oct 21 2011 | C. R. Bard, Inc. | Systems and methods for ultrasound-based medical device assessment |
11419517, | Jun 12 2009 | Bard Access Systems, Inc. | Apparatus and method for catheter navigation using endovascular energy mapping |
11529070, | Nov 26 2007 | C. R. Bard, Inc. | System and methods for guiding a medical instrument |
11621518, | Oct 16 2018 | Bard Access Systems, Inc. | Safety-equipped connection systems and methods thereof for establishing electrical connections |
11707205, | Nov 26 2007 | C. R. Bard, Inc. | Integrated system for intravascular placement of a catheter |
11759166, | Sep 20 2019 | Bard Access Systems, Inc. | Automatic vessel detection tools and methods |
11779240, | Nov 26 2007 | C. R. Bard, Inc. | Systems and methods for breaching a sterile field for intravascular placement of a catheter |
11877810, | Jul 21 2020 | Bard Access Systems, Inc. | System, method and apparatus for magnetic tracking of ultrasound probe and generation of 3D visualization thereof |
11890139, | Sep 03 2020 | Bard Access Systems, Inc. | Portable ultrasound systems |
6857196, | Apr 30 2003 | Method and apparatus for measuring a intracorporal passage image | |
6991386, | Jul 02 2003 | ADTEC ENGINEERING CO , LTD | Image forming apparatus and image forming method |
7190814, | Oct 13 2000 | Omron Corporation | Image comparison apparatus and method for checking an image of an object against a stored registration image |
7236091, | Feb 10 2005 | PINC Solutions | Position-tracking system |
7245215, | Feb 10 2005 | PINC Solutions | Position-tracking device for position-tracking system |
7266454, | Jul 01 2003 | Nissan Motor Co., Ltd. | Obstacle detection apparatus and method for automotive vehicle |
7274400, | Jan 28 2000 | FUJIFILM Corporation | Digital camera and composition assist frame selecting method for digital camera |
7274819, | May 31 2001 | Canon Kabushiki Kaisha | Pattern recognition apparatus using parallel operation |
7281871, | Jul 02 2003 | ADTEC ENGINEERING CO , LTD | Image forming apparatus and image forming method |
7286164, | Jan 19 2001 | Ricoh Company, Ltd. | Method of and unit for inputting an image, and computer product |
7319776, | Nov 26 2001 | Fujitsu Limited | Image processing method and image processing program |
7321305, | Jul 05 2005 | PINC Solutions | Systems and methods for determining a location of an object |
7561721, | Feb 02 2005 | THE BANK OF NEW YORK MELLON, AS ADMINISTRATIVE AGENT | System and method for range measurement of a preceding vehicle |
7609877, | Feb 09 2006 | NAVY, UNITED STATES OF AMERICA, THE, AS REPRESENTED BY THE SECRETARY OF THE | Tactical image parameter adjustment method for stereo pair correlation |
7623681, | Dec 07 2005 | THE BANK OF NEW YORK MELLON, AS ADMINISTRATIVE AGENT | System and method for range measurement of a preceding vehicle |
7636043, | Jun 29 2006 | Google Technology Holdings LLC | Method and apparatus for improving the accuracy of security systems |
7643653, | Feb 04 2000 | CheckVideo LLC | System for automated screening of security cameras |
7650030, | Dec 03 2004 | SRI International | Method and apparatus for unsupervised learning of discriminative edge measures for vehicle matching between non-overlapping cameras |
7683939, | Jan 19 2001 | Ricoh Company, Ltd. | Method of and unit for inputting an image, and computer product |
7822224, | Jun 22 2005 | CheckVideo LLC | Terrain map summary elements |
7841300, | Nov 08 2002 | BIOPAR LLC | System for uniquely identifying subjects from a target population |
7876361, | Jul 26 2005 | Honeywell International Inc | Size calibration and mapping in overhead camera view |
7884858, | Mar 30 2001 | Intel Corporation | Object trackability via parametric camera tuning |
8019164, | Jan 29 2007 | HITACHI HIGH-TECH CORPORATION | Apparatus, method and program product for matching with a template |
8090147, | Dec 19 2006 | Fujitsu Ten Limited | Matching object and reference pattern images using search window size widths and heights to predict object collision |
8113151, | Mar 02 2010 | Biopar, LLC | System for uniquely identifying subjects from a target population |
8144207, | Nov 15 2006 | Nikon Corporation | Computer program product for photographic subject tracking, photographic subject tracking device, and camera |
8228347, | May 08 2006 | C R BARD, INC | User interface and methods for sonographic display device |
8254670, | Feb 25 2009 | Toyota Motor Corporation | Self-learning object detection and classification systems and methods |
8345923, | Feb 04 2000 | CheckVideo LLC | System for automated screening of security cameras |
8355048, | Mar 16 2007 | Nikon Corporation | Subject tracking computer program product, subject tracking device and camera |
8416987, | Dec 26 2008 | Canon Kabushiki Kaisha | Subject tracking apparatus and control method therefor, image capturing apparatus, and display apparatus |
8428357, | Dec 07 2010 | Movement Training Systems LLC | Systems and methods for performance training |
8432417, | May 08 2006 | C. R. Bard, Inc. | User interface and methods for sonographic display device |
8438997, | Nov 08 2002 | Biopar, LLC | System for uniquely identifying subjects from a target population |
8451338, | Mar 28 2008 | Massachusetts Institute of Technology | Method and apparatus for motion invariant imaging |
8457358, | May 30 2003 | Microsoft Technology Licensing, LLC | Head pose assessment methods and systems |
8509482, | Dec 21 2009 | Canon Kabushiki Kaisha | Subject tracking apparatus, subject region extraction apparatus, and control methods therefor |
8509543, | Jun 08 2009 | Nikon Corporation | Subject tracking device and camera |
8526672, | Jun 15 2009 | Canon Kabushiki Kaisha | Object detection apparatus and method thereof |
8571261, | Apr 22 2009 | CheckVideo LLC | System and method for motion detection in a surveillance video |
8587667, | Jul 09 2007 | International Business Machines Corporation | Beyond field-of-view tracked object positional indicators for television event directors and camera operators |
8594371, | Apr 08 2009 | Nikon Corporation | Subject tracking device and camera |
8614741, | Mar 31 2003 | CU CODING LIMITED | Method and apparatus for intelligent and automatic sensor control using multimedia database system |
8682034, | Feb 04 2000 | CheckVideo LLC | System for automated screening of security cameras |
8798387, | Nov 11 2010 | Panasonic Intellectual Property Corporation of America | Image processing device, image processing method, and program for image processing |
8826863, | Nov 08 2002 | Biopar, LLC | System for uniquely identifying subjects from a target population |
8873798, | Feb 05 2010 | Rochester Institute of Technology | Methods for tracking objects using random projections, distance learning and a hybrid template library and apparatuses thereof |
8937630, | May 08 2006 | C. R. Bard, Inc. | User interface and methods for sonographic display device |
9025824, | Dec 07 2010 | Movement Training Systems LLC | Systems and methods for evaluating physical performance |
9171195, | Jun 16 2014 | Amazon Technologies, Inc | Recognizing three-dimensional objects |
9230175, | Apr 22 2009 | CheckVideo LLC | System and method for motion detection in a surveillance video |
9295227, | Nov 08 2002 | Biopar, LLC | System for uniquely identifying subjects from a target population |
9693535, | Nov 08 2002 | Biopar, LLC | System for uniquely identifying subjects from a target population |
9712798, | Apr 01 2014 | Canon Kabushiki Kaisha | Imaging apparatus and image processing system processing image based on detection result of partial region |
9741129, | Apr 14 2008 | GRASS VALLEY CANADA | Technique for automatically tracking an object by a camera based on identification of an object |
9949720, | Oct 21 2011 | C R BARD, INC | Systems and methods for ultrasound-based medical device assessment |
RE45551, | Aug 31 2004 | PANASONIC INTELLECTUAL PROPERTY MANAGEMENT CO., LTD. | Surveillance recorder and its method |
RE47442, | Apr 26 2001 | LG Electronics Inc. | Method and apparatus for assisting data input to a portable information terminal |
Patent | Priority | Assignee | Title |
5554983, | Apr 24 1992 | Hitachi, Ltd. | Object recognition system and abnormality detection system using image processing |
5757287, | Apr 24 1992 | Hitachi, Ltd. | Object recognition system and abnormality detection system using image processing |
6075557, | Apr 17 1997 | Sharp Kabushiki Kaisha | Image tracking system and method and observer tracking autostereoscopic display |
JP1013821, |
Executed on | Assignor | Assignee | Conveyance | Frame | Reel | Doc |
Jun 01 2000 | ITO, WATARU | Hitachi Denshi Kabushiki Kaisha | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 010873 | /0057 | |
Jun 01 2000 | UEDA, HIROTADA | Hitachi Denshi Kabushiki Kaisha | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 010873 | /0057 | |
Jun 13 2000 | Hitachi Denshi Kabushiki Kaisha | (assignment on the face of the patent) | / |
Date | Maintenance Fee Events |
Aug 02 2007 | M1551: Payment of Maintenance Fee, 4th Year, Large Entity. |
Jul 21 2011 | M1552: Payment of Maintenance Fee, 8th Year, Large Entity. |
Aug 30 2011 | ASPN: Payor Number Assigned. |
Sep 11 2015 | REM: Maintenance Fee Reminder Mailed. |
Feb 03 2016 | EXP: Patent Expired for Failure to Pay Maintenance Fees. |
Date | Maintenance Schedule |
Feb 03 2007 | 4 years fee payment window open |
Aug 03 2007 | 6 months grace period start (w surcharge) |
Feb 03 2008 | patent expiry (for year 4) |
Feb 03 2010 | 2 years to revive unintentionally abandoned end. (for year 4) |
Feb 03 2011 | 8 years fee payment window open |
Aug 03 2011 | 6 months grace period start (w surcharge) |
Feb 03 2012 | patent expiry (for year 8) |
Feb 03 2014 | 2 years to revive unintentionally abandoned end. (for year 8) |
Feb 03 2015 | 12 years fee payment window open |
Aug 03 2015 | 6 months grace period start (w surcharge) |
Feb 03 2016 | patent expiry (for year 12) |
Feb 03 2018 | 2 years to revive unintentionally abandoned end. (for year 12) |