A traffic lane marking line recognition system for vehicle including a traffic lane marking line recognizing device configured to recognize at least a traffic lane marking line in broken (e.g., white) line on a road surface in a camera photographed image, and an image compositing device which composites images photographed at different time points to elongate the traffic lane marking line in the photographed image, wherein the images are composited in a processing stage in such a manner that no change will be imparted to at least a shape of the traffic lane marking line in the photographed image in a traffic lane marking line recognition processing, specifically in a processing stage prior to edge detection. With this, a distant traffic lane marking line is prevented from being chipped away when compositing the photographed images, thereby enabling accurate and unerring recognition of the traffic lane marking line.
|
4. A vehicle comprising:
image photographing means for photographing a range including a road surface in a direction of travel of the vehicle;
traffic lane marking line recognizing means configured to be capable of recognizing at least a traffic lane marking line in broken line on the surface of road in a photographed image photographed by the image photographing means by detecting edges in the image and by conducting a hough transformation on the detected edges; and
image compositing means for compositing a plurality of images photographed at different time points by the image photographing means to elongate the traffic lane marking line in the photographed images;
wherein the image compositing means selects at least one of the plurality of images and composites the plurality of images in a processing stage prior to the edge detection in a traffic lane marking line recognition processing of the traffic lane marking line recognizing means.
3. A vehicle comprising:
image photographing means for photographing a range including a road surface in a direction of travel of the vehicle;
traffic lane marking line recognizing means configured to be capable of recognizing at least a traffic lane marking line in broken line on the road surface in a photographed image photographed by the image photographing means; and
image compositing means for compositing a plurality of images photographed at different time points by the image photographing means to elongate the traffic lane marking line in the photographed images;
wherein the image compositing means selects at least one of the plurality of images and composites the plurality of images in a processing stage in such a manner that no change will be imparted to at least a shape of the traffic lane marking line in the photographed image in a traffic lane marking line recognition processing of the traffic lane marking line recognizing means.
2. A system for recognizing a traffic lane marking line for a vehicle comprising image photographing means for photographing a range including a road surface in a direction of travel of the vehicle; traffic lane marking line recognizing means configured to be capable of recognizing at least a traffic lane marking line in broken line on the road surface in a photographed image photographed by the image photographing means by detecting edges in the image and by conducting a hough transformation on the detected edges; and image compositing means for compositing a plurality of images photographed at different time points by the image photographing means to elongate the traffic lane marking line in the photographed images;
wherein
the image compositing means selects at least one of the plurality of images and composites the plurality of images in a processing stage prior to the edge detection in a traffic lane marking line recognition processing of the traffic lane marking line recognizing means.
1. A system for recognizing a traffic lane marking line for a vehicle comprising: image photographing means for photographing a range including a road surface in a direction of travel of the vehicle; traffic lane marking line recognizing means configured to be capable of recognizing at least a traffic lane marking line in broken line on the road surface in a photographed image photographed by the image photographing means; and image compositing means for compositing a plurality of images photographed at different time points by the image photographing means to elongate the traffic lane marking line in the photographed images;
wherein
the image compositing means selects at least one of the plurality of images and composites the plurality of images in a processing stage in such a manner that no change will be imparted to at least a shape of the traffic lane marking line in the photographed image in a traffic lane marking line recognition processing of the traffic lane marking line recognizing means.
|
The present application is a U.S. National phase of, and claims priority based on PCT/JP2005/008937 (published as WO 2005/111937 A1), which, in turn, claims priority from Japanese patent application 2004-148577, filed May 19, 2004. The entire disclosure of each of the referenced priority documents is incorporated herein by reference.
This invention relates to a traffic lane marking line recognition system for vehicle.
Road surfaces (roads) on which vehicles drive are marked or installed with various road markings such as white lines, yellow lines and cat's eyes constituting traffic lane marking lines. Conventionally the recognition of white lines and other such traffic lane marking lines has, as taught by Patent Reference 1, been done by subjecting an image photographed or taken by image photographing means such as a CCD camera and an image processing ECU to differentiation and binarization processing to detect edges in the image and subjecting point sequences of the detected edge (white line candidate point sequences) to Hough transformation to extract approximated linear components.
Further, when, for example, a white line marked on the road surface is itself physically worn through or partially chipped away, or the white line is indistinct in the photographed image owing to low contrast such as may occur during night driving, the white line cannot be accurately recognized, and, as set out in Patent Reference 2, a technique has been proposed for, in such a case, forming a window in the photographed image, judging the degree of white line wear-off from the density of the white line candidate points in the window, and when wear-off is found, superimposing/compositing a white line candidate point sequences extracted from a photographed image taken a given time earlier on/with the white line candidate point sequences extracted from the current white line image, determining a straight line approximating the composited white line candidate point sequences, and recognizing it as the white line.
As mentioned above, in the technique set out in Patent Reference 2, a chipped-away portion or an indistinct portion is covered or complemented by compositing (superimposing) the white line candidate point sequences extracted from the current white line image with a white line candidate point sequences extracted from a photographed image taken a given time earlier to elongate the white line candidate point sequences, determining a straight line approximating the elongated white line candidate point sequences, and recognizing it as the white line. However, since the superimposing is conducted on the edge image after edge detection, the accuracy of traffic lane marking line (white line) recognition is not necessarily satisfactory.
That is, since the size of a traffic lane marking line distant from the subject vehicle in the photographed image is smaller than that of a closer traffic lane marking line, a case arises in which a distant traffic lane marking line can not be recognized as edges. As a result, the recognition accuracy is not necessarily satisfactory, since point sequences corresponding to a distant traffic lane marking line are chipped away and are not elongated in the composited image.
Therefore, an object of this invention is to overcome the aforesaid drawbacks and provide a traffic lane marking line recognition system for vehicle configured such that point sequences corresponding to a distant traffic lane marking line are prevented from being chipped away in a composited image when compositing photographed images, thereby enabling accurate and unerring recognition of the traffic lane marking line.
In order to achieve the object, according to a first aspect of the invention there is provided a system for recognizing a traffic lane marking line for a vehicle having image photographing means for photographing a range including a road surface in a direction of travel of the vehicle, traffic lane marking line recognizing means configured to be capable of recognizing at least a traffic lane marking line in broken line on the road surface in a photographed image photographed by the image photographing means, and image compositing means for compositing a plurality of images photographed at different time points by the image photographing means to elongate the traffic lane marking line in the photographed images; wherein the image compositing means composites the plurality of images in a processing stage in such a manner that no change will be imparted to at least a shape of the traffic lane marking line in the photographed image in a traffic lane marking line recognition processing of the traffic lane marking line recognizing means.
Further, according to a second aspect of the invention there is provided a system for recognizing a traffic lane marking line for a vehicle having image photographing means for photographing a range including a road surface in a direction of travel of the vehicle, traffic lane marking line recognizing means configured to be capable of recognizing at least a traffic lane marking line in broken line on the road surface in a photographed image photographed by the image photographing means by detecting edges in the image and by conducting a Hough transformation on the detected edges, and image compositing means for compositing a plurality of images photographed at different time points by the image photographing means to elongate the traffic lane marking line in the photographed images; wherein the image compositing means composites the plurality of images in a processing stage prior to the edge detection in a traffic lane marking line recognition processing of the traffic lane marking line recognizing means.
Since the traffic lane marking line recognition system for vehicle according to the first aspect is configured such that the plurality of images are composited in a processing stage in such a manner that no change will be imparted to at least a shape of the traffic lane marking line in the photographed image in a traffic lane marking line recognition processing, it becomes possible to detect a distant traffic lane marking line unerringly, and to elongate the traffic lane marking line in appearance, thereby improving recognition accuracy.
Further, segments of the traffic lane marking line are sometimes worn through or partially chipped away, and in some cases, segments are temporality painted short in length as a stopgap measure during construction work or for other such reason. In such cases, however, it becomes possible to recognize the traffic lane marking line accurately. In addition, the direction of the traffic lane marking line can be recognized more accurately by the elongated traffic lane marking line, than by immediately recognizing broken lines (dotted lines) painted on roads, thereby enabling more accurate recognition of the traffic lane marking line.
Since the traffic lane marking line recognition system for vehicle according to the second aspect is configured such that the plurality of images are composited in a processing stage prior to the edge detection in a traffic lane marking line recognition processing, it also becomes possible to detect a distant traffic lane marking line unerringly, and to elongate the traffic lane marking line in appearance, thereby improving recognition accuracy.
It should be noted that, the phrase “direction of travel of the vehicle” recited in claims is used to mean not only the forward direction of the vehicle when traveling forward, but also to recognize the road surface behind the vehicle photographed by an image photographing means for photographing a range including the road surface behind the vehicle when the vehicle is traveling forward. Thus, the phrase “direction of travel of the vehicle” is used to mean the fore-aft direction of the vehicle.
The best mode for implementing the traffic lane marking line recognition system for a vehicle according to this invention will be explained with reference to the attached drawings in the following.
The symbol 10 in the drawing designates a camera equipped with an image pickup device such as a CCD, C-MOS or the like and mounted inside a passenger compartment 12a to face in the direction of travel of a vehicle 12, which camera 10 photographs a region including the road surface in the direction of travel of the vehicle 12. An image processing ECU (Electronic Control Unit) is accommodated in a case 10a of the camera 10. The image processing ECU inputs images (pictures) outputted from the camera 10 representing information on the lane forward of the vehicle and is equipped with a hardware-implemented image processing IC (Integrated Circuit) that performs image processing explained later. The camera 10 and the image processing ECU correspond to the aforesaid image photographing means.
Note that in this embodiment “traffic lane marking line” means a road marking for separating vehicle passageways (lanes), such as a solid or broken white or yellow line that is applied to the surface of the road by painting, or cat's eyes or the like installed at intervals on the road surface. A vehicle passageway partitioned by a traffic lane marking line or lines is called a “lane.”
A steering wheel 16 installed at a driver's seat in the passenger compartment 12a of the vehicle 12 is connected to a rack shaft 20 through a rack-and-pinion type steering gear and the rack shaft 20 is connected to driven wheels 24 through tie rods 22. An electric power steering (EPS) mechanism 30 including an electric motor 26 is disposed on the rack shaft 20 and the rack shaft 20 is reciprocated by rotation of the electric motor 26.
The driven wheels 24 and the free wheels (not shown) each has a wheel-speed sensor (only two shown) 32 in the vicinity thereof that produces an output or signal once per predetermined angle of rotation, i.e., output indicative of the travel speed (vehicle speed) of the vehicle 12. A yaw rate sensor 34 is disposed at the middle of the vehicle 12 (near the rear axle) and produces an output or signal corresponding to the yaw rate (angular velocity) about the vertical axis (gravity axis) at the center of gravity of the vehicle 12. Further, a steering angle sensor 36 is provided near the steering wheel 16 and produces an output or signal corresponding to the amount of rotation of the steering wheel 16 manipulated by the driver, i.e., the steering angle.
A control ECU (Electronic Control Unit) 40 similarly equipped with a microcomputer is provided at a suitable location of the vehicle 12. The control ECU 40 inputs the outputs of the image processing ECU and the aforesaid sensors, calculates a steering force required for, inter alia, lane keep assist control for driving along the traffic lane or lane deviation prevention control for preventing unintentional deviation from the traffic lane, and converts it into command values to output. The control ECU 40 is connected to an EPSECU 42 that controls the operation of the EPS 30. The EPSECU 42 is also equipped with a microcomputer, exchanges data with the control ECU 40, and operates the electric motor 26 based on the command value outputted from the control ECU 40.
As illustrated, the control ECU 40 inputs the output of the image photographing means composed of the camera 10 and the aforesaid image processing ECU (designated by symbol 44) and the output of the yaw rate sensor 34 etc., and outputs lane keep assist control and other command values to the EPSECU 42 by communication to operate the electric motor 26. Further, although omitted in
This will be explained in the following: In step S10 (“step” is omitted hereinafter), the image pickup device of the camera 10 produces an image signal. The image pickup device of the camera 10 is provided with an imaging region or range made up of n rows vertical×m columns horizontal pixels. On the factory production line, for example, the camera 10 has its optical axis aligned in a predetermined direction including the traffic lane ahead.
Next, in S12, frame composition processing is conducted, i.e., the image signal outputted from the image pickup device is inputted and composited. Specifically, the image signal inputted in the current processing cycle and an image signal inputted and stored in the memory of the image processing ECU 44 in an earlier processing cycle, in other words, the current image (frame) and a past image (frame) are composited. Note that since features that characterize this embodiment reside in the frame composition, this will be explained in detail below. Processing from S12 to S16 is performed by the hardware-implemented image processing IC in the image processing ECU 44.
Next, in S14, edge detection processing including well-known differentiation processing and ensuing binarization processing is conducted to produce from the inputted image an edge image like that shown in
Next, in S18, as shown in
Next, in S20, when a plurality of candidates are present as lane candidates (e.g., when, as shown in
Next, in S22, vehicle control is implemented. That is, based on the estimated lane location, the command value for vehicle control, such as the aforesaid lane keep assist control or lane deviation prevention control, is outputted to the EPSECU 42 to control the operation of the electric motor 26. Note that the processing of S12 to S16 is performed by the aforesaid hardware-implemented image processing IC in the image processing ECU 44, the processing of S18 and S20 by the image processing ECU 44, and the processing from S22 onward by the control ECU 40.
The characterizing features of this embodiment reside in the various frame composition shown in
That is, the image to be composited with the photographed image inputted at the present time (current processing cycle) t (shown in
In the ensuing S102, brightness correction of the selected earlier or past frame is conducted. That is, the photographed image inputted at the present time and the image photographed at an earlier time point and selected in accordance with the vehicle speed in S100 were photographed at different time points, meaning that the brightness of the photographed object (road surface) may differ owing to the effect of shading and the like and that if composition should be conducted in a state where the brightness of the road surface portion of one image is brighter than that of the white lines or other such traffic lane marking lines of the other image, there would be a risk of the white lines being buried in the road surface in the composited image, and therefore, in order to prevent this, processing is performed to match the brightness of the image photographed at the earlier time point with the brightness of the image photographed at the present time point such that the brightness of the images at the present and earlier time points are made equal. Note that the details thereof will be explained below.
In S104, the composition start point is determined. That is, it may happen that the posture of the vehicle 12 differ from the initial camera parameter conditions (at the time of the optical axis alignment of the camera 10) owing to vehicle loading or some other cause; in other words, a posture change may occur in the static pitching direction. As this posture change may vertically shift the photographic range of the camera 10, the composition start (reference) point is learned and corrected (determined) to take the amount of this shift into account at the time of image composition. Since the region above the road surface is ordinarily not needed for traffic lane marking line recognition, this processing is performed by bringing the positions of the horizons at the present and earlier time points into registration.
Next, in S106, horizontal pixel correction is conducted. That is, depending on the image-photographing time points, the posture and angle of the vehicle 12 relative to a lane may differ between the present and earlier time points, so that cases may arise in which offset occurs by the amount of such position and/or angle. In such a case, if the images are composited as they are, the horizontal position and angular direction of the traffic lane marking line will of course also shift. In order to prevent this, the horizontal position and angular direction of traffic lane marking line in the image photographed at the earlier time point is corrected by an amount proportional to the change in vehicle posture between the present and earlier time points. Note that the details thereof will be explained below.
Next, in S108, vertical row correction is conducted. That is, it is conceivable that the pitching condition of the vehicle 12 may differ between the present and earlier time points, so the dynamic pitch angle variation between the two time points is determined from the change in the position of the horizon found from the images, and the vertical (up/down) direction in the image is corrected to obtain a corrected image like that shown in
Next, advancing to S110, the present time point's and corrected earlier time point's photographed images are composited to obtain an image wherein, as shown in
The features characterizing this embodiment will be successively explained in the following; the first characterizing feature is, as has been explained regarding S100 in
This will be explained in the following: In the prior art of Patent Reference 2, the earlier image to be composited with the present image is an image photographed a given time earlier. However, the travel distance of the vehicle 12 differs with the vehicle speed (driving speed), which of course means that the photographic range of the camera (image pickup device) 10 similarly moves in the forward direction, so that the positions of traffic lane marking lines in the image also move with travel of the vehicle 12; in other words, the higher the vehicle speed, the greater is the movement of the photographed traffic lane marking line positions. Therefore, the amount and rate by which the traffic lane marking lines are elongated in the composited image differs depending on after how much of a time interval the earlier image is selected and composited with the present image.
Regarding traffic lane marking lines, there are known ones of broken line configuration composed of periodically repeated white line segments (colored segments) and blank (uncolored segments of asphalt or the like). From the aspect of durability, the white line segments of such a traffic lane marking line sometimes wear through or partially chip away, and in some cases, the white line segments are temporarily painted short in length as a stopgap measure during construction work or for other such reason.
Further, while only natural, recognition of traffic lane marking lines by image recognition amounts to driving lane (subject vehicle lane) recognition, and since the direction of the traffic lane marking line can be recognized more accurately in proportion as the length of the white line segments of the traffic lane marking lines is longer, composition is preferably done so as to elongate the white line segments in appearance, irrespective of presence/absence of the aforesaid white line segment wear and chipping. Taking this point into account, in this embodiment the image concerned is selected from among images (past frames) photographed earlier by a time interval determined in accordance with the vehicle speed (driving speed). Note that this is not limited to white lines but is also the same for yellow lines.
In view of what the inventors learned empirically, it is possible to recognize traffic lane marking lines with good accuracy and recognize the subject vehicle's traffic lane unerringly provided that the length of (proportion accounted for by) the white line segments relative to the total length of the white line segments and blanks of the traffic lane marking lines is equal to or greater than around 30%. The length of an elongated white line segment varies in accordance with the vehicle speed as mentioned earlier, and, therefore, the determination as to how many frames earlier in the image-photographing cycle should be used is made in accordance with the controlled speed range of the vehicle control systems that utilize recognized traffic lane marking line data, including, inter alia, the lane keep assist control or lane deviation prevention control discussed at the beginning of the specification.
Specifically, the image-photographing cycle is 33.3 milliseconds, so that if the vehicle speed (driving speed) is 60 km/h or faster, for example, it suffices to use the image three frames earlier to secure the required amount of white line segment elongation. Even at the image two frames earlier, a vehicle speed of 90 km/h or faster suffices, although the amount of white line segment elongation will be somewhat insufficient, and even in the vehicle speed range of 60-90 km/h, the amount of white line segment elongation decreases but the length elongated is 1 meter or greater and the proportion of the total length accounted for by the white line segments also increases to about 30%, so that the traffic lane marking lines can be recognized with good accuracy to enable the traffic lane of the subject vehicle to be reliably recognized.
Therefore, at least one photographed image taken in the cycle matching the time interval determined in accordance with the detected vehicle speed (driving speed) is selected as the photographed image to be composited with a given (present) photographed image. Note that it is also possible to switch the image used in the composition to two frames earlier or three frames earlier with the vehicle speed value imparted with hysteresis at, for example, 105 km/h or 110 km/h, or, if the vehicle speed is 60 km/h or greater, to use only the image three frames earlier, because when using the image three frames earlier, the length of the white line segments relative to the total length of the white line segments and blanks of the traffic lane marking line becomes 30% or greater.
Thus, in this embodiment the time of taking a given photographed image among the photographed images (specifically, the present image, still more specifically, the image photographed in the current traffic lane marking line recognition processing cycle) is defined or determined as the reference, and with respect to the given photographed image, there is selected as the photographed image to be composited with the photographed image at least one photographed image taken earlier by a time interval determined in accordance with the detected vehicle speed (driving speed) (more specifically, at least one photographed image taken in the cycle corresponding to the time interval determined in accordance with the detected vehicle speed (driving speed)), so that, irrespective of how high or low the vehicle speed is, it becomes possible to optimally determine the amount of apparent elongation of the white line segments of the traffic lane marking lines or the elongation ratio thereof and to enhance the recognition accuracy of the traffic lane marking lines to enable unerring recognition of the traffic lane of the subject vehicle.
Further, a configuration is adopted such that there is selected at least one photographed image taken earlier by a time interval that is longer in proportion as the detected vehicle speed is lower. That is, since the travel distance of the vehicle 12 becomes shorter with decreasing vehicle speed, an image photographed or inputted and memorized at a time point farther in the past is selected when the vehicle speed is low such that the amount of elongation or the elongation rate of the length of the traffic lane marking lines in the image after composition will not be insufficient. On the other hand, since the travel distance of the vehicle 12 becomes longer with increasing vehicle speed, an image photographed or inputted and memorized at an earlier time point near the present time point is selected.
Further, there is selected at least one photographed image taken earlier by a time interval such that the length of the colored segments relative to the total length of the white line segments (colored segments) and blanks (uncolored segments) of the traffic lane marking lines is equal to or greater than 30%, so that even when the traffic lane marking lines are worn through or chipped away, or short in length because of construction work or the like, it becomes possible to optimally determine the amount of apparent elongation of the white line segments of the traffic lane marking lines or the elongation ratio thereof and to enhance the recognition accuracy of the traffic lane marking lines to enable unerring recognition of the traffic lane of the subject vehicle.
Further, the image on which the composition is based is made an image photographed in the current traffic lane marking line recognition processing cycle, so that image composition can be performed based on up-to-date information. Further, there is selected at least one photographed image taken in the cycle corresponding to the time interval determined in accordance with the detected vehicle speed, so that image composition can similarly be performed based on up-to-date information.
In this embodiment, the second characterizing feature is, as is explained regarding S102 in
This will be explained: In order to recognize the white line segments of the traffic lane marking lines with good accuracy, ordinarily, as shown in
Note that in this specification“brightness” is used in the meaning of including all photographic conditions and the like, including brilliance, luminosity, density, and shutter speed, iris and image signal output amplifier gain and the like adjusted in accordance with the detected image brightness.
In normal driving conditions, the brightness of the road surface changes from moment to moment because of the effect of various environmental changes, such as the shadows of buildings etc., wetting of the road surface, and strong sunshine. The aforesaid adjustment of photographic conditions is conducted for clearly detecting the white line segments despite these brightness differences.
Consideration is now given to the compositing of the present and earlier images in the case where the road surface brightness differs at the image-photographing time points. Since the white line segments are brighter than the remaining road surface (asphalt, concrete) in the same image, at the time of compositing, the brighter at the same point in the image (same location) is assumed to be a white line segment and composition is conducted by selecting it.
As mentioned above, the shutter speed and the like are adjusted based on the detected brightness to make the brightness between the photographed images the same, but when variations of light and shade actually occur on the road surface, the shutter speed, iris and amplifier gain are not changed in one stroke but, for preventing hunting and other reasons, are gradually changed, with the result that brightness does not become the same between the photographed images but becomes somewhat different, so that cases arise in which the white line segments in one photographed image become darker than the road surface (asphalt portions) in the other photographed image.
A in
At this time point, there is no change between A and A′, and if at this stage, as indicated by the broken line with accompanying remark “W/O CORRECTION” in
In view of this point, in this embodiment the brightness of the photographed image is detected from the brightness or the like of a predetermined road surface region in the photographic range (hatched region in
This will be explained in the following: First, in S200, S202, the photographic conditions at the time of image-photographing, i.e., the photographic conditions comprising the shutter speeds S0, S1, irises I0, I1 and amplifier gains G0, G1 indicative of the image signal amplification factors are memorized. The detected photographed image brightness, specifically the brightness of a predetermined region of the road surface (hatched region in
Note that, as is well known, image brightness can be adjusted also by the iris adjustment because it varies the amount of light taken in, as shown in
Next, in S204, the brightness ratio between the present time point and earlier time point images is calculated based on the memorized photographic condition using the equation shown there, and in S206, the brightness of the earlier time point image is multiplied by the calculated brightness ratio to carry out correction for making the brightness of the earlier time point image the same as the brightness of the present time point image.
More specifically, the brightness ratios of a plurality of photographed images are detected based on at least one of the shutter speed and iris and the brightness is corrected based on at least one of the ratio and difference of one or both of the shutter speed and iris at the image-photographing time point of each of the photographed images, more specifically, based on both ratios thereof. Further, the brightness ratios of the multiple photographed images are also detected based on the amplification factors.
Regarding the brightness, note that the brightness of the predetermined region of the road surface indicated by the hatched region in
As a result, in the processing of S110 of
When the brightness is corrected by this processing, then, as shown by graph b on the right side of
As set out in the foregoing, this embodiment is configured to detect the brightnesses of the photographed images photographed by the image photographing means, to perform correction to make the brightness in multiple photographed images the same, taking one image among the multiple photographed images as the reference, and, after the brightness has been corrected, to composite the multiple images, whereby, as a result of the brightness of the two images being the same, the white line segment in one of the images at the earlier time point and present time point does not become buried in the road surface of the other image and the white line segment can be favorably elongated. Therefore, the traffic lane marking line recognition accuracy can be enhanced and the traffic lane of the subject vehicle can be reliably recognized.
Further, since the image constituting the reference for composition is the image of the current traffic lane marking line recognition processing cycle, the images can be composited while making brightness equal based on up-to-date information to achieve improved traffic lane marking line recognition accuracy, thereby enabling reliable recognition of the subject vehicle's traffic lane. Further, by detecting the brightness of a predetermined region of the road surface, the brightness detection can be performed using a common yardstick to improve the traffic lane marking line recognition accuracy.
Further, brightness detection and correction are simple because the brightness detection and correction are performed based on at least one of the ratio and difference of one or both of the shutter speed and iris at the image-photographing time point of each of the multiple photographed images and on the amplifier gain (amplification factor). Since the correction is made so as to make the brightnesses equal at the same location in multiple images, the brightness can be corrected with still better accuracy.
Further, a configuration is adopted wherein multiple photographed images are composited by selecting from among the pixels constituting the same area of the multiple images that is brighter in brightness, and therefore, the traffic lane marking line can be still more accurately recognized from the composited image, whereby the traffic lane of the subject vehicle can be recognized still more reliably.
In this embodiment, the third characterizing feature is, as is explained regarding S106 and S108 in
This will be explained: The posture (angle) of the vehicle 12 relative to the traffic lane marking line and the distance to the traffic lane marking line in the horizontal direction sometime differ between the present time point and earlier time point, so that the angular direction of the vehicle 12 with respect to the horizontal direction and traffic lane marking line is shifted between the images photographed at the two time points, which of course means that if the images are composited as they are, an offset condition will occur in the horizontal position and angle of the traffic lane marking line.
In order to prevent this, it is necessary to correct the horizontal position and angular direction of the traffic lane marking line in the image at the earlier time point by an amount corresponding to the change in the vehicle posture between the present and earlier time points, and further, since it is conceivable that the pitching condition of the vehicle 12 may differ between the present time point and earlier time point, it is necessary to determine the pitch angle change between the two time points from the change in the horizon position and the like obtained from the images and correct vertical direction position in the images.
On this point, the prior art taught by Patent reference 2 merely detects the amount of vehicle yaw rate change, i.e., presence/absence of vehicle rotation (turning) movement, and imparts correction to the image in correspondence to the amount of rotation movement, if any, so that no correction of the horizontal position is carried out and, therefore, there is a disadvantage in that the traffic lane marking line is not suitably elongated in the composited image. Further, the technology of Patent reference 2 is disadvantageous in that it does not take pitch change into consideration.
In this embodiment, the traffic lane marking lines on the left and right of the vehicle 12 in
In order to composite the image photographed and memorized at exposure time t−1, which is the earlier time point, with the image photographed at the present time point, i.e., the exposure time t0, taking the traffic lane marking lines (lane), point sequences and line segments obtained at exposure time t−2 as the references, the distances L0, L1 and angles θ0, θ1 are determined, the deviations or changes between L0 and L1 and between θ0 and θ1 are determined, and once the changes in the photographed images caused by the changes in the relative position and angle of the vehicle 12 relative to the traffic lane marking lines between times t0 and t−1 have been corrected, the images at the two time points are composited.
This figure shows the processing for determining the lateral position deviation between exposure times t0 and t−1 (amount of past (time t−1) lateral vehicle movement relative to present time t0) ΔL and the angle deviation relative to the traffic lane marking lines or the reference line therebetween (past (time t−1) angle deviation relative to present time t0) Δθ, and then determining the amount of horizontal (lateral) pixel position correction in the images from the determined deviations.
Although the details will be explained below, in S300, the deviations ΔL, Δθ are determined, and in S302 the pixel position correction amount on the image plane at compositing is determined from the deviations ΔL, Δθ, and correlation of the camera coordinate system (U, V) and the actual plane coordinate system (X, Y) is performed as pre-compositing hardware-based processing. This processing of S302 will be explained in detail below.
First, in S400, the pitch direction change Δθpit between exposure times t0 and t−1 is determined, and in S402 the amount of vertical row position correction in the image plane at the time of composition is determined based on the change thereof, and correlation of the camera coordinate system (U, V) and the actual plane coordinate system (X, Y) is also performed as pre-compositing hardware-based processing.
Explanation will be made with reference to
The position and direction of the vehicle 12 at exposure time t−1 are converted into values on the time t−2-based X, Y coordinate system using the estimated trajectory (locus) of the vehicle between exposure times t0 and t−1 determined from the outputs of the wheel-speed sensor 32 and the yaw rate sensor 34. At this time, the coordinates of the position of the vehicle 12 on the detection result coordinate system X, Y at exposure time t−1 are (x, y). In (1) of S500 in
In (2) of S500, as shown in
Similarly, in (3) of S500, the distance L0 between the vehicle position at exposure time t0 and the straight line connecting the point sequences is determined taking exposure time t−2 as the reference. Note that in the interest of simplicity, the vehicle position at exposure time t0, point sequence Pk+2, and so on are omitted in
Then, in (4) of S500, the lateral direction position deviation ΔL between exposure times t0 and t−1 is determined by the subtraction as L0−L−1. The sign of ΔL of course becomes positive or negative depending on the vehicle position at the exposure time points.
Next, the angle deviation Δθ of the vehicle relative to the reference line between exposure times t0 and t−1 i.e., the point sequences and the straight line connecting them, is determined.
First, in (1) of S502 in
Further, when the straight lines, determined at exposure time t−2, connecting the point sequences Pi, Pi+1, Pi+2 where the vehicle is positioned at two times, i.e., exposure times t0 and t−1, differ so that the straight lines are not a single straight line, it is necessary to determine the angle deviation θR between the straight lines, as shown in
In (2) of S502 in
y=x·tan Δθ+ΔL.
In S600 of
Next, in S602, the line positions V5 m and V30 m in the image plane corresponding to the distances X=5 m and X=30 m in the actual plane are determined. The image plane coordinate system is shown in
Next, in S604, the lateral pixel positions U5 m and U30 m in the image plane that correspond to the lateral corrective distances Y5 m and Y30 m obtained in S600 are determined.
Next, in S606, the straight line passing through the two points of the image plane coordinates (U5 m, V5 m) and (U30 m, V30 m) obtained in S606 is determined. In this embodiment, the composition positions at the time of compositing images photographed at different time points are determined such that six points x0, x1, . . . , x6 on the V axis of the image plane fall on this straight line. That is, in order to correct the position and angle change relative to the reference line in the actual plane at time points t−1, t0, the straight line indicating the position and angle of the vehicle at time t−1 in the actual plane (coincident with the X axis as the reference) is made equal to the straight line indicating the position and angle at time t0 shown in
When vehicle lateral direction position and angle correction relative to the traffic lane marking lines is imparted, as in
In this embodiment, the fore-aft direction angles θ of the vehicle 12 relative to the traffic lane marking lines and the distances L between the traffic lane marking lines and the vehicle 12 in multiple photographed images are determined, the angle and position deviations (Δθ, ΔL) between the respective photographed images are determined, and the multiple images are composited after having been corrected such that the angle and distance of the vehicle relative to the traffic lane marking line are equal to each other in the images based on the determined angle and position deviations, whereby the traffic lane marking lines can be reliably elongated in appearance in the composited image.
Further, the pitch angles of the vehicle 12 at image-photographing time points t−1 and t0 of multiple photographed images are detected to determine the pitch angle change (Δθpit) between the photographed images, and the multiple images are composited after being corrected such that the pitch angles are equal to each other in the images based on the determined pitch angle change, whereby the traffic lane marking lines can be still more reliably elongated in appearance.
Note that, instead of the foregoing, it is acceptable to composite by adjusting the compositing positions between the multiple images based on the ascertained angle and position deviations to the condition of being photographed at the same vehicle angle and position relative to the traffic lane marking lines. That is, rather than move the images themselves as set out in the foregoing, it is acceptable to adjust the compositing positions (superimposition positions).
To explain with reference to
In this embodiment, the fourth characterizing feature is that in the composition processing of S110 in
This will be explained:
In the prior art processing of
Next position correction processing is performed based on the vehicle posture change to obtain the position-corrected image shown in image a′ in
In contrast to this, as shown in
Thus, in this embodiment, the configuration is adopted such that the photographed images are composited in a processing stage in such a manner that no change will be imparted to at least the shape of the traffic lane marking lines in the photographed image in the traffic lane marking line recognition processing, more specifically, before edge detection, i.e., such that the present time point and earlier time point images are composited at the original image stage prior to edge detection processing, so that distant traffic lane marking lines can be detected, thereby making it possible to extend the traffic lane marking lines in appearance and improve recognition accuracy.
Note that although it has been defined as being before edge detection processing, it suffices for the shape and profile of the traffic lane marking lines in the photographed original images to be in a retained condition and since, as has been mentioned above, edge detection processing is ordinarily constituted of differentiation processing, binarization processing and edge detection, the effect of the present invention can be obtained if compositing is done before binarization processing even if not before edge detection processing, i.e., before differentiation processing.
Note that although this invention has been explained taking traffic lane marking line recognition by edge detection processing and Hough transformation as an example, this invention is not limited thereto and by elongating the traffic lane marking line segments with respect to another method, e.g., a method using pattern matching such as shown in
That is, a method is known wherein the traffic lane marking lines are established beforehand by defining shape patterns in an image photographed by an on-board camera as multiple templates such as shown in
Further, in this embodiment, although the number of images to be composited in the processing of S100 is two, it can as necessary be a larger plural number. In addition, in this embodiment, selection is made in accordance with vehicle speed from among images periodically photographed, but it also acceptable, for example, to provide two or more image photographing means and select in accordance with the vehicle speed from among images not photographed in the same common cycle.
Further, in the correction of S108, it is acceptable to make the correction by, instead of using the change in the horizon obtained from the images, installing a pitch angle sensor at an appropriate location of the vehicle 12 and determining dynamic offset from the output thereof.
Further, although a configuration is adopted whereby the camera 10 photographs a range including the road surface ahead of the vehicle, it is of course acceptable to adopt a configuration provided with an image photographing means (camera) for photographing a range including the road surface behind the vehicle when the vehicle 12 is traveling forward, wherein the traffic lane marking lines behind the vehicle are recognized from the photographed images taken by the image photographing means. Therefore, in this specification “direction of travel of the vehicle” is used to mean the fore-aft direction of the vehicle.
As stated above, the embodiment is thus configured to provide a system for recognizing a traffic lane marking line for a vehicle 12 including image photographing means (camera 10, image processing ECU 44) for photographing a range including a road surface in a direction of travel of the vehicle, traffic lane marking line recognizing means (image processing ECU 44, S12 to S16) configured to be capable of recognizing at least a traffic lane marking line in broken line on the road surface in a photographed image photographed by the image photographing means, and image compositing means (image processing ECU 44, S12) for compositing a plurality of images photographed at different time points by the image photographing means to elongate the traffic lane marking line in the photographed images; wherein the image compositing means composites the plurality of images in a processing stage in such a manner that no change will be imparted to at least a shape of the traffic lane marking line in the photographed image in a traffic lane marking line recognition processing of the traffic lane marking line recognizing means (S12).
Further, the embodiment is configured to provide a system for recognizing a traffic lane marking line for a vehicle 12 including image photographing means (camera 10, image processing ECU 44) for photographing a range including a road surface in a direction of travel of the vehicle, traffic lane marking line recognizing means (image processing ECU 44, S12 to S16) configured to be capable of recognizing at least a traffic lane marking line in broken line on the road surface in a photographed image photographed by the image photographing means by detecting edges in the image and by conducting a Hough transformation on the detected edges, and image compositing means (image processing ECU 44, S12) for compositing a plurality of images photographed at different time points by the image photographing means to elongate the traffic lane marking line in the photographed images; wherein the image compositing means composites the plurality of images in a processing stage prior to the edge detection (S14) in a traffic lane marking line recognition processing of the traffic lane marking line recognizing means.
According to this invention, since it is configured such that the plurality of images are composited in a processing stage in such a manner that no change will be imparted to at least a shape of the traffic lane marking line in the photographed image in a traffic lane marking line recognition processing, it becomes possible to detect a distant traffic lane marking line unerringly, and to elongate the traffic lane marking line in appearance, thereby improving recognition accuracy. Further, segments of the traffic lane marking line are sometimes worn through or partially chipped away, and in some cases, segments are temporality painted short in length as a stopgap measure during construction work or for other such reason. In such cases, however, it becomes possible to recognize the traffic lane marking line accurately. In addition, the direction of the traffic lane marking line can be recognized more accurately by the elongated traffic lane marking line, than by immediately recognizing broken lines (dotted lines) painted on roads, thereby enabling to provide a traffic lane marking line recognition system for vehicle that can recognize the traffic lane marking line more accurately.
Although there have been described what are the present exemplary embodiments of the invention, it will be understood that variations and modifications may be made thereto within the spirit and scope of the appended claims.
Kobayashi, Sachio, Ikeda, Tetsuo, Unoura, Kiyozumi
Patent | Priority | Assignee | Title |
10048691, | Nov 17 2015 | Ford Global Technologies, LLC | Distinguishing lane markings for a vehicle to follow |
11120527, | May 18 2018 | Denso Corporation | Vehicular camera and camera system |
8107683, | Apr 28 2005 | HONDA MOTOR CO , LTD | Method and system for in-vehicle image processing |
8174374, | Jun 30 2009 | Mitsubishi Electric Research Laboratories, Inc | Method and system for coding digital information in lane markings using an optical sensor |
8175806, | Jan 29 2007 | Toshiba Solutions Corporation; Toshiba Digital Solutions Corporation | Car navigation system |
8224031, | Apr 26 2005 | Subaru Corporation | Road line recognition apparatus |
8378799, | Jun 30 2009 | Mitsubishi Electric Research Laboratories, Inc | Method and system for coding information on a roadway surface subject to motion blur |
8520954, | Mar 03 2010 | Denso Corporation | Apparatus for detecting lane-marking on road |
8594380, | Oct 22 2008 | NEC Corporation | Lane marking detection apparatus, lane marking detection method, and lane marking detection program |
9516277, | May 02 2012 | GM Global Technology Operations LLC | Full speed lane sensing with a surrounding view system |
9538144, | May 02 2012 | GM Global Technology Operations LLC | Full speed lane sensing using multiple cameras |
9544549, | Nov 12 2010 | VALEO Schalter und Sensoren GmbH | Method for generating an image of the surroundings of a vehicle and imaging device |
9829888, | Nov 17 2015 | Ford Global Technologies, LLC | Distinguishing lane markings for a vehicle to follow |
Patent | Priority | Assignee | Title |
5359666, | Sep 28 1988 | Honda Giken Kogyo Kabushiki Kaisha | Driving way judging device and method |
5555312, | Jun 25 1993 | Fujitsu Limited | Automobile apparatus for road lane and vehicle ahead detection and ranging |
6091833, | Aug 28 1996 | MATSUSHITA ELECTRIC INDUSTRIAL CO , LTD | Local positioning apparatus, and a method therefor |
7006667, | Nov 22 2000 | Nissan Motor Co., Ltd. | Apparatus and method for detecting road white line for automotive vehicle |
7034861, | Jul 07 2000 | Panasonic Corporation | Picture composing apparatus and method |
7317813, | Jun 13 2001 | Denso Corporation; Nippon Soken, Inc | Vehicle vicinity image-processing apparatus and recording medium |
20050031169, | |||
JP10320559, | |||
JP2001236506, | |||
JP2002029348, | |||
JP2002092796, | |||
JP2002197469, | |||
JP2004110521, | |||
JP6042261, |
Executed on | Assignor | Assignee | Conveyance | Frame | Reel | Doc |
May 17 2005 | Honda Motor Co., Ltd. | (assignment on the face of the patent) | / | |||
Oct 26 2006 | IKEDA, TETSUO | HONDA MOTOR CO , LTD | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 018624 | /0083 | |
Oct 26 2006 | KOBAYASHI, SACHIO | HONDA MOTOR CO , LTD | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 018624 | /0083 | |
Oct 26 2006 | UNOURA, KIYOZUMI | HONDA MOTOR CO , LTD | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 018624 | /0083 |
Date | Maintenance Fee Events |
Feb 01 2012 | M1551: Payment of Maintenance Fee, 4th Year, Large Entity. |
Feb 17 2016 | M1552: Payment of Maintenance Fee, 8th Year, Large Entity. |
Apr 20 2020 | REM: Maintenance Fee Reminder Mailed. |
Oct 05 2020 | EXP: Patent Expired for Failure to Pay Maintenance Fees. |
Date | Maintenance Schedule |
Sep 02 2011 | 4 years fee payment window open |
Mar 02 2012 | 6 months grace period start (w surcharge) |
Sep 02 2012 | patent expiry (for year 4) |
Sep 02 2014 | 2 years to revive unintentionally abandoned end. (for year 4) |
Sep 02 2015 | 8 years fee payment window open |
Mar 02 2016 | 6 months grace period start (w surcharge) |
Sep 02 2016 | patent expiry (for year 8) |
Sep 02 2018 | 2 years to revive unintentionally abandoned end. (for year 8) |
Sep 02 2019 | 12 years fee payment window open |
Mar 02 2020 | 6 months grace period start (w surcharge) |
Sep 02 2020 | patent expiry (for year 12) |
Sep 02 2022 | 2 years to revive unintentionally abandoned end. (for year 12) |