An adhered substance detection apparatus includes a controller configured to function as a determination part, an extractor, and a detector. The determination part determines a representative edge direction using a predetermined angle range as a unit for each pixel area of a plurality of pixel areas of a photographic image photographed by a photographing device, the representative edge direction being determined for each of the pixel areas based on an edge angle of each pixel included in the pixel area. The extractor extracts an array pattern in which a plurality of the pixel areas having a same representative edge direction are continuously arranged along a predetermined scanning direction based on the representative edge directions of the pixel areas determined by the determination part. The detector detects whether an adhered substance area exists on a lens of the photographing device based on the array pattern extracted by the extractor.
|
1. An adhered substance detection apparatus comprising a computer having a hardware processor configured to:
determine a representative edge direction using a predetermined angle range as a unit for each pixel area of a plurality of pixel areas of a photographic image photographed by a photographing device, the representative edge direction being determined for each of the pixel areas based on an edge angle of each pixel included in the pixel area;
extract an array pattern in which a plurality of the pixel areas having a same representative edge direction are continuously arranged along a predetermined scanning direction based on the representative edge directions of the pixel areas determined by the hardware processor; and
detect whether an adhered substance area exists on a lens of the photographing device based on the array pattern extracted by the hardware processor, wherein
the hardware processor extracts, as the array pattern, a first array pattern along a first scanning direction and a second array pattern along a second scanning direction perpendicular to the first scanning direction, and the hardware processor detects whether the adhered substance area exists based on an intersection between the first array pattern in the first scanning direction and the second array pattern in the second scanning direction.
8. An adhered substance detection method comprising the steps of:
(a) determining, by a hardware processor of a computer, a representative edge direction using a predetermined angle range as a unit for each pixel area of a plurality of pixel areas of a photographic image photographed by a photographing device, the representative edge direction being determined for each of the pixel areas based on an edge angle of each pixel included in the pixel area;
(b) extracting, by the hardware processor, an array pattern in which a plurality of the pixel areas having a same representative edge direction are continuously arranged along a predetermined scanning direction based on the representative edge directions of the pixel areas determined in step (a); and
(c) detecting, by the hardware processor, whether an adhered substance area exists on a lens of the photographing device for each of the pixel areas based on the array pattern extracted in step (b), wherein
the hardware processor, in the step (b), extracts, as the array pattern, a first array pattern along a first scanning direction and a second array pattern along a second scanning direction perpendicular to the first scanning direction, and the hardware processor, in the step (c), detects whether the adhered substance area exists based on an intersection between the first array pattern in the first scanning direction and the second array pattern in the second scanning direction.
2. The adhered substance detection apparatus according to
the first and second array patterns are extracted when two pixel areas positioned on opposite sides of each of the first and second array patterns along the first and second scanning directions respectively have a value for the representative edge directions that is adjacent to a value of the representative edge direction of the first and second array patterns.
3. The adhered substance detection apparatus according to
the hardware processor detects whether the adhered substance area exists for each attention area formed of a plurality of the pixel areas, and determines whether or not the attention area is the adhered substance area based on the first and second array patterns included in an extension area obtained by extending the attention area.
4. The adhered substance detection apparatus according to
the hardware processor detects whether the adhered substance area exists based on types of the representative edge directions of the first and second array patterns that form the intersection.
5. The adhered substance detection apparatus according to
the hardware processor detects whether the adhered substance area exists based on a total sum of products of the first and second array patterns that form the intersection.
6. The adhered substance detection apparatus according to
the hardware processor determines a first representative edge direction using a first angle range as a unit and a second representative edge direction using a second angle range as a unit for each of the pixel areas, and the hardware processor extracts the first and second array patterns for each of the first representative edge direction and the second representative edge direction.
7. The adhered substance detection apparatus according to
the hardware processor determines whether or not each of the representative edge directions satisfies a detection condition, and when a predetermined number or more of the representative edge directions satisfy the detection condition, the adhered substance area is detected to exist.
|
The invention relates to an adhered substance detection apparatus and an adhered substance detection method.
Conventionally, there has been known an adhered substance detection apparatus that detects an adhered substance adhered to a lens of a camera based on a photographic image photographed by the camera mounted on a vehicle, or the like. The adhered substance detection apparatus detects an adhered substance, for example, based on a difference between time-series photographic images.
However, in a conventional technology, there has been room for improvement in terms of detecting an adhered substance early and with high accuracy.
According to one aspect of the invention, an adhered substance detection apparatus includes a controller configured to function as a determination part, an extractor, and a detector. The determination part determines a representative edge direction using a predetermined angle range as a unit for each pixel area of a plurality of pixel areas of a photographic image photographed by a photographing device, the representative edge direction being determined for each of the pixel areas based on an edge angle of each pixel included in the pixel area. The extractor extracts an array pattern in which a plurality of the pixel areas having a same representative edge direction are continuously arranged along a predetermined scanning direction based on the representative edge directions of the pixel areas determined by the determination part. The detector detects whether an adhered substance area exists on a lens of the photographing device based on the array pattern extracted by the extractor.
As a result, it is possible to detect the adhered substance early and with high accuracy.
Therefore, an object of the invention is to provide an adhered substance detection apparatus and an adhered substance detection method capable of detecting the adhered substance early and with high accuracy.
These and other objects, features, aspects and advantages of the invention will become more apparent from the following detailed description of the invention when taken in conjunction with the accompanying drawings.
Hereinafter, embodiments of an adhered substance detection apparatus and an adhered substance detection method disclosed in the present application will be described in detail with reference to accompanying drawings. The invention is not limited to the embodiments described below.
First, an outline of the adhered substance detection method according to the embodiment will be described with reference to
The adhered substance is not limited to snow, and may be, for example, light-colored dirt, or the like. In other words, although, in the photographic image I, the adhered substance prevents an object from being reflected, some light may transmit through the adhered substance and a small luminance change may be caused by a light transmission variation.
Here, in a conventional adhered substance detection apparatus, there has been a technology that detects an adhered substance based on a difference between time-series photographic images. However, in the conventional technology, for example, when the whole surface of the lens is covered with snow, or the like, the difference between the images is hardly caused so that there is a possibility that the adhered substance cannot be detected. Furthermore, conventionally, since the time-series photographic images are required, it takes time to detect the adhered substance.
Therefore, in the adhered substance detection method according to the embodiment, an adhered substance is detected using an edge angle that is detected from one photographic image I. Specifically, as illustrated in
For example, the pixel area 100 is an area in which the pixels PX are arranged in a 4×4 matrix (16 pixels PX in total). In an upper stage of
The representative edge direction is an edge direction representing the respective edge angles of 4×4 pixels PX. In
Subsequently, in the adhered substance detection method according to the embodiment, an array pattern in which the pixel areas 100 having the same representative edge direction are continuously arranged along a predetermined scanning direction is extracted (a step S2).
When the array pattern 200-0 is not particularly distinguished from the array patterns 200-1, an array pattern 200 obtained by deleting a last digit in the numerical sequence is described. The scanning direction is not limited to the horizontal direction and may be a vertical direction. A horizontal direction array pattern 200 is combined with a vertical direction array pattern 200 to detect an adhered substance area. This point will be described in detail later with reference to
Subsequently, in the adhered substance detection method according to the embodiment, the adhered substance area on the lens of the camera, to which the adhered substance is adhered, is detected based on the array pattern 200 extracted in the step S2 (a step S3).
For example, in the adhered substance detection method according to the embodiment, the adhered substance area is detected for each division area Sa obtained by dividing the photographic image I into a plurality of areas. The division area Sa is, for example, an area including hundred (10×10) pixel areas 100. Since the pixel area 100 consists of 16 (4×4) pixels PX, the division area Sa consists of 1600 (16×100) pixels PX.
For example, when snow is detected as the adhered substance, in an area (hereinafter, referred to as a “snow-covered area”) in which the lens is covered with snow, an edge gently changes so that the edge angle gently changes. That is, the snow-covered area has a feature that the pixel areas 100 having the same representative edge direction are likely to be continuously arranged.
In the snow-covered area, since snow accumulating on the lens has a concave and convex shape, the array patterns 200 based on plural types of the representative edge directions tend to appear in one division area Sa. Furthermore, since a size of snow (a size of the concave and convex shape) is limited to a predetermined size, the snow-covered area has a feature that a length of each of the array patterns 200 falls within a predetermined range.
In other words, in the snow-covered area, the array patterns 200 based on the plural types of the representative edge directions are included in one division area Sa. On the other hand, in an area other than the snow-covered area (hereinafter, referred to as a “background area”), since the edge of each pixel PX changes due to a background image, it is difficult to acquire the feature of the snow-covered area described above.
Therefore, in the adhered substance detection method according to the embodiment, as one example of detection conditions, the adhered substance area is detected for the division area Sa that includes plural types of the array patterns 200, each of which satisfies a predetermined length condition. As a result, in the adhered substance detection method according to the embodiment, the snow-covered area can be detected as the adhered substance area.
As described above, in the adhered substance detection method according to the embodiment, by detecting the adhered substance based on the array pattern 200 in which the pixel areas 100 having the same representative edge direction are continuously arranged, even when the difference between the images is not caused, it is possible to detect the adhered substance with high accuracy.
In the adhered substance detection method according to the embodiment, it is possible to detect the adhered substance using one photographic image I. Therefore, in the adhered substance detection method according to the embodiment, it is possible to detect the adhered substance early and with high accuracy.
Next, a configuration of the adhered substance detection apparatus 1 according to the embodiment will be described with reference to
The camera 10 is, for example, an in-vehicle camera including a lens, such as a fish-eye lens, and an imaging element, such as a CCD (Charge Coupled Device) or a CMOS (Complementary Metal Oxide Semiconductor). The camera 10 is, for example, provided at each position capable of photographing front, rear and side images of a vehicle, and outputs the photographed photographic image I to the adhered substance detection apparatus 1.
The various devices 50 acquire a detection result of the adhered substance detection apparatus 1 to perform various control of the vehicle. The various devices 50 include, for example, a display device that informs a user that an adhered substance is adhered to the lens of the camera 10 and instructs the user to remove the adhered substance from the lens, a removing device that removes the adhered substance from the lens by spraying a fluid, air, etc. onto the lens, and a vehicle control device that controls autonomous driving, or the like.
As illustrated in
Here, the adhered substance detection apparatus 1 includes a computer having, for example, a CPU (Central Processing Unit), a ROM (Read Only Memory), a RAM (Random Access Memory), a data flash, an input/output port, and the like, and various circuits.
The CPU of the computer, for example, reads out and executes a program stored in the ROM so as to function as the image acquisition part 21, the determination part 22, the extractor 23 and the detector 24 of the controller 2.
At least one or all of the image acquisition part 21, the determination part 22, the extractor 23 and the detector 24 of the controller 2 may be constituted of hardware such as an ASIC (Application Specific Integrated Circuit) and a FPGA (Field Programmable Gate Array).
The memory 3 corresponds to the RAM or data flash, for example. The RAM and data flash may store the representative edge direction information 31, the threshold value information 32, or information on various programs. The adhered substance detection apparatus 1 may acquire the programs or information described above via computers connected via a wired or wireless network or a portable recording medium.
The representative edge direction information 31 is information related to the representative edge direction using the predetermined angle range as a unit. Specific examples of the representative edge direction information 31 will be described later with reference to
The controller 2 determines the representative edge direction for each of the pixel areas 100 of the photographic image I, extracts the array pattern 200 in which the pixel areas 100 having the same representative edge direction are continuously arranged, and detects the adhered substance area based on the extracted array pattern 200.
The image acquisition part 21 acquires the photographic image I photographed by the camera 10. The image acquisition part 21 performs a grayscale conversion of expressing each pixel of the acquired photographic image I in gray scales from white to black depending on the luminance, performs a smoothing process on each pixel, and outputs the processed pixel to the determination part 22. For example, an averaging filter or an arbitrary smoothing filter, such as a Gaussian filter, may be used for the smoothing process. The gray scale conversion and the smoothing process may be omitted.
The determination part 22 determines the representative edge direction for each of the pixel areas 100 of the photographic image I acquired from the image acquisition part 21. Here, the determination process of the representative edge direction by the determination part 22 will be specifically described with reference to
Each of
Subsequently, the determination part 22 calculates a vector V including information of the edge angle and edge strength of the pixel PX using a trigonometric function based on the detected strength of the edge ex in the X-axis direction and the detected strength of the edge ey in the Y-axis direction. Specifically, an angle θ between the vector V and the X-axis on the positive direction side will be referred to as an edge angle, and the length L of the vector V will be referred to as an edge strength of each pixel.
Subsequently, the determination part 22 determines the representative edge direction in the pixel area 100 based on the calculated vector V of each pixel PX. Specifically, as illustrated in
Specifically, when the edge angle in the vector V is within an angle range of −45° or more and less than 45°, the determination part 22 classifies the edge angle into the group “(0)”. When the edge angle in the vector V is within an angle range of 45° or more and less than 135°, the determination part 22 classifies the edge angle into the group “(1)”. When the edge angle in the vector V is within an angle range of 135° or more and less than 180° or within an angle range of −180° or more and less than −135°, the determination part 22 classifies the edge angle into the group “(2)”. When the edge angle in the vector V is within an angle range of −135° or more and less than −45°, the determination part 22 classifies the edge angle into the group “(3)”.
As shown in a lower stage of
The frequency of the histogram is calculated by summing up the edge strengths of the pixels PX which are classified into a same angle range among the pixels PX within the pixel area 100. Specifically, the frequency of the histogram that belongs to the group (class) “(0)” will be considered. For example, there are three pixels PX which are classified into the group (class) “(0)”, and the edge strengths of the three pixels PX are 10, 20, and 30, respectively. In this case, the frequency of the histogram that belongs to the group (class) “0” is calculated by the following equation: 10+20+30=60. The frequency of the histogram that belongs to the group (class) “(1)”, “(2)”, or “(3)” is calculated in a similar manner.
Based on the calculated histogram, the determination part 22 determines the edge strength. Specifically, when the frequency of the class whose frequency is the highest in the histogram is the predetermined threshold value THa or more, the frequency corresponding to the class is regarded as the edge strength. That is, a determination process of a representative edge strength by the determination part 22 can be regarded as a process of extracting a feature related to the edge strength within the pixel area 100 corresponding to the representative edge direction.
On the other hand, when the frequency of the class whose frequency is the highest is less than the predetermined threshold value THa, the determination part 22 regards the representative direction in the pixel area 100 as “invalid”, in other words, “no representative direction”. As a result, when there is a large variation in the edge angles of each pixel PX, it is possible to prevent an erroneous determination of a specific representative edge direction.
As described above, the determination part 22 associates each of the representative edge directions respectively different in the first angle range and the second angle range with each of the pixel areas 100. As a result, since a data amount used for detection of the adhered substance area can be increased, it is possible to improve detection accuracy of the adhered substance area.
The determination process of the determination part 22 shown in
Referring back to
Specifically, the extractor 23 scans a plurality of the pixel areas 100 of the photographic image I that are arranged in a two-dimensional (horizontal and vertical) array in horizontal and vertical directions, and extracts the array pattern 200. The horizontal and vertical directions are one example of first and second scanning directions. The scanning direction is not limited to the horizontal and vertical directions, but may be an oblique direction.
Here, when two pixel areas 100 positioned on opposite sides of an array composed of the plurality of the pixel areas 100 along the scanning direction have the representative edge directions (group “(0)” and group “(2)”) adjacent to the representative edge direction (group “(1)”) of the array pattern 200, the extractor 23 extracts the array as the array pattern 200.
As shown in
Specifically, as shown in an upper stage of
On the other hand, as shown in a lower stage of
As described above, since, in the snow-covered area, the edge angle gently changes so that the edge angle continuously changes, the adjacent pixel areas 100 hardly have the representative edge directions opposite to each other.
When at least one of the two pixel areas 100 positioned on opposite sides of the array pattern has the representative edge direction opposite to that of the array pattern, the array pattern needs not be extracted.
That is, the extractor 23 extracts only the array pattern 200 having the representative edge direction that continuously changes along the scanning direction, and excludes the array pattern having the representative edge direction that irregularly changes along the scanning direction from extracted targets. As a result, it becomes possible to improve extraction accuracy of the array pattern 200 corresponding to the snow-covered area.
As described above, since, in the snow-covered area, the representative edge direction continuously changes, the snow-covered area has the feature that the length of one array pattern falls within the predetermined range. As a result, the extractor 23 can extract only the array pattern 200 whose length falls within the predetermined range.
Here, the predetermined range is, for example, a range from 24 pixels PX (three pixel areas 100) to 100 pixels PX (twenty-five pixel areas 100), but is not limited thereto and may be changed arbitrarily.
As a result, it is possible to extract only the array pattern 200 corresponding to the snow-covered area. In other words, since the array pattern that is inconsistent with the feature of the snow-covered area is not extracted, it is possible to improve detection accuracy of the snow-covered area as the adhered substance area.
Referring back to
Each of
As shown in
Subsequently, the detector 24 calculates a product of a number of the pixel areas of both array patterns 200-1 that form the intersection CP.
This means, as the product of each intersection CP increases, the pixel areas 100 having the same representative edge direction are continuously distributed in a wider range. That is, by calculating the product of each intersection CP, it is possible to easily calculate the aggregate (chunk) size of the pixel areas 100 having the same representative edge direction.
The detector 24 may exclude the array pattern 200 that extends either in the vertical direction or in the horizontal direction on a basis of each of the pixel areas 100 by detecting the adhered substance area based on the intersection CP of the array pattern 200. That is, the detector 24 uses only the array pattern 200 extending radially for the detection of the adhered substance area on the basis of each of the pixel areas 100. The detector 24 associates the intersection CP with the representative edge direction and intersection information indicating the above-mentioned product.
The detector 24 may count the number of the pixel areas 100 forming one chunk of the same representative edge direction instead of the product of the intersection CP. In this case, for example, a sum of the number of the pixel areas 100 included in the array pattern 200-1 extracted along the horizontal direction and the number of the pixel areas 100 included in the array pattern 200-1 perpendicular to the above-mentioned array pattern 200-1 may be calculated as a size of one chunk of the same representative edge direction.
Subsequently, as shown in
Here, the attention area SaT is one division area Sa, and the extension area Ea is 3×3 division areas Sa centered on the attention area Sat. The extension area Ea is centered on the attention area SaT, and may be another area that is wider than the attention area SaT.
The detector 24 determines whether or not the attention area SaT is the adhered substance area based on the intersection information included in the extension area Ea. Specifically, as shown in
When there are the intersections CP associated with all types of the representative edge directions in the extension area Ea, and both the total value of the products in the first angle range and the total value of the products in the second angle range exceed a predetermined total threshold value, the detector 24 detects the attention area SaT as the adhered substance area.
As described above, by including the types of the representative edge directions associated with the intersections CP in the detection condition, it is possible to easily determine the feature of the snow-covered area in which a wide variety of the representative edge directions are included.
When the products of the respective representative edge directions in the first angle range and the second angle range are equal to “1” or more, and the respective total values of the products in the first angle range and the second angle range exceed the predetermined total threshold value, the detector 24 detects the attention area SaT as the adhered substance area.
That is, when the extension area Ea includes the intersections CP based on all types of the representative edge directions, and is an aggregate of the division areas Sa in which chunks of the respective representative edge directions have a certain size, the detector 24 detects the attention area SaT as the adhered substance area.
As described above, by deriving the sum of the products in the respective representative edge directions, it is possible to easily understand percentages of the pixel areas 100 having each of the representative edge directions in the attention area SaT.
When the detection condition is satisfied in both the first angle range and the second angle range, the attention area SaT is detected as the adhered substance area. As a result, it is possible to reduce erroneous detection of the adhered substance area as compared to when the detection condition is satisfied either in the first angle range or in the second angle range, the attention area SaT is detected as the adhered substance area.
The detector 24 determines whether or not the attention area SaT is the adhered substance area based on the intersections CP included in the extension area Ea obtained by extending the attention area SaT. This is because if the attention area SaT is the snow-covered area, there is a high possibility that an area around the attention area SaT is also the snow-covered area.
In other words, there is a low possibility that only one attention area SaT becomes the snow-covered area. That is, by determining whether or not the attention area SaT is the adhered substance area based on the extension area Ea obtained by extending the attention area SaT, it is possible to improve the detection accuracy of the snow-covered area.
Next, a processing procedure executed by the adhered substance detection apparatus 1 according to the embodiment will be described with reference to
As shown in
Subsequently, the adhered substance detection apparatus 1 sets the attention area SaT (a step S105), and determines whether or not there are eight types of the intersections CP associated with different representative edge directions in the extension area Ea including the attention area SaT (a step S106).
In the determination process of the step S106, when there are eight types of the intersections CP (Yes in the step S106), the adhered substance detection apparatus 1 determines whether or not the total value of the products in the first angle range is greater than the total threshold value (a step S107).
In the determination process of the step S107, when the total value of the products exceeds the total threshold value (Yes in the step S107), the adhered substance detection apparatus 1 then determines whether or not the total value of the products in the second angle range is greater than the total threshold value (a step S108).
In the determination process of the step S108, when the total value of the products exceeds the total threshold value (Yes in the step S108), the adhered substance detection apparatus 1 detects the attention area SaT as the adhered substance area (a step S109).
The adhered substance detection apparatus 1 determines whether or not each process of the step S105 to the step S109 has been terminated for all of the division areas Sa (a step S110), and when each process has been terminated for all of the division areas Sa (Yes in the step S110), the adhered substance detection apparatus 1 ends the process.
On the other hand, in the determination process of the step S110, when each process has not been terminated for all of the pixel areas 100 (No in the step S110), the adhered substance detection apparatus 1 moves to the step S105.
In the determination process of the step S106, when there are seven or less types of the intersections (No in the step S106), and when the total value of the products is equal to or less than the total threshold value in each determination process of the step S107 and the step S108 (No in the step S107 and the step S108), the adhered substance detection apparatus 1 moves to the step S110.
As described above, the adhered substance detection apparatus 1 according to the embodiment includes the determination part 22, the extractor 23, and the detector 24. The determination part 22 determines the representative edge direction using the predetermined angle range as a unit for each of the pixel areas 100 based on the edge angle of each pixel PX included in the pixel area 100 of the photographic image I photographed by the photographing device.
The extractor 23 extracts the array pattern 200 in which the pixel areas 100 having the same representative edge direction are continuously arranged along the predetermined scanning direction based on the representative edge direction determined by the determination part 22. The detector 24 detects the adhered substance area on the lens of the photographing device, to which an adhered substance is adhered, based on the array pattern 200 extracted by the extractor 23. Therefore, the adhered substance detection apparatus 1 according to the embodiment can detect the adhered substance early and with high accuracy.
In the embodiment described above, a case in which the respective representative edge directions are determined in both the first angle range and the second angle range has been described. However, the representative edge directions may be determined either in the first angle range or in the second angle range.
As shown in
For example, when the sum of the products exceeds a predetermined value, the detector 24 determines that it is “OK”, and when the sum of the products is lower than the predetermined value, the detector 24 determines that it is “NG”. For example, when there are three or more types of the representative edge directions whose determination results are “OK”, the detector 24 detects the attention area SaT as an adhered substance area Da.
On the other hand, when there are less than three types of the representative edge directions whose determination results are “OK”, the detector 24 does not detect the attention area SaT as the adhered substance area Da.
That is, when the attention area SaT includes an aggregate in which chunks of plural types of the representative edge directions have a predetermined size, the detector 24 detects the attention area SaT as the adhered substance area. Even in such a case, it is possible to detect the adhered substance area rapidly.
For example, even when the products associated with the intersection CP are less than a threshold value separately set, the detector 24 may add the products to the sum of the products. That is, when the products are too large, the products need not be used for the above determination. As a result, since the above determination can be performed by using only the intersections CP similar to a feature of a snow-covered area, it is possible to improve detection accuracy of the snow-covered area.
The above-described embodiment and modification shows an example in which the representative edge direction (in each pixel PX having an angle range of 0° to 360°) is divided into four directions by each angle range of 90°. However, the angle range is not limited to 90°. For example, the representative edge direction may be divided into six directions by each angle range of 60°.
A first representative edge direction uses the first angle range as a unit and a second representative edge direction uses the second range as a unit. Widths of the respective angle ranges for the first representative edge direction and the second representative edge direction may be different from each other. For example, the first representative edge direction may be divided into four directions by each angle range of 90°, and the second representative edge direction may be divided into six directions by each angle range of 60°.
For the first representative edge direction and the second representative edge direction, respective boundaries of two adjacent angle ranges are deviated by an angle of 45° from each other. However, the deviated angle may exceed 45°, or may be less than 45°. It is possible to arbitrarily set the boundaries of two adjacent angle ranges for the first representative edge direction and the second representative edge direction.
In the above-described embodiment, although the photographic image I photographed by the camera to be mounted on the vehicle is used, the photographic image I photographed by a security camera or a camera on a street light may be used. That is, the photographic image I photographed by a camera with a lens to which an adhered substance can be adhered may be used.
It is possible for a person skilled in the art to easily come up with more effects and modifications. Thus, a broader modification of this invention is not limited to specific description and typical embodiments described and expressed above. Therefore, various modifications are possible without departing from the general spirit and scope of the invention defined by claims attached and equivalents thereof.
While the invention has been shown and described in detail, the foregoing description is in all aspects illustrative and not restrictive. It is therefore understood that numerous other modifications and variations can be devised without departing from the scope of the invention.
Kono, Takashi, Yamamoto, Daisuke, Ikeda, Nobuhisa, Asayama, Nobunori, Tani, Yasushi, Oki, Tomokazu, Kamibayashi, Teruhiko
Patent | Priority | Assignee | Title |
11620833, | Sep 23 2020 | Toyota Jidosha Kabushiki Kaisha | Vehicle driving support device |
Patent | Priority | Assignee | Title |
10227062, | Sep 23 2016 | Fujitsu Ten Limited | Liquid-droplet detecting apparatus, liquid-droplet detecting method, and computer-readable recording medium |
20140010408, | |||
20140029008, | |||
20150172582, | |||
20150339535, | |||
20160364620, | |||
20190041849, | |||
DE102014219424, | |||
JP2010014494, | |||
JP2012038048, | |||
JP201238048, | |||
JP2018072312, | |||
JP201872312, |
Executed on | Assignor | Assignee | Conveyance | Frame | Reel | Doc |
Oct 10 2019 | IKEDA, NOBUHISA | DENSO TEN Limited | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 050899 | /0407 | |
Oct 10 2019 | ASAYAMA, NOBUNORI | DENSO TEN Limited | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 050899 | /0407 | |
Oct 10 2019 | KONO, TAKASHI | DENSO TEN Limited | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 050899 | /0407 | |
Oct 10 2019 | TANI, YASUSHI | DENSO TEN Limited | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 050899 | /0407 | |
Oct 10 2019 | YAMAMOTO, DAISUKE | DENSO TEN Limited | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 050899 | /0407 | |
Oct 10 2019 | OKI, TOMOKAZU | DENSO TEN Limited | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 050899 | /0407 | |
Oct 10 2019 | KAMIBAYASHI, TERUHIKO | DENSO TEN Limited | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 050899 | /0407 | |
Oct 31 2019 | DENSO TEN Limited | (assignment on the face of the patent) | / |
Date | Maintenance Fee Events |
Oct 31 2019 | BIG: Entity status set to Undiscounted (note the period is included in the code). |
Date | Maintenance Schedule |
Nov 09 2024 | 4 years fee payment window open |
May 09 2025 | 6 months grace period start (w surcharge) |
Nov 09 2025 | patent expiry (for year 4) |
Nov 09 2027 | 2 years to revive unintentionally abandoned end. (for year 4) |
Nov 09 2028 | 8 years fee payment window open |
May 09 2029 | 6 months grace period start (w surcharge) |
Nov 09 2029 | patent expiry (for year 8) |
Nov 09 2031 | 2 years to revive unintentionally abandoned end. (for year 8) |
Nov 09 2032 | 12 years fee payment window open |
May 09 2033 | 6 months grace period start (w surcharge) |
Nov 09 2033 | patent expiry (for year 12) |
Nov 09 2035 | 2 years to revive unintentionally abandoned end. (for year 12) |