A contrast-based image fusion system and method of processing multiple images to form a processed or fused image including regions selected from one or more images. images are divided into image regions. Portions of the images are filtered if necessary. A contrast map is generated for each image via a convolution kernel resulting in a contrast map with contrast values for each image region. contrast values are compared and image regions are selected based on a selection criteria or process such as greater or maximum contrast. The selected image regions form the fused image. If necessary, the luminance of one or more portions of the fused image is adjusted. One sensor is selected as a reference sensor, and an average intensity of each region of the reference sensor image is determined across the reference sensor image. The intensity of one or more regions in the final image is adjusted by combining the determined average luminance values and intensity values of the final image.
|
0. 65. A method of forming a processed image using a plurality of images, each image generated by a respective sensor, comprising:
dividing each image into a plurality of image regions;
generating a contrast map for each image, each contrast map including a contrast value for each image region;
applying a selection process to said contrast value for selecting an image region for use in said processed image; and
forming said processed image with the selected image regions
wherein dividing the images into the plurality of image regions further comprises dividing each image on a pixel-by-pixel basis, into blocks of pixels, or into arbitrary shaped regions, and
wherein applying the selection process includes comparing competing contrast values of corresponding image regions from respective images.
0. 101. A system for combining a plurality of images to form a final image, comprising:
a plurality of sensors that generate respective images;
a processor configured to
divide each image into a plurality of image regions,
generate a contrast map for each image, each contrast map including a contrast value for each image region,
apply a selection criterion to said contrast value for selecting an image region for use in said processed image, and
form said processed image with the selected image regions;
wherein the processor is further configured to divide the images into the plurality of image regions by dividing each image on a pixel-by-pixel basis, into blocks of pixels, or into arbitrary shaped regions, and
wherein the process is further configured to apply the selection process by comparing competing contrast values of corresponding image regions from respective images.
27. A method of forming a processed image using a plurality of images, each image generated by a respective sensor, comprising:
dividing each image into a plurality of image regions;
generating a contrast map for each image, each contrast map including a contrast value for each image region;
applying a selection process to said contrast value for selecting an image region for use in said processed image; and
forming said processed image with the selected image regions;
wherein generating the contrast map further comprises performing a convolution to determine the contrast value of the contrast map;
wherein performing the convolution further comprises performing the convolution with a kernel kc,
wherein [{kc*S1(x,y), kc*S2(x,y)}] represents the convolution;
S1 represents image regions of a first image;
S2 represents image regions of a second image; and
(x,y) represent spatial coordinates of the images.
1. A method of forming a processed image using a plurality of images, each image generated by a respective sensor, comprising:
dividing each image into a plurality of image regions;
generating a contrast map for each image, each contrast map including a contrast value for each image region;
applying a selection process to said contrast value for selecting an image region for use in said processed image; and
forming said processed image with the selected image regions,
wherein contrast values of contrast maps of respective first, second, and third sensors are compared together to form said processed image, the method further comprising:
identifying contrast values from first and second sensor images to form an intermediate contrast map;
wherein applying the selection process comprises applying a selection process to the contrast values of the intermediate contrast map and contrast values of a contrast map of a third sensor image.
59. A system for combining a plurality of images to form a final image, comprising:
a plurality of sensors that generate respective images; and
a processor configured to
divide each image into a plurality of image regions,
generate a contrast map for each image, each contrast map including a contrast value for each image region,
apply a selection criterion to said contrast value for selecting an image region for use in said processed image, and
form said processed image with the selected image regions;
wherein the processor is configured to generate the contrast map by performing a convolution to determine the contrast value of the contrast map; and
wherein the processor is configured to perform the convolution with a kernel kc,
wherein [{kc*S1(x,y), kc*S2(x,y)}] represents the convolution;
S1 represents image regions of a first image;
S2 represents image regions of a second image; and
(x,y) represent spatial coordinates of the images.
33. A system for combining a plurality of images to form a final image, comprising:
a plurality of sensors that generate respective images;
a processor configured to
divide each image into a plurality of image regions,
generate a contrast map for each image, each contrast map including a contrast value for each image region,
apply a selection criterion to said contrast value for selecting an image region for use in said processed image, and
form said processed image with the selected image regions;
wherein contrast values of contrast maps of a respective first, second, and third sensors are compared together to form the final image;
wherein the processor is further configured to identify contrast values from first and second sensor images to form an intermediate contrast map; and
wherein the processor applies the selection criterion by applying a selection process to the contrast values of the intermediate contrast map and contrast values of a contrast map of a third sensor image.
0. 100. A method of forming a processed image using a plurality of images, each image generated by a respective sensor, comprising:
comparing contrast values of contrast maps of the images by
defining a plurality of sets of corresponding image regions in the plurality of images;
generating contrast maps with contrast values for the sets of corresponding image regions in the plurality of images;
identifying, for each set of corresponding image regions, one contrast value as a maximum contrast value, and
selecting image regions corresponding to the maximum contrast values,
forming a processed image using the selected image regions; and
adjusting an intensity of at least one portion of the processed image by:
determining at least one intensity value for the at least one portion of the processed image,
selecting one of the sensors as a reference sensor,
determining an average intensity value for one or more regions of an image generated by the reference sensor, and
adjusting the at least one intensity value for the at least one portion of the processed image in accordance with the determined average intensity value.
0. 87. A method of forming a processed image using a plurality of images, each image generated by a respective sensor to form a processed image, comprising:
filtering portions of one or more images;
comparing contrast values of the images by
dividing each image into a plurality of image regions,
generating a contrast map for each image, each contrast map including contrast values for each image region of each image;
comparing contrast values in each contrast map of the image regions,
identifying maximum contrast values based on the comparison of contrast values, and
selecting image regions corresponding to the maximum contrast values,
forming the processed image with the selected image regions; and
adjusting an intensity of one or more portions of the processed image by selecting one sensor as a reference sensor,
determining at least one average intensity value for one or more regions of the reference sensor image across the reference sensor image, and
adjusting the intensity of one or more regions in the processed image by combining the determined average intensity values and intensity values of the processed image;
wherein generating the contrast map further comprises performing a convolution to determine the contrast value of the contrast map.
0. 119. A system for forming a processed image using a plurality of images, comprising:
a first sensor that generates a first image;
a second sensor that generates a second image,
wherein the first and second images are divided into a plurality of image regions;
a processor configured to
filter one or more portions of one or more images;
compare contrast values of the images by
dividing each image into a plurality of image regions,
generating a contrast map for each image, each contrast map including contrast values for each image region of each image,
comparing contrast values in each contrast map of the image regions,
identifying maximum contrast values,
selecting image regions corresponding to the maximum contrast values, and
forming the processed image with the selected image regions; and
adjust an intensity of one or more regions of the final image by
selecting one sensor as a reference sensor,
determining at least one average intensity value for each region of the reference sensor image across the reference sensor image, and
adjusting the intensity of one or more regions in said processed image by combining the determined average intensity values and intensity values of the final image;
wherein the processor is further configured to divide the images into the plurality of image regions by dividing each image on a pixel-by-pixel basis, into blocks of pixels, or into arbitrary shaped regions, and
wherein the processor is further configured to apply the selection process by comparing competing contrast values of corresponding image regions from respective images.
2. The method of
4. The method of
5. The method of
6. The method of
7. The method of
8. The method of
9. The method of
10. The method of
11. The method of
13. The method of
14. The method of
16. The method of
17. The method of
selecting one sensor as a reference sensor;
determining at least one average intensity value for each region of the reference sensor image; and
adjusting the intensity of one or more regions in said processed image by combining the determined average intensity values and intensity values of said processed image.
18. The method of
19. The method of
20. The method of
21. The method of
22. The method of
23. The method of
24. The method of
25. The method of
26. The method of
28. The method of
30. The method of
31. The method of
32. The method of
34. The system of
36. The system of
37. The system of
38. The system of
39. The system of
40. The system of
41. The system of
42. The system of
43. The system of
45. The system of
46. The system of
47. The system of
48. The system of
49. The system of
select one sensor as a reference sensor,
determine at least one average intensity value for each region of the reference sensor image across the reference sensor image, and
adjust the intensity of one or more regions in said processed image by combining the determined average intensity values and intensity values of said processed image.
50. The system of
51. The system of
52. The system of
53. The system of
54. The system of
55. The system of
56. The system of
57. The system of
58. The system of
60. The system of
62. The system of
63. The system of
64. The system of
0. 66. The method of
each sensor detects a range of wavelengths that is different than a range of wavelengths detected by the other sensors such that each of the images is associated with a different range of wavelengths;
dividing each image comprises defining sets of corresponding image regions by dividing each image into a plurality of image regions each of which is substantially aligned with and substantially overlaps with one image region in each of the other images; and
selecting an image region for use in the processed image comprises comparing, for each set of corresponding image regions, the contrast values for each image region in the set and selecting one image region from the set based on the comparing.
0. 67. The method of
0. 68. The method of
0. 69. The method of
0. 70. The method of
0. 71. The method of
0. 72. The method of
0. 73. The method of
0. 74. The method of
selecting one sensor as a reference sensor;
determining an average intensity value of each region of the reference sensor image; and
adjusting the intensity value of one or more regions in said processed image by combining the determined average intensity values and intensity values of said processed image.
0. 75. The method of
0. 76. The method of
0. 77. The method of
determining an average intensity value of the reference sensor image on a pixel-by-pixel basis; and
adjusting the intensity value of said processed image on a pixel-by-pixel basis.
0. 78. The method of
0. 79. The method of
0. 80. The method of
0. 81. The method of
0. 82. The method of
0. 83. The method of
0. 84. The method of
0. 85. The method of
0. 86. The method of
0. 88. The method of
0. 89. The method of
0. 90. The method of
0. 91. The method of
0. 92. The method of
0. 93. The method of
0. 94. The method of
0. 95. The method of
0. 96. The method of
0. 97. The method of
0. 98. The method of
0. 99. The method of
0. 102. The system of
0. 103. The system of
0. 104. The system of
0. 105. The system of
0. 106. The system of
0. 107. The system of
0. 108. The system of
0. 109. The system of
0. 110. The system of
0. 111. The system of
0. 112. The system of
0. 113. The system of
0. 114. The system of
0. 115. The system of
0. 116. The system of
0. 117. The system of
0. 118. The system of
|
The present invention relates generally to imaging systems and methods, and more particularly, to an imaging system and method that selectively fuse or combine regions of images from two or more sensors to form a single, processed image.
Image fusion generally refers to combining or merging portions of two or more images into a single processed image. Image fusion is commonly used when two or more detectors are used in generating an image, whereby the image displayed to a user or provided to an automated processing system is combined from information provided by each of the sensors.
One manner in which known systems combine images from different sensors is by merely adding the two images together on a pixel by pixel basis. Thus, for example, for rendering a two-dimensional (2-D) processed image of pixels arranged in an n×m matrix wherein each pixel position is identified by the position (x,y), a value or data in pixel (1,1) of the first image is added to the data or value in pixel (1,1) in the second image, a value or data in pixel (1,2) of the first image is added to the value or data in pixel (1,2) of the second image, and so on for each pixel through pixel (n,m) of both images. Other known systems perform a variant of this technique and calculate the average of the values in each pixel instead of adding the two values. Thus, the final image contains averaged pixel values.
These systems, however, have a number of shortcomings. First, known image fusion techniques typically result in undesirable and unnecessary distortion. For example, if a portion of an image is clear and understandable by a user, while the corresponding portion of a second image is blurry, then adding or averaging pixel values can distort the clear image into one that is less clear. This undesirable effect is the result of incorporating elements of the blurry pixel(s) into the clear pixel(s) through addition or averaging. As a further example, adding unnecessary background regions to a bright image region can decrease the contrast and quality of the bright image region. For example, if regions of two images have high dominance or are bright, then adding two bright regions together can result in a final image that is “overexposed” or too bright. This results in a saturated image. Finally, averaging two dim image regions can result in a relatively dim image, and image regions that were originally dim can have their brightness further reduced.
Other known systems have attempted to overcome these shortcomings using techniques that identify patterns in images and forming a fused image on the basis of patterns. Each source or original image is decomposed into multiple, lower resolution images using filters with different bandwidths (e.g., based on Gaussian roll-off or a Laplacian “pyramid” approach). The pyramid approach is based on using different resolutions for different image regions—coarse features are analyzed at low resolution, and fine features are analyzed at high resolution. These systems, however, are also deficient in that the complete image from each sensor is received before the process of constructing a pyramid can begin. This requirement typically results in a time lag of at least one image from the slowest sensor. Such a time lag is unacceptable in sensors placed on fast moving platforms, such as aircraft or other vehicles, or more generally where real-time operation is desired.
Other known systems use a technique in which the Laplacian method is modified and source images are decomposed into patterns which are assigned saliency values or weights. A pattern is “salient” if it carries information that is useful to understanding the image. A final image is formed on the basis of “weighted” patterns. These techniques, however, can also be deficient in that they typically involve analyzing and assigning saliency weights to each pixel or region of the entire image. Then, the entire image is processed. Thereafter, the salient patterns are selected. As a result, excessive time is wasted analyzing regions of entire images and their corresponding saliency values.
These shortcomings are particularly problematic when known image systems are used in connection with time sensitive activities, e.g., landing an airplane, driving a tank, etc. In these situations, it is desirable that clear images be generated quickly. Known techniques, however, typically cannot generate quality images within these time constraints or typically do so only after full images are available for processing.
Accordingly, a need exists for a method and system that effectively and efficiently select useful, pertinent or relevant information from source images to form a more informative or useful processed image which includes relevant, pertinent and useful information from each of the source images in a time efficient manner. Further, it is desirable to apply the selective image fusion technique to a variety of detectors or image generators to provide flexibility for use in different applications.
The present invention provides a method and system for selectively combining regions of images generated by different sensors (also herein referred to as sensor or source images) to form a processed or fused image using the relevant information from the sensor images. The method and system are implemented by dividing each sensor image into image regions, and generating for each image region a map of contrast values by means of for example, a convolution. The map of contrast values for one sensor image is then compared to the corresponding map of contrast values for the other sensor image. Between or among the compared contrast values, one contrast value is selected based on a selection criterion, which can be, for example, the greater of the two or more contrast values compared. The image regions corresponding to the selected contrast values are then used to form the processed image. According to the present invention the image regions can be divided on a pixel-by-pixel basis, based on groups of pixels, or based on arbitrarily shaped regions.
In yet further accordance with the invention, each sensor detects a different wavelength. Also in accordance with the present invention, images from different types, numbers, and combinations of sensors can be processed. Sensors that can be used include infrared (IR), radio-frequency sensors (e.g., active sensors such as radar, or passive sensors such as radiometers)”.
In still further accordance with the present invention, image regions from a plurality of sensors are combined to form the processed image.
In further accordance with the present invention, contrast maps for images from a first sensor and a second sensor are combined to form an intermediate contrast map, which is then compared with a contrast map of third image to form the processed image.
In further accordance with the invention, the image fusion method and system are used in connection with directing a moving vehicle such as an aircraft, watercraft, automobile, or train.
In further accordance with the invention, the intensity or luminance of one or more image sections is adjusted across the processed image. One sensor is selected as a reference sensor, and an average intensity of regions of the reference sensor image is determined. The intensity of the same or corresponding region or an adjacent region in the processed image is adjusted by combining the determined average luminance values of the reference image and intensity values of the processed image.
Also in accordance with the invention, the method and system are implemented to filter portions of the sensor images before contrast comparisons are performed.
In the following description of embodiments of the invention, reference is made to the accompanying drawings which form a part hereof, and which is shown by way of illustration specific embodiments in which the invention may be practiced. It is to be understood that other embodiments may be utilized as structural changes may be made without departing from the scope of the present invention.
With reference to
The processor 110 is programmed to selectively combine regions from each image 104, 106 into a processed or fused image 115. More specifically, the processor 110 compares regions of each image 104, 106, and selects image regions based on a selection criterion, for example, a comparison of contrast values representing the apparent difference in brightness between light and dark areas of sensor images. The processor can be programmed to consider different selection criteria including, but not limited to, the greater or maximum contrast values of each comparison. Thus, the processing system essentially extracts the desirable regions or regions of choice based on the selection criterion from one or more or all of the images. The selected regions are pieced together to form the fused image 115 (much like a jigsaw puzzle is formed from multiple pieces, except that each piece of the puzzle can be selected from multiple sources). The “puzzle pieces” or image regions can come from a single image, some of the images, or all of the images. The fused image 115 is then presented to the pilot or user through the visual display 120. The fused image can also be provided to an image processor or computer for further processing.
While
The technique of fusing or selectively combining portions of images 104, 106 into a processed image 115 is illustrated in the flow diagrams of
Turning now to
Those persons of skill in the art will recognize that the image fusion method and system can be used in many different environments and applications that process multiple images. For example, besides an aircraft (e.g. an airplane, jet, helicopter, etc.) the method and system can be implemented in other moving vehicles such as a watercraft, an automobile, or a train. Moreover, the image fusion method and system can be used to display images from medical instruments (which use, e.g., ultrasound, infrared, laser imaging or tomography sensors), and surveillance systems. Indeed, many applications can benefit from the selective fusion of image regions to form a processed or fused image that includes relevant information or information of choice from each sensor image.
However, for purposes of explanation, this specification primarily refers to images related to an aircraft. Such images may be related to landing, taxiing, takeoff, or cruising of the aircraft and in connection with applications to prevent Controlled Flight Into Terrain (CFIT). As a specific example of how the system can be used in aircraft applications, this specification refers to processing images generated by a radar sensor and an IR sensor. However, as will be explained, many different types, numbers, and combinations of sensors and sensor images can be processed. Accordingly, the example system and method explained in this specification can be used with many different applications.
Images and Sensors
Turning now to
More specifically,
In this instance both the IR sensor and the radar sensor generally provide the same or overlapping fields of view such that objects or conditions visible in both fields of view may be better detected by one sensor than the other sensor. Those of ordinary skill in the art will recognize that the system and method can be applied to images with different degrees of overlap or fields of view, as later described. Moreover, while the described embodiment provides a specific example of a system including radar and IR sensors and images, different types, numbers, and combinations of sensors and images can be utilized. For example, the system can also be used with ultraviolet (UV) sensors, one example UV sensor being available from Pulnix America, Inc., Sunnyvale, Calif. Further, one of the sensors can be based on an active or passive radio-frequency (RF) system such as an imaging radar or radiometer, operating in various RF bands including but not limited to 10, 35, 76, 94, and 220 GHz, one example of such a sensor being available from TRW, Inc., Redondo Beach, Calif. As a further example, a sensor can be an ultrasonic sensor, such as those ultrasonic sensors utilized in medical imaging applications available from General Electric Medical Systems Division, Waukesha, Wis. A sensor can also be a visible band sensor, e.g., a low-light level visible band sensor, Charged Coupled Device (CCD), or color or grayscale camera which can use natural or artificial illumination, available from Panasonic, Inc., Secaucus, N.J.
Further, the image fusion system can be configured to process images from a plurality of sensors, e.g., three, four, or other numbers of sensors. One possible combination of sensors includes two IR sensors and a radar sensor. The images from all of the images can be jointly processed and selectively combined into a processed image. For example, images A, B, and C can be selectively combined into processed or fused image D. Alternatively, two sensor images can be processed, the result of which is processed with a third sensor image to form a processed or fused image or its representative contrast map. For example, images A and B are combined into image C or an intermediate contrast map C that is subsequently selectively combined with image D or contrast map D to form fused image E or further intermediate contrast map, and so on, until all of the images are processed to form a fused image. Indeed, different combinations of different number of sensor images can be processed with different iterations of comparisons as desired or needed.
The selection of the type of the sensors may depend on the conditions and environment in which the sensor is used. As previously discussed, one type of sensor may be better suited for one environment, whereas another sensor may be better suited for a different environment. More specifically, certain types of sensors may provide clearer images depending on whether the environment is daylight, night, fog, rain, etc. and depending on whether the image is distant or near. For example, radar sensors typically provide better images in fog conditions compared to IR sensors, but may lack the photograph-like qualities of IR images.
Comparing Contrast Values of Image Regions
Image region contrast values are compared (step 204) by dividing images into regions, generating contrast maps based on the defined regions, and comparing the corresponding contrast map values using a selection criterion or criteria. The comparison is based on aligned or pre-registered images or images arranged to permit comparison of related image regions. Thus, if images that do not overlap are processed, they are pre-registered or aligned such that related regions are compared as described in further detail below. Contrast values are then selected (step 206), for example, on a selection criterion favoring the greater or maximum contrast values. Other selection criteria may also be utilized, for example, temporal persistence, brightness, color, etc.
Dividing Images Into Regions
Initially, sensor sensors are divided into sensor regions as illustrated in
For purposes of explanation,
However, those skilled in the art will recognize that the system and method can be utilized with different numbers, types, and combinations of sensor images having different degrees of overlap depending on the location, position, field of view, and detection capabilities of a sensor. In cases involving different degrees of overlap, the image regions can be aligned or pre-registered such that the comparisons can be performed.
For example, sensors can be positioned closely together (e.g., near the front or bottom of the aircraft) to detect essentially the same images, such as the runway scene illustrated in
As a further example, one sensor may detect a first image whereas a different sensor may detect most of the first image, but additional scene elements as well. This may occur when, for example, sensors are positioned apart from each other or are positioned to have different fields of view. In this instance, the selection process may be applied to some or all of the overlapping regions. The image regions are processed by application of a selection process or criteria such as contrast comparisons. The competing regions are compared, and the image regions are selected to form the processed or fused image. The image regions that are not overlapping or are not competing can be processed in different ways depending on, e.g., the quality of the source and fused or processed images, the types of sensors, and user and system needs. For example, non-overlapping images can be added to the processed image as filler or background. Alternatively, non-overlapping regions can be discarded and precluded from inclusion in the processed for fused image. In some cases, the overlapping regions may not be processed depending on the particular system and application.
Thus, the method and system can be utilized with images having different degrees of overlap and image regions having different degrees of alignment. The overlap and alignment variations may result from sensors having different detection capabilities and positions. However, for purposes of explanation, this specification and supporting figures refer to and illustrate images having a high degree of overlap with aligned, corresponding image regions. As a result, most or all of the image regions are competing image regions and processed with the selection criterion. However, the method and system can be configured to process other image region configurations having different degrees of overlap, alignment, and correspondence.
Generating Contrast Maps
As shown in
For this example radar map, the contrast values in the general top and bottom portions 702, 706 of the image/map are of a relatively lower value, and the contrast values in the general middle portion 704 are of a relatively higher value. For the example IR map, the contrast values in the general middle portion 714 are of a relatively lower value and the contrast values in the general top and bottom portions 712, 716 are relatively higher in value.
In accordance with the present invention, contrast maps including contrast values for each image region are generated via, e.g., a convolution with an appropriate kernel. One example convolution and kernel that can be utilized is a 2-dimensional (3×3) normalized convolution kernel:
Kc*S1(x,y),Kc*S2(x,y)
where
denotes a convolution;
x,y are spatial coordinates of the image, ranging from 0 to the image width (w) and height (h), respectively;
The example kernel Kc includes values that reflect a distance metric from its center. A contrast map is generated including contrast values for each image region of each image as a result of the convolution.
The processor can execute the convolution with a program in C-code or another programming language, or in dedicated integrated circuit hardware. Real-time implementation of the convolution can be achieved through the use of a Digital Signal Processor (DSP), Field Programmable Gate Arrays (FPGAs), Application Specific Integrated Circuits (ASICs) or other hardware-based means.
Selection of Contrast Values
Each image region associated with a selected contrast value is selected from each image and then combined (or “pieced together”) with other such selected image regions to form the processed or fused image, e.g., the fused image illustrated in FIG. 5C. Thus, in this example, the criteria for selecting image regions based on maximum contrast values can be stated as follows:
Fmax-con(x,y)=max{Kc*S1(x,y), Kc*S2(x,y)}
where the “maximum criteria” operation is performed or an arbitrary shaped region. Thus, the selection of image regions based on maximum contrast essentially serves as a pixel value resulting in a fused image that includes a combination or superset of image regions from different images. The image regions may, as a result of the selection process, be selected all from a single image or from multiple images depending on the content and contrast values of the images. Some sensor images may not contribute any image regions to the fused or processed image. For example, if a first image has all of its contrast values identified or selected, then the processed or fused image will be the same as the first image. As a further example, if contrast values are selected from second and third images but not the first image, then the fused image includes regions from the second and third images but not the first image. Thus, in the processed image having image regions A, B and C, image region A may be from sensor image 1, image region B may from sensor image 2 and image region C may be from sensor image 3.
The previously described example involving the application of a convolution results in the generation of two contrast maps. Indeed, other numbers and combinations of convolutions can be performed to generate multiple contrast maps for use in multiple comparisons or multiple sensor images. For example, referring to
Kc*S1 (x,y), Kc*S2 (x,y), Kc*S3 (x,y)
where the third sensor S3 is also an IR sensor, for example. Those persons of ordinary skill in the art will recognize that different kernels can be used with the same or different sensors. Thus, a process involving three convolutions can use, for example three different convolution kernels.
Then, corresponding contrast values of the three images are compared 930, and contrast values are selected 940 based on a selection criterion. The image regions selected 945 correspond to the selected contrast values. The selection of image regions based on maximum contrast value criteria can be expressed as follows:
Fmax-con(x,y)=max{Kc*S1(x,y), Kc*S2 (x,y), Kc*S3 (x,y)}
The selected regions from one or more of the sensor images are then pieced together to form a processed or fused image 950. Thus, in this example, all of the corresponding contrast values are compared together (three contrast values compared at the same time) to select the image region(s) having the maximum contrast value.
In an alternative embodiment, multiple iterations of convolutions can be performed to generate respective contrast maps, the values of which are compared in iterations to eventually form a processed image. For example, referring to
Thus, for example, a comparison 100 of contrast values in contrast maps 920 and 921 is performed resulting in a selection of a set of contrast values 1010 based on, e.g., greater or maximum contrast. The selected contrast values are selectively combined to form an intermediate image or contrast map 1030.
Contrast values in contrast map 1030 are then compared 1040 to contrast values in contrast map 922 from the third image 902. The contrast values are selected or identified 1050, and the image regions corresponding to the selected contrast values are selected 1055. The selected regions form the processed or fusedimage 1060. Those skilled in the art will recognize that different numbers of iterations or comparisons of different numbers of contrast maps can be performed with the same or different convolution kernel. Thus, the present image processing system and method provide flexible image fusion that is adaptable to different applications using convolution.
Correcting Luminance of Fused Image
The luminance or brightness of the fused image can be corrected or adjusted, if desirable, depending on the types of sensors utilized and the quality of the resulting sensor and fused images. Luminance correction is particularly useful when the fused image is not sufficiently clear to the pilot.
In the example involving radar and IR images, there may be noticeable artifacts in the fused image, as shown in FIG. 5C. The artifacts result from the brightness or luminance of the fused image being inconsistent, resulting in discontinuous luminance across the fused image. In this particular example, high-contrast regions selected from the radar image (central horizontal band in this example) are generally darker relative to the high-contrast regions from the IR image. The luminance distribution of the resulting processed or fused image varies between the luminance of the two input sensors. For example, the darker band across the center of the image is generally selected from the radar image, which, in that region, has higher contrast, but lower luminance than the IR image. This reduces the overall clarity of the fused image.
The luminance distribution within the fused image can be adjusted to generate a clearer fused image. Luminance adjustment is performed by determining average luminance values in regions of an image generated by a reference sensor, and adjusting the luminance of regions of the fused image based on the corresponding determined values. In the example images of
Reference Sensor
Luminance correction can be performed by selecting one sensor as a reference sensor and adjusting the luminance of the fused image to match or approximate the luminance distribution of the reference sensor. The reference sensor can be arbitrarily selected or based on the expected utility of a sensor in a particular situation. For example, a radar sensor generally provides more image detail in low-visibility conditions than an IR sensor. However, an IR sensor typically provides a more natural or photographic image, at least at close range.
For purposes of explanation, this specification describes the IR sensor I(x,y) as the reference sensor for luminance distribution, to capture the natural-looking characteristics of images from that sensor. However, the radar sensor or other sensors can be the reference sensor.
Determining Average Luminance
Adjusting luminance involves determining the average intensity in the reference sensor in specific image regions, such as, for example, strips along each image cross-section parallel the scene horizon. The scene horizon refers to the “actual” real-world horizon. The scene horizon may be at an angle relative to the image horizontal during a roll, bank or other motion of an aircraft.
The average luminance of each such strip of the reference sensor image is determined. Then, luminance values obtained from the determination are added to each corresponding strip of the fused image to adjust the luminance of the fused image. Further, if necessary, the degree of luminance can be weighted for a particular luminance adjustment effect. The weight λ can be used to reduce the effect of the luminance compensation, although a value of λ=1 has been determined to provide a sufficiently clear adjusted fused image in most situations.
Thus, the manner in which image is adjusted in a fused image can be expressed as follows:
where
Those persons of ordinary skill in the art will recognize that the reference sensor image can be sampled along different cross sections besides a horizontal cross section, and with different segments besides a strip across the image. The selection of the cross section and sampling segment may depend on various factors, including the types of sensors, sensor images, orientation of images, and application of the system or method. However, for purposes of explanation, this specification refers to cross-sectional sampling of strips of the reference sensor image, and correcting corresponding strips in the processed image.
An example of applying luminance adjustment is illustrated in FIG. 11. The runway scene portrayed in the fused image 1100 before luminance correction includes a number of artifacts that distort the processed or fused image. As a result, the runway scene is somewhat unclear, particularly in the middle portion of the image. The image 1110 represents the same image 1100 after luminance correction and selecting the IR sensor as the reference sensor.
As can be seen by comparing images 1100 (before luminance correction) and 1110 (after luminance correction), the luminance compensated image demonstrates less striking luminance variations in elevation, which otherwise tend to produce a noisy image. The result is a clearer, processed or fused image.
Luminance correction of the fused image can be performed by correcting different strips or regions of the fused image. For example, the mean luminance of the reference sensor is determined for an image line or strip in the reference sensor image. The determined mean luminance value from the reference sensor image is processed with, e.g. the previously stated luminance adjustment expression, to add it to each pixel in the corresponding fused image line or strip.
In an alternative embodiment, processing efficiency can be increased by using the mean or determined luminance value from one line of the reference sensor image and applying it as a correction to a line in the processed or fused image that is adjacent to a line in the fused image corresponding to the determined line in the reference sensor image (e.g., the next line above or below the corresponding determined line). Applying luminance values to the following line is generally acceptable since the mean typically does not substantially vary between successive image lines. However, this technique can be applied to adjust the next line above or below the subject line, or a number of lines separated from the reference line depending on luminance variation.
Luminance correction can also be adapted to situations in which the scene horizon is not parallel to the image horizontal, e.g., when an aircraft rolls or banks to one side. In this case, the scene horizontal angle and elevation are generally known from aircraft orientation sensors. Luminance correction can be calculated from the reference sensor, stored as a two-dimensional lookup table. The correction obtained from the lookup table is applied on a pixel-by-pixel basis to the fused image. In order to minimize latency and processing time, table values can be applied to the current frame based on values calculated during the previous frame, if sufficient memory storage resources for the full-image lookup table are available. These requirements can be approximately equal to the image frame size, for example, 320×240 bytes for an 8-bit per pixel sensor or other sizes depending on the details of the image produced by each sensor.
Spatial Pre-Filtering of Sensor Images
Regions or portions or sensor images can also be filtered to simplify processing of comparing contrast values and application of the selection criteria. The filtered regions can be represented as a number less than one to de-emphasize their contribution to the fused image, or a zero to remove them from contributing at all of the fused image, to simplify and reduce processing time.
Image regions that can be filtered include portions of the images that will not be included in the fused image, e.g., regions above a radar horizon in the case of a radar sensor. If a radar sensor is utilized, there is typically no useful information above the radar horizon (i.e., beyond the detection limit of the radar sensor) and little or no information in the near field (at least at higher altitudes). IR sensors are typically most effective at shorter ranges (near field), especially in weather conditions where the far-field cannot be detected due to the sensor's inability to penetrate obscurants such as rain or fog. Thus, with the example radar and IR sensors, radar image regions above the radar horizon and in the near field can be pre-filtered, and IR image regions in the far field can be pre-filtered. Other fields and regions may be suitable for filtering depending on the sensors, resulting images generated thereby and the needs of the user or system.
A general spatial filter is illustrated in
Once technique for filtering image regions is performed by selecting space-dependent α and β weighting functions. Continuing with the example involving radar and IR images, the weighting functions can be selected to overweight the radar image contribution in those regions where the radar signal is strongest, and, overweight the IR signal everywhere else.
The weighting function can be implemented through a spatial filter or other smoothing function that does not introduce unnecessary artifacts, e.g., a one-dimensional Gaussian weighting function as follows:
α(x,y)=αMe−b
β(x,y)=α1(1−e−b
where:
αM and αI determine the maximum amplitude of the Gaussian function (usually 1, but other values can also be used to overweight one sensor, or to compensate for the pedestal values, PM and PI);
bM and bI determine the Gaussian function width, i.e., the region of interest of the sensor or the region where the sensor information is clustered; and
y0 shifts the center of the Gaussian function vertically up and down in the image as required.
More detailed examples of such weighting functions are illustrated in
Referring to
Specifically, one example of a radar filter is configured with full-contrast cycle: 100% transparency at its maximum, in the center of the image and 0% at the upper and lower edges of the image. The example filter 1300 is constructed with a standard deviation of 50 pixels. Different filter configurations and functions can be utilized depending on the sensor used and the desired filtering effect.
Weighting sensor images in this manner essentially pre-selects image regions that contain useful and relevant information, and are therefore candidates for inclusion in the fused image. In addition, by filtering out regions where little information is available, processing time can be reduced.
The pre-selection or filtering of image regions is further illustrated in
Similarly,
For optimal filtering, the weighting function should account for state or operating parameters depending on the needs and design of the specific system. For example, as illustrated in
Thus, for example,
Combination of Pre-Filtering, Contrast-Based Image Fusion, and Luminance Correction
Depending on the sensors and resulting quality of sensor and fused images, the spatial pre-filtering and/or luminance correction processes can be applied to images as part of the image fusion processing.
If only contrast-based image fusion and luminance correction are performed, they will usually be completed in the recited order. If all three processes are performed, spatial pre-filtering will typically be performed first, then contrast-based sensor fusion, and finally luminance correction. These sequences typically result in more effective fused images while reducing processing time. Luminance correction should normally follow both pre-filtering and contrast-based fusion to most closely achieve the desired luminance distribution and to prevent image luminance distribution from changing as a result of subsequent processing. By applying these techniques in this manner, system performance is enhanced by minimizing pipeline delays and data latency. These enhancements can be particularly useful in time-intensive situations that involve the images, e.g., airborne, pilot-in-the-loop applications, or other applications that use real-time image processing.
Although references have been made in the foregoing description to a preferred embodiment, persons of ordinary skill in the art of designing image processing systems will recognize that insubstantial modifications, alterations, and substitutions can be made to the preferred embodiment described without departing from the invention as claimed in the accompanying claims.
Thus, while the preferred embodiment is primarily described as processing two images from radar and IR sensors in connection with an aircraft, those skilled in the art will recognize that images from other types, combinations, and numbers of sensors can be utilized. For example, instead of two sensors, the system can be implemented with three, four, five, or other numbers of sensors. Moreover, instead of a radar and an IR sensor, the system can process images from the same type of sensors at different wavelengths, ultraviolet (UV) sensors, sensors based on an active or passive radio-frequency (RF) system; an ultrasonic sensor, a visible band sensor, e.g., a low-light level visible band sensor, Charge Coupled Device (CCD), or a color or gray-scale camera. Moreover, persons of ordinary skill in the art will appreciate that the present image fusion system and method can be used in other applications besides processing aircraft images. For example, the system and method can be used in connection with other moving vehicles, medical procedures, surveillance, and other monitoring and image processing applications involving multiple images or sensors. Additionally, persons of ordinary skill in the art will recognize that a fused or processed image can be formed based on various selection criteria or processes, greater or maximum contrast values being example criteria.
Patent | Priority | Assignee | Title |
10198803, | May 01 2007 | Pictometry International Corp. | System for detecting image abnormalities |
10679331, | May 01 2007 | Pictometry International Corp. | System for detecting image abnormalities |
11100625, | May 01 2007 | Pictometry International Corp. | System for detecting image abnormalities |
11514564, | May 01 2007 | Pictometry International Corp. | System for detecting image abnormalities |
8385672, | May 01 2007 | Pictometry International Corporation | System for detecting image abnormalities |
8692804, | Aug 24 2010 | QUANTA COMPUTER INC. | Optical touch system and method |
9053558, | Jul 26 2013 | Method and system for fusing multiple images | |
9262818, | May 01 2007 | Pictometry International Corp. | System for detecting image abnormalities |
9299130, | Mar 15 2013 | Trustees of Tufts College | Methods and apparatus for image processing and analysis |
9633425, | May 01 2007 | Pictometry International Corp. | System for detecting image abnormalities |
9712350, | Nov 20 2014 | AT&T Intellectual Property I, L.P. | Transmission device with channel equalization and control and methods for use therewith |
9959609, | May 01 2007 | Pictometry International Corporation | System for detecting image abnormalities |
Patent | Priority | Assignee | Title |
4661986, | Jun 27 1983 | RCA Corporation | Depth-of-focus imaging process method |
4674125, | Jun 27 1983 | RCA Corporation | Real-time hierarchal pyramid signal processing apparatus |
5325449, | May 15 1992 | Sarnoff Corporation | Method for fusing images and apparatus therefor |
5510772, | |||
5686960, | Jan 14 1992 | REMOLEAUX TECHNOLOGIES AG, LLC | Image input device having optical deflection elements for capturing multiple sub-images |
5778108, | Jun 07 1996 | HEWLETT-PACKARD DEVELOPMENT COMPANY, L P | Method and system for detecting transitional markers such as uniform fields in a video signal |
5881163, | Mar 30 1993 | U.S. Philips Corporation | X-ray examination apparatus with an imaging arrangement having a plurality of image sensors |
6137498, | Jan 02 1997 | Runaway Technology, Inc. | Digital composition of a mosaic image |
6163309, | Jan 16 1998 | THRESHOLDS UNLIMITED, INC | Head up display and vision system |
6201899, | Oct 09 1998 | Sarnoff Corporation | Method and apparatus for extended depth of field imaging |
6208766, | Feb 09 1994 | Deutsche Aerospace AG | Process for influencing the quality images furnished by an electronic imaging system |
6320979, | Oct 06 1998 | Canon Kabushiki Kaisha | Depth of field enhancement |
6898331, | Aug 28 2002 | BAE SYSTEMS CONTROLS INC | Image fusion system and method |
20020025154, |
Executed on | Assignor | Assignee | Conveyance | Frame | Reel | Doc |
Apr 20 2006 | BAE Systems Controls, Inc. | (assignment on the face of the patent) | / | |||
Dec 27 2006 | BAE SYSTEMS AIRCRAFT CONTROLS INC | BAE SYSTEMS CONTROLS INC | MERGER SEE DOCUMENT FOR DETAILS | 024358 | /0986 | |
Dec 31 2006 | BAE SYSTEMS AIRCRAFT CONTROLS INC | BAE SYSTEMS CONTROLS INC | MERGER SEE DOCUMENT FOR DETAILS | 022100 | /0916 |
Date | Maintenance Fee Events |
Jan 07 2013 | REM: Maintenance Fee Reminder Mailed. |
Feb 08 2013 | M1552: Payment of Maintenance Fee, 8th Year, Large Entity. |
Feb 08 2013 | M1555: 7.5 yr surcharge - late pmt w/in 6 mo, Large Entity. |
Dec 30 2016 | REM: Maintenance Fee Reminder Mailed. |
May 24 2017 | EXP: Patent Expired for Failure to Pay Maintenance Fees. |
Date | Maintenance Schedule |
Jul 20 2013 | 4 years fee payment window open |
Jan 20 2014 | 6 months grace period start (w surcharge) |
Jul 20 2014 | patent expiry (for year 4) |
Jul 20 2016 | 2 years to revive unintentionally abandoned end. (for year 4) |
Jul 20 2017 | 8 years fee payment window open |
Jan 20 2018 | 6 months grace period start (w surcharge) |
Jul 20 2018 | patent expiry (for year 8) |
Jul 20 2020 | 2 years to revive unintentionally abandoned end. (for year 8) |
Jul 20 2021 | 12 years fee payment window open |
Jan 20 2022 | 6 months grace period start (w surcharge) |
Jul 20 2022 | patent expiry (for year 12) |
Jul 20 2024 | 2 years to revive unintentionally abandoned end. (for year 12) |