One or more techniques and/or systems are provided for image blending and/or facilitating image transitions. In an example, a map interface displays map information, such as of a town, through a visualization. A first image having a first level of detail and a first image type (e.g., satellite imagery of the town) may be displayed through the visualization. While zoomed into the town, a second image having a second level of detail and a second image type (e.g., aerial imagery of the town) may be displayed through the visualization. Instead of merely transitioning the visualization from displaying the first image to displaying the second image (during zooming), which may otherwise provide a visually abrupt transition, one or more intermediate blended images, having intermediate levels of detail between the first image and the second image, may be generated and displayed during the transition between the first image and the second image.
|
1. A method for image blending, comprising:
generating an intermediate blended image between a first image, having a first image type and a first level of detail, and a second image having a second image type and a second level of detail, the intermediate blended image having an intermediate level of detail between the first level of detail and the second level of detail, the generating comprising:
performing structure transfer upon the first image to generate a new image;
downsampling the second image to generate a downsampled image having a downsampled size corresponding to a size of the new image;
computing a color difference between the new image and the downsampled image;
upsampling the color difference to generate an upsampled color difference; and
blending a gradient image using an interpolation weight and the upsampled color difference to generate the intermediate blended image, the blending the gradient image comprising determining a blending amount metric based upon a gradient value of the gradient image and then blending the gradient image based upon the blending amount metric; and
displaying the intermediate blended image through a visualization during a transition between display of the first image and display of the second image.
10. A method for image blending, comprising:
generating an intermediate blended image between a first image, having a first image type and a first level of detail, and a second image having a second image type and a second level of detail, the intermediate blended image having an intermediate level of detail between the first level of detail and the second level of detail, the generating comprising:
performing structure transfer upon the first image to generate a new image;
downsampling the second image to generate a downsampled image having a downsampled size corresponding to a size of the new image;
computing a color difference between the new image and the downsampled image;
upsampling the color difference to generate an upsampled color difference; and
blending a gradient image using an interpolation weight and the upsampled color difference to generate the intermediate blended image, the blending the gradient image comprising building a gaussian pyramid based upon successive fine-to-coarse downsampling performed upon the second image and then determining a blending amount metric based upon a pyramid level of the gradient image within the gaussian pyramid; and
displaying the intermediate blended image through a visualization during a transition between display of the first image and display of the second image.
15. A method for image blending, comprising:
generating an intermediate blended image between a first image, having a first image type and a first level of detail, and a second image having a second image type and a second level of detail, the intermediate blended image having an intermediate level of detail between the first level of detail and the second level of detail, the generating comprising:
performing structure transfer upon the first image to generate a new image;
downsampling the second image to generate a downsampled image having a downsampled size corresponding to a size of the new image;
computing a color difference between the new image and the downsampled image;
upsampling the color difference to generate an upsampled color difference; and
blending a gradient image using an interpolation weight and the upsampled color difference to generate the intermediate blended image, the blending the gradient image comprising: either determining a horizontal partial derivative with respect to an x-axis, the horizontal partial derivative indicative of a horizontal color difference between pixels along the x-axis or determining a vertical partial derivative with respect to a y-axis, the vertical partial derivative indicative of a vertical color difference between pixels along the y-axis, and then blending the gradient image based upon either the horizontal partial derivative or vertical partial derivative; and
displaying the intermediate blended image through a visualization during a transition between display of the first image and display of the second image.
2. The method of
3. The method of
generating a set of intermediate blended images between the first level of detail and the second level of detail; and
displaying the set of intermediate blended images through the visualization during the transition between display of the first image and display of the second image.
4. The method of
performing successive fine-to-coarse downsampling upon the second image to generate the downsampled image.
5. The method of
building a gaussian pyramid based upon the successive fine-to-coarse downsampling; and
utilizing the gaussian pyramid to generate the downsampled image.
6. The method of
7. The method of
8. The method of
determining a horizontal partial derivative with respect to an x-axis, the horizontal partial derivative indicative of a horizontal color difference between pixels along the x-axis; and
blending the gradient image based upon the horizontal partial derivative.
9. The method of
determining a vertical partial derivative with respect to a y-axis, the vertical partial derivative indicative of a vertical color difference between pixels along the y-axis; and
blending the gradient image based upon the vertical partial derivative.
11. The method of
12. The method of
generating a set of intermediate blended images between the first level of detail and the second level of detail; and
displaying the set of intermediate blended images through the visualization during the transition between display of the first image and display of the second image.
13. The method of
building a gaussian pyramid based upon successive fine-to-coarse downsampling performed on the second image; and
utilizing the gaussian pyramid to generate the downsampled image.
14. The method of
either determining a horizontal partial derivative with respect to an x-axis, the horizontal partial derivative indicative of a horizontal color difference between pixels along the x-axis or determining a vertical partial derivative with respect to a y-axis, the vertical partial derivative indicative of a vertical color difference between pixels along the y-axis; and
blending the gradient image based upon either the horizontal partial derivative or vertical partial derivative.
16. The method of
generating a set of intermediate blended images between the first level of detail and the second level of detail; and
displaying the set of intermediate blended images through the visualization during the transition between display of the first image and display of the second image.
17. The method of
building a gaussian pyramid based upon successive fine-to-coarse downsampling performed on the second image; and
utilizing the gaussian pyramid to generate the downsampled image.
18. The method of
19. The method of
|
Many applications and/or websites provide information through map interfaces. For example, a running website may display running routes through a running map visualization; a mobile map app may display driving directions on a road map visualization; etc. Such applications and/or websites may allow a user to zoom-in, zoom-out, and/or rotate a viewing angle of a map visualization.
This summary is provided to introduce a selection of concepts in a simplified form that are further described below in the detailed description. This summary is not intended to identify key factors or essential features of the claimed subject matter, nor is it intended to be used to limit the scope of the claimed subject matter.
Among other things, one or more systems and/or techniques for image blending and/or for facilitating image transitions are provided herein. In an example of image blending, an intermediate blended image between a first image, having a first image type and a first level of detail, and a second image having a second image type and a second level of detail may be generated. The intermediate blended image may have an intermediate level of detail between the first level of detail and the second level of detail. In an example of generating the intermediate blended image, structure transfer may be performed upon the first image to generate a new image. The second image may be downsampled to generate a downsampled image having a downsampled size corresponding to a size of the new image. A color difference between the new image and the downsampled image may be computed. The color difference may be upsampled to generate an upsampled color difference. A gradient image may be blended using an interpolation weight and the upsampled color difference to generate the intermediate blended image. The intermediate blended image may be displayed through a visualization during transition between display of the first image and display of the second image.
In an example of facilitating image transitions, a second image, having a second image type and a second level of detail, may be subsampled to generate a subsampled image. The subsampled image may have a subsampled size corresponding to a first image having a first image type and a first level of detail. A color distance metric between the subsampled image and the first image may be determined. A structure similarity metric between the subsampled image and the first image may be determined. A transition blending metric may be generated based upon the color distance metric and the structure similarity metric. A set of images may be sorted based upon the transition blending metric to generate a sorted set of images ordered from smooth transitions to abrupt transitions between displaying the first image and displaying the second image.
To the accomplishment of the foregoing and related ends, the following description and annexed drawings set forth certain illustrative aspects and implementations. These are indicative of but a few of the various ways in which one or more aspects may be employed. Other aspects, advantages, and novel features of the disclosure will become apparent from the following detailed description when considered in conjunction with the annexed drawings.
The claimed subject matter is now described with reference to the drawings, wherein like reference numerals are generally used to refer to like elements throughout. In the following description, for purposes of explanation, numerous specific details are set forth to provide an understanding of the claimed subject matter. It may be evident, however, that the claimed subject matter may be practiced without these specific details. In other instances, structures and devices are illustrated in block diagram form in order to facilitate describing the claimed subject matter.
One or more techniques and/or systems for image blending and/or facilitating image transitions are provided herein. Users may desire to zoom into and/or out of a map without abrupt visual transitions that may occur when transitioning from a first map zoom level derived from a first type of image such as satellite imagery and a second map zoom level derived from a second type of image such as aerial imagery (e.g., a user may zoom into a map of the United States from a satellite view, derived from satellite imagery, to an aerial view derived from aerial imagery). However, many client devices and/or map providers may lack processing power, storage, and/or bandwidth to provide smooth visual transitions from the first map zoom level to the second map zoom level due to the different types of imagery data sources and different types of imagery being accessed, retrieved, and/or utilized for the transition. Accordingly, as provided herein, one or more intermediate blended images, having an intermediate level of detail between a first level of detail of a first image (e.g., a satellite image) and a second level of detail of a second image (e.g., an aerial image), may be generated for display during a transition between display of the first image and display of the second image. Because the one or more intermediate blended images are generated for display during the transition, a visually smooth transition may be provided.
An embodiment of image blending is illustrated by an exemplary method 100 of
At 104, an intermediate blended image between a first image having a first image type and a first level of detail (e.g., the satellite image) and a second image having a second image type and a second level of detail (e.g., the aerial image) may be generated. The intermediate blended image may have an intermediate level of detail between the first level of detail and the second level of detail. In an example, the first image may correspond to a resolution above a high detail threshold for the first image type (e.g., a highest resolution satellite image before a level of detail transition to aerial imagery). In an example, the second image may correspond to a resolution below a low detail threshold for the second image type (e.g., a lowest resolution aerial image after a level of detail transition from the satellite imagery).
In an example of the generating, structure transfer is performed upon the first image to generate a new image, at 106. The new image may be generated, and the structure from the first image, such as the satellite image, may be transferred to the new image at the first level of detail of the first image (e.g., such that the new image has a similar resolution as the first image). In an example, structure may correspond to color differences (e.g., information related to different colors of red). At 108, the second image may be downsampled to generate a downsampled image having a downsampled size corresponding to a size of the new image. In an example, successive fine-to-coarse downsampling may be performed upon the second image (e.g., until the downsampled size of the downsampled image corresponds to the size of the new image, such as about 256 pixels by about 256 pixels) to generate the downsampled image. In an example, a Gaussian pyramid may be built based upon the successive fine-to-coarse downsampling. The Gaussian pyramid may be utilized to generate the downsampled image.
At 110, a color difference may be computed between the new image and the downsampled image (e.g., a Landsat or satellite color level difference). At 112, the color difference may be upsampled to generate an upsampled color difference. In an example of upsampling, an interpolation weight is calculated based upon a number of first levels of detail (e.g., a number of satellite levels of detail) and/or a number of second levels of detail (e.g., a number of aerial levels of detail). At 114, a gradient image may be blended using the interpolation weight and/or the upsampled color difference to generate the intermediate blended image. In an example, a blending amount metric (e.g., indicative of an amount of blending to perform based upon an amount of difference, such as color difference, between the new image and the downsampled image) may be determined based upon a gradient value of the gradient image. In this way, the gradient image may be blended based upon the blending amount metric (e.g., an amount of blending between the first image and the second image). In another example, the blending amount metric may be determined based upon a pyramid level of the gradient image within the Gaussian pyramid. In an example of blending, a horizontal partial derivative may be determined with respect to an x-axis. The horizontal partial derivative may be indicative of a horizontal color difference between pixels along the x-axis. A vertical partial derivative may be determined with respect to a y-axis. The vertical partial derivative may be indicative of a vertical color difference between pixels along the y-axis. The gradient image may be blended based upon the horizontal partial derivative and/or the vertical partial derivative.
In this way, the intermediate blended image may be generated based upon blending the gradient image. The intermediate blended image may cover an amount of ground area corresponding to an amount of ground area covered by the second image. In an example, a set of intermediate blended images between the first level of detail and the second level of detail may be generated, where the set of intermediate blended images covers an amount of ground area covered by a corresponding set of second images (e.g., set of aerial images). At 116, the intermediate blended image (e.g., and/or other intermediate blended images within the set of intermediate blended images) may be displayed through the visualization during a transition between display of the first image and display of the second image. Displaying the intermediate blended image during the transition may provide a relatively smoothly visual transition compared to an abrupt change that may otherwise occur from a direct transition from the first image to the second image. At 118, the method ends.
An embodiment of facilitating image transitions is illustrated by an exemplary method 500 of
In an example of sorting the set of images, the second image, having a second image type and the second level of detail, may be subsampled to generate a subsampled image, at 504. The subsampled image may have a subsampled size corresponding to a size of the first image having a first image type and the first level of detail. At 506, a color distance metric between the subsampled image and the first image may be determined (e.g., distance between DC components (e.g., Fourier transforms) of the subsampled image and the first image, which may corresponding to a normalized color distance). At 508, a structure similarity metric between the subsampled image and the first image may be determined (e.g., structure similarity between DC components of the subsampled image and the first image, which may correspond to texture differences).
At 510, a transition blending metric may be generated based upon the color distance metric and/or the structure similarity metric (e.g., the structure similarity metric may be subtracted from the color distance metric). The larger the transition blending metric, the more dissimilarity between the subsampled image and the first image. The transition blending metric may indicate how many levels of detail to blend between the first level of detail of the first image and the second level of detail of the second image.
At 512, the set of images may be sorted based upon the transition blending metric to create a sorted set of images ordered from smooth transitions to abrupt transitions between displaying the first image and displaying the second image. In an example, a transition outlier threshold may be defined based upon a user threshold for outliers. The set of images may be sorted based upon the transition outlier threshold. The sorted set of images may be represented as a sorted array of quadkeys, where a quadkey corresponds to an image within the sorted set of images. In this way, the sorted set of images may comprise intermediate blended images sorted based upon transitional smoothness, which may aid a user in efficiently identifying potential issues with intermediate blended images. At 514, the method ends.
According to an aspect of the instant disclosure, a method for image blending is provided. The method includes generating an intermediate blended image between a first image, having a first image type and a first level of detail, and a second image having a second image type and a second level of detail. The intermediate blended image may have an intermediate level of detail between the first level of detail and the second level of detail. The generating comprises performing structure transfer upon the first image to generate a new image. The second image may be downsampled to generate a downsampled image having a downsampled size corresponding to a size of the new image. A color difference between the new image and the downsampled image may be computed. The color difference may be upsampled to generate an upsampled color difference. A gradient image may be blended using an interpolation weight and the upsampled color difference to generate the intermediate blended image. The intermediate blended image may be displayed through a visualization during a transition between display of the first image and display of the second image.
According to an aspect of the instant disclosure, a method for facilitating image transitions is provided. The method includes subsampling a second image, having a second image type and a second level of detail, to generate a subsampled image. The subsampled image may have a subsampled size corresponding to a size of a first image having a first image type and first level of detail. A color distance metric between the subsampled image and the first image may be determined. A structure similarity metric between the subsampled image and the first image may be determined. A transition blending metric may be generated based upon the color distance metric and the structure similarity metric. A set of images may be sorted based upon the transition blending metric to generate a sorted set of images ordered from smooth transitions to abrupt transitions between displaying the first image and displaying the second image.
According to an aspect of the instant disclosure, a system for image blending is provided. The system includes an image blending component configured to generate an intermediate blended image between a first image, having a first image type and a first level of detail, and a second image having a second image type and a second level of detail. The intermediate blended image may have an intermediate level of detail between the first level of detail and the second level of detail. The image blending component may be configured to perform structure transfer upon the first image to generate a new image. The image blending component may be configured to downsample the second image to generate a downsampled image having a downsampled size corresponding to a size of the new image. The image blending component may be configured to compute a color difference between the new image and the downsampled image. The image blending component may be configured to upsample the color difference to generate an upsampled color difference. The image blending component may be configured to blend a gradient image using an interpolation weight and the upsampled color difference to generate the intermediate blended image. The system includes an image visualization component configured to display the intermediate blended image through a visualization during a transition between display of the first image and display of the second image.
According to an aspect of the instant disclosure, a means for image blending is provided. An intermediate blended image between a first image, having a first image type and a first level of detail, and a second image, having a second image type and a second level of detail, may be generated by the means for image blending. The intermediate blended image may have an intermediate level of detail between the first level of detail and the second level of detail. The generating comprises performing structure transfer upon the first image to generate a new image. The second image may be downsampled by the means for image blending to generate a downsampled image having a downsampled size corresponding to a size of the new image. A color difference between the new image and the downsampled image may be computed by the means for image blending. The color difference may be upsampled by the means for image blending to generate an upsampled color difference. A gradient image may be blended by the means for image blending using an interpolation weight and the upsampled color difference to generate the intermediate blended image. The intermediate blended image may be displayed by the means for image blending through a visualization during a transition between display of the first image and display of the second image.
According to an aspect of the instant disclosure, a means for facilitating image transitions is provided. A second image, having a second image type and a second level of detail, may be subsampled by the means for facilitating image transitions to generate a subsampled image. The subsampled image may have a subsampled size corresponding to a first image having a first image type and first level of detail. A color distance metric between the subsampled image and the first image may be determined by the means for facilitating image transitions. A structure similarity metric between the subsampled image and the first image may be determined by the means for facilitating image transitions. A transition blending metric may be generated by the means for facilitating image transitions based upon the color distance metric and the structure similarity metric. A set of images may be sorted by the means for facilitating image transitions based upon the transition blending metric to generate a sorted set of images ordered from smooth transitions to abrupt transitions between displaying the first image and displaying the second image.
Still another embodiment involves a computer-readable medium comprising processor-executable instructions configured to implement one or more of the techniques presented herein. An example embodiment of a computer-readable medium or a computer-readable device is illustrated in
Although the subject matter has been described in language specific to structural features and/or methodological acts, it is to be understood that the subject matter defined in the appended claims is not necessarily limited to the specific features or acts described above. Rather, the specific features and acts described above are disclosed as example forms of implementing at least some of the claims.
As used in this application, the terms “component,” “module,” “system”, “interface”, and/or the like are generally intended to refer to a computer-related entity, either hardware, a combination of hardware and software, software, or software in execution. For example, a component may be, but is not limited to being, a process running on a processor, a processor, an object, an executable, a thread of execution, a program, and/or a computer. By way of illustration, both an application running on a controller and the controller can be a component. One or more components may reside within a process and/or thread of execution and a component may be localized on one computer and/or distributed between two or more computers.
Furthermore, the claimed subject matter may be implemented as a method, apparatus, or article of manufacture using standard programming and/or engineering techniques to produce software, firmware, hardware, or any combination thereof to control a computer to implement the disclosed subject matter. The term “article of manufacture” as used herein is intended to encompass a computer program accessible from any computer-readable device, carrier, or media. Of course, many modifications may be made to this configuration without departing from the scope or spirit of the claimed subject matter.
Although not required, embodiments are described in the general context of “computer readable instructions” being executed by one or more computing devices. Computer readable instructions may be distributed via computer readable media (discussed below). Computer readable instructions may be implemented as program modules, such as functions, objects, Application Programming Interfaces (APIs), data structures, and the like, that perform particular tasks or implement particular abstract data types. Typically, the functionality of the computer readable instructions may be combined or distributed as desired in various environments.
In other embodiments, device 812 may include additional features and/or functionality. For example, device 812 may also include additional storage (e.g., removable and/or non-removable) including, but not limited to, magnetic storage, optical storage, and the like. Such additional storage is illustrated in
The term “computer readable media” as used herein includes computer storage media. Computer storage media includes volatile and nonvolatile, removable and non-removable media implemented in any method or technology for storage of information such as computer readable instructions or other data. Memory 818 and storage 820 are examples of computer storage media. Computer storage media includes, but is not limited to, RAM, ROM, EEPROM, flash memory or other memory technology, CD-ROM, Digital Versatile Disks (DVDs) or other optical storage, magnetic cassettes, magnetic tape, magnetic disk storage or other magnetic storage devices, or any other medium which can be used to store the desired information and which can be accessed by device 812. Computer storage media does not, however, include propagated signals. Rather, computer storage media excludes propagated signals. Any such computer storage media may be part of device 812.
Device 812 may also include communication connection(s) 826 that allows device 812 to communicate with other devices. Communication connection(s) 826 may include, but is not limited to, a modem, a Network Interface Card (NIC), an integrated network interface, a radio frequency transmitter/receiver, an infrared port, a USB connection, or other interfaces for connecting computing device 812 to other computing devices. Communication connection(s) 826 may include a wired connection or a wireless connection. Communication connection(s) 826 may transmit and/or receive communication media.
The term “computer readable media” may include communication media. Communication media typically embodies computer readable instructions or other data in a “modulated data signal” such as a carrier wave or other transport mechanism and includes any information delivery media. The term “modulated data signal” may include a signal that has one or more of its characteristics set or changed in such a manner as to encode information in the signal.
Device 812 may include input device(s) 824 such as keyboard, mouse, pen, voice input device, touch input device, infrared cameras, video input devices, and/or any other input device. Output device(s) 822 such as one or more displays, speakers, printers, and/or any other output device may also be included in device 812. Input device(s) 824 and output device(s) 822 may be connected to device 812 via a wired connection, wireless connection, or any combination thereof. In one embodiment, an input device or an output device from another computing device may be used as input device(s) 824 or output device(s) 822 for computing device 812.
Components of computing device 812 may be connected by various interconnects, such as a bus. Such interconnects may include a Peripheral Component Interconnect (PCI), such as PCI Express, a Universal Serial Bus (USB), firewire (IEEE 1394), an optical bus structure, and the like. In another embodiment, components of computing device 812 may be interconnected by a network. For example, memory 818 may be comprised of multiple physical memory units located in different physical locations interconnected by a network.
Those skilled in the art will realize that storage devices utilized to store computer readable instructions may be distributed across a network. For example, a computing device 830 accessible via a network 828 may store computer readable instructions to implement one or more embodiments provided herein. Computing device 812 may access computing device 830 and download a part or all of the computer readable instructions for execution. Alternatively, computing device 812 may download pieces of the computer readable instructions, as needed, or some instructions may be executed at computing device 812 and some at computing device 830.
Various operations of embodiments are provided herein. In one embodiment, one or more of the operations described may constitute computer readable instructions stored on one or more computer readable media, which if executed by a computing device, will cause the computing device to perform the operations described. The order in which some or all of the operations are described should not be construed as to imply that these operations are necessarily order dependent. Alternative ordering will be appreciated by one skilled in the art having the benefit of this description. Further, it will be understood that not all operations are necessarily present in each embodiment provided herein. Also, it will be understood that not all operations are necessary in some embodiments.
Further, unless specified otherwise, “first,” “second,” and/or the like are not intended to imply a temporal aspect, a spatial aspect, an ordering, etc. Rather, such terms are merely used as identifiers, names, etc. for features, elements, items, etc. For example, a first object and a second object generally correspond to object A and object B or two different or two identical objects or the same object.
Moreover, “exemplary” is used herein to mean serving as an example, instance, illustration, etc., and not necessarily as advantageous. As used herein, “or” is intended to mean an inclusive “or” rather than an exclusive “or”. In addition, “a” and “an” as used in this application are generally be construed to mean “one or more” unless specified otherwise or clear from context to be directed to a singular form. Also, at least one of A and B and/or the like generally means A or B and/or both A and B. Furthermore, to the extent that “includes”, “having”, “has”, “with”, and/or variants thereof are used in either the detailed description or the claims, such terms are intended to be inclusive in a manner similar to the term “comprising”.
Also, although the disclosure has been shown and described with respect to one or more implementations, equivalent alterations and modifications will occur to others skilled in the art based upon a reading and understanding of this specification and the annexed drawings. The disclosure includes all such modifications and alterations and is limited only by the scope of the following claims. In particular regard to the various functions performed by the above described components (e.g., elements, resources, etc.), the terms used to describe such components are intended to correspond, unless otherwise indicated, to any component which performs the specified function of the described component (e.g., that is functionally equivalent), even though not structurally equivalent to the disclosed structure. In addition, while a particular feature of the disclosure may have been disclosed with respect to only one of several implementations, such feature may be combined with one or more other features of the other implementations as may be desired and advantageous for any given or particular application.
Mehta, Bimal K., Schickler, Wolfgang, Muktinutalapati, Kartik, Liu, Dingding, Barclay, Thomas Donald, Kaowthumrong, Khomkrit
Patent | Priority | Assignee | Title |
Patent | Priority | Assignee | Title |
5187754, | Apr 30 1991 | Intel Corporation | Forming, with the aid of an overview image, a composite image from a mosaic of images |
6469710, | Sep 25 1998 | Microsoft Technology Licensing, LLC | Inverse texture mapping using weighted pyramid blending |
7991226, | Oct 12 2007 | Pictometry International Corporation | System and process for color-balancing a series of oblique images |
8189959, | Apr 17 2008 | Microsoft Technology Licensing, LLC | Image blending using multi-splines |
8340415, | Apr 05 2010 | Microsoft Technology Licensing, LLC | Generation of multi-resolution image pyramids |
8340423, | Sep 29 2009 | GOOGLE LLC | Enhancing digital image mosaics using region-statistics |
8374428, | Dec 05 2010 | Microsoft Technology Licensing, LLC | Color balancing for partially overlapping images |
8379972, | Dec 01 2009 | Adobe Inc | Color decontamination for image compositing |
8547389, | Apr 05 2010 | Microsoft Technology Licensing, LLC | Capturing image structure detail from a first image and color from a second image |
8577139, | Apr 28 2011 | National Central University | Method of orthoimage color correction using multiple aerial images |
8625890, | Oct 17 2011 | GOOGLE LLC | Stylizing geographic features in photographic images based on image content |
8649596, | Oct 12 2007 | PICTOMETRY INTERNATIONAL CORP | System and process for color-balancing a series of oblique images |
8665266, | Jun 23 2010 | The United States of America, as represented by the Secretary of the Navy | Global visualization process terrain database builder |
20070024931, | |||
20110243438, | |||
20130002810, | |||
20130321399, | |||
20130322702, | |||
20150363641, |
Executed on | Assignor | Assignee | Conveyance | Frame | Reel | Doc |
Aug 26 2014 | MEHTA, BIMAL K | Microsoft Corporation | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 033912 | /0909 | |
Aug 28 2014 | LIU, DINGDING | Microsoft Corporation | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 033912 | /0909 | |
Sep 25 2014 | KAOWTHUMRONG, KHOMKRIT | Microsoft Corporation | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 033912 | /0909 | |
Oct 07 2014 | MUKTINUTALAPATI, KARTIK | Microsoft Corporation | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 033912 | /0909 | |
Oct 08 2014 | Microsoft Technology Licensing, LLC | (assignment on the face of the patent) | / | |||
Oct 08 2014 | BARCLAY, THOMAS DONALD | Microsoft Corporation | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 033912 | /0909 | |
Oct 08 2014 | SCHICKLER, WOLFGANG | Microsoft Corporation | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 033912 | /0909 | |
Jul 02 2015 | Microsoft Corporation | Microsoft Technology Licensing, LLC | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 036100 | /0048 |
Date | Maintenance Fee Events |
Feb 03 2021 | M1551: Payment of Maintenance Fee, 4th Year, Large Entity. |
Date | Maintenance Schedule |
Aug 15 2020 | 4 years fee payment window open |
Feb 15 2021 | 6 months grace period start (w surcharge) |
Aug 15 2021 | patent expiry (for year 4) |
Aug 15 2023 | 2 years to revive unintentionally abandoned end. (for year 4) |
Aug 15 2024 | 8 years fee payment window open |
Feb 15 2025 | 6 months grace period start (w surcharge) |
Aug 15 2025 | patent expiry (for year 8) |
Aug 15 2027 | 2 years to revive unintentionally abandoned end. (for year 8) |
Aug 15 2028 | 12 years fee payment window open |
Feb 15 2029 | 6 months grace period start (w surcharge) |
Aug 15 2029 | patent expiry (for year 12) |
Aug 15 2031 | 2 years to revive unintentionally abandoned end. (for year 12) |