A method, apparatus and computer program product are provided for object identification and location correlation based on received images. A method is provided including receiving identity data associated with an object derived from one or more of a plurality of images determining an identity of an object in one or more of the plurality of images, receiving location information associated with a camera which captured the plurality of images, and correlating the identified object with a location for each of the respective images of the plurality of images.
|
1. A method comprising:
receiving identity data associated with an object derived from at least one image of a sequence of two or more images;
receiving location information associated with a camera which captured the sequence of images;
correlating the identified object with a location for a respective image of the sequence of images;
determining an edge of a field of view of the camera by which the identified object exits the field of view in the sequence of two or more images; and
determining a location offset of the identified object relative to the camera based on the determined edge of the field of view.
7. An apparatus comprising at least one processor and at least one memory including computer program code, the at least one memory and computer program code configured to, with the processor, cause the apparatus to at least:
receive identity data associated with an object derived from at least one image of a sequence of two or more images;
receive location information associated with a camera which captured the sequence of images;
correlate the identified object with a location for a respective image of the sequence of images;
determine an edge of a field of view of the camera by which the identified object exits the field of view in the sequence of two or more images; and
determine a location offset of the identified object relative to the camera based on the determined edge of the field of view.
13. A computer program product comprising at least one non-transitory computer-readable storage medium having computer-executable program code portions stored therein, the computer-executable program code portions comprising program code instructions configured to:
receive identity data associated with an object derived from at least one image of a sequence of two or more images;
receive location information associated with a camera which captured the sequence of images;
correlate the identified object with a location for a respective image of the sequence of images;
determine an edge of a field of view of the camera by which the identified object exits the field of view in the sequence of two or more images; and
determine a location offset of the identified object relative to the camera based on the determined edge of the field of view.
2. The method of
determining a location of the object based on the time stamp associated with the location information and the time stamp associated with a respective image.
3. The method of
4. The method of
determining a probable object location based on one or more correlated identified object locations.
5. The method of
updating or generating map data associated with the probable location of the object.
6. The method of
receiving the sequence of images including at least one object; and
determining an identity of an object in one or more of the sequence of images.
8. The apparatus of
determine a location of the object based on the time stamp associated with the location information and the time stamp associated with a respective image.
9. The apparatus of
10. The apparatus of
determine a probable object location based on one or more correlated identified object locations.
11. The apparatus of
update or generate map data associated with the probable location of the object.
12. The apparatus of
receive the sequence of images including at least one object; and
determine an identity of an object in one or more of the sequence of images.
14. The computer program product of
determine a location of the object based on the time stamp associated with the location information and the time stamp associated with the image.
15. The computer program product of
16. The computer program product of any of
determine a probable object location based on one or more correlated identified object locations.
17. The computer program product of
update or generate map data associated with the probable location of the object.
18. The computer program product of
receive the sequence of images including at least one object; and
determine an identity of an object in one or more of the sequence of images.
19. The method of
determining the location offset based at least in part on a focal length associated with the determined edge of the field of view.
20. The method of
|
An example embodiment of the present invention relates to map rendering and, more particularly, to object identification and location correlation based on images.
Current digital maps are typically built and updated using vehicles with expensive, high accuracy sensors, such as light detection and ranging (LIDAR), radar, multiple cameras, accelerometers, or the like. These mapping vehicles may be continuously deployed, driving on roads in order to map new geographical areas and update data for areas that have been previously digitally mapped. However, the road information, e.g. number of lanes, new lanes, closed lanes traffic signs, road surface conditions, or the like, may change faster than the mapping vehicles deployment rate. Therefore, digital maps may suffer from delayed updates.
A method and apparatus are provided in accordance with an example embodiment for object identification and location correlation based on images. In an example embodiment, a method is provided that includes receiving identity data associated with an object derived from one or more of a plurality of images, receiving location information associated a camera which captured the plurality of images, and correlating the identified object with a location for a respective image of the plurality of images.
In an example embodiment, the method also includes determining one or more images of the plurality of images in which the object exits the field of view and correlating the identified object with the location is based on the one or more images in which the object leaves the field of view. In some example embodiments of the method, respective images of the plurality of images include a time stamp and the location information includes a time stamp. The method also includes determining a location of the object based on the time stamp associated with the location information and the time stamp associated with the image.
In some example embodiments of the method, the determining the identity of the object further comprises applying pattern recognition to the image. In an example embodiment, the method also includes determining a portion of the one or more images in which the object exits the field of view and correlating the identified object with the location is further based on the portion of the one or more images in which the object exits the field of view.
In an example embodiment, the method also includes determining a probable object location based on one or more correlated identified object locations. In some example embodiments, the method also includes updating or generating map data associated with the probable location of the object. In an example embodiment, the method also includes, receiving the plurality of images including at least one object and determining an identity of an object in one or more of the plurality of images,
In another example embodiment an apparatus is provided including at least one processor and at least one memory including computer program code, with the at least one memory and computer program code configured to, with the processor, cause the apparatus to at least receive identity data associated with an object derived from one or more of a plurality of images, receive location information associated with a camera which captured the plurality of images, and correlate the identified object with a location for a respective image of the plurality of images.
In an example embodiment of the apparatus, the at least one memory and the computer program code are further configured to determine one or more images of the plurality of images in which the object exits the field of view, and correlating the identified object with the location is based on the one or more images in which the object leaves the field of view. In some example embodiments of the apparatus, respective images of the plurality of images include a time stamp and the location information includes a time stamp and the at least one memory and the computer program code are further configured to determine a location of the object based on the time stamp associated with the location information and the time stamp associated with the image.
In some example embodiments of the apparatus, the determining the identity of the object further comprises applying pattern recognition to the image. In an example embodiment of the apparatus, the at least one memory and the computer program code are further configured to determine a portion of the one or more images in which the object exits the field of view and correlating the identified object with the location is further based on the portion of the one or more images in which the object exits the field of view.
In an example embodiment of the apparatus, the at least one memory and the computer program code are further configured to determine a probable object location based on one or more correlated identified object locations. In some example embodiments of the apparatus, the at least one memory and the computer program code are further configured to update or generate map data associated with the probable location of the object. In an example embodiment of the apparatus, the at least one memory and the computer program code are further configured to receive the plurality of images including at least one object and determine an identity of an object in one or more of the plurality of images,
In a further example embodiment, a computer program product is provided including at least one non-transitory computer-readable storage medium having computer-executable program code portions stored therein, with the computer-executable program code portions comprising program code instructions configured to receive identity data associated with an object derived from one or more of a plurality of images, receive location information associated with a camera which captured the plurality of images, and correlate the identified object with a location for a respective image of the plurality of images.
In an example embodiment of the computer program product, the computer-executable program code portions further comprise program code instructions configured to determine one or more images of the plurality of images in which the object exits the field of view and correlating the identified object with the location is based on the one or more images in which the object leaves the field of view. In some example embodiments of the computer program product, respective images of the plurality of images include a time stamp and the location information includes a time stamp and the computer-executable program code portions further comprise program code instructions configured to determine a location of the object based on the time stamp associated with the location information and the time stamp associated with the image.
In some example embodiments of the computer program product, the determining the identity of the object further comprises applying pattern recognition to the image. In an example embodiment of the computer program product, the computer-executable program code portions further comprise program code instructions configured to determine a portion of the one or more images in which the object exits the field of view and correlating the identified object with the location is further based on the portion of the one or more images in which the object exits the field of view.
In an example embodiment of the computer program product, the computer-executable program code portions further comprise program code instructions configured to determine a probable object location based on one or more correlated identified object locations. In some example embodiments of the computer program product, the computer-executable program code portions further comprise program code instructions configured to update or generate map data associated with the probable location of the object. In an example embodiment of the computer program product, the computer-executable program code portions further comprise program code instructions configured to receive the plurality of images including at least one object and determine an identity of an object in one or more of the plurality of images
In yet a further example embodiment, an apparatus is provided including means for receiving identity data associated with an object derived from one or more of a plurality of images, means for receiving location information associated with a camera which captured the plurality of images, and means for correlating the identified object with a location for a respective image of the plurality of images.
In an example embodiment, the apparatus also includes means for determining one or more images of the plurality of images in which the object exits the field of view and correlating the identified object with the location is based on the one or more images in which the object leaves the field of view. In some example embodiments of the apparatus respective images of the plurality of images include a time stamp and the location information includes a time stamp, wherein the apparatus further comprises means for determining a location of the object based on the time stamp associated with the location information and the time stamp associated with the image.
In an example embodiment of the apparatus the means for determining the identity of the object further comprises means for applying pattern recognition to the image. In some example embodiments, the apparatus also includes means for determining a portion of the one or more images in which the object exits the field of view and correlating the identified object with the location is further based on the portion of the one or more images in which the object exits the field of view.
In some example embodiments, the apparatus also includes means for determining a probable object location based on one or more correlated identified object locations. In an example embodiment, the apparatus also includes means for updating or generating map data associated with the probable location of the object. In some example embodiments, the apparatus also includes means for receiving the plurality of images including at least one object and means for determining an identity of an object in one or more of the plurality of images,
Having thus described example embodiments of the invention in general terms, reference will now be made to the accompanying drawings, which are not necessarily drawn to scale, and wherein:
Some embodiments of the present invention will now be described more fully hereinafter with reference to the accompanying drawings, in which some, but not all, embodiments of the invention are shown. Indeed, various embodiments of the invention may be embodied in many different forms and should not be construed as limited to the embodiments set forth herein; rather, these embodiments are provided so that this disclosure will satisfy applicable legal requirements. Like reference numerals refer to like elements throughout. As used herein, the terms “data,” “content,” “information,” and similar terms may be used interchangeably to refer to data capable of being transmitted, received and/or stored in accordance with embodiments of the present invention. Thus, use of any such terms should not be taken to limit the spirit and scope of embodiments of the present invention.
Additionally, as used herein, the term ‘circuitry’ refers to (a) hardware-only circuit implementations (for example, implementations in analog circuitry and/or digital circuitry); (b) combinations of circuits and computer program product(s) comprising software and/or firmware instructions stored on one or more computer readable memories that work together to cause an apparatus to perform one or more functions described herein; and (c) circuits, such as, for example, a microprocessor(s) or a portion of a microprocessor(s), that require software or firmware for operation even if the software or firmware is not physically present. This definition of ‘circuitry’ applies to all uses of this term herein, including in any claims. As a further example, as used herein, the term ‘circuitry’ also includes an implementation comprising one or more processors and/or portion(s) thereof and accompanying software and/or firmware. As another example, the term ‘circuitry’ as used herein also includes, for example, a baseband integrated circuit or applications processor integrated circuit for a mobile phone or a similar integrated circuit in a server, a cellular network device, other network device, and/or other computing device.
As defined herein, a “computer-readable storage medium,” which refers to a non-transitory physical storage medium (for example, volatile or non-volatile memory device), can be differentiated from a “computer-readable transmission medium,” which refers to an electromagnetic signal.
A method, apparatus and computer program product are provided in accordance with an example embodiment for object identification and location correlation based on received images.
The UE 104 may be a mobile computing device such as a laptop computer, tablet computer, mobile phone, smart phone, navigation unit, personal data assistant, or the like. Additionally or alternatively the UE 104 may be a fixed computing device, such as a personal computer, computer workstation, kiosk, office terminal computer or system, or the like. The map server 102 may be one or more fixed or mobile computing devices.
In some embodiments the camera 110 and the location sensor 112 may be associated with a vehicle, such as an automobile, or a pedestrian. The camera 110 may be a fixed image camera a video camera, a radar, or LIDAR capable of generating an image, or the like. The camera 110 may be associated with other sensors, such as accelerometers. In an example embodiment, the camera 110 may be a single front facing camera attached or otherwise mounted or placed on the vehicle or pedestrian. In some example embodiments, the camera 110 and/or the location sensor 112 may be an original equipment manufacturer (OEM) system.
The map server 102 may receive one or more images from the camera 112 or from the image database 106. Additionally or alternatively, the map server 102 or image database 106 may receive the one or more images from a camera 112 associated with the UE 104. The images may be associated with a time stamp indicating the time at which the image was captured. The image may include at least one object, such as a traffic sign, person, vehicle, structure, lane indicator, or the like. The map server 102 may store the one or more images in the image database 106 for later processing or for system analysis.
The map server 102 may receive location information from a location sensor, such as a global position sensor (GPS), proximity sensor, such as radio frequency identification (RFID), or the like, or from the image data base. The location information may be associated with the camera 110 which captured the image. The location data may be associated with a time stamp indicating the time of the location information.
The map server 102 may apply pattern recognition techniques, such as a probabilistic classifier to the one or more images. The pattern recognition may identify one or more objects in the one or more images. For example, the pattern recognition may identify a street sign, a person, a structure, a lane indicator, a vehicle, or the like.
In an example embodiment, an OEM (original equipment manufacturer) server or UE 102 may receive the images and identify objects within the images. The map server 102 may receive identity data associated with an object derived from one or more images.
The map server 102 may correlate location information with an identified object. In an embodiment, the image and location information may be captured at the same time as indicated by the respective time stamps. In an instance in which the time stamps of the image and the location data are not the same, a location may be interpolated.
In an example embodiment, the map server may use location information associated with a time stamp prior to the image time stamp and location information associated with a time stamp after the image time stamp. The map server 102 may determine an interpolated location by averaging of the two locations, e.g. a location halfway between the two locations. In some example embodiments, the map server 102 may determine an interpolated location by using a weighted average of the two locations. For example, if the two location information time stamps are 5 seconds apart and the image time stamp correlates with second 4, the second location would be weighted more heavily than the first location such as by weighting the second location by 80% and the first location by 20%. The determined location may be closer to the second location than the first location.
In an example embodiment in which the camera 110 may be associated with other sensors, such as accelerometers, vehicle speed indicators, or the like, the indications of the rate of travel of the camera 110 may be used to interpolate the location of the camera at the time of capturing the image. For example, if a sensor indicates that the vehicle is traveling at a rate of 60 miles per hour (MPH) and the time stamp associated with the image is approximately 0.5-1 second after the time stamp associated with the location information, the interpolated location may be 1 foot from the location information. In an example embodiment, the interpolated direction of travel may be based on the last two or more location information points, for example the 1 foot would be in the same direction of travel as indicated by the two prior location information points. In some example embodiments, the map server 102 may receive and use map data for location interpolation. The map server 102 may determine a path of travel, such as a road, highway, trail, or the like, and plot the interpolated distance from the last location information received on the determined path.
In an example embodiment, the map server 102 may identify an image from the one or more images in which the identified object leaves the field of view of the image, e.g. an image at which the object is at the edge of the field of view and not in the subsequent image, an image that the object is partially in the field of view of the image and not in the subsequent image, or an image which the object is not present in the field of view of the image, but was present in the previous image. The identified image may be used for correlation of location information to identified objects, since the object is in a known location in the field of view, e.g. corresponding to the exit of the peripheral field of view.
In an example embodiment, the map server 102 may determine the angle of observation for an object in one or more images. The map server 102 may correlate the location information to the identified objects by triangulating or interpolating the object location based on the angle of observation for the one or more images and the location information.
In some example embodiments, the map server 102 may determine the portion of the image at which the identified object exited the field of view, for example top, left, right, bottom, or median positions, such as top left, bottom right, or the like. The map server may determine that the object is above, below, left, or right of the path of travel based on the portion of the image at which the object exits the field of view. For example, the offset may be applied in the direction of travel, or ahead of the location information, in an instance in which the object exits the top or bottom of the image, indicating that the object is above or below the vehicle or pedestrian path. In an instance in which the object exits the right or left of the field of view of the image, the offset may be applied to the left or right respectively of the location information.
The map server 102 may determine a location offset for the identified object. The location offset may be based on the focal length of the image, which may be a standard focal length or reported as a portion of the image. In some embodiments, the offset may be, additionally or alternatively, based on the placement, e.g. height, angle, or the like, of the camera on a vehicle or pedestrian. The location offset may be based, at least in part, on the focal length associated with the edge of the field of view. The portion of the image at which the identified object exits the field of view of the image may influence the direction of the location offset. For example, an offset of 5 feet in front of the camera location may be applied in an instance in which the object exited the field of view in the top middle. In an instance in which the object exited the field of view in the right middle position of the image, the location offset may be 5 feet in front of the camera location and 20 feet to the right of the camera location.
In an example embodiment, the location offset may be a predetermined projection distance, such as 10 feet, 20 feet, 50 feet, or any other distance, from a location plot cluster, discussed below. In an instance in which the offset is a predetermined projection distance, the predetermined projection distance may be a set value for objects identified as left, right, top, e.g. gantry mounted, or the like. For example, in an instance in which there is a median and break down lane, the left predetermined projection distance may be 10 feet, the right predetermined projection distance may be 20 feet, and the top predetermined projection distance may be 30 feet.
In an example embodiment, the map server may correlate the location information to an identified object for each observation, e.g. camera reporting occurrence. The camera reporting occurrences may be associated with a single camera 110 at different reporting occurrences, such as different days, hours apart, or the like, or associated with two or more cameras. The map server 102 may correlate an identified object with a location, by plotting an object location on a map, based on each of the camera reporting occurrences. In an example embodiment, the map server 102 may utilize map information and travel path information associated with the camera 110 to clarify plot locations occurring in intersections, or other complex topologies. For example, the map server may plot objects on the side of a road rather than in the middle of the road.
The map server 102 may determine the location of the identified object based on the plotted locations of the identified object. The map server 102 may determine the identified object probable location based on cluster analysis of the plotted location of the object on one or more maps received from the map database 108. In an example embodiment, the map server 102 may apply a density based cluster algorithm to plots of the identifier object to determine the location, e.g. probable location, of the object.
In an example embodiment, the map server 102 may determine the probable location of the identified object using a derived probability density function corresponding to a single observation, e.g. image from a single vehicle for a single pass of the vehicle by a sign. In some example embodiments, the map server 102 may have multiple observations, e.g. images, such as from several vehicles or several passes by the object location, and the map server 102 may generate a combined probability density for the object. The map server 102 may refine the probability density by performing a time dependent determination based on images in which the object is observed, and drives of vehicles by the same location where the object is not observed, which may account for newly added or removed objects. The map server 102 may determine the probable object location in an instance in which the combined probability density at that location satisfies a predetermined threshold.
The map server 102 may generate map data associated with the object based on the probable location. For example, the map server 102 may plot the probable location of the object on a highway map, trail map, or the like.
In some example embodiments, object changes may be determined and map data updated accordingly. For example, additional objects, removed objects, changes to an object, such as change of speed limit on a traffic sign, or the like. In an example embodiment, updates to changed objects may be time dependent. The time interval sensitivity may depend on the amount of available camera reporting occurrences, which may be a function of the frequency of traffic in the geographic area under investigation and the fraction of vehicles or pedestrians equipped with a camera 110 and location sensor 112. The camera reporting occurrences may be approximately the same for a given geographic area from day to day. Fluctuations due to weekend traffic or holidays may be controlled by removal or adjusting the weight of data collected during these periods. The variation in camera reporting occurrences from day to day may be controlled by a Pission distribution with a standard deviation given by the square root of the expected number of camera reporting occurrences. In an instance in which an expected value is not determined, an observed average may be used. For example, assuming 100 vehicles equipped with a camera 110 and location sensor 112 observe and report a specific traffic speed sign in a certain day, and assuming that the average number of camera reporting occurrences is approximately 100, the standard deviation would be 10 camera reporting occurrences. A new sign may be plotted, a sign removed or changed, in an instance in which a significant cluster is different than the object data present in previous maps, e.g. 11 or more camera reporting occurrences in the previous example.
In an example embodiment, the map server may determine an object placement efficiency. The object placement efficiency may be the number of identified objects within a predetermined radius of the true location of the object, such as 100 meters, 20 meters, or the like, divided by the total number of identifiable objects. The actual position or true position of the identified object may be entered by a user or determined by precision maps.
The map server 102 may cause the updated map to be stored in the map database 108 and/or cause the updated map data to be transmitted to a UE 104 to be displayed on a user interface. For example, the map may be displayed for navigational and traffic purposes.
A map server 102 may include or otherwise be associated with an apparatus 200 as shown in
As noted above, the apparatus 200 may be embodied by map server 102. However, in some embodiments, the apparatus may be embodied as a chip or chip set. In other words, the apparatus may comprise one or more physical packages (for example, chips) including materials, components and/or wires on a structural assembly (for example, a baseboard). The structural assembly may provide physical strength, conservation of size, and/or limitation of electrical interaction for component circuitry included thereon. The apparatus may therefore, in some cases, be configured to implement an embodiment of the present invention on a single chip or as a single “system on a chip.” As such, in some cases, a chip or chipset may constitute means for performing one or more operations for providing the functionalities described herein.
The processor 202 may be embodied in a number of different ways. For example, the processor may be embodied as one or more of various hardware processing means such as a coprocessor, a microprocessor, a controller, a digital signal processor (DSP), a processing element with or without an accompanying DSP, or various other processing circuitry including integrated circuits such as, for example, an ASIC (application specific integrated circuit), an FPGA (field programmable gate array), a microcontroller unit (MCU), a hardware accelerator, a special-purpose computer chip, or the like. As such, in some embodiments, the processor may include one or more processing cores configured to perform independently. A multi-core processor may enable multiprocessing within a single physical package. Additionally or alternatively, the processor may include one or more processors configured in tandem via the bus to enable independent execution of instructions, pipelining and/or multithreading.
In an example embodiment, the processor 202 may be configured to execute instructions stored in the memory device 204 or otherwise accessible to the processor. Alternatively or additionally, the processor may be configured to execute hard coded functionality. As such, whether configured by hardware or software methods, or by a combination thereof, the processor may represent an entity (for example, physically embodied in circuitry) capable of performing operations according to an embodiment of the present invention while configured accordingly. Thus, for example, when the processor is embodied as an ASIC, FPGA or the like, the processor may be specifically configured hardware for conducting the operations described herein. Alternatively, as another example, when the processor is embodied as an executor of software instructions, the instructions may specifically configure the processor to perform the algorithms and/or operations described herein when the instructions are executed. However, in some cases, the processor may be a processor of a specific device (for example, a mobile terminal or a fixed computing device) configured to employ an embodiment of the present invention by further configuration of the processor by instructions for performing the algorithms and/or operations described herein. The processor may include, among other things, a clock, an arithmetic logic unit (ALU) and logic gates configured to support operation of the processor.
The apparatus 200 of an example embodiment may also include a communication interface 206 that may be any means such as a device or circuitry embodied in either hardware or a combination of hardware and software that is configured to receive and/or transmit data from/to a communications device in communication with the apparatus, such as to facilitate communications with one or more user equipment 110, utility device, or the like. In this regard, the communication interface may include, for example, an antenna (or multiple antennas) and supporting hardware and/or software for enabling communications with a wireless communication network. Additionally or alternatively, the communication interface may include the circuitry for interacting with the antenna(s) to cause transmission of signals via the antenna(s) or to handle receipt of signals received via the antenna(s). In some environments, the communication interface may alternatively or also support wired communication. As such, for example, the communication interface may include a communication modem and/or other hardware and/or software for supporting communication via cable, digital subscriber line (DSL), universal serial bus (USB) or other mechanisms.
The apparatus 200 may also include a user interface 208 that may, in turn, be in communication with the processor 202 to provide output to the user and, in some embodiments, to receive an indication of a user input. As such, the user interface may include a display and, in some embodiments, may also include a keyboard, a mouse, a joystick, a touch screen, touch areas, soft keys, one or more microphones, a plurality of speakers, or other input/output mechanisms. In one embodiment, the processor may comprise user interface circuitry configured to control at least some functions of one or more user interface elements such as a display and, in some embodiments, a plurality of speakers, a ringer, one or more microphones and/or the like. The processor and/or user interface circuitry comprising the processor may be configured to control one or more functions of one or more user interface elements through computer program instructions (for example, software and/or firmware) stored on a memory accessible to the processor (for example, memory device 204, and/or the like).
Referring now to
As shown in block 504 of
In an example embodiment, blocks 502 and 504 may be performed by an OEM server or UE, such as UE 102. In an instance in which the OEM server performs blocks 502 and 504, the processor 202 may receive identity data associated with an object derived from one or more of the plurality of images. The processor may receive the identity data from a communications interface 206, which in turn, receives the identity information from the OEM server.
As shown at block 506, of
As shown at block 508 of
As shown at block 510 of
As shown at block 512 of
In an example embodiment, the location offset may be a predetermined projection distance, such as 10 feet, 20 feet, 50 feet, or any other distance, from a location plot cluster, discussed at block 516. In an instance in which the offset is a predetermined projection distance, the predetermined projection distance may be a set value objects identified as left, right, top, e.g. gantry mounted, or the like. For example, in an instance in which there is a median and break down lane, the left predetermined projection distance may be 10 feet, the right predetermined projection distance may be 20 feet, and the top predetermined projection distance may be 30 feet.
As shown at block 514 of
In an instance in which the time stamp associated with the image and the time stamp associated with the location information are different, the processor 202 may interpolate the location of the camera at the time the image was captured. The camera 110 may be associated with other sensors, such as accelerometers, vehicle speed indicators, or the like, and the indications of the rate of travel of the camera 110 may be used to interpolate the location of the camera at the time of capturing the image. For example, if a sensor indicates that the vehicle is traveling at a rate or 60 miles per hour (MPH) and the time stamp associated with the image is approximately 1 0.5-1 second after the time stamp associated with the location information, the interpolated location may be 1 foot from the location information. In an example embodiment, the interpolated direction of travel may be based on the last two or more location information points, for example the 1 foot would be in the same direction of travel as indicated by the two prior location information points. In some example embodiments, the processor 202 may receive and use map data, from a memory 204, such as map database 108, for location interpolation. The processor 202 may determine a path of travel, such as a road, highway, trail, or the like, and plot the interpolated distance from the last location information received on the determined path.
In some example embodiments, the processor 202 may apply the location offset to the location, for example the distance in front of the camera the object is likely to be based on focal length, mounting angle, mounting height or the like. Additionally or alternatively, the processor may apply a location offset based on the portion of the image in which the object exits the field of view. In an instance in which the object exits the image field of view at the top middle the offset may be 5 feet in front of the camera location. In an instance in which the object leaves the image field of view at the left middle, the offset may be 20 feet to the left and 5 feet in front of the camera location.
In an example embodiment, the processor 202 may determine the angle of observation for an object in one or more images. The processor 202 may correlate the location information to the identified objects by triangulating or interpolating the object location based on the angle of observation for the one or more images and the location information.
As shown at block 516 of
In an example embodiment, the processor 202 may determine the probable location of the identified object using a derived probability density function corresponding to a single observation, e.g. image from a single vehicle for a single pass of the vehicle by a sign. In some example embodiments, the processor 202 may have multiple observations, e.g. images, such as from several vehicles or several passes by the object location, and the map server 102 may generate a combined probability density for the object. The processor 202 may refine the probability density by performing a time dependent determination based on images in which the object is observed, and drives of vehicles by the same location where the object is not observed, which may account for newly added or removed objects. The processor 202 may determine the probable object location in an instance in which the combined probability density at that location satisfies a predetermined threshold.
In an example embodiment, the processor 202 may determine the probable location by applying the predetermined projection distance location offset to the plotted location cluster center.
As shown in block 518 of
As shown in block 520 of
In some example embodiments, the processor 202 may cause the map data to be stored in the memory, e.g. map database 108 and/or cause the communications interface 206, to cause the map data to be transmitted to a UE 104 to be displayed on a user interface. For example, the map may be displayed for navigational and traffic purposes.
Crowd sourcing object location information from vehicle and pedestrian cameras and sensors may allow for the map server to receive millions of observations, e.g. camera reporting occurrences. These camera reporting occurrences may be used to determine accurate locations of objects as quickly as they are reported. Generating or updating object locations based on vehicle sensor information may allow for maps to be updated for new objects or changed objects more quickly and efficiently than the current mapping vehicle deployment methods. In some examples, the map data may be updated in real time or near real time.
As described above,
Accordingly, blocks of the flowchart support combinations of means for performing the specified functions and combinations of operations for performing the specified functions for performing the specified functions. It will also be understood that one or more blocks of the flowchart, and combinations of blocks in the flowchart, can be implemented by special purpose hardware-based computer systems which perform the specified functions, or combinations of special purpose hardware and computer instructions.
In some embodiments, certain ones of the operations above may be modified or further amplified. Furthermore, in some embodiments, additional optional operations may be included, such as illustrated by the dashed outline of block 506, 508, 510, 512, 516, 518, and 520 in
Many modifications and other embodiments of the inventions set forth herein will come to mind to one skilled in the art to which these inventions pertain having the benefit of the teachings presented in the foregoing descriptions and the associated drawings. Therefore, it is to be understood that the inventions are not to be limited to the specific embodiments disclosed and that modifications and other embodiments are intended to be included within the scope of the appended claims. Moreover, although the foregoing descriptions and the associated drawings describe example embodiments in the context of certain example combinations of elements and/or functions, it should be appreciated that different combinations of elements and/or functions may be provided by alternative embodiments without departing from the scope of the appended claims. In this regard, for example, different combinations of elements and/or functions than those explicitly described above are also contemplated as may be set forth in some of the appended claims. Although specific terms are employed herein, they are used in a generic and descriptive sense only and not for purposes of limitation.
Giurgiu, Gavril Adrian, Jang, Dongwook, Anastassov, Anton
Patent | Priority | Assignee | Title |
10755118, | Jul 31 2018 | HERE Global B.V. | Method and system for unsupervised learning of road signs using vehicle sensor data and map data |
Patent | Priority | Assignee | Title |
6449384, | Oct 23 1998 | GOOGLE LLC | Method and apparatus for rapidly determining whether a digitized image frame contains an object of interest |
8086071, | Oct 30 2007 | HERE GLOBAL B V | System and method for revealing occluded objects in an image dataset |
20020047895, | |||
20020126913, | |||
20050216186, | |||
20050278111, | |||
20060271287, | |||
20070136259, | |||
20070140595, | |||
20100061591, | |||
20120155744, | |||
20130190981, | |||
20130287253, |
Executed on | Assignor | Assignee | Conveyance | Frame | Reel | Doc |
Dec 31 2014 | HERE Global B.V. | (assignment on the face of the patent) | / | |||
Jan 12 2015 | ANASTASSOV, ANTON | HERE GLOBAL B V | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 034920 | /0712 | |
Jan 12 2015 | GIURGIU, GAVRIL ADRIAN | HERE GLOBAL B V | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 034920 | /0712 | |
Jan 12 2015 | JANG, DONGWOOK | HERE GLOBAL B V | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 034920 | /0712 |
Date | Maintenance Fee Events |
Jul 30 2020 | M1551: Payment of Maintenance Fee, 4th Year, Large Entity. |
Jul 31 2024 | M1552: Payment of Maintenance Fee, 8th Year, Large Entity. |
Date | Maintenance Schedule |
Feb 14 2020 | 4 years fee payment window open |
Aug 14 2020 | 6 months grace period start (w surcharge) |
Feb 14 2021 | patent expiry (for year 4) |
Feb 14 2023 | 2 years to revive unintentionally abandoned end. (for year 4) |
Feb 14 2024 | 8 years fee payment window open |
Aug 14 2024 | 6 months grace period start (w surcharge) |
Feb 14 2025 | patent expiry (for year 8) |
Feb 14 2027 | 2 years to revive unintentionally abandoned end. (for year 8) |
Feb 14 2028 | 12 years fee payment window open |
Aug 14 2028 | 6 months grace period start (w surcharge) |
Feb 14 2029 | patent expiry (for year 12) |
Feb 14 2031 | 2 years to revive unintentionally abandoned end. (for year 12) |