Image feature trackability ranking systems and methods are disclosed. A method of establishing a trackability ranking order from tracked image features within a training video sequence at a tracking analysis device includes establishing a tracking region within the training video sequence using a feature detection algorithm. Trajectories of tracked image features within the tracking region are compiled using a feature tracking algorithm. Saliency metrics are assigned to each one of the trajectories of tracked image features based on one or more feature property measurements within the tracking region, and a trackability ranking algorithm that is a function of the saliency metrics and a defined feature trajectory ranking associated with the training video sequence is determined, the trackability ranking algorithm being usable for ranking, based on trackability, tracked image features within another video sequence.
|
39. A method for establishing a trackability ranking order from tracked observed image features, the method comprising:
accessing a database storing a trackability ranking algorithm established using tracked training image features within a training video sequence;
establishing a tracking region within a query video sequence using a feature detection algorithm;
compiling trajectories of tracked observed image features within the tracking region using a feature tracking algorithm;
assigning saliency metrics to (1) each one of the trajectories of the tracked observed image features based on one or more feature property measurements within the tracking region, and (2) a surrounding environment of the tracked observed image features in the tracking region;
determining a trackability ranking order for the observed image features within the query video sequence using the trackability ranking algorithm, wherein the trackability ranking algorithm comprises at least sorting the trajectories of the tracked observed image features by a result of a linear combination of the saliency metrics; and
selecting a subset of the tracked observed image features according to the trackability ranking order to facilitate at least one of an image-based recognition search, a video segmentation, or a pose estimation and mapping system within another video sequence according to the trackability ranking order using the trackability ranking algorithm.
1. A tracking analysis device comprising:
a communication interface configured to access a database storing a trackability ranking algorithm established using tracked training image features within a training video sequence; and
a processor coupled with the database via the communication interface and configured to:
establish a tracking region within a query video sequence using a feature detection algorithm;
compile trajectories of tracked observed image features within the tracking region using a feature tracking algorithm;
assign saliency metrics to (1) each one of the trajectories of the tracked observed image features based on one or more feature property measurements within the tracking region, and (2) a surrounding environment of the tracked observed image features in the tracking region;
determine a trackability ranking order for the observed image features within the query video sequence using the trackability ranking algorithm, wherein the trackability ranking algorithm comprises at least sorting the trajectories of the tracked observed image features by a result of a linear combination of the saliency metrics; and
select a subset of the tracked observed image features according to the trackability ranking order to facilitate at least one of an image-based recognition search, a video segmentation, or a pose estimation and mapping system within another video sequence according to the trackability ranking order using the trackability ranking algorithm.
40. A computer program product embedded in a non-transitory computer readable medium comprising instructions executable by a computer processor for establishing a trackability ranking order from tracked observed image features, the instructions being executable by a computer processor to execute processing comprising:
accessing a database storing a trackability ranking algorithm established using tracked training image features within a training video sequence;
establishing a tracking region within a query video sequence using a feature detection algorithm;
compiling trajectories of tracked observed image features within the tracking region using a feature tracking algorithm;
assigning saliency metrics to (1) each one of the trajectories of the tracked observed image features based on one or more feature property measurements within the tracking region, and (2) a surrounding environment of the tracked observed image features in the tracking region;
determining a trackability ranking order for the observed image features within the query video sequence using the trackability ranking algorithm, wherein the trackability ranking algorithm comprises at least sorting the trajectories of the tracked observed image features by a result of a linear combination of the saliency metrics; and
selecting a subset of the tracked observed image features according to the trackability ranking order to facilitate at least one of an image-based recognition search, a video segmentation, or a pose estimation and mapping system within another video sequence according to the trackability ranking order using the trackability ranking algorithm.
41. A vehicle, comprising:
a tangible, non-transitory, computer-readable memory having a trackability ranking algorithm stored thereon, the trackability ranking algorithm established using tracked training image features within a training video sequence;
one or more optical sensors configured to capture a query video sequence; and
at least one tracking analysis device comprising a processor communicatively coupled with the tangible, non-transitory computer-readable memory and the one or more optical sensors, which, upon execution of software instructions, is configured to:
establish a tracking region within the query video sequence using a feature detection algorithm;
compile trajectories of tracked observed image features within the tracking region using a feature tracking algorithm;
assign saliency metrics to (1) each one of the trajectories of the tracked observed image features based on one or more feature property measurements within the tracking region, and (2) a surrounding environment of the tracked observed image features in the tracking region;
determine a trackability ranking order for the observed image features within the query video sequence using the trackability ranking algorithm, wherein the trackability ranking algorithm comprises at least sorting the trajectories of the tracked observed image features by a result of a linear combination of the saliency metrics; and
select a subset of the tracked observed image features according to the trackability ranking order to facilitate at least one of an image-based recognition search, a video segmentation, or a pose estimation and mapping system within another video sequence according to the trackability ranking order using the trackability ranking algorithm.
42. A server comprising:
a tangible, non-transitory, computer-readable memory having a ranking module stored thereon, the ranking module comprising a trackability ranking algorithm comprising a data structure having weights associated with training image feature saliency metrics that are established using tracked training image features within a training video sequence; and
a processor communicatively coupled with the tangible, non-transitory computer-readable memory which, upon execution of software instructions, is configured to:
install the ranking module on at least one tracking device such that the at least one tracking device, upon execution of the ranking module, is configured to:
establish a tracking region within a query video sequence using a feature detection algorithm;
compile trajectories of tracked observed image features within the tracking region using a feature tracking algorithm;
assign saliency metrics to (1) each one of the trajectories of the tracked observed image features based on one or more feature property measurements within the tracking region, and (2) a surrounding environment of the tracked observed image features in the tracking region;
determine a trackability ranking order for the observed image features within the query video sequence using the trackability ranking algorithm, wherein the trackability ranking algorithm comprises at least sorting the trajectories of the tracked observed image features by a result of a linear combination of the saliency metrics; and
select a subset of the tracked observed image features according to the trackability ranking order to facilitate at least one of an image-based recognition search, a video segmentation, or a pose estimation and mapping system within another video sequence according to the trackability ranking order using the trackability ranking algorithm.
3. The device of
4. The device of
5. The device of
6. The device of
7. The device of
8. The device of
9. The device of
10. The device of
11. The device of
12. The device of
13. The device of
15. The method of
16. The method of
17. The method of
18. The method of
19. The method of
20. The method of
21. The method of
22. The method of
23. The method of
24. The device of
25. The device of
26. The device of
27. The device of
28. The device of
29. The device of
30. The device of
31. The device of
33. The device of
34. The device of
35. The device of
36. The method of
37. The method of
38. The device of
|
This application is a continuation of U.S. application Ser. No. 14/795,515, filed Jul. 9, 2015, which claims the benefit of U.S. Provisional Application No. 62/022,558, filed Jul. 9, 2014. The entire contents of these applications are hereby incorporated herein by reference.
This disclosure relates generally to image-based object tracking technologies.
The background description includes information that may be useful in understanding the present invention. It is not an admission that any of the information provided herein is prior art or relevant to the presently claimed invention, or that any publication specifically or implicitly referenced is prior art.
Tracking objects in video streams requires detecting trackable image features in the video data, then detecting movement of the image features from one frame of the video to another. Feature detection algorithms can be quite slow relative to the typical frame rate of video streams. As the frame rate increases or the frame size increases, the feature detection algorithms and tracking algorithms can become computationally bound. In such cases, image features cannot be tracked at high frame rates, which reduces tracking fidelity. Additionally, typical feature detection algorithms are based on properties of static images that are independent of the tracking algorithm applied.
It would be useful for a feature tracking device to be able to determine which image features are most suitable for tracking in order to select an optimal subset of available image features for tracking at any given time. Additionally, it would be useful rank image features to be used for tracking according to how well image features with certain image properties track in practice.
Previously, others have put forth some effort toward determining how to rank image features for tracking. For example, U.S. patent application publication No. 2013/0287256 to Narashima et al., titled “Method and System for Real-Time Image Recognition on a Mobile Device” (filed Apr. 29, 2013), describes ranking features points to determine which features are stable in an image. However, the Narashima approach occurs during the act of image recognition, which further increases the computational cost to the device.
Another example includes European patent application publication no. EP 2 339 537 to BenHimane et al., titled “Method of Determining Reference Features for use in an Optical Object Initialization Tracking Process and Object Initialization Tracking Method” (filed Dec. 23, 2009). The BenHimane approach compares current image features to reference image features for object tracking. However, the BenHimane approach only describes an object-specific training procedure for selecting reference features for a single target object. Also, the BenHimane approach fails to appreciate that features matched from a static image can behave differently when tracked continuously through time.
Other examples of previous attempts to rank features for tracking include:
While each of these methods relate to selecting reference features, they do not adequately address analyzing image feature properties to determine, for example, how well the image features may be tracked through a video sequence. Further, the current methods of selecting reference features do not include analyzing image feature properties to determine a trackability ranking algorithm (i.e., a ranking strategy) through which only selected image features of relevance are tracked (e.g., features that satisfy a trackability ranking threshold), thereby reducing computational overhead and increasing tracking efficiency, e.g., with respect to frame rate.
Thus, there is still a need for computationally efficient methods that allow for determining image feature trackability, and for determining ranking strategies based on image feature trackability. The inventive subject matter provides apparatuses, systems and methods in which image features can be analyzed to determine trackability. The trackability information can be leveraged to establish a trackability ranking algorithm/ranking strategy through which selected features can be tracked in query video sequences.
In an embodiment, a method of establishing a trackability ranking order from tracked image features within a training video sequence at a tracking analysis device includes establishing a tracking region within a training video sequence using a feature detection algorithm. The training video sequence may comprise at least one of stereoscopic, motion-capture, time-lapsed, 3D and computer-rendered video data. The tracking region may span across multiple frames of the training video sequence. Establishing the tracking region may include locating the tracking region based on at least one image feature identified using the feature detection algorithm. The feature detection algorithm may comprise at least one of a features from accelerated segment test (FAST) algorithm, Harris corner algorithm, difference of Gaussian algorithm, level curve curvature algorithm, Laplacian of Gaussian algorithm, determinant of Hessian algorithm and maximally stable extremal regions (MSER) algorithm. Trajectories of tracked image features are compiled within the tracking region using a feature tracking algorithm. Saliency metrics are assigned to each one of the trajectories of tracked image features based on one or more feature property measurements within the tracking region, and a trackability ranking algorithm that is a function of the saliency metrics and a defined feature trajectory ranking associated with the training video sequence is determined, the trackability ranking algorithm being usable for ranking, based on trackability, tracked image features within another video sequence.
In some embodiments, the multiple frames of the training video sequence may comprise at least two sequential frames.
In some embodiments, the tracking region may comprise an image patch in at least one frame of the training video sequence, or a set of image patches that includes the image patch. The set of image patches may represent the image patch over a time interval.
In some embodiments, the tracking region may comprise at least a 3D region of the training video sequence, wherein the 3D region includes a time dimension, or at least a 4D region of the training video sequence, wherein the 4D region includes a depth dimension.
In some embodiments, the tracked image features may be identified according to an additional feature detection algorithm. The additional feature detection algorithm may be different from the feature detection algorithm used for establishing the tracking region within the training video sequence, and may comprise at least one a difference of Gaussian algorithm, Harris algorithm, FAST algorithm, edge detection algorithm, level curve curvature algorithm, Laplacian of Gaussian algorithm, determinant of Hessian algorithm and MSER algorithm.
In some embodiments, the defined feature trajectory ranking may be based on at least one of a track length, tracking error, tracking path, track depth of one or more training image features within the training video sequence, a number of frames, a length of time, an error, a confidence level, a momentum, a curvature, a patch count, a measure of photometric variance over time, and a line-of-sight measure.
In some embodiments, the trackability ranking algorithm comprises a linear ranking function, and the ordering of tracked image features may comprise sorting trajectories by a result of a linear combination of the saliency metrics. The linear ranking function may be learned using a ranking support vector machine (SVM).
In some embodiments, the linear ranking function may operate on a vector of saliency metrics expanded via a kernel map.
In some embodiments, the trackability ranking algorithm may comprise a multi-dimensional ranking function. The multi-dimensional ranking function may be learned using a ranking support vector machine (SVM).
In some embodiments, a trajectory of a tracked image feature may include at least one of a number of frames, a length of time, an error, a confidence level, a momentum, a curvature, a patch count, a measure of photometric variance over time, and a line-of-sight measure.
In some embodiments, the saliency metrics may comprise multi-valued metrics, and the one or more feature property measurements may include at least one of a Harris score, a pixel variance, a FAST score, a normalized cross-correlation (NCC) response in a neighborhood around the tracking region, a difference in NCC responses between inner and outer radii at a point location, a Harris score computed on NCC responses in a neighborhood around the tracking region, a number of frames, a length of time, a confidence level, a momentum, a curvature, a patch count, and a measure of photometric variance over time. The saliency metrics may be calculated from a mean or variance of one or more of the feature property measurements over a time interval.
In some embodiments, a trackability ranking order may be determined for tracked image features within a query video sequence using the trackability ranking algorithm, and a subset of the tracked image features within the query video sequence may be selected according to the trackability ranking order to facilitate at least one of an image-based recognition search, a video segmentation, or a pose estimation and mapping system. The subset of the tracked image features may be selected to satisfy at least one trackability ranking threshold, or at least one feature count requirement, wherein tracked image features are prioritized for selection according to the trackability ranking order. The image-based recognition search may be at least one of an image-based object recognition search and an image-based activity recognition search.
In some embodiments, a query tracking region may be established within the query video sequence that comprises an image patch in at least one frame of the query video sequence. The query tracking region may comprise a set of image patches that includes the image patch. The set of image patches may represent the image patch over a time interval. The tracked image features may be identified within the query video sequence according to an additional feature detection algorithm, which may be different from the feature detection algorithm used for establishing the tracking region within the training video sequence. The additional feature detection algorithm may comprise at least one a difference of Gaussian algorithm, Harris algorithm, FAST algorithm, edge detection algorithm, level curve curvature algorithm, Laplacian of Gaussian algorithm, determinant of Hessian algorithm and MSER algorithm.
Various objects, features, aspects and advantages of the inventive subject matter will become more apparent from the following detailed description of preferred embodiments, along with the accompanying drawing figures in which like numerals represent like components.
While the invention is described with reference to the above drawings, the drawings are intended to be illustrative, and other embodiments are consistent with the spirit, and within the scope, of the invention.
The various embodiments now will be described more fully hereinafter with reference to the accompanying drawings, which form a part hereof, and which show, by way of illustration, specific examples of practicing the embodiments. This description may, however, be embodied in many different forms and should not be construed as limited to the embodiments set forth herein; rather, these embodiments are provided so that this description will be thorough and complete, and will fully convey the scope of the invention to those skilled in the art. Among other things, this description may be embodied as methods or devices. Accordingly, any of the various embodiments herein may take the form of an entirely hardware embodiment, an entirely software embodiment or an embodiment combining software and hardware aspects. The following description is, therefore, not to be taken in a limiting sense.
Throughout the description and claims, the following terms take the meanings explicitly associated herein, unless the context clearly dictates otherwise:
The phrase “in one embodiment” as used herein does not necessarily refer to the same embodiment, though it may. Thus, as described below, various embodiments of the invention may be readily combined, without departing from the scope or spirit of the invention.
As used herein, the term “or” is an inclusive “or” operator, and is equivalent to the term “and/or,” unless the context clearly dictates otherwise.
The term “based on” is not exclusive and allows for being based on additional factors not described, unless the context clearly dictates otherwise.
As used herein, and unless the context dictates otherwise, the term “coupled to” is intended to include both direct coupling (in which two elements that are coupled to each other contact each other) and indirect coupling (in which at least one additional element is located between the two elements). Therefore, the terms “coupled to” and “coupled with” are used synonymously. Within the context of a networked environment where two or more components or devices are able to exchange data, the terms “coupled to” and “coupled with” are also used to mean “communicatively coupled with”, possibly via one or more intermediary devices.
In addition, throughout the description, the meaning of “a,” “an,” and “the” includes plural references, and the meaning of “in” includes “in” and “on.”
The following discussion provides many example embodiments of the inventive subject matter. Although some of the various embodiments presented herein constitute a single combination of inventive elements, it should be appreciated that the inventive subject matter is considered to include all possible combinations of the disclosed elements. As such, if one embodiment comprises elements A, B, and C, and another embodiment comprises elements B and D, then the inventive subject matter is also considered to include other remaining combinations of A, B, C, or D, even if not explicitly discussed herein.
As used in the description herein and throughout the claims that follow, when a system, engine, server, device, module, or other computing element is described as being configured to perform or execute functions on data in a memory, the meaning of “configured to” or “programmed to” is defined as one or more processors or cores of the computing element being programmed by a set of software instructions stored in the memory of the computing element to execute the set of functions on target data or data objects stored in the memory.
It should be noted that any language directed to a computer should be read to include any suitable combination of computing devices, including servers, interfaces, systems, databases, agents, peers, engines, controllers, modules, or other types of computing device structures operating individually or collectively. One skilled in the art should appreciate the computing devices comprise a processor configured to execute software instructions stored on a tangible, non-transitory computer readable storage medium (e.g., hard drive, FPGA, PLA, solid state drive, RAM, flash, ROM, etc.). The software instructions configure or program the computing device to provide the roles, responsibilities, or other functionality as discussed below with respect to the disclosed apparatus. Further, the disclosed technologies can be embodied as a computer program product that includes a non-transitory computer readable medium storing the software instructions that causes a processor to execute the disclosed steps associated with implementations of computer-based algorithms, processes, methods, or other instructions. In some embodiments, the various servers, systems, databases, or interfaces exchange data using standardized protocols or algorithms, possibly based on HTTP, HTTPS, AES, public-private key exchanges, web service APIs, known financial transaction protocols, or other electronic information exchanging methods. Data exchanges among devices can be conducted over a packet-switched network, such as the Internet, a LAN, WAN, VPN, or other type of packet-switched network, a circuit-switched network, cell-switched network; or other type of network.
All publications identified herein are incorporated by reference to the same extent as if each individual publication or patent application were specifically and individually indicated to be incorporated by reference. Where a definition or use of a term in an incorporated reference is inconsistent or contrary to the definition of that term provided herein, the definition of that term provided herein applies and the definition of that term in the reference does not apply.
The following description includes information that may be useful in understanding the present invention. It is not an admission that any of the information provided herein is prior art or relevant to the presently claimed invention, or that any publication specifically or implicitly referenced is prior art.
In some embodiments, the numbers expressing quantities of ingredients, properties such as concentration, reaction conditions, and so forth, used to describe and claim certain embodiments of the invention are to be understood as being modified in some instances by the term “about.” Accordingly, in some embodiments, the numerical parameters set forth in the description and attached claims are approximations that can vary depending upon the desired properties sought to be obtained by a particular embodiment. In some embodiments, the numerical parameters should be construed in light of the number of reported significant digits and by applying ordinary rounding techniques. Notwithstanding that the numerical ranges and parameters setting forth the broad scope of some embodiments of the invention are approximations, the numerical values set forth in the specific examples are reported as precisely as practicable. The numerical values presented in some embodiments of the invention may contain certain errors necessarily resulting from the standard deviation found in their respective testing measurements.
Unless the context dictates the contrary, all ranges set forth herein should be interpreted as being inclusive of their endpoints and open-ended ranges should be interpreted to include only commercially practical values. Similarly, all lists of values should be considered as inclusive of intermediate values unless the context indicates the contrary.
As used in the description herein and throughout the claims that follow, the meaning of “a,” “an,” and “the” includes plural reference unless the context clearly dictates otherwise. Also, as used in the description herein, the meaning of “in” includes “in” and “on” unless the context clearly dictates otherwise.
The recitation of ranges of values herein is merely intended to serve as a shorthand method of referring individually to each separate value falling within the range. Unless otherwise indicated herein, each individual value is incorporated into the description as if it were individually recited herein. All methods described herein can be performed in any suitable order unless otherwise indicated herein or otherwise clearly contradicted by context. The use of any and all examples, or exemplary language (e.g. “such as”) provided with respect to certain embodiments herein is intended merely to better illuminate the invention and does not pose a limitation on the scope of the invention otherwise claimed. No language in the description should be construed as indicating any non-claimed element essential to the practice of the invention.
Groupings of alternative elements or embodiments of the invention disclosed herein are not to be construed as limitations. Each group member can be referred to and claimed individually or in any combination with other members of the group or other elements found herein. One or more members of a group can be included in, or deleted from, a group for reasons of convenience and/or patentability. When any such inclusion or deletion occurs, the description herein is deemed to contain the group as modified, thus fulfilling the written description of all Markush groups used in the appended claims.
One should appreciate that the disclosed techniques provide many advantageous technical effects including reducing computational overhead on a device recognizing an activity. Thus, a computing device is enabled to process digital imagery data by tracking a subset of image features selected using a trackability ranking algorithm. The computing device is able to process substantial amounts of digital data well beyond the capability of a human being.
The following disclosed techniques seek to determine which types of image features have greater trackability relative to other image features based on selected feature properties. It should be appreciated that the disclosed techniques treat image features in a generic sense as there is no requirement that training image features from which trackability ranking algorithms are generated be bound to specific objects. Thus, the disclosed training techniques could be performed once from training video sequences and then applied to query video sequences across many domain-specific applications.
The following techniques leverage a ranking support vector machine (SVM) algorithm adapted from an algorithm developed by Joachims Thorsen in 2003 (see Joachims, T. (2003), “Optimizing Search Engines using Clickthrough Data”, Proceedings of the ACM Conference on Knowledge Discovery and Data Mining and URL en.wikipedia.org/wiki/Ranking_SVM). It should be appreciated that ranking SVM was developed specifically for ranking results of search engines rather than ranking image feature trackability. Thus, the various embodiments herein apply ranking SVM beyond its originally intended purpose.
Methods and systems for establishing a trackability ranking order from tracked image features within a training video sequence can be implemented and accessed in a variety of ways.
Within the environment 100, one or more servers 106A-C can establish a trackability ranking order from tracked image features within a training video sequence. For example, server 106A may establish a tracking region within the training video sequence using a feature detection algorithm, compile trajectories of tracked image features within the tracking region using a feature tracking algorithm, assign saliency metrics to each one of the trajectories of tracked image features based on one or more feature property measurements within the tracking region, determine a trackability ranking algorithm that is a function of the saliency metrics and a defined feature trajectory ranking associated with the training video sequence, and provide the trackability ranking algorithm to one or more client devices 104A-E.
Client devices, such as one or more of client devices 104A-E, can access a trackability ranking order or ranking strategy established within environment 100 via network 102. Examples of client devices 104A-E can include cell phones, kiosks, personal data assistants, tablet computers, toys, vehicles, web cameras, or other types of computing devices. In an embodiment, a client device, e.g., client device 104A, may determine a trackability ranking order for tracked image features within a query video sequence using the trackability ranking algorithm, and select a subset of the tracked image features within the query video sequence according to the trackability ranking order to facilitate at least one of an image-based recognition search, a video segmentation, or a pose estimation and mapping system.
In a further embodiment, one or more servers (e.g., server 106A) can host or be in communication with apparatuses, e.g., tracking analysis device 110 and image capture device 120, persistent storage devices 130A-B, and main memory device 140, for establishing a trackability ranking order from tracked image features that is accessible by user devices 104A-E.
In an embodiment, tracking analysis device 110 is configured to establish a trackability ranking order from tracked image features within a training video sequence (e.g., obtained from image capture device 120) by establishing a tracking region 152 within a training video sequence using a feature detection algorithm. Training video sequence 150 may comprise at least one of stereoscopic, motion-capture, time-lapsed, 3D and computer-rendered video data. The tracking region may span across multiple frames of the training video sequence. In some embodiments, the multiple frames of the training video sequence may comprise at least two sequential frames.
Establishing the tracking region may include locating the tracking region based on at least one image feature identified using the feature detection algorithm. For example, the feature detection algorithm may comprise at least one of a features from accelerated segment test (FAST) algorithm, Harris corner algorithm, difference of Gaussian algorithm, level curve curvature algorithm, Laplacian of Gaussian algorithm, determinant of Hessian algorithm and maximally stable extremal regions (MSER) algorithm.
In some embodiments, the tracking region may comprise an image patch in at least one frame of the training video sequence, or a set of image patches that includes the image patch. For example, the set of image patches may represent the image patch over a time interval.
Further, the tracking region may comprise at least a 3D region of the training video sequence, wherein the 3D region includes a time dimension, or at least a 4D region of the training video sequence, wherein the 4D region includes a depth dimension.
Tracking analysis device 110 is further configured to compile trajectories 154 of tracked image features within the tracking region using a feature tracking algorithm. In some embodiments, a trajectory of a tracked image feature may include at least one of a number of frames, a length of time, an error, a confidence level, a momentum, a curvature, a patch count, a measure of photometric variance over time, and a line-of-sight measure. In some embodiments, the tracked image features may be identified according to an additional feature detection algorithm. The additional feature detection algorithm may be different from the feature detection algorithm used for establishing the tracking region within the training video sequence, and may comprise at least one a difference of Gaussian algorithm, Harris algorithm, FAST algorithm, edge detection algorithm, level curve curvature algorithm, Laplacian of Gaussian algorithm, determinant of Hessian algorithm and MSER algorithm.
Tracking analysis device 110 is further configured to assign saliency metrics 156 to each one of the trajectories of tracked image features based on one or more feature property measurements within the tracking region. In some embodiments, the saliency metrics may comprise multi-valued metrics, and the one or more feature property measurements may include at least one of a Harris score, a pixel variance, a FAST score, a normalized cross-correlation (NCC) response in a neighborhood around the tracking region, a difference in NCC responses between inner and outer radii at a point location, a Harris score computed on NCC responses in a neighborhood around the tracking region, a number of frames, a length of time, a confidence level, a momentum, a curvature, a patch count, and a measure of photometric variance over time. The saliency metrics may also have appended to them some subset of the trajectory properties mentioned previously. The saliency metrics may be calculated from a mean or variance of one or more of the feature property measurements over a time interval.
Tracking analysis device 110 is further configured to determine a trackability ranking algorithm 158 that is a function of the saliency metrics and a defined feature trajectory ranking associated with the training video sequence. In an embodiment, the defined feature trajectory ranking may be based on at least one of a track length, tracking error, tracking path, track depth of one or more training image features within the training video sequence, a number of frames, a length of time, an error, a confidence level, a momentum, a curvature, a patch count, a measure of photometric variance over time, and a line-of-sight measure. In some embodiments, the trackability ranking algorithm comprises a linear ranking function, and the ordering of tracked image features may comprise sorting trajectories by a result of a linear combination of the saliency metrics. The linear ranking function may be learned using a ranking support vector machine (SVM). In some embodiments, the linear ranking function may operate on a vector of saliency metrics expanded via a kernel map. Further, the trackability ranking algorithm may comprise a multi-dimensional ranking function. The multi-dimensional ranking function may be learned using a ranking support vector machine (SVM).
The trackability ranking algorithm may be stored (e.g., within a persistent storage device 130A-B and/or main memory device 140) for ranking, based on trackability, tracked image features within another video sequence, such as a query video sequence. For example, the stored trackability ranking algorithm may be used to facilitate at least one of an image-based recognition search, a video segmentation, or a pose estimation and mapping system at one or more of client devices 104A-E via network 102.
In the various embodiments, tracking analysis device 110 may obtain at least one training video sequence having a defined feature trajectory ranking associated with a set of training features in the training video sequence. Tracking analysis device 110 can obtain the training video sequence through various techniques. For example, tracking analysis device 110 may be configured to obtain the training video sequence from image capture device 120. In some embodiments, tracking analysis device 110 may comprise one or more optical sensors (e.g., camera sensors) configured to directly capture the training video sequence. In other embodiments, tracking analysis device 110 may be configured to download or upload the training video sequence, such as via network 102. For example, tracking analysis device 110 could obtain the training video sequence from one or more remote (e.g., internet-based) video services such as, for example, YouTube®, Vimeo®, etc.
Training video sequence 225 can represent various forms of video data, possibly including stereoscopic video, motion capture video as referenced above, time-lapsed video, 3D video, 4D video (e.g., 3D content that changes with time), or other types of video data. It should be appreciated that training video sequence 225 could also represent different domains including games, sports, medical imaging or video, surveillance, television programming, web programming, dynamic big-data visualizations, or other domains.
Training video sequence 225 comprises a digital representation of scene 210 where frames 220 include one or more observable training features 230. Training features 230 represent known or previously understood features considered present within one or more of frames 220. For example, the circle feature 230 could represent a gradient-based feature present in Frames 220-1, 220-2 and 220-3, but not present in frame 220-N. Training video sequence 225 includes one or more defined feature trajectories 235. Each of the trajectories can be characterized by a set of tuples that include one or more of a feature identifier (e.g., descriptor, GUID, etc.), pixel coordinate (e.g., X, Y, possibly Z, etc.), frame number, time stamp, or other feature tracking information.
In an embodiment, training video sequence 225 is associated with a defined feature tracking ranking 240. Defined feature tracking ranking 240 represents a “ground truth” of how well each feature tracked through the video sequence 225. Defined feature tracking ranking 240 can be established through various techniques. In some embodiments, video sequence 225 can be instantiated or otherwise constructed with desired tracking properties. For example, video sequence 225 may be a video game graphic animation having defined feature properties. In other embodiments, video sequence 225 may be captured based on motion capture techniques where features and their tracking properties are well understood. In an optional embodiment, training video sequence 225 could be analyzed with respect to one or more feature detection algorithms to identify the features and trajectories of the features could be ranked according to an automatic criteria such as track length. Then, mechanical turk workers could rank the trajectories of the features, thereby validating the defined feature trajectory ranking 240.
In the example shown, defined feature trajectory ranking 240 ranks trajectories 235 by track length as measured in the number of frames where the feature is present. With respect to a traditional Ranking SVM use case for search results, track length would correspond to a search query while defined feature trajectory ranking 240 would correspond to the ranking method applied to the trajectories associated with the “query” of track length. It should be appreciated that the ranking of the trajectories can be according to other metrics beyond number of frames. For example, other ranking metrics may include track errors, feature repeatability, tracking confidence scores, track path, a track depth, time or duration of tracks, photometric variance of a patch over the lifetime of the trajectory, or other factors. Although a Ranking SVM embodiment may leverage a single query (e.g., track length), other Ranking SVM embodiments may leverage additional queries (e.g., errors, depth, etc.).
It should be appreciated that training video sequence 225 represents the ground truth for tracking information by which the disclosed techniques determine trackability ranking algorithm as discussed below. The following steps also apply to training video sequence 225.
At step 162, a tracking region is established within the training video sequence using a feature detection algorithm.
In an embodiment, tracking region 340 may comprise a static shape (e.g., circle, ellipse, bound box, etc.) that does not change in size or shape from frame to frame. In other embodiments, tracking region could change size or shape to account for changing properties of video sequence 225 from frame-to-frame. For example, tracking region 340 could be a circle within frame 220-1, but change to a reduced size ellipse in frame 220-2 to account for, e.g., a change in scale of the image data, a change in depth information, or a change in image capturing device orientation relative to the scene.
In some embodiments, tracking regions are located based on a feature identified using a feature detection algorithm, as in step 164. In the example shown, tracking region 340 is located based on a detected corner. For example, tracking region 340 could comprise a patch that is identified in frame 220-1 using a FAST algorithm. The feature detection algorithm could include one or more of a FAST algorithm, a Harris corner algorithm, a difference of Gaussian algorithm, or other known feature detection algorithms.
Tracking region 340 can span across multiple frames of video sequence 220. Typically, tracking region 340 will span across at least two sequential frames, e.g., frames 220-1 to 220-2. However, it is also possible that tracking region 340 could comprise interruptions. For example, a foreground object could temporarily occlude features within tracking region 340. Thus, tracking region 340 could comprise discontinuous sets of frame sequences. Thus, tracking region 340 can comprise one or more image patches or sets of image patches projected through time.
In view that tracking region 340 can comprise image patches through time, it could be considered a multi-dimensional data construct. For example, tracking region 340 can be a 3D region subtending video sequence 220, wherein region 340 includes a time dimension (e.g., time stamps, duration, frames, etc.). Further, tracking region 340 could be a 4D region having time and depth dimensional components beyond height and width properties. When multiple tracking regions 340 are present, the regions 340 can overlap each other or be restricted from overlapping, if desired.
At step 164, trajectories of tracked image features are compiled within the tracking region using a feature tracking algorithm. At this point, the patches associated with regions 340 are analyzed to identify tracked features, which could be identified according to at least one additional feature detection algorithm. The tracked features could include a heterogeneous mix of features (e.g., Difference of Gaussian, Harris, FAST, Edges, Level curve curvature, Laplacian of Gaussian, Determinant of Hessian, MSER, or other features). The one or more feature detection algorithms used to identify the tracked features are not required to be the same as the feature detection algorithm used to construct tracking region 340. However, the tracked features may correspond with the training features referenced above.
Tracking analysis device 110 compiles a set of trajectories comprising one or more tracked feature trajectories as represented by trajectories 450A and 450B, collectively referred to as trajectories 450. Each of trajectories 450 can include information relating to how a feature is tracked from frame to frame. For example, trajectories 450 can include a start frame, an end frame, a pixel coordinate, depth, number of frames, discontinuity in frame sequences, an error, a momentum, a path curvature, a patch count, a line of sight measure, or other track-related information. For example, the trajectory associated with the cross feature might be a NULL set because the feature is not present in more than one frame. Trajectory 450B may include feature information for only two frames while trajectory 450A may include track information spanning many frames. It should be appreciated that trajectories 450 have similar characteristics to the trajectories 235 associated with the training features in video sequence 225.
At step 166, saliency metrics are assigned to each one of the trajectories of tracked image features based on feature property measurements within the tracking region. With respect to an embodiment employing Ranking SVM, the set of saliency metrics represent the mapping function Φ(q, d) where q represents track length or other trajectory property and d represents the ranking of trajectories 450, which could include ranking by track length.
The saliency metrics are derived according to one or more saliency measures relating to the corresponding tracked features 450. The saliency measures represent functions applied to tracked features 450 or their surrounding environments in tracked regions 340. Example saliency measures include a Harris score, a pixel variance, a FAST score, a normalized cross-correlation (NCC) response in a neighborhood around a feature, a difference in NCC scores between concentric radii in a neighborhood around the feature, gradients, or other properties of the feature. It should be appreciated that the set of saliency metrics could grow over time as new feature properties are added. If new properties are added, then step 166 can be repeated to create new vector information. The saliency metrics can be single valued or multi-valued. For example, a multi-valued saliency measure can include an edge curvature as well as an error or confidence level of the curvature. In such case, each value of the saliency metric can be incorporated into the vector.
At step 168, a trackability ranking algorithm that is a function of the saliency metrics and a defined feature trajectory ranking associated with the training video sequence is determined, the trackability ranking algorithm being usable for ranking, based on trackability, tracked image features within another video sequence. The trackability ranking algorithm (i.e., ranking strategy) determines how tracked features 430 should be ranked. With respect to an embodiment that employs Ranking SVM, the trackability ranking algorithm is a retrieval function of weights. The trackability ranking algorithm may comprise a linear ranking function (e.g., a model learned for a single query type in Ranking SVM, etc.), which comprises a linear combination of saliency metrics. Further, the trackability ranking algorithm could comprise a multi-dimensional ranking function (e.g., a model learned from different query types in Ranking SVM, etc.). Even further, to achieve nonlinear level of performance in learning the trackability ranking algorithm while maintaining the speed and scalability of a linear Ranking SVM framework, a kernel map (such as described in A. Vedaldi and A. Zisserman, “Efficient Additive Kernels via Explicit Feature Maps”, in Pattern Analysis and Machine Intelligence (PAMI), 34(3), 2012), may be applied to the set of saliency metrics defined by the mapping function Φ(q, d).
At step 170, at least one of an image-based recognition search, a video segmentation, or a pose estimation and mapping system is facilitated using the trackability ranking algorithm. For example, tracking analysis device 110 may store the trackability ranking algorithm as a ranking strategy on a computer readable medium. The ranking strategy can then be deployed in a tracking client device 104 (e.g., a cell phone, tablet, surveillance camera, toy, vehicle, etc.), which leverages the trackability ranking algorithm to determine which features should be tracked according to the trackability ranking. Therefore, the trackability ranking algorithm can be considered as a data structure having weights associated with saliency metrics. In field use, the tracking device can identify a set of features in an image, and then calculate saliency scores of each feature. The trackability ranking can then be calculated by taking a dot product of the feature's saliency scores and the weight vector. In the case of nonlinear approximation via kernel maps, the vector of saliency scores can be expanded using a kernel map to match the dimension of the learned weight vector, before the dot product calculation. The result of the trackability ranking algorithm is a scalar rank for each feature where the larger the rank value indicates a feature having greater trackability. It should be noted that the tracked features used to establish the trackability ranking algorithm do not have to correspond to observed features in the field. Rather, the trackability ranking algorithm is determined based on the properties of the tracked features or based on ingested known object images. In field use, the feature ranking is based on observed features having similar properties as the training sample.
With respect to object recognition and tracking, the tracking device can be provisioned with one or more key frames comprising image information of a known object from specific points-of-view (PoV). Each key frame can be divided into a grid of cells, e.g., a five-by-five grid. Each cell can include a sorted list of features expected to be present in the cell where the list of features have been ranked by the ranking strategy. This approach is considered advantageous for several reasons. First, the tracking device can calculate object identification features directly from the key frame rather than retrieving the features or their descriptors from a remote host, especially when the key frame has a smaller data size than the corresponding features or descriptors. Second, the cells of the grid can be annotated with feature trackability information, which reduces computation time and ensures only trackable features are monitored. Such a technique may provide for processing video data at typical frame rates during field use.
At step 502, a tracking region is established within a training video sequence according to first feature detection algorithm, the training video sequence having a defined feature trajectory ranking associated with a set of training features in the training video sequence.
At step 504, a set of feature trajectories of tracked features within the tracking region is compiled according to at least one feature tracking algorithm.
At step 506, each feature trajectory of the set of feature trajectories is assigned a set of saliency metrics derived from a respective training feature's trajectory through the training video sequence and based on feature property measures applied to the tracking region.
At step 508, a ranking strategy is defined that orders the tracked features according to a trackability ranking function of the saliency metrics and the defined feature trajectory ranking.
At step 510, the ranking strategy is stored on a computer readable medium, e.g., to facilitate at least one of an image-based recognition search, a video segmentation, or a pose estimation and mapping system.
As described above in
High-level flow diagrams of exemplary client-server operations that may be used to implement systems, apparatus and methods described herein are illustrated in
One skilled in the art will appreciate that the exemplary client-server relationship illustrated in
Systems, apparatus, and methods described herein may be implemented using a computer program product tangibly embodied in an information carrier, e.g., in a non-transitory machine-readable storage device, for execution by a programmable processor; and the method steps described herein, including one or more of the steps of
A high-level block diagram of an exemplary apparatus that may be used to implement systems, apparatus and methods described herein is illustrated in
Processor 710 may include both general and special purpose microprocessors, and may be the sole processor or one of multiple processors of apparatus 700. Processor 710 may comprise one or more central processing units (CPUs), for example. Processor 710, persistent storage device 720, and/or main memory device 730 may include, be supplemented by, or incorporated in, one or more application-specific integrated circuits (ASICs) and/or one or more field programmable gate arrays (FPGAs).
Persistent storage device 720 and main memory device 730 each comprise a tangible non-transitory computer readable storage medium. Persistent storage device 720, and main memory device 730, may each include high-speed random access memory, such as dynamic random access memory (DRAM), static random access memory (SRAM), double data rate synchronous dynamic random access memory (DDR RAM), or other random access solid state memory devices, and may include non-volatile memory, such as one or more magnetic disk storage devices such as internal hard disks and removable disks, magneto-optical disk storage devices, optical disk storage devices, flash memory devices, semiconductor memory devices, such as erasable programmable read-only memory (EPROM), electrically erasable programmable read-only memory (EEPROM), compact disc read-only memory (CD-ROM), digital versatile disc read-only memory (DVD-ROM) disks, or other non-volatile solid state storage devices.
Input/output devices 790 may include peripherals, such as a printer, scanner, display screen, etc. For example, input/output devices 790 may include a display device such as a cathode ray tube (CRT), plasma or liquid crystal display (LCD) monitor for displaying information (e.g., a plurality of image transformations for selection) to a user, a keyboard, and a pointing device such as a mouse or a trackball by which the user can provide input to apparatus 700.
Any or all of the systems and apparatus discussed herein, including tracking analysis device 110 and image capture device 120 may be performed by, and/or incorporated in, an apparatus such as apparatus 700.
One skilled in the art will recognize that an implementation of an actual computer or computer system may have other structures and may contain other components as well, and that
Patent | Priority | Assignee | Title |
10540772, | Jul 09 2014 | NANT HOLDINGS IP, LLC | Feature trackability ranking, systems and methods |
Patent | Priority | Assignee | Title |
6618490, | Sep 16 1999 | COMPAQ INFORMATION TECHNOLOGIES GROUP, L P | Method for efficiently registering object models in images via dynamic ordering of features |
6795567, | Sep 16 1999 | HEWLETT-PACKARD DEVELOPMENT COMPANY, L P | Method for efficiently tracking object models in video sequences via dynamic ordering of features |
9984473, | Jul 09 2014 | NANT VISION, INC | Feature trackability ranking, systems and methods |
20130215264, | |||
20130287256, | |||
20140063275, | |||
20140328570, | |||
EP2339537, |
Executed on | Assignor | Assignee | Conveyance | Frame | Reel | Doc |
Aug 17 2015 | NANT VISION, INC | NANT HOLDINGS IP, LLC | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 045748 | /0229 | |
Aug 27 2015 | WNUK, KAMIL | NANT VISION, INC | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 045748 | /0226 | |
May 08 2018 | NANT HOLDINGS IP, LLC | (assignment on the face of the patent) | / |
Date | Maintenance Fee Events |
May 08 2018 | BIG: Entity status set to Undiscounted (note the period is included in the code). |
Aug 17 2022 | M1551: Payment of Maintenance Fee, 4th Year, Large Entity. |
Date | Maintenance Schedule |
Feb 26 2022 | 4 years fee payment window open |
Aug 26 2022 | 6 months grace period start (w surcharge) |
Feb 26 2023 | patent expiry (for year 4) |
Feb 26 2025 | 2 years to revive unintentionally abandoned end. (for year 4) |
Feb 26 2026 | 8 years fee payment window open |
Aug 26 2026 | 6 months grace period start (w surcharge) |
Feb 26 2027 | patent expiry (for year 8) |
Feb 26 2029 | 2 years to revive unintentionally abandoned end. (for year 8) |
Feb 26 2030 | 12 years fee payment window open |
Aug 26 2030 | 6 months grace period start (w surcharge) |
Feb 26 2031 | patent expiry (for year 12) |
Feb 26 2033 | 2 years to revive unintentionally abandoned end. (for year 12) |