A method of recognizing a feature of an image may include receiving an input image including an object; extracting first feature information using a first layer of a neural network, the first feature information indicating a first feature corresponding to the input image among a plurality of first features; extracting second feature information using a second layer of the neural network, the second feature information indicating a second feature among a plurality of second features, the indicated second feature corresponding to the first feature information; and recognizing an element corresponding to the object based on the first feature information and the second feature information.
|
10. A method of verifying a feature of an image, the method comprising:
receiving a first image including a first object;
performing a filtering of information corresponding to the first image;
extracting, based on the filtering of information corresponding to the first image, first feature information using a first layer of a neural network;
extracting, based on the filtering of information corresponding to the first image, second feature information using a second layer of the neural; and
determining, using the first feature information and the second feature information of the first image and first feature information and second feature information associated with a second image, whether the first object of the first image is similar to a second object of the second image,
wherein the second feature information of the first image corresponds to a different level of the neural network from the first feature information of the first image.
1. A method of recognizing a feature of an image, the method comprising:
receiving an input image including an object;
performing a filtering of information corresponding to the input image;
extracting, based on the filtering of information corresponding to the input image, first feature information using a first layer of a neural network,
extracting, based on the filtering of information corresponding to the input image, second feature information using a second layer of the neural network; and
recognizing an element corresponding to the object based on the first feature information and the second feature information,
wherein the recognizing comprises:
generating, using the first feature information and the second feature information, a feature vector;
performing a comparison operation based on the generated feature vector and a reference vector corresponding to each of a plurality of elements; and
recognizing the element corresponding to the object based on a result of the comparison operation, and
wherein the second feature information corresponds to a different level of the neural network from than the first feature information.
8. An apparatus for recognizing a feature of an image, the apparatus comprising:
a memory storing computer-readable instructions; and
one or more processors configured to execute the computer-readable instructions such that the one or more processors are configured to,
receive an input image including an object;
perform a filtering of information corresponding to the input image;
extract, based on the filtering of information corresponding to the input image, first feature information using a first layer of a neural network;
extract, based on the filtering of information corresponding to the input image, second feature information using a second layer of the neural network;
generating, using the first feature information and the second feature information, a feature vector;
performing a comparison operation based on the generated feature vector and a reference vector corresponding to each of a plurality of elements; and
recognize an element corresponding to the object based on a result of the comparison operation,
wherein the second feature information corresponds to a different level of the neural network from the first feature information.
2. The method of
generating the feature vector based on values corresponding to all nodes in the first layer and values corresponding to all nodes in the second layer.
3. The method of
wherein the recognizing comprises calculating a feature value corresponding to the element based on weights between nodes included in the neural network.
4. The method of
wherein the extracting of the second feature information includes performing the filtering of the first feature information using a filter with a second size.
5. The method of
6. The method of
7. The method of
wherein m is an integer equal to or greater than 3, and
wherein the recognizing of the element includes recognizing the element based on the first feature information through the m-th feature information.
9. The apparatus of
extract third feature information using a third layer, the third feature information indicating a third feature among a plurality of third features, the indicated third feature corresponding to the second feature information, and
recognize the element corresponding to the object based on the first feature information, the second feature information, and the third feature information.
11. The method of
generating a feature vector associated with the first image based on the first feature information and the second feature information of the first image;
generating a feature vector associated with the second image based on the first feature information and the second feature information associated with the second image; and
performing a comparison operation based on the feature vector associated with the first image and the feature vector associated with the second image, and determining whether the first object is similar to the second based on a result of the comparison operation.
12. The method of
wherein the generating of the feature vector associated with the second image includes generating the feature vector associated with the second image based on feature values calculated based on the first feature information and the second feature information associated with the second image.
13. The method of
wherein the extracting of the second feature information includes extracting second feature information associated with the second image, using the second layer.
14. The method of
wherein the extracting of the second feature information includes extracting second feature information associated with the second image, using a layer of the neural network that is separated from the second layer.
|
This U.S. non-provisional patent application is a continuation of U.S. application Ser. No. 14/958,030, filed on Dec. 3, 2015, which claims the priority benefit of Korean Patent Application No. 10-2014-0180213 and Korean Patent Application No. 10-2015-0138491, respectively filed on Dec. 15, 2014 and Oct. 1, 2015, in the Korean Intellectual Property Office, the entire contents of each of which are incorporated herein by reference.
At least one example embodiment relates to an object recognition method and apparatus, and a recognizer learning method and apparatus.
In a transition to the 21st century information society, information on particular organizations and personal information may have significant importance. To protect the above important information, various types of passwords are used, and other technologies for verifying identity are desperately desired. Among these technologies, face recognition technologies have been evaluated as the most convenient and competitive identity verification methods because a user does not need to take a particular motion or action, and moreover a user's identity can be verified while the user does not recognize it.
Currently, a face recognition technology including recognizing a face by applying a principal component analysis (PCA) to a face image is being used frequently. The PCA refers to a scheme of projecting image data onto a low-dimensional eigenvector space while reducing or, alternatively, minimizing a loss of unique information of an image so as to reduce information. A method of extracting a principal feature vector of a face and recognizing the face using a pattern classifier learned using a principal component vector extracted from a preregistered image has been used frequently as a face recognition method using the PCA. However, by using the method to recognize a face with a large amount of information, a recognition speed and reliability may be reduced, and a satisfactory face recognition result may not be obtained in a change in a pose or a facial expression, even though a feature robust against illumination is obtained based on a selection of a PCA basis vector.
A face recognition performance varies depending on a performance of a pattern classifier to distinguish a registered face from an unregistered face. To learn the pattern classifier, an artificial neural network may be used.
According to at least some example embodiments, a method of recognizing a feature of an image includes receiving an input image including an object; extracting first feature information using a first layer of a neural network, the first feature information indicating a first feature corresponding to the input image among a plurality of first features; extracting second feature information using a second layer of the neural network, the second feature information indicating a second feature among a plurality of second features, the indicated second feature corresponding to the first feature information; and recognizing an element corresponding to the object based on the first feature information and the second feature information.
According to at least some example embodiments, an apparatus for recognizing a feature of an image includes a memory storing computer-readable instructions; and one or more processors configured to: execute the computer-readable instructions such that the one or more processors are configured to, receive an input image including an object; extract first feature information using a first layer, the first feature information indicating a first feature among a plurality of first feature information, the indicated first feature corresponding to the input image; extract second feature information using a second layer, the second feature information indicating a second feature among a plurality of second features, the indicated second feature corresponding to the first feature information; and recognize an element corresponding to the object based on the first feature information and the second feature information.
According to at least some example embodiments, a method of learning a feature to recognize an image includes receiving a training element and a training image associated with a training object; and learning a parameter of a recognizer such that the recognizer recognizes the training element from the training image, the recognizer being configured to recognize a plurality of elements from first feature information extracted from an input image using a first layer of a neural network and second feature information extracted using a second layer of the neural network.
According to at least some example embodiments, an apparatus for learning a feature to recognize an image includes a memory storing computer-readable instructions; and one or more processors configured to execute the computer-readable instructions such that the one or more processors are configured to: receive a training element and a training image associated with a training object; implement a recognizer; and learn a parameter of the recognizer such that the recognizer recognizes the training element from the training image, the recognizer being configured to recognize a plurality of elements from first feature information extracted from an input image using a first layer of a neural network and second feature information extracted using a second layer of the neural network.
According to at least some example embodiments, a method of verifying a feature of an image may include receiving a first image including a first object; extracting first feature information using a first layer of a neural network, the first feature information indicating a first feature among a plurality of first features, the indicated first feature corresponding to the first image; extracting second feature information using a second layer of the neural network, the second feature information indicating a second feature among a plurality of second features, the indicated second feature corresponding to the first feature information; and determining whether the object of the first image is similar to a second object of a second image, based on the first feature information and the second feature information of the first image and based on first feature information and second feature information associated with the second image.
According to at least some example embodiments, an apparatus for verifying a feature of an image includes a memory storing computer-readable instructions; an one or more processors configured to execute the computer-readable instructions such that the one or more processors are configured to: receive a first image including a first object; extract first feature information using a first layer of a neural network, the first feature information indicating a first feature among a plurality of first features, the first feature corresponding to the first image; extract second feature information using a second layer of a neural network, the second feature information indicating a second feature among a plurality of second features, the second feature corresponding to the first feature information among; and determine whether the first object is similar to a second object of a second image, based on the first feature information and the second feature information of the first image and based on first feature information and second feature information associated with the second image.
According to at least some example embodiments, a method of learning a feature to verify an image includes receiving a pair of training images, and training information corresponding to the pair of training images; and learning a parameter of a verifier so that a result of comparing, by the verifier, the training images corresponds to the training information, the verifier being configured to determine whether two input images are similar to each other, based on first feature information extracted from the two input images using a first layer of a neural network and based on second feature information extracted using a second layer of the neural network.
According to at least some example embodiments, an apparatus for learning a feature to verify an image includes a memory storing computer-readable instructions; and one or more processors configured to execute the computer-readable instructions such that the one or more processors are configured to: receive a pair of training images, and training information corresponding to the pair of training images, and learn a parameter of a verifier so that a result of comparing, by the verifier, the training images corresponds to the training information, the verifier being configured to determine whether two input images are similar to each other, based on first feature information extracted from the two input images using a first layer of a neural network and based on second feature information extracted using a second layer of the neural network.
The above and other features and advantages of example embodiments of the inventive concepts will become more apparent by describing in detail example embodiments of the inventive concepts with reference to the attached drawings. The accompanying drawings are intended to depict example embodiments of the inventive concepts and should not be interpreted to limit the intended scope of the claims. The accompanying drawings are not to be considered as drawn to scale unless explicitly noted.
Detailed example embodiments of the inventive concepts are disclosed herein. However, specific structural and functional details disclosed herein are merely representative for purposes of describing example embodiments of the inventive concepts. Example embodiments of the inventive concepts may, however, be embodied in many alternate forms and should not be construed as limited to only the embodiments set forth herein.
Accordingly, while example embodiments of the inventive concepts are capable of various modifications and alternative forms, embodiments thereof are shown by way of example in the drawings and will herein be described in detail. It should be understood, however, that there is no intent to limit example embodiments of the inventive concepts to the particular forms disclosed, but to the contrary, example embodiments of the inventive concepts are to cover all modifications, equivalents, and alternatives falling within the scope of example embodiments of the inventive concepts. Like numbers refer to like elements throughout the description of the figures.
It will be understood that, although the terms first, second, etc. may be used herein to describe various elements, these elements should not be limited by these terms. These terms are only used to distinguish one element from another. For example, a first element could be termed a second element, and, similarly, a second element could be termed a first element, without departing from the scope of example embodiments of the inventive concepts. As used herein, the term “and/or” includes any and all combinations of one or more of the associated listed items.
It will be understood that when an element is referred to as being “connected” or “coupled” to another element, it may be directly connected or coupled to the other element or intervening elements may be present. In contrast, when an element is referred to as being “directly connected” or “directly coupled” to another element, there are no intervening elements present. Other words used to describe the relationship between elements should be interpreted in a like fashion (e.g., “between” versus “directly between”, “adjacent” versus “directly adjacent”, etc.).
The terminology used herein is for the purpose of describing particular embodiments only and is not intended to be limiting of example embodiments of the inventive concepts. As used herein, the singular forms “a”, “an” and “the” are intended to include the plural forms as well, unless the context clearly indicates otherwise. It will be further understood that the terms “comprises”, “comprising,”, “includes” and/or “including”, when used herein, specify the presence of stated features, integers, steps, operations, elements, and/or components, but do not preclude the presence or addition of one or more other features, integers, steps, operations, elements, components, and/or groups thereof.
It should also be noted that in some alternative implementations, the functions/acts noted may occur out of the order noted in the figures. For example, two figures shown in succession may in fact be executed substantially concurrently or may sometimes be executed in the reverse order, depending upon the functionality/acts involved.
Example embodiments of the inventive concepts are described herein with reference to schematic illustrations of idealized embodiments (and intermediate structures) of the inventive concepts. As such, variations from the shapes of the illustrations as a result, for example, of manufacturing techniques and/or tolerances, are to be expected. Thus, example embodiments of the inventive concepts should not be construed as limited to the particular shapes of regions illustrated herein but are to include deviations in shapes that result, for example, from manufacturing.
Hereinafter, in the present disclosure, deep learning refers to a set of algorithms in machine learning that attempt to model high-level abstractions using a combination of multiple non-linear transformations. The deep learning may be, for example, machine learning to train a computer to recognize a human way of thinking. The abstractions refer to a process of extracting core data from data.
A deep learning structure 100 of
The term ‘processor’, as used herein, may refer to, for example, a hardware-implemented data processing device having circuitry that is physically structured to execute desired operations including, for example, operations represented as code and/or instructions included in a program. Examples of the above-referenced hardware-implemented data processing device include, but are not limited to, a microprocessor, a central processing unit (CPU), a processor core, a multi-core processor; a multiprocessor, an application-specific integrated circuit (ASIC), and a field programmable gate array (FPGA). Processors executing program code are programmed processors, and thus, are special-purpose computers.
A feature may be core data learned through an abstraction of data (for example, training images). In the present disclosure, a feature may include, for example, a feature image learned through an abstraction of an arbitrary image. The feature image may be learned as an image generated by performing a convolution filtering of a training image using a predetermined or, alternatively, desired number of filters with a predetermined or, alternatively, desired size. A number of learned feature images may correspond to the predetermined or, alternatively, desired number of the filters.
For example, the low-level feature 110 may be represented based on one of low-level feature images 111, the mid-level feature 120 may be represented based on one of mid-level feature images 121, and the high-level feature 130 may be represented based on one of high-level feature images 131. The low-level feature images 111 may be images learned by performing a convolution filtering of a training image, and the mid-level feature images 121 may be images learned by performing another convolution filtering of the low-level feature images 111. The high-level feature images 131 may be images learned by performing still another convolution filtering of the mid-level feature images 121.
A result of the abstraction of the input image 101 in the deep learning structure 100 may be represented as feature information indicating a feature corresponding to the input image 101 from features corresponding to each of the levels. The feature information may include, for example, a feature value indicating an arbitrary feature image. The deep learning structure 100 may extract feature information corresponding to each of a plurality of levels (e.g., low, mid and high levels) based on a pre-trained layer corresponding to each of the plurality of levels.
For example, the deep learning structure 100 may extract low-level feature information indicating the low-level feature 110 corresponding to the input image 101, may extract mid-level feature information indicating the mid-level feature 120 corresponding to the low-level feature information, and may extract high-level feature information indicating the high-level feature 130 corresponding to the mid-level feature information.
A recognition/verification module 140 in the deep learning structure 100 may perform a recognition and verification based on feature information of a last level, through sequential abstractions for each of the levels. For example, in
In a deep learning structure 200 of
As described above, all feature information output for each of the layers may be utilized and thus, a recognition rate and a verification rate of an image may be ensured.
The deep learning structure 200 may be applied to recognize and verify various input images 101. For example, the input image 101 may include an image associated with an object (for example, an image representing a shape of an object). The object may include, for example, an animal, an inanimate object, or a human (for example, a human face, or a human body) included in a region of interest (ROI) of an image. For example, the deep learning structure 200 may be used to recognize a human face and to perform a recognition and authentication of a user. Also, the deep learning structure 200 may be used to search for and manage a considerable amount of content (for example, multimedia including a picture or video), automatically.
The deep learning structure 200 may be implemented in software or hardware, for example, a chip, and may be mounted in an electronic device. The electronic device may include, for example, a mobile device (for example, a mobile phone, or a smartphone), a home appliance (for example, a TV), and the like.
The deep learning structure 200 may be applied to an apparatus for recognizing an image, and an apparatus for verifying an image (hereinafter, referred to as an “image recognition apparatus,” and an “image verification apparatus,” respectively). The image recognition apparatus and the image verification apparatus may be trained based on training data. The training data may include, for example, training information, a training element and a training image associated with a training object.
In an example, training data used to train the image recognition apparatus may include a training image, and a training element associated with the training image. The image recognition apparatus may be trained to allow the training element to be output from the training image. In this example, the training element may be a value indicating a training object included in the training image. The above example will be further described with reference to
In another example, training data used to train the image verification apparatus may include a pair of training images and training information. The image verification apparatus may be trained to allow the training information to be output from the pair of training images. For example, the image verification apparatus may be trained to associate the training information with the pair of training images. In this example, the training information may be a value indicating whether the training images include identical training objects. The above example will be further described with reference to
A deep learning structure according to at least one example embodiment may include an artificial neural network. Examples of such a neural network include, but are not limited to, a convolutional neural network (CNN) and a deep convolutional neural network (DCNN). An example of the DCNN will be further described with reference to
A neural network may include an input layer, a hidden layer, and an output layer. Each layer may include a plurality of nodes, and nodes between neighboring layers may be connected to each other with a connection weight. Each node may operate based on an activation model. An output value corresponding to an input value may be determined based on the activation model. An output value of an arbitrary node may be input to a node that is included in a next layer and that is connected to the arbitrary node. The node in the next layer may receive inputs of values output from a plurality of nodes. When the output value of the arbitrary node is input to the node in the next layer, a connection weight may be applied. The node in the next layer may output a value corresponding to the input value to a node that is included in a layer after the next layer and that is connected to the node, based on the activation model.
The output layer may include nodes corresponding to a plurality of elements. The nodes in the output layer may output feature values corresponding to the plurality of elements. Feature values output from the neural network may be transformed to elements using linear classifiers for the plurality of elements, which will be further described below.
Referring to
In operation 320, a first extractor of the feature recognition apparatus may extract first feature information from the input image using a first layer. For example, the feature recognition apparatus may extract the first feature information indicating a first feature corresponding to the input image among first features, using the first layer.
The first layer may be, for example, a layer that is included in a deep learning network and that is pre-trained on first features based on training data.
The first feature information may indicate a first feature corresponding to the input image. For example, the first feature information may include a feature value indicating at least one first feature corresponding to the input image among a plurality of first features. Additionally, the first feature information may be provided as a feature matrix and a feature vector including a plurality of feature values indicating the first features. The first feature may include a feature image learned through an abstraction of a training image included in training data.
In operation 330, the feature recognition apparatus may extract second feature information from the first feature information, using a second layer. For example, the feature recognition apparatus may extract the second feature information indicating a second feature corresponding to the first feature information among second features, using the second layer. The second feature information may correspond to a higher level of image complexity than the first feature information.
The second layer may be, for example, a layer that is included in a deep learning network and that is pre-trained on second features based on the first features.
The second feature information may indicate a second feature corresponding to the first feature information. For example, the second feature information may include a feature value indicating at least one second feature corresponding to arbitrary first feature information among a plurality of second features. The second feature may include a feature image learned through an abstraction of the first feature.
In operation 340, the feature recognition apparatus may recognize an element corresponding to the input image, based on the first feature information and the second feature information. The element may include, for example, a value indicating an object included in reference images that are pre-learned from training data and stored.
For example, the feature recognition apparatus may generate, using a feature vector generator, a feature vector based on the first feature information and the second feature information, and may recognize, using a recognizer, an element corresponding to an object by comparing the generated feature vector to a pre-stored reference vector corresponding to each of a plurality of elements. For example, the pre-stored reference vector may be stored before the generation of the feature vector. In this example, the feature vector may be data including a plurality of feature values in the form of vectors.
Referring to
The image receiver 410 may receive an input image 401 including an object. The input image 401 may be a preprocessed image.
The first extractor 421 may extract first feature information indicating a first feature corresponding to the input image 401 among first features, using a first layer. The first layer may be, for example, a layer pre-trained on the first features based on training data.
The second extractor 422 may extract second feature information indicating a second feature corresponding to the first feature information among second features, using a second layer. The second layer may be, for example, a layer pre-trained on the second features based on the first features.
The recognizer 430 may recognize an element corresponding to the object, based on the first feature information and the second feature information. For example, the recognizer 430 may include a feature vector generator (not shown) configured to generate a feature vector based on the first feature information and the second feature information. In this example, the recognizer 430 may recognize the element corresponding to the object by comparing the generated feature vector to a pre-stored reference vector corresponding to each of a plurality of elements. For example, as is discussed above, the pre-stored reference vector may be stored before the generation of the feature vector.
For example, the recognizer 430 may include a neural network, and may calculate a feature value corresponding to an element based on pre-learned weights between nodes included in the neural network.
An element 409 recognized by the recognizer 430, as a result of recognition of the input image 410 by the feature recognition apparatus 400, may include a value indicating a reference image or object recognized to correspond to the input image 401 by the recognizer 430 among pre-learned reference images.
For example, when three reference objects are learned, elements 409 recognized by the recognizer 430 may be represented as a value indicating a first reference object, a value indicating a second reference object, and a value indicating a third reference object. When an input image includes the second reference object, recognized elements may be represented by values of “0.1,” “0.8,” “0.1.” In other words, an element corresponding to an object determined to be most similar to the second reference object may have a highest value.
While, for the purpose of simplicity, feature extractor 420 is illustrated in
For example, the m-th extractor may extract m-th feature information indicating an m-th feature corresponding to (m−1)-th feature information among m-th features, using an m-th layer. The m-th layer may be, for example, a layer pre-trained on the m-th features based on (m−1)-th features. According to at least one example embodiment, higher levels of feature information may be associated with higher levels of image complexity in comparison to lower levels of feature information.
In this example, the recognizer 430 may recognize an element corresponding to an object included in an input image, based on the first feature information through n-th feature information.
The m-th features may be obtained by performing a higher abstraction of the (m−1)-th features. For example, the m-th layer may correspond to a level higher than a level corresponding to an (m−1)-th layer. A feature recognition apparatus according to at least one example embodiment may be used to recognize an object included in an image, using features in an abstraction process (for example, a first feature through an (n−1)-th feature) as well as a feature corresponding to a last level (for example, an n-th feature).
Referring to
The m-th extractor may extract m-th feature information indicating an m-th feature corresponding to a combination of first feature information through (m−1)-th feature information among m-th features, using an m-th layer that is pre-trained on the m-th features based on first features through (m−1)-th features. The combination of the first feature information through the (m−1)-th feature information may be a vector in which values output from the first extractor 421 through an (m−1)-th extractor are arranged, and may have the same dimension as a sum of a number of nodes in an output layer of the first extractor 421 through a number of nodes in an output layer of the m-th extractor. The m-th extractor may adjust, based on a number of nodes in the m-th layer, a dimension of feature information generated by combining the first feature information through the (m−1)-th feature information. For example, the m-th extractor may include a fully connected layer, a pooling layer and a convolution layer, in addition to the m-th layer. In this example, the m-th extractor may process the feature information generated by combining the first feature information through the (m−1)-th feature information, using a corresponding layer (for example, the fully connected layer, the pooling layer and the convolution layer), and may adjust the dimension of the feature information generated by combining the first feature information through the (m−1)-th feature information to be the same as a dimension of the m-th layer (for example, the number of the nodes in the m-th layer).
For example, the fourth extractor may extract fourth feature information using a fourth layer. In this example, the fourth feature information may indicate a fourth feature corresponding to a combination of first feature information, second feature information and third feature information among fourth features. The fourth layer may be, for example, a layer pre-trained on the fourth features based on first features, second features and third features. When the first extractor 421, the second extractor 422 and the third extractor are assumed to have “n1” output nodes, “n2” output nodes and “n3” output nodes, respectively, a dimension of the combination of the first feature information through the third feature information may be a sum of “n1”, “n2” and “n3.” When a dimension of the fourth layer is “m4,” the fourth extractor may adjust the dimension of the combination to be “m4,” for example, using a fully connected layer, a pooling layer and a convolution layer. The dimension represents a number of data output from or input to a layer.
The feature recognition apparatus of
The feature extractor 620 may include a first extractor 621, a second extractor 622, a third extractor 623, a fourth extractor 624, and a fifth extractor 625.
Each of the first extractor 621 through the fifth extractor 625 may include at least one layer, for example, a convolution layer, a pooling layer, and the like. Each of the first extractor 621 through the fifth extractor 625 may necessarily include a convolution layer. For example, the first extractor 621 may include a first layer, and the second extractor 622 may include a second layer. Each of the first layer and the second layer may include a convolution layer, a pooling layer, and the like.
The convolution layer may be used to perform a convolution filtering for filtering of information extracted by a previous extractor or a previous layer, using a filter with a predetermined or, alternatively, desired size (for example, 8×8). In
As a result of the convolution filtering, the same number of filtering images as a number of filters included in the convolution layer may be generated. The convolution layer may be formed with nodes included in the filtering images. Each of the nodes in the convolution layer may receive a value obtained by performing a filtering of an area with a predetermined or, alternatively, desired size in a feature image of a previous extractor or a previous layer. Exceptionally, each of nodes in the convolution layer in the first extractor 621 may receive a value obtained by performing a filtering of an input image. An activation model of each of the nodes in the convolution layer may be, for example, a rectifier linear unit (ReLU). The ReLU may refer to a model to output “0” in response to an input equal to or less than “0,” and to output a value in linear proportion to an input exceeding “0.”
The pooling layer may be used to extract representative values from feature images of a previous layer, through pooling. In
Nodes between layers adjacent to the convolution layer and the pooling layer may be partially connected to each other, and may share a connection weight.
Filters of a convolution layer included in the second extractor 622 may perform a filtering of a complex edge, in comparison to the filters of the convolution layer included in the first extractor 621. A pooling layer included in the second extractor 622 may be used to extract representative values from filtering images (for example, first feature information) obtained by the filtering by the convolution layer in the second extractor 622 through pooling. As described above, in layers included in the second extractor 622, feature information with a higher complexity in comparison the first extractor 621 may be extracted. For example, feature information corresponding to a middle level of an input image may be extracted. The middle level may have, for example, a middle complexity.
A convolution layer of the third extractor 623 may be used to perform a convolution filtering using filters with predetermined or, alternatively, desired sizes. Each of the filters may perform a filtering of a predetermined or, alternatively, desired edge. Filters of the convolution layer of the third extractor 623 may perform a filtering of a complex edge, in comparison to the filters of the convolution layer of the second extractor 622.
Through a similar operation to the above-described operation, the fourth extractor 624 and the fifth extractor 625 may extract feature information with a higher complexity, in comparison to previous extractors. As shown in
However, a number of extractors, and a configuration of layers in an extractor are not limited to those described above, and may be changed based on a design.
The recognizer 630 may include at least one layer, for example, a fully connected layer that is denoted as “FULLY CON” in
Nodes between layers adjacent to the fully connected layer may be fully connected to each other, and a connection weight may be individually set. For example, each of fully connected layers may include 2,048 nodes. Additionally, a model regularization algorithm, for example, a dropout may be applied to the fully connected layers. In the dropout, 50% of nodes may not randomly participate in learning in a current epoch. The above-described configuration of layers, the number of nodes, and the sizes of the filters are merely examples to facilitate understanding of the present disclosure, and may be changed. According to an example embodiment, nodes between a fully connected layer and layers included in all extractors of the feature extractor 620 may be fully connected to each other.
For example, the recognizer 630 may generate a feature vector based on values corresponding to all nodes in each of layers included in the feature extractor 620, using the above-described fully connected layer. In this example, the values may include, for example, values corresponding to all nodes in the first layer and values corresponding to all nodes in the second layer.
The recognizer 630 may further include a loss layer (not shown), and may calculate losses corresponding to a plurality of elements using the loss layer. The loss layer may include linear classifiers corresponding to the plurality of elements. The loss layer may predict the plurality of elements from outputs of fully connected layers, using the linear classifiers, and may calculate losses by comparing the predicted elements to actual elements.
The losses may be back-propagated to the fully connected layers and convolution layers, using a back propagation scheme. Based on the back-propagated losses, connection weights in the convolution layers and the fully connected layers may be updated. The configuration of the layers, the activation mode, and a related algorithm described above with reference to
Referring to
In operation 720, a learner of the apparatus may learn a parameter of a recognizer so that the recognizer may recognize a training element from a training image. The recognizer may be configured to recognize a plurality of elements from first feature information and second feature information. For example, the first feature information may be extracted from an input image using a first layer, and the second feature information may be extracted using a second layer. Learning of the parameter of the recognizer will be further described with reference to
The apparatus 810 of
The receiver 811 may receive training data 801. The training data 801 may include a training element and a training image associated with a training object.
The learner 812 may learn a parameter to be provided to the feature recognition apparatus 400 of
In the present disclosure, a parameter may include, for example, a structure of a neural network (for example, a type and a number of layers), a connection weight between nodes, and the like.
For example, the learner 812 may learn a parameter of a first layer of the first extractor 421 so that first features may be extracted from a training image. Additionally, the learner 812 may learn a parameter of a second layer of the second extractor 422 so that second features may be extracted from at least one of the first features. Furthermore, the learner 812 may learn a parameter of the recognizer 430 so that the recognizer 430 may recognize the training element from the first features and the second features.
According to an example embodiment, the learner 812 may calculate a loss corresponding to the training element. When the recognizer 430 includes a neural network, the learner 812 may learn weights between nodes included in the neural network, based on the loss.
For example, the learner 812 may calculate losses between predicted elements and actual elements, because the learner 812 knows training elements labeled to the training image.
The learner 812 may update the parameter to be provided to the recognizer 430 using a back propagation scheme, so that the losses may be reduced. For example, the learner 812 may propagate the losses in an inverse direction, that is, a direction from an output layer to an input layer through a hidden layer in the neural network. When the losses are propagated in the inverse direction, connection weights between nodes may be updated so that the losses may be reduced. As described above, the learner 812 may update the parameter to be provided to the recognizer 430, based on losses corresponding to a plurality of elements. The completely updated parameter may be used in a next learning epoch, and the above-described multitask learning operation may be repeatedly performed until the losses are less than a predetermined or, alternatively, desired threshold. The learned parameter may be provided to the recognizer 430.
The learner 812 may allow the feature recognition apparatus 400 to store a final parameter (for example, a connection weight) of each of the first extractor 421, the second extractor 422 and the recognizer 430.
The structure of the feature extractor 420 of
The training data of
The training image 901 may be associated with a training object. In
According to an example embodiment, a plurality of training images 901 may be provided for a single training object. In
The training element 909 may be labeled to the training object. When the plurality of training images 901 include identical training objects, the same training element 909 may be labeled to the plurality of training images 901. For example, in
Referring to
In operation 1020, a first extractor of the feature verification apparatus may extract first feature information from the input image, using a first layer. The first layer may be, for example, a layer pre-trained on first features based on training data. The first feature information may indicate a first feature corresponding to the input image among the first features.
In operation 1030, a second extractor of the feature verification apparatus may extract second feature information from the first feature information, using a second layer. The second layer may be, for example, a layer pre-trained on second features based on the first features. The second feature information may indicate a second feature corresponding to the first feature information among the second features.
In operation 1040, a verifier of the feature verification apparatus may compare the input image to another image, based on the first feature information and the second feature information of the input image and first feature information and second feature information associated with the other image. For example, the verifier may determine whether the object of the input image is similar to an object of the other image, based on the first feature information and the second feature information of the input image and the first feature information and the second feature information associated with the other image.
The feature verification apparatus 1100 of
The image receiver 1110 may receive an input image 1101 and another image 1102. The other image 1102 may be compared to the input image 1101 and may include an object identical to or different from an object of the input image 1101. For example, the image receiver 1110 may sequentially or simultaneously the input image 1101 and the other image 1102.
The first extractor 1121 may extract first feature information indicating a first feature corresponding to the input image 1101 among first features, using a first layer. The first extractor 1121 may also extract first feature information associated with the other image 1102, using the first layer.
The second extractor 1122 may extract second feature information indicating a second feature corresponding to the first feature information among second features, using a second layer. The second extractor 1122 may also extract second feature information associated with the other image 1102, using the second layer.
The verifier 1130 may compare the first feature information and the second feature information of the input image 1101 to the first feature information and the second feature information associated with the other image 1102, and may determine whether the object of the input image 1101 is similar to or identical to the object of the other image 1102. The verifier 1130 may output, as a verification result 1109, information indicating whether the object of the input image 1101 is identical to the object of the other image 1102. For example, a value of “1” may be output in response to the objects being identical to each other, and a value of “−1” may be output in response to the objects being different from each other.
For example, the verifier 1130 may generate a feature vector associated with the input image 1101 based on the first feature information and the second feature information of the input image 1101, and may generate a feature vector associated with the other image 1102 based on the first feature information and the second feature information associated with the other image 1102. The verifier 1130 may compare the feature vector associated with the input image 1101 to the feature vector associated with the other image 1102, and may determine whether the object of the input image 1101 is similar to the object of the other image 1102.
To determine whether the objects are similar to each other, the verifier 1130 may calculate a similarity between the generated feature vectors. When the similarity exceeds a predetermined or, alternatively, desired threshold similarity, the verifier 1130 may determine that the object of the input image 1101 and the object of the other image 1102 are identical to each other. The similarity between the feature vectors may be calculated, for example, as a level to which feature values of the feature vectors and histograms are similar to each other.
Additionally, the verifier 1130 may calculate a similarity between a set of recognized elements, that is, a result of a recognition of the input image 1101 by the recognizer 430 of
The feature extractor 1120 may include the first extractor 1121 and the second extractor 1122, as shown in
The verifier 1130 may determine whether an object of an input image is similar to or identical to an object of another image, based on first feature information through n-th feature information of each of the input image and the other image.
The feature verification apparatus 1100 may be used to determine whether an object included in an input image is similar to or identical to an object included in another image, based on all of features (for example, a first feature through an (n−1)-th feature) in an abstraction process as well as a feature corresponding to a last level (for example, an n-th feature).
However, the configuration of the feature extractor 1120 is not limited to that of
Referring to
Each of the first extractor 1311 through the third extractor 1313 in the first feature extractor 1310 and the first extractor 1321 through the third extractor 1323 in the second feature extractor 1320 may be trained using a method of
The first feature extractor 1310 may extract first feature information through third feature information associated with an input image. The second feature extractor 1320 may extract first feature information through third feature information associated with another image.
For example, the second feature extractor 1320 may extract the first feature information associated with the other image, using a layer (for example, a layer corresponding to the first extractor 1321 of the second feature extractor 1320) that is separated from a first layer corresponding to the first extractor 1311 and that is trained in the same way as the first layer. The second feature extractor 1320 may also extract the second feature information associated with the other image, using a layer (for example, a layer corresponding to the second extractor 1322 of the second feature extractor 1320) that is separated from a second layer corresponding to the second extractor 1312 and that is trained in the same way as the second layer. The third feature information associated with the other image may also be extracted similarly to the first feature information and the second feature information.
The verifier 1330 may compare a feature vector generated based on the first feature information through the third feature information extracted by the first feature extractor 1310 to a feature vector generated based on the first feature information through the third feature information extracted by the second feature extractor 1320, and may determine whether an object included in the input image is identical to an object included in the other image.
The feature extractor 1120 and the verifier 1130 may be configured in various ways.
Referring to
In
The feature vector generator 1432 may generate a feature vector corresponding to an input image and a feature vector corresponding to another image, may perform a linear transformation, and may transform a dimension of each of the feature vectors. For example, the feature vector generator 1432 may reduce a dimension of the feature vector. The feature vector generator 1532 may perform the same function as that of the feature vector generator 1432.
The verifier 1130 of
The verifier 1130 of
The verifier 1130 of
A number and a configuration of extractors included in the feature extractor 1120 of
Referring to
In operation 1620, a learner of the apparatus may learn a parameter of a verifier based on the training data. The verifier may be configured to compare an input image to another image based on first feature information and second feature information extracted using a first layer and a second layer.
For example, the learner may learn the parameter of the verifier so that a result of comparing, by the verifier, the training images may correspond to the training information. In this example, the verifier may be configured to determine whether two input images are similar to each other, based on first feature information extracted from the two input images using the first layer and based on second feature information extracted using the second layer.
The training information may include information indicating whether the training images include identical objects. For example, when the training images include identical objects, the training information may indicate “1.” When the training images include different objects, the training information may indicate “−1.” The above-described training information is merely an example and accordingly, may include all information indicating true T corresponding to a pair of training images with identical objects and false F corresponding to a pair of training images with different objects.
The apparatus 1710 may include a receiver 1711 and a learner 1712.
The receiver 1711 may receive training data. The training data may include a pair of training images and training information.
The learner 1712 may learn a parameter to be provided to the feature verification apparatus 1100 of
For example, the learner 1712 may learn a parameter of a first layer of the first extractor 1121 so that first features may be extracted from each of the training images, and may learn a parameter of a second layer of the second extractor 1122 so that second features may be extracted from the first features. Also, the learner 1712 may learn a parameter of the verifier 1130 so that the verifier 1130 may output training information based on the first features and second features corresponding to each of the training images.
The learner 1712 may calculate a loss corresponding to the training information. When the verifier 1130 includes a neural network, the learner 1712 may learn weights between nodes included in the neural network, based on the loss.
For example, the learner 1712 may calculate losses between actual training information and a verification result as a result output through the learner 1712 (for example, information indicating whether two input images are identical to each other), because the learner 1712 knows training information labeled to the pair of the training images.
The learner 1712 may update a parameter to be provided to the verifier 1130 so that losses may be reduced using a back propagation scheme, similarly to a learning process for recognition of
Referring to
The pair 1801 of training images may include identical or different training objects. In
The pair 1801 of training images may be, for example, a pair of images with identical or different objects. For example, a pair of an image of an A's face #1 and an image of an A's face #2, or a pair of an image of the A's face #1 and an image of a B's face #1 may be formed as the pair 1801. For example, numerals of the same face may indicate different face states (for example, face angles, facial expressions, and the like).
The training information 1809 may be labeled to the pair 1801 of training images. When the pair 1801 of training images includes an identical training object, the training information 1809 of “1” may be labeled to the pair 1801 of training images. When the pair 1801 of training images includes different training objects, the training information 1809 of “−1” may be labeled to the pair 1801 of training images.
Additionally, the pair 1801 of training images may not be limited to those shown in
According to at least one example embodiment of the inventive concepts, any or all of the units, structures, apparatuses and modules described herein (including, for example, any or all of the deep learning structures 100 and 200, the feature recognition apparatus 400, the feature verification apparatuses 1100 and 1300, and the apparatuses 810 and 1710) may be implemented by one or more circuits or circuitry (e.g., hardware) specifically structured to carry out some or all of the operations described herein as being performed by any or all of the above-referenced units, structures, apparatuses and modules described herein (or an element thereof). According to at least one example embodiment of the inventive concepts, any or all of the above-referenced units, structures, apparatuses and modules described herein may include or be implemented by a memory and one or more processors executing computer-readable code (e.g., software) that is stored in the memory and includes instructions corresponding to some or all of the operations described herein as being performed by any or all of the above-referenced units, structures, apparatuses and modules described herein (or an element thereof). According to at least one example embodiment of the inventive concepts, any or all of the above-referenced units, structures, apparatuses and modules described herein may be implemented by, for example, a combination of the above-referenced hardware and processors executing computer-readable code. For example, the hardware components may include microphones, amplifiers, band-pass filters, audio to digital convertors, and processing devices. A processing device may be implemented using one or more hardware devices configured to carry out and/or execute program code by performing arithmetical, logical, and input/output operations. The processing device(s) may include, for example, one or more processors, a controller and an arithmetic logic unit, a digital signal processor, a microcomputer, a field programmable array, a programmable logic unit, a microprocessor or any other device capable of responding to and executing instructions in a defined manner.
The processing device may run an operating system (OS) and one or more software applications that run on the OS. The processing device also may access, store, manipulate, process, and create data in response to execution of the software. For purpose of simplicity, the description of a processing device is used as singular; however, one skilled in the art will appreciated that a processing device may include multiple processing elements and multiple types of processing elements. For example, a processing device may include multiple processors or a processor and a controller. In addition, different processing configurations are possible, such a parallel processors.
The software may include a computer program, a piece of code, an instruction, or some combination thereof, to independently or collectively instruct and/or configure the processing device to operate as desired, thereby transforming the processing device into a special purpose processor. Software and data may be embodied permanently or temporarily in any type of machine, component, physical or virtual equipment, computer storage medium or device, or in a propagated signal wave capable of providing instructions or data to or being interpreted by the processing device. The software also may be distributed over network coupled computer systems so that the software is stored and executed in a distributed fashion. The software and data may be stored by one or more non-transitory computer readable recording mediums.
The methods according to the above-described example embodiments may be recorded in non-transitory computer-readable media including program instructions to implement various operations of the above-described example embodiments. The media may also include, alone or in combination with the program instructions, data files, data structures, and the like. The program instructions recorded on the media may be those specially designed and constructed for the purposes of example embodiments. Examples of non-transitory computer-readable media include magnetic media such as hard disks, floppy disks, and magnetic tape; optical media such as CD-ROM discs, DVDs, and/or Blue-ray discs; magneto-optical media such as optical discs; and hardware devices that are specially configured to store and perform program instructions, such as read-only memory (ROM), random access memory (RAM), flash memory (e.g., USB flash drives, memory cards, memory sticks, etc.), and the like. Examples of program instructions include both machine code, such as produced by a compiler, and files containing higher level code that may be executed by the computer using an interpreter. The above-described devices may be configured to act as one or more software modules executed by a hardware-implemented processor in order to perform the operations of the above-described example embodiments.
Example embodiments of the inventive concepts having thus been described, it will be obvious that the same may be varied in many ways. Such variations are not to be regarded as a departure from the intended spirit and scope of example embodiments of the inventive concepts, and all such modifications as would be obvious to one skilled in the art are intended to be included within the scope of the following claims.
Choi, Chang Kyu, Han, Seungju, Han, Jaejoon, Suh, Sungjoo
Patent | Priority | Assignee | Title |
11216694, | Aug 08 2017 | SAMSUNG ELECTRONICS CO , LTD | Method and apparatus for recognizing object |
Patent | Priority | Assignee | Title |
5774633, | Mar 13 1989 | Hitachi, Ltd. | Supporting neural network method for process operation |
6038338, | Feb 03 1997 | NAVY, UNITED STATES OF AMERICA AS REPRESENTED BY THE SECRETARY OF THE NAVY | Hybrid neural network for pattern recognition |
7788196, | Sep 09 2003 | Semeion | Artificial neural network |
8386401, | Sep 10 2008 | DIGITAL INFUZION, INC | Machine learning methods and systems for identifying patterns in data using a plurality of learning machines wherein the learning machine that optimizes a performance function is selected |
8615476, | Apr 15 2009 | University of Southern California | Protecting military perimeters from approaching human and vehicle using biologically realistic neural network |
9330171, | Oct 17 2013 | GOOGLE LLC | Video annotation using deep network architectures |
9400955, | Dec 13 2013 | Amazon Technologies, Inc.; Amazon Technologies, Inc | Reducing dynamic range of low-rank decomposition matrices |
9530047, | Nov 30 2013 | BEIJING SENSETIME TECHNOLOGY DEVELOPMENT CO , LTD | Method and system for face image recognition |
20100321229, | |||
20130290222, | |||
20140180989, | |||
20140229158, | |||
20160117587, | |||
20160180151, | |||
KR100716376, | |||
KR101254181, | |||
KR19990052660, | |||
KR20050043869, | |||
KR20120052610, | |||
WO2016025189, |
Executed on | Assignor | Assignee | Conveyance | Frame | Reel | Doc |
May 30 2019 | Samsung Electronics Co., Ltd. | (assignment on the face of the patent) | / |
Date | Maintenance Fee Events |
May 30 2019 | BIG: Entity status set to Undiscounted (note the period is included in the code). |
Date | Maintenance Schedule |
Aug 17 2024 | 4 years fee payment window open |
Feb 17 2025 | 6 months grace period start (w surcharge) |
Aug 17 2025 | patent expiry (for year 4) |
Aug 17 2027 | 2 years to revive unintentionally abandoned end. (for year 4) |
Aug 17 2028 | 8 years fee payment window open |
Feb 17 2029 | 6 months grace period start (w surcharge) |
Aug 17 2029 | patent expiry (for year 8) |
Aug 17 2031 | 2 years to revive unintentionally abandoned end. (for year 8) |
Aug 17 2032 | 12 years fee payment window open |
Feb 17 2033 | 6 months grace period start (w surcharge) |
Aug 17 2033 | patent expiry (for year 12) |
Aug 17 2035 | 2 years to revive unintentionally abandoned end. (for year 12) |