The characteristic data for determining the characteristics of the transfer functions (for example, sigmoid functions) of the neurons of the hidden layer and the output layer (the gradients of the sigmoid functions) of a neural network are learned and corrected in a manner similar to the correction of weighting data and threshold values. Since at least one characteristic data which determines the characteristics of the transfer function of each neuron is learned, the transfer function characteristics can be different for different neurons in the network independently of the problem and/or the number of neurons, and be optimum. Accordingly, a learning with high precision can be performed in a short time.

Patent
   5280564
Priority
Feb 20 1991
Filed
Feb 18 1992
Issued
Jan 18 1994
Expiry
Feb 18 2012
Assg.orig
Entity
Large
14
10
EXPIRED
1. A neural network comprising a plurality of neurons each performing a signal processing corresponding to a neural element, said neurons being hierarchically connected in the order of at least one input layer, at least one hidden layer and an output layer, and said input layer, hidden layer and output layer each consisting of at least one neuron, individual neurons of said hidden layer and output layer being operative to correct the data weighted by multiplying the outputs of the preceding input layer or hidden layer by predetermined weighting data with a predetermined threshold value, respectively, and to substitute the data after the correction in a predetermined transfer function to calculate output data, said neural network including:
error data calculation means operative to use a predetermined cost function to calculate error data from the output data of the output layer,
weighting data correction means operative to partially differentiate said cost function with each of a plurality of weighting variables to obtain partial differentiated coefficients, and to correct said weighting data according to said partial differential coefficients.
threshold value correction means operative to partially differentiate said cost function with each of a plurality of threshold variables to obtain partial differentiated coefficients, and to correct each said threshold value according to said partial differential coefficients, and
characteristic data correction means for partially differentiating said cost function, to obtain partial differentiated coefficients, with at least one characteristic variable that determines the characteristics of the respective transfer functions of said hidden layer and said output layer, and operative to correct said characteristic data according to said partial differential coefficients.
2. A neural network as set forth in claim 1 wherein said characteristic data correction means corrects said characteristic data using the partial differential coefficients obtained by partially differentiating said cost function with said at least one characteristic variable, and the previous values of said partial differential coefficients.
3. A neural network as set forth in claim 2 wherein said cost function is an error function, and said error data calculation means compares the data output from said output layer with training data corresponding to the input data supplied to said input layer and applies the comparison result to said error function, thereby to calculate error data.
4. A neural network as set forth in claim 2 wherein said weighting data correction means uses the partially differentiated value of said cost function by said weighting variables and the previous value of said partially differentiated value to correct said weighting data.
5. A neural network as set forth in claim 2 wherein said threshold value correction means uses the partially differentiated variables of said cost function by said threshold value and the previous value of said partially differentiated value to correct said threshold value.
6. A neural network as set forth in claim 1 wherein said transfer function is a sigmoid function and said characteristic variable is the gradient of the sigmoid function.
7. A neural network as set forth in claim 1 wherein said transfer function is a power function (axb +c) and said characteristic variable is at least one constant of said power function.
8. A neural network as set forth in claim 1 wherein said transfer function is a trigonometric function {a Sin(cx)+b Cos(dx)} and said characteristic variable is at least one constant of said trigonometric function.
9. A neural network as set forth in claim 1 wherein said cost function is an error function, and said error data calculation means compares the data output from said output layer with training data corresponding to the input data supplied to said input layer and applies the comparison result to said error function, thereby to calculate error data.
10. A neural network as set forth in claim 1 wherein said weighting data correction means uses the partially differentiated value of said cost function by said weighting variables and the previous value of said partially differentiated value to correct said weighting data.
11. A neural network as set forth in claim 1 wherein said threshold value correction means uses the partially differentiated value of said cost function by said threshold variables and the previous value of said partially differentiated value to correct said threshold value.

1. Field of the Invention

The present invention is related to a neural network and particularly to a neural network which can learn with high precision and at a high speed.

2. Description of the Prior Art

Neural networks which perform signal operations in parallel using neurons that process signals in a similar manner to neural elements can be provided with many functions such as

(1) a pattern recognition function,

(2) an interpolation function, and

(3) predicting and optimizing multivariable functions,

and have attracted considerable attention as new information processing means not only in the manufacturing industry but also in a variety of fields such as medical science and economic forecast.

In a neural network of a hierarchical structure consisting of an input layer, a hidden layer and an output layer, a learning algorithm of back propagation using training data has been known.

The learning algorithm based on the conventional back propagation operation is described below.

FIG. 14 shows an example of the neural network of a three-layer structure, wherein the input layer consists of P neurons 100(p) (p=1∼P), the hidden layer consists of Q neurons 200(q) (q=1∼Q) and the output layer consists of R neurons 300(r) (r=1∼R). In this example, P=4, Q=6 and R=2. The numbers of neurons included in the input and output layers depend on the number of input and output data, and the number of neurons of the hidden layer can arbitrarily be set.

FIG. 15 is a conceptual diagram showing a neuron (unit) which is an element of a neural network. Each neuron fetches data obtained by respectively multiplying a plurality of supplied data X(1)∼X(n) by each of weights W(1)∼W(n) as input data, and calculates a value "x" by subtracting a threshold value θ from the total sum ΣW(n)X(n) of them. And, as a transfer function f(x) with this "x" as a variable, a function which can be differentiated by "x" is used. A sigmoid function, a typical example of transfer function, is described in equation (1) in FIG. 24A.

The relationships between "x" and f(x) when coefficient "k" of the sigmoid function (hereinafter referred to as gradient "k") is assumed to be 0.1, 1.0 and 10 are shown in FIG. 16, and the differentiated value of the sigmoid function is shown in equation (2) in FIG. 24A.

With reference to FIG. 17, a learning algorithm by the conventional back propagation is now described. XO1(1), XO1(2), . . . XO1(P) are output data of the individual neurons 100(p) in the input layer, XO2(1), XO2(2), . . . XO2(Q) are output data of the individual neurons 200(q) in the hidden layer, XO3(1), XO3(2), . . . XO3(R) are output data of the individual neurons 300(r) in the output layer, XI2(1), XI2(3), . . . XI2(Q) are input data of the individual neurons 200(q) in the hidden layer, XI3(1), XI3(2), . . . XI3(R) are input data of the individual neurons 300(r) in the output layer, and T(1), T(2), . . . T(R) are training data.

W12 is weighting data for obtaining input data of the hidden layer from output data of the input layer, and W12(1,1) is weighting data when the output XO(1) is given to XI2(1,1) and WI2(2,2) is weighting data when the output XO1(2) is given to XI2(2) In general, W12(P,Q) is weighting data when XO1(P) is given to XI2(Q). Accordingly, although not shown, for instance, the weighting data when XO1(2) is given to XI2(1) is W12(2,1), and that when XO1(P) is given to XI2(1) is W12(p,1).

Similarly, W23 is weighting data for obtaining input data of the output layer from output data of the hidden layer, and in general, W23(Q,R) is weighting data when XO2(Q) is given to XI3(R). θ2(1), θ2(2), . . . θ2(Q) are threshold values of the hidden layer, and θ3(1), θ3(2), . . . θ3(R) are threshold values of the output layer.

XI2(1), input data of the hidden layer, is expressed by equation (3).

XI2(1)=XO1(1)×W12(1,1)+XO1(2)×W12(2,1) +. . . +XO1(P)×W12(P,1) (3)

Similarly, XI2(Q) is expressed by equation (4).

XI2(Q)=XO1(1)×W12(1,Q)+XO1(2)×W12(2,Q) +. . . +XO1(P)×W12(P,Q) (4)

XI3(1), input data of the output layer, is expressed by equation (5).

XI3(1)=XO2(1)×W23(1,1)+XO2(2)×W23(2,1) +. . . +XO2(Q)×W23(Q,1) (5)

Similarly, XI3(R) is expressed by equation (6).

XI3(R)=XO2(1)×W23(1,R)+XO2(2)×W23(2,R) +. . . +XO2(Q)×W23(Q,R) (6)

In the hidden layer, output data XO2(1) is obtained from input data XI2(1) by the equation (7). ##EQU1##

Similarly, the calculation for obtaining XO2(Q) from XI2(Q) is based on equation (8).

XO2(Q)=f{XI2(Q)-θ2(Q)} (8)

The calculation for obtaining XO3(1) from input data XI3(1)and the calculation for obtaining XO3(R) from input data XI3(R) are based on equations (9) and (10), respectively.

XO3(1)=f{XI3(1)-θ3(1)} (9)

XO3(R)=f{XI3(R)-θ3(R)} (10)

When output data XO3(r) (r=1∼R) have been obtained from the R neurons constituting the output layer for data XO1(p) (p=1∼P) given to the P neurons constituting the input layer, as described above, error data E are calculated on the basis of a predetermined cost function using the output data XO3(r) (r=1∼R) and training data T(r) (r=1∼R). When training data are used as in this example, such an error function as shown in equation (11) of FIG. 24A can be used as a cost function. Then, equations (12) to (15) are calculated using the error or cost function E.

Equations (12) to (15) of FIG. 24A represent partial differentiations of the error or cost function E by weighting variables W23(q,r), threshold variables θ3(r), weighting variables W12(p,q) and threshold variables θ2(q), respectively. That is, in equation (12), E is partially differentiated by all the combinations of W23(q,r) (all the combinations of q=1∼Q and r=1∼R, namely, W23(1,1), W23(2,1), . . . , W23(Q,1), W23(1,2), W23(2,2), . . . , W23(Q,2), . . . , W23(1,R), W23(2,R), . . . , W23(Q,R)). Similarly for equations (13) to (15), E is partially differentiated using all of the threshold variables θ3(r), weighting variables W12(p,q) and threshold variables θ2(q).

Then, the changing amount of current processing of each weighting data or threshold value ΔW12(p,q)0, Δθ2(q)0, ΔW23(q,r)0 and Δθ3(r)0 are determined from equations (16) to (19).

The various changing amounts in one time before (previous), two times before, . . . , and N times before precessings are discriminated by replacing the suffix 0 for the changing amount in current processing shown in equations (16) to (19) with 1∼N, as shown in equation (20). Further, the various changing coefficients in current, one time before, two times before, . . . , and N times before processings are determined by an equation (21).

These changing coefficients α0∼αN and β0∼βN may be preset to any values. The accumulated changing amounts ΔW12(p,q), Δθ2(q), ΔW23(q,r) and Δθ3(r) used in current processing are calculated from equations (22) to (25).

The accumulated changing amounts ΔW12(p,q), Δθ2(q), ΔW23(q,r) and Δθ3(r) calculated in this way are added to the weighting data and threshold value data W12(p,q), θ2(q), W23(q,r) and θ3(r), respectively, to correct them.

If there are "A" sets of combinations of input data and training data, those processings are repeated "A" times on all sets of said combinations. With this, one-time learning operation is finished. Predetermined times of such learnings are subsequently executed.

For example, for the learning of a problem of exclusive-OR (XOR) as shown in Table 1 described at the end of the specification, the numbers of neurons are two for the input layer and one for the output layer, and the hidden layer may have any number of neurons. In this example, the sets of the combinations of input data and training data is four.

As mentioned above, in the conventional back propagation operation, only weighting data and threshold values were subject to change by learning. The above described learning process is further described with reference to FIG. 18.

FIG. 18 is a flowchart showing an example of the learning algorithm according to the traditional back propagation operation.

In step S1, an input data pattern is selected (in the example of Table 1, one of four sets of input data is selected), and the selected data is supplied to the input layer.

In step S2, predetermined calculations are performed in the input hidden and output layers using the input data. By this, data (calculation result) is output from the output layer.

In step S3, the output data is compared with the training data corresponding to the selected input data, and error data "E" is calculated by equation (11).

In step S4, accumulated changing amounts ΔW12(p,q), Δθ2(q), ΔW23(q,r) and Δθ3(r) are calculated by equations (22) to (25).

In step S5, the calculated accumulated changing amounts are added to W12(p,q), θ2(q), W23(q,r) and θ3(r), which are weighting data and threshold value data, respectively, thereby to correct them.

In step S6, it is determined whether or not all of the input patterns have been selected. If all of the input patterns have not yet been selected, the process returns to step S1, and the process moves to step S7 if they have all been selected.

In step S7, it is determined that one learning operation has been completed.

In step S8, it is determined whether or not the learning has been completed predetermined times, and the process returns to step S1 if it has not been completed and terminates if it has been completed. Alternatively, in step S8, it may by determined whether or not output data can be obtained with precision higher than a pre-determined value, that is, whether or not the error data "E" has become smaller than a predetermined one.

The functions of the above described neural network is described more specifically using FIG. 19 in which input data and training data as shown in Table 1 are stored in an input/training data memory means 11, for instance. A selection means 12 sequentially transfers input data from the memory means 11 to neurons 100(p) of the input layer, and simultaneously sequentially outputs training data corresponding to the transferred input data to error calculation means 13.

Weighting calculation means 51 multiplies each output signal XO1(p) of neurons 100(p) by weighting data W12(p,q) stored in weighting data memory means 21 and outputs the result to each neuron 200(q) of the hidden layer. Similarly, weighting calculation means 52 multiplies each output signal XO2(q) of neurons 200(q) by weighting data W23(q,r) stored in weighting data memory means 22 and outputs the result to each neuron 300(r) of the output layer.

Transfer function calculation means 61 performs the calculations shown in equations (7), (8), etc. in each neuron 200(q), using the input data and the threshold values θ2(q) stored in threshold value memory means 31. Similarly, transfer function calculation means 62 performs the calculations shown in equations (9), (10), etc. within each neuron 300(r), using the input data and the threshold values θ3(r) stored in threshold value memory means 32.

The error calculation means 13 calculates error data "E" by performing the calculation of equation (11) using the training data and the data output from the output layer.

Accumulated changing amount calculation means 20 uses error data "E" and weighting data W12(p,q) to calculate accumulated changing amount ΔW12(p,q). In addition, using the error data "E" and weighting data W23(q,r), it calculates accumulated changing amount ΔW23(q,r).

Accumulated changing amount calculation means 30 uses error data "E" and threshold values θ2(q) to calculate accumulated changing amount Δθ2(q). Also, using the error data "E" and threshold values θ3(r), it calculates accumulated changing amount Δθ3(r).

The accumulated changing amounts ΔW12(p,q), Δθ2(q), ΔW23(q,r) and Δθ3(r) calculated in the individual accumulated changing amount calculation means 20 and 30 are added, in the adding means 23, 33, 24 and 34, respectively, to W12(p,q), θ2(q), W23(q,r) and θ3(r) which are the weighting data and threshold values respectively stored in the memory means 21, 31, 22 and 32. The weighting data and threshold values after the additions or change are stored again in the memory means 21, 31, 22 and 32.

In such conventional back propagation, the learning speed and precision of a neural network largely varied depending on the characteristics (in the case of a sigmoid function, the value of gradient "k") of the transfer function of individual neurons (in the above described example, sigmoid function), and sometimes the learning did not advance or the error could not converge.

A table 2 shows the result obtained by the learning of the problem of exclusive-OR as shown in Table 1 by a neural network which uses a sigmoid function as a transfer function according to the prior art. The neural network used in the learning comprises two neurons for the input layer, three neurons for the hidden layer and one neuron for the output layer. Table 2 shows the error sum of squares after the learning was performed ten thousand times respectively for the gradients "k" of the sigmoid function being fixed at 0.1, 1.0, 2.0, 3.0 and 10∅ The error sum of squares is the error data for one-time learning, namely, the mean value of four kinds of E's calculated when all of the four input patterns are supplied, and it is defined by equation (11).

As mentioned above, the learning speed and precision of the neural network greatly changes when the gradient "k" of the sigmoid function varies. Accordingly, a learning with high precision can be performed in a short time if the gradient "k" is set to an optimum value.

However, only the fact that the optimum value of the gradient "k" depends on the problem and the number of the neurons constituting the neural network has been revealed by the studies which have been made up to now, and the rule of trial and error must be applied for each problem to find the optimum value of "k", which is cumbersome and time-consuming. In addition, sufficient precision cannot be obtained unless the value of "k" is set at a different value for each neuron.

It is the object of the present invention to provide a neural network wherein the characteristics of the transfer function (gradient "k" for a sigmoid function, and various coefficients for other power or trigonometric function) can automatically be optimized independent of the problem to be solved and the number of the neurons constituting the network.

In the present invention, weighting data and threshold values are changed through learning, and at least one characteristic data (such as a coefficient) for determining the characteristics of the transfer function of each neuron is also changed through learning.

By changing each characteristic data through learning, the characteristics of each transfer function is optimized so as to be different for each neuron.

FIG. 1 is a functional block diagram of an embodiment of the present invention.

FIG. 2 is a conceptual diagram for explaining a part of the learning algorithm of the embodiment of the present invention.

FIG. 3 is a flowchart showing the learning algorithm of the embodiment of the present invention.

FIG. 4 is a graph showing the convergence process when the problem of XOR was learned using the embodiment of the present invention.

FIG. 5 is a graph showing the convergence process when the problem of XOR was learned using conventional approach.

FIG. 6 is a graph showing the convergence process when the problem of XOR was learned according to another embodiment of the present invention.

FIG. 7 is a graph showing the convergence process when the problem of multiplication was learned using the embodiment of the present invention.

FIG. 8 is a graph showing the process when the problem of multiplication was learned using the conventional approach.

FIG. 9 is a graph showing an example of a special problem.

FIG. 10 is a graph showing the result when the example special problem of FIG. 9 was learned using the embodiment of the present invention.

FIG. 11 is a graph showing the convergence process when the example special problem of FIG. 9 was learned using the one embodiment of the present invention.

FIG. 12 is a graph showing the result when the example special problem of FIG. 9 was learned using the conventional approach.

FIG. 13 is a graph showing the process when the example special problem of FIG. 9 was learned using the conventional approach.

FIG. 14 is a diagram showing an example of the neural network of a three-layer structure.

FIG. 15 is a conceptual diagram showing a neuron which is an element of the neural network.

FIG. 16 is a representation showing an example characteristic of a sigmoid function.

FIG. 17 is a conceptual diagram for explaining the main part of the learning algorithm according to the conventional back propagation.

FIG. 18 is a flowchart showing an example of the learning algorithm according to the conventional back propagation.

FIG. 19 is a functional block diagram of the conventional neural network.

FIG. 20 is a graph showing the positions of the input points of a periodic function to be interpolated and the interpolated periodic function waveform.

FIG. 21 is a graph showing the convergence process when the periodic function problem was learned using still another embodiment of the present invention.

FIG. 22 is a graph showing the convergence process when the same periodic function problem as FIG. 21 was learned using the conventional approach.

FIG. 23 is a periodic function waveform of the learning result by the approach shown in FIG. 22.

FIGS. 24A-24D describe equations 1, 2, 11-34 which are used for the present invention.

The present invention is now described in detail with reference to the drawings.

FIG. 2 is a conceptual diagram for explaining the main part of the learning algorithm according to one embodiment of the present invention. In the figure, the same symbols as in FIG. 17 represent the same or equivalent portions.

The transfer function of the individual neurons constituting the hidden layer is a sigmoid function expressed by equation (1), and k2(q) (q=1∼Q) represents the gradient of the sigmoid function similarly to "k" in the equation (1). The variable "x" of the sigmoid function of the hidden layer is {XI2(q)-θ2(q)} (q=1∼Q).

Similarly, the transfer function of the individual neurons constituting the output layer is also a sigmoid function, and k3(r) (r=1∼R) represents the gradient thereof. The variable "x" of the sigmoid function of the output layer is {X13(r)∼θ3(r)} (r=1∼R).

In a neural network having the above-mentioned construction, error data "E" is calculated from equation (11). Using the error or cost function "E", equations (12) to (15) are calculated. Of course, in this case, the calculations are performed for all the combinations of each of variables W23(q,r), θ3(r), W12(p,q) and θ2(q) (p=1∼P, q=1∼Q, r=1∼R) as described above. Equations (26) and (27) are also calculated. These calculations are also performed for all the combinations of each of variables k3(r) and k2(q) (r=1∼R, q=1∼Q).

Then, the changing amounts of current processing of the respective weighting data and threshold values ΔW12(p,q)0, Δθ2(q)0, ΔW23(q,r)0 and Δθ3(r)0 are determined by equations (16) to (19).

Similarly, the changing amounts of current processing Δk2(q)0 and Δk3(r)0 for the respective gradients k2 and k3 are determined by equations (28) and (29). In addition, the various changing amounts in the one-time-before (previous), two-time-before, . . . , N-time-before processings are discriminated by replacing the suffix 0 of the changing amount of current processing by 1∼N, as shown in equations (20) to (30).

The various changing coefficients in the current, one-time-before, two-time-before, . . . , N-time-before processings are determined as shown in equations (21) and (31). Each of the changing coefficients α0∼αN, β0∼βN and γ0∼γN may be preset to any value. The accumulated changing amounts ΔW12(p,q), Δθ2(q), ΔW23(q,r) and Δθ3(r) used in current processing are calculated by equations (22) to (25), and the accumulated changing amounts Δk2(q) and Δk3(r) are calculated from equations (32) and (33).

The accumulated changing amounts ΔW12(p,q), Δθ2(q), ΔW23(q,r), Δθ3(r), Δk2(q) and Δk3(r) calculated in the described method are respectively added to the weighting data and threshold values W12(p,q), θ2(q), W23(q,r) and θ3(r), and gradients k2(q) and k3(r) to correct them. In equations (22) to (25) and equations (32) and (33), each accumulated changing amount may be calculated using the changing amount of current processing and that of at least one time before processing.

If there are "A" sets of combinations of the input data and training data, these processings are performed A-times for all the combinations. With this, one-time learning is completed. Such learning is subsequently executed predetermined times.

The learning algorithm in the embodiment of the present invention changes or corrects the weighting data, threshold values and gradients at step S5A, as shown in the flowchart of FIG. 3. Of course, in the previous step S4A, the accumulated changing amounts Δk2(q) and Δk3(r) for changing gradients are also calculated in addition to the accumulated amounts calculated in the prior art.

If "x" and "k" of the sigmoid function expressed by equation (1) are variables, partial differentiations by "x" and "k" can be expressed by equation (34).

The embodiment of the present invention is described more specifically with reference to the functional block diagram of FIG. 1 in which the same symbols as in FIG. 19 represent the same or equivalent portions.

In FIG. 1, the transfer function calculation means 61 and 62 perform calculations using sigmoid functions as described in connection with FIG. 19. The gradients k2(q) and k3(r) of the sigmoid functions are stored in the gradient memory means 41 and 42, respectively.

Accumulated changing amount calculation mean 40 calculates accumulated changing amount Δk2(q) using the error data "E" and gradient k2(q). In addition, using the "E" and gradient k3(r), the accumulated changing amount Δk3(r) is calculated. The accumulated changing amounts Δk2(q) and Δk3(r) are added to the gradients k2(q) and k3(r) stored in memory means 41 and 42 in adding means 43 and 44. The gradients corrected by the addition are newly stored in the memory means 41 and 42 for gradient correction.

The present inventors calculated various problems using the neural network according to the learning algorithm as described above. Several calculation results are shown below.

(1) XOR

FIG. 4 shows a convergence state, which was obtained when the problem of XOR shown in Table 1 was learned by the approach of the present invention, using a neural network consisting of a two-neuron input layer, three-neuron hidden layer and one-neuron output layer. The transfer function of each neuron is a sigmoid function, and in FIG. 4, the relationship between the number of times of learning (abscissa) and error sum of squares (ordinate) is shown. For the purpose of comparison the relationship between the number of times of learning (abscissa) and error sum of squares ordinate) according to the conventional method, in which the gradient "k" of the sigmoid function is not learned, is shown in FIG. 5. FIG. 5 is an example for k=2.

In Table 3, learning precision obtained by ten thousand times learnings, that is the error sum of squares, and calculation time required for that are shown. For calculation, a workstation of 20 [MIPS] was used.

Table 4 shows the number of times of learning required before the error sum of squares 1×10-4 was obtained and the calculation time spent for that. As seen from these tables, regarding the problem of XOR, the learning precision could be increased by 10 to 100 times for the same number of times of learning as compared with the conventional method in accordance with the embodiment of the present invention, and the number of times of learning for obtaining the same learning precision can be decreased by at least 1/3 to 1/4 times (1/2 to 1/3 times in calculation time) as compared with the conventional method.

(2) Multiplication

FIG. 7 shows a learning process when a problem of multiplication as shown in Table 7 was learned by the technique of the present invention using a neural network, which consists of a two-neuron input layer, a nine-neuron hidden layer and a single-neuron output layer. The transfer function of each neuron is a sigmoid function. In FIG. 7, the relationship between the number of times of learning (abscissa) and error sum of squares (ordinate) is shown as in FIGS. 4 and 5.

The relationship between the number of times of learning (abscissa) and error sum of squares (ordinate) according to the conventional approach is shown for comparison in FIG. 8. FIG. 8 is an example for k=2.

As seen from the comparison of the two figures, even a problem for which learning precision was hardly increased by the conventional approach can be learned up to considerable precision in the present invention.

(3) Special problem

The result of a learning of a functional relation as shown in FIG. 9 is described below. The network used consists of one neuron for the input layer, seven neurons for the hidden layer and one neuron for the output layer, and the transfer function of each neuron is a sigmoid function.

Table 5 shows the calculation result of the embodiment of the present invention, in which error sums of squares are shown with a changing coefficient of α=0.9 and the changing coefficients β and γ being parameters.

In table 5, each of marks `×` indicates that convergence was not achieved because of a too large error and learning could not be made at last, each of marks `Δ` indicates that the converged error sum of squares was in the order of 0.1×10-5, each of marks `◯` indicates that the converged error sum of squares was in the order of 0.1×10-6, and a mark `⊚` indicates that the converged error sum of squares was the smallest and its value was 0.278×10-6. The input and output data for the best case are shown in FIG. 10. In FIG. 11, the relationship between the number of times of learning (abscissa) and error sum of squares (ordinate) is shown, as in FIGS. 4 and 5.

For the purpose of comparison, the calculation result by the conventional approach is shown in Table 6. Table 6 shows error sums of squares when the changing coefficients α and β are parameters. The relation between the input and output data for α=0.1 and β=0.1 which provided the smallest error sum of squares is shown in FIG. 12. In FIG. 13, the relationship between the number of times of learning (abscissa) and error sum of squares (ordinate) is shown.

It will be realized from the foregoing that the problem of FIG. 9 can be learned by the present invention with considerable high precision, though it can be hardly learned by the conventional learning technique.

Now, description is made of another embodiment in which a power function is used as the transfer function.

(4) XOR of Table 1

The neural network used consists of two neurons for the input layer, two neurons for the hidden layer and one neuron for the output layer, and the transfer function of each neuron is a power function (axb +c). Two coefficients b1 and b2 of the power functions for the two neurons of the hidden layer were set at 1.0 and 2.0, respectively, and constants a1 and a2, and c1 and c2 were learned. FIG. 6 shows the attenuation state of the error sum of squares (ordinate) vs. the number of times of learning (abscissa). Table 8 shows the result of a thousand times learnings. The error sum of squares for this was 1.40×10-13, and the required calculation time was 0.93 sec. For the purpose of comparison, the result of a thousand times learnings performed in the conventional method using a sigmoid function is shown in Table 9. The error sum of squares for said result was 7.31×10-5, and the required calculation time was 1.15 sec. It is understood from these comparisons that a learning with higher precision can be accomplished by the present invention in a shorter time.

(5) Interpolation of periodic function

Description is made of a further embodiment of the present invention in which a trigonometric function is used as the transfer function. The exercise is to obtain a periodic function which interpolates a periodic function, in which output values (ordinate) are given corresponding to discrete values of input (abscissa) as shown in Table 10 and by black points in FIG. 20. The network used consists of one neuron for the input layer, two neurons for the hidden layer and one neuron for the output layer, the transfer function of each neuron is a trigonometric function {a Sin(cx)+b Sin(dx)}, and all constants in the trigonometric function were learned.

The relationship between the number of times of learning (abscissa) and error sum of squares (ordinate) is shown in FIG. 21. The error sum of squares after ten thousand times learnings was 6.07×10-7, and the required time was 88 sec. The waveform of the periodic function obtained by the learning is shown by a dotted line in FIG. 20. For comparison, the result of a hundred thousand times learnings by the conventional method using a sigmoid function, which is similar to FIG. 21, is shown in FIG. 22. The error sum of squares after the hundred thousand times learnings was 1.13×10-2, and the required time was 21 sec. It is seen from FIG. 22 that the error sum of squares does not readily decrease. Also, the waveform of the periodic function obtained by the prior art learning, which corresponds to FIG. 20, is shown in FIG. 23. As seen from the comparison of FIGS. 20 and 23, and/or FIGS. 21 and 22, a learning with higher precision can be performed by the present invention in a shorter time.

Although, in the above description, the learning (changing) of the weighting data, threshold values and gradients are performed by adding accumulated changing amounts of past changing amounts to the weighting data, threshold values and gradients, as shown in equations (22) to (25) and equations (32) and (33), the learning may be made using only the changing amounts of current processing.

Moreover, in the above described embodiments, the changing coefficients α, β and γ are common for the weighting data, threshold values and gradients of each of the hidden and output layers, but coefficients different among layers and/or neurons may be used.

In the above described embodiments, if there are a plurality of input data and training data as shown in Table 1 and FIG. 6, the combination of input data and training data are sequentially selected out of them and are calculated as shown in FIG. 3, and it is determined that one learning has been completed when the calculations have been performed using all the sets of input data and training data. As a result, all combinations of input and training data are evenly used to perform the learning. However, for instance, data may be selected at random from a plurality of input and training data to perform a learning, without evenly using data.

Although the weighting data, threshold values and gradients were to be corrected each time error data is detected in the foregoing description, the changing amounts (or the accumulated changing amounts) used for correction may be stored, and when the learning has been completed a predetermined number of times, they may be collectively added to the weighting data, threshold values and gradients to make data correction.

Although the foregoing description has been made to the effect that error data "E" is calculated using training data with an error function as shown in equation (11) being as a cost function, it is also possible that an appropriate cost function is used instead of training data and the learning is made so that the calculation result of the cost function becomes a predetermined value (for instance, the smallest value).

The above described various calculations can be executed by software using a microcomputer, but they may be accomplished instead by hardware circuit networks.

As apparent from the above description, in accordance with the present invention, at least one characteristic data which determines the characteristics of the transfer function of each neuron is to be learned, and thus the transfer function characteristics are different for each neuron independently of the nature of a problem and/or the number of neurons and become optimum. Accordingly, a learning with high precision can be performed in a short time.

If it is taken into consideration that neural networks have been studied on the analogy of the neural circuit networks of living things and the individual nerve cells of a living thing are not always uniform in the same and single living body, it can be said that the learning system of the present invention is quite appropriate.

TABLE 1
______________________________________
Training
Input Data
X (1) X (2) Y
______________________________________
0 1 1
1 0 1
0 0 0
1 1 0
______________________________________
TABLE 2
______________________________________
Gradient of Error Sum of Squares
Sigmoid Function
After 10000 Learnings
______________________________________
0.1 2.5 × 10-1
1.0 1.6 × 10-4
2.0 3.3 × 10-5
3.0 2.3 × 10-5
10.0 5.0 × 10-1
______________________________________
TABLE 3
______________________________________
Calculation
Error Sum of Squares
Time [sec]
______________________________________
Conventional Method
0.326 × 10-4
5.45
Present Invention
0.506 × 10-5
5.84
______________________________________
TABLE 4
______________________________________
Number of Times of
Calculation
Learning Time [sec]
______________________________________
Conventional Method
3592 1.99
Present Invention
1184 0.73
______________________________________
TABLE 5
______________________________________
α = 0.9
Changing Changing Coefficient γ
Coefficient
0.1 0.3 0.5 0.7 0.9
______________________________________
β
0.1 Δ Δ
x x x
0.3 Δ x x x x
0.5 x x ∘
0.7 x x ∘
0.9 x x x Δ
Δ
______________________________________
TABLE 6
______________________________________
Changing Changing Coefficient β
Coefficient
0.1 0.3 0.5 0.7 0.9
______________________________________
α
0.1 0.147 0.159 0.171 0.172
0.173
0.3 0.159 0.161 0.167 0.173
0.178
0.5 0.171 0.172 0.173 0.174
0.177
0.7 0.175 0.176 0.176 0.177
0.177
0.9 0.178 0.178 0.179 0.179
0.179
______________________________________
TABLE 7
______________________________________
Input Training
X (1) X (2) Data Y
______________________________________
0 0 0
1 0 0
2 0 0
3 0 0
4 0 0
5 0 0
0 1 0
1 1 1
2 1 2
3 1 3
4 1 4
5 1 5
0 2 0
1 2 2
2 2 4
3 2 6
4 2 8
5 2 10
0 3 0
1 3 3
2 3 6
3 3 9
4 3 12
5 3 15
0 4 0
1 4 4
2 4 8
3 4 12
4 4 16
5 4 20
0 5 0
1 5 5
2 5 10
3 5 15
4 5 20
5 5 25
______________________________________
TABLE 8
______________________________________
Input Output
X (1) X (2) Y
______________________________________
0.00000 0.00000 0.00000
0.00000 1.00000 1.00000
1.00000 0.00000 1.00000
1.00000 1.00000 0.00000
______________________________________
TABLE 9
______________________________________
Input Output
X (1) X (2) Y
______________________________________
0.00000 0.00000 0.00985
0.00000 1.00000 0.97118
1.00000 0.00000 0.97214
1.00000 1.00000 0.03446
______________________________________
TABLE 10
______________________________________
Training
Input Data
______________________________________
-10.000 -9.372 -8.743 -8.115 -7.487 -6.858 -6.230 -5.602 -4.973 -4.345
-3.717 -3.088 -2.460 -1.832 -1.204 -0.575 +0.053 +0.681 +1.310 +1.938
-1.542 +0.185 +1.623 +0.388 -2.165 -2.785 -0.788 +1.014
+0.204 -1.907 -2.110 +0.239 +2.291 +1.607 -0.516 -0.867
+1.212 +2.899 +1.791 -0.774
##STR1##
+2.566 -1.542
+3.195 +0.185
+3.823 +1.623
+4.451 +0.388
+5.080 -2.165
+5.708 -2.785
+6.336 -0.788
+6.965 +1.014
+7.593 +0.204
+8.221 -1.907
+8.850 -2.110
+9.478 +0.239
______________________________________

Watanabe, Sei, Shiomi, Kazuyuki

Patent Priority Assignee Title
10915298, Oct 08 2019 Current mode multiply-accumulate for compute in memory binarized neural networks
11144316, Apr 17 2018 Current-mode mixed-signal SRAM based compute-in-memory for low power machine learning
11488007, Dec 06 2019 International Business Machines Corporation Building of custom convolution filter for a neural network using an automated evolutionary process
11593664, Jul 10 2019 STMICROELECTRONICS ROUSSET SAS Method and computational tool for determining transfer functions between pairs of successive layers of a neural network
11610104, Dec 30 2019 Asynchronous analog accelerator for fully connected artificial neural networks
11615256, Dec 30 2019 FAR, ALI TASDIGHI Hybrid accumulation method in multiply-accumulate for machine learning
5452402, Nov 26 1992 Matsushita Electric Industrial Co., Ltd. Neural network circuit for adaptively controlling the coupling of neurons
5486999, Apr 20 1994 Apparatus and method for categorizing health care utilization
5630019, May 23 1992 Kabushiki Kaisha Topcon Waveform evaluating apparatus using neural network
5649067, Dec 16 1992 U.S. Philips Corporation Neural device and method of constructing the device
5796923, Dec 27 1991 Daikin Industries, Ltd Methods and apparatus for estimating unknown quantities
5802507, Dec 16 1992 U.S. Philips Corporation Method for constructing a neural device for classification of objects
5812992, May 24 1995 GOOGLE LLC Method and system for training a neural network with adaptive weight updating and adaptive pruning in principal component space
6513023, Oct 01 1999 The United States of America as represented by the Administrator of the National Aeronautics and Space Administration Artificial neural network with hardware training and hardware refresh
Patent Priority Assignee Title
4945494, Mar 02 1989 Texas Instruments Incorporated; TEXAS INSTRUMENTS INCORPORATED, 13500 NORTH CENTRAL EXPRESSWAY, DALLAS, TEXAS 75265 A CORP OF DELAWARE Neural network and system
5043913, Jun 14 1988 Mitsubishi Denki Kabushiki Kaisha Neural network
5046020, Dec 16 1988 DIRECT RADIOGRAPHY CORP Distributed parallel processing network wherein the connection weights are generated using stiff differential equations
5113483, Jun 15 1990 Stovokor Technology LLC Neural network with semi-localized non-linear mapping of the input space
5119469, May 17 1989 WACHOVIA BANK, NATIONAL Neural network with weight adjustment based on prior history of input signals
5134685, Feb 06 1990 Westinghouse Electric Corporation Neural node, a netowrk and a chaotic annealing optimization method for the network
5167006, Dec 29 1989 Ricoh Company, Ltd. Neuron unit, neural network and signal processing method
5187680, Jun 15 1989 General Electric Company Neural net using capacitive structures connecting input lines and differentially sensed output line pairs
JP336659,
JP390956,
///
Executed onAssignorAssigneeConveyanceFrameReelDoc
Jan 27 1992SHIOMI, KAZUYUKIHONDA GIKEN KOGYO KABUSHIKI KAISHA A COMPANY OF JAPANASSIGNMENT OF ASSIGNORS INTEREST 0060200584 pdf
Jan 27 1992WATANABE, SEIHONDA GIKEN KOGYO KABUSHIKI KAISHA A COMPANY OF JAPANASSIGNMENT OF ASSIGNORS INTEREST 0060200584 pdf
Feb 18 1992Honda Giken Kogyo Kabushiki Kaisha(assignment on the face of the patent)
Date Maintenance Fee Events
Apr 15 1997M183: Payment of Maintenance Fee, 4th Year, Large Entity.
Apr 15 1997R283: Refund - Payment of Maintenance Fee, 4th Yr, Small Entity.
Jul 09 1997LSM1: Pat Hldr no Longer Claims Small Ent Stat as Indiv Inventor.
Jun 28 2001M184: Payment of Maintenance Fee, 8th Year, Large Entity.
Aug 03 2005REM: Maintenance Fee Reminder Mailed.
Jan 18 2006EXP: Patent Expired for Failure to Pay Maintenance Fees.


Date Maintenance Schedule
Jan 18 19974 years fee payment window open
Jul 18 19976 months grace period start (w surcharge)
Jan 18 1998patent expiry (for year 4)
Jan 18 20002 years to revive unintentionally abandoned end. (for year 4)
Jan 18 20018 years fee payment window open
Jul 18 20016 months grace period start (w surcharge)
Jan 18 2002patent expiry (for year 8)
Jan 18 20042 years to revive unintentionally abandoned end. (for year 8)
Jan 18 200512 years fee payment window open
Jul 18 20056 months grace period start (w surcharge)
Jan 18 2006patent expiry (for year 12)
Jan 18 20082 years to revive unintentionally abandoned end. (for year 12)