An information processing apparatus includes: model learning means for self-organizing, on the basis of a state transition model having a state and state transition to be learned by using time series data as data in time series, an internal state from an observation signal obtained by a sensor; and controller learning means for performing learning for allocating a controller, which outputs an action, to each of transitions of a state or each of transition destination states in the state transition model indicating the internal state self-organized by the model learning means.
|
0. 29. An information processing apparatus comprising:
at least one processing unit to:
receive time series data for learning;
store a state transition model for modeling the time series data, the state transition model including state information and state transition information;
estimate a parameter of the state transition model using the time series data; and
adjust structure of the state transition model under conditions of sparse structure,
wherein the structure of the state transition model is adjusted by, with reference to a first state included in the structure of the state transition model and set as a target, removing a second merging state, in which the second merging state is included in the structure of the state transition model, from the structure of the state transition model.
0. 28. A non-transitory storage medium on which is recorded a program executable by a computer, the program comprising:
receiving time series data for learning;
storing a state transition model for modeling the time series data, the state transition model including state information and state transition information;
initializing structure of the state transition model to sparse structure;
estimating a parameter of the state transition model using the time series data; and
adjusting the structure of the state transition model by, with reference to a first state included in the structure of the state transition model and set as a target, removing a second merging state, in which the second merging state is included in the structure of the state transition model, from the structure of the state transition model.
0. 27. A method comprising:
receiving time series data for learning;
storing a state transition model for modeling the time series data, the state transition model including state information and state transition information;
initializing structure of the state transition model to sparse structure;
estimating a parameter of the state transition model using the time series data; and
adjusting the structure of the state transition model by, with reference to a first state included in structure of the state transition model and set as a target, removing a second merging state, in which the second merging state is included in the structure of the state transition model, from the structure of the state transition model,
wherein the receiving, the storing, the initializing, the estimating and the adjusting are by a processing unit.
6. An information processing method comprising the steps of:
self-organizing, on the basis of a state transition model having a state and state transition to be learned by using time series data as data in time series, an internal state from an observation signal obtained by a sensor;
performing learning for allocating a controller, which outputs an action, to each of transitions of a state or each of transition destination states in the state transition model indicating the self-organized internal state self-organized;
initializing structure of the state transition model to sparse structure;
adjusting the time series data used for the learning according to progress of the learning and outputting time series data after the adjustment;
estimating a parameter of the state transition model using the time series data after adjustment; and
adjusting the structure of the state transition model.
0. 20. An information processing apparatus comprising:
a data receiving unit that receives time series data for learning;
a model storing unit that stores a state transition model for modeling the time series data, the state transition model including state information and state transition information;
an initial-structure setting unit that initializes structure of the state transition model to sparse structure;
a parameter estimating unit that estimates a parameter of the state transition model using the time series data; and
a structure adjusting unit that adjusts the structure of the state transition model by, with reference to a first state included in the structure of the state transition model and set as a target, removing a second merging state, in which the second merging state is included in the structure of the state transition model, from the structure of the state transition model.
19. An information processing apparatus comprising:
a model learning unit self-organizing, on the basis of a state transition model having a state and state transition to be learned by using time series data as data in time series, an internal state from an observation signal obtained by a sensor;
a controller learning unit performing learning for allocating a controller, which outputs an action, to each of transitions of a state or each of transition destination states in the state transition model indicating the internal state self-organized by the model learning unit;
an initial-structure setting unit initializing structure of the state transition model to sparse structure;
a data adjusting unit adjusting the time series data used for the learning according to progress of the learning and outputting time series data after the adjustment;
a parameter estimating unit estimating a parameter of the state transition model using the time series data after adjustment; and
a structure adjusting unit adjusting the structure of the state transition model.
1. An information processing apparatus comprising:
model learning means for self-organizing, on the basis of a state transition model having a state and state transition to be learned by using time series data as data in time series, an internal state from an observation signal obtained by a sensor;
controller learning means for performing learning for allocating a controller, which outputs an action, to each of transitions of a state or each of transition destination states in the state transition model indicating the internal state self-organized by the model learning means;
initial-structure setting means for initializing structure of the state transition model to sparse structure;
data adjusting means for adjusting the time series data used for the learning according to progress of the learning and outputting time series data after the adjustment;
parameter estimating means for estimating a parameter of the state transition model using the time series data after adjustment; and
structure adjusting means for adjusting the structure of the state transition model.
2. An information processing apparatus according to
planning means for planning a path for attaining a target as a transition sequence of a state on the state transition model indicating the internal state self-organized by the model learning means; and
execution managing means for invoking, for each of transitions included in the path planned by the planning means, the controller allocated by the controller learning means to manage execution of an action along the path.
3. An information processing apparatus according to
the model learning means self-organizes, independently for each of plural modals, an internal state from an observation signal obtained by a sensor of a modal corresponding thereto on the basis of state transition models, and
the information processing apparatus further includes causality means for estimating causality of transition in one state transition model and a state of another state transition model among the state transition models for each of the plural modals respectively indicating the internal state self-organized by the model learning means.
4. An information processing apparatus according to
5. An information processing apparatus according to
the controller learning means, the planning means, and the execution managing means execute respective kinds of processing to realize the target spontaneously set by the setting means.
7. An information processing apparatus according to
8. An information processing apparatus according to
9. An information processing apparatus according to
10. An information processing apparatus according to
11. An information processing apparatus according to
12. An information processing apparatus according to
13. An information processing apparatus according to
detecting means for detecting an event that occurs immediately preceding state transition that occurs in a first set including events exclusive to one another, the event being an event in a second set, which is a single or plural other sets including events exclusive to one another; and
estimating means estimating, with the state transition set as a result event and the event in the second set detected by the detecting means set as a cause event, causality between the events included in the different sets.
14. An information processing apparatus according to
15. An information processing apparatus according to
the detecting means detects, for a first event immediately preceding occurrence of the state transition and a second event in the second set that occurs simultaneously with the first event, a first number of times the first and second events simultaneously occur immediately preceding the state transition and a second number of times the first and second events simultaneously occurs, and
the estimating means calculates a conditional probability concerning the state transition by dividing the first number of times detected by the detecting means by the second number of times detected by the detecting means.
16. An information processing apparatus according to
17. An information processing apparatus according to
18. An information processing apparatus according to
0. 21. The information processing apparatus of claim 20, wherein the state transition model is based on an HMM (Hidden Markov model).
0. 22. The information processing apparatus of claim 20, further comprising:
a data adjusting unit that adjusts the time series data according to progress of learning.
0. 23. The information processing apparatus of claim 22, wherein the structure adjusting unit adjusts the structure based on adjustment conducted by the data adjusting unit.
0. 24. The information processing apparatus of claim 22, wherein the data adjusting unit adjusts the time series data by adjusting a sampling rate for the time series data.
0. 25. The information processing apparatus of claim 20, wherein the structure adjusting unit adjusts the structure by at least limiting a transition of the state transition model.
0. 26. The information processing apparatus of claim 20, wherein the structure of the state transition model is adjusted by the removing of the second merging state and adding, with reference to the first state, a third state to the structure of the state transition model, according to progress of the learning.
0. 30. The information processing apparatus of claim 29, wherein the structure is adjusted by at least limiting a transition of the state transition model.
|
The present application claims priority from Japanese Patent Application Nos. JP 2008-064993, JP 2008-064994, and JP 2008-064995, all filed in the Japanese Patent Office on Mar. 13, 2008, the entire contents of which are incorporated herein by reference.
1. Field of the Invention
The present invention relates to an information processing apparatus, an information processing method, and a computer program, and, more particularly to an information processing apparatus, an information processing method, and a computer program that can self-organize an internal state to create an environment model.
2. Description of the Related Art
In recent years, researches and developments have been actively performed concerning reinforcement learning. The reinforcement learning means a method of mechanical learning for autonomously acquiring an optimum behavior on the basis of actual experiences and returns. Mechanical learning for learning by trial and error, relying only on returns from an environment, a control method for attaining the returns is referred to as reinforcement learning in a broad sense (see, for example, “Reinforcement Learning” Richard S. Sutton, Andrew G. Barto, translated by Sadayoshi Mikami and Masaaki Minakawa, Morikita Publishing. The reinforcement learning have been applied to various Markov decision problems having finite numbers of states and finite numbers of behaviors such as acquisition of strategies in games and achieved successes.
However, there are still a large number of problems in applying the reinforcement learning to various difficult problems in the real world.
One significant problem among the problems is that returns, behaviors, and environment models are learned all together. In the reinforcement learning, an environment is acquired in a form of prediction of a return that can be received, i.e., a value. In other words, it is only learned how much value a present state has. A change in a state is not modeled. In architectures such as SARSA and actor critic, a value function depending on a present actor (behavior determination) is learned. Therefore, it is necessary to learn an actor and a value function from the beginning every time a purpose (a return) changes. However, originally, an environment model indicating how an environment changes when a user behaves toward the environment how should be able to be used in common even if a purpose changes. Therefore, in solving various problems, it is more efficient to plan a behavior on the basis of an environment prediction model.
However, it is difficult to say that the technique in the past can sufficiently set problems given to the reinforcement learning. It is possible to solve problems on the basis of a framework of the reinforcement learning if a human sorts out the problems in advance and passes the problems to a machine as Markov decision problems having finite numbers of states and finite numbers of behaviors as explained above. However, in problems in the real world, there are many unknown matters, for example, how many states are sufficient, how many pieces a behavior is divided into, whether a Markov decision process is adopted in the first place, and whether given information is enough. Therefore, in the past, the human sorts out a large number of problems and solve the problems before causing the machine to learn the problems.
Therefore, a framework for the machine itself to classify and solve the problems in this way is necessary. For this purpose it is demanded that the machine itself self-organizes an internal state to create an environment model. However, in the present situation, such a demand is not sufficiently met.
Therefore, it is desirable to make it possible to self-organize an internal state to create an environment model.
According to an embodiment of the present invention, there is provided an information processing apparatus including: model learning means for self-organizing, on the basis of state transition model having a state and state transition to be learned by using time series data as data in time series, an internal state from an observation signal obtained by a sensor; and controller learning means for performing learning for allocating a controller, which outputs an action, to each of transitions of a state or each of transition destination states in the state transition model indicating the internal state self-organized by the model learning means.
The information processing apparatus further includes: planning means for planning a path for attaining a target as a transition sequence of a state on the state transition model indicating the internal state self-organized by the model learning means; and execution managing means for invoking, for each of transitions included in the path planned by the planning means, the controller allocated by the controller learning means to manage execution of an action along the path.
The model learning means self-organizes, independently for each of plural modals, an internal state from an observation signal obtained by a sensor of a modal corresponding thereto on the basis of state transition models. The information processing apparatus further includes causality means for estimating causality of transition in one state transition model and a state of another state transition model among the state transition models for each of the plural modals respectively indicating the internal state self-organized by the model learning means.
The execution managing means causes, when it is difficult to directly control an internal state of a predetermined modal among the plural modals respectively indicating the internal state self-organized by the model learning means, the planning means to recursively execute planning to control the internal state on the basis of the causality estimated by the causality means.
The information processing apparatus further includes setting means for spontaneously setting a target from the internal state self-organized by the model learning means. The controller learning means, the planning means, and the execution managing means execute respective kinds of processing to realize the target spontaneously set by the setting means.
According to another embodiment of the present invention, there are provided an information processing method and a computer program corresponding to the information processing apparatus according to the embodiment explained above.
In the information processing apparatus, the information processing method, and the computer program according to the embodiments, an internal state is self-organized from an observation signal obtained by a sensor on the basis of a Markov model. In the Markov model indicating the self-organized internal state, learning for allocating a controller, which outputs an action, is allocated to each of transitions of a state.
As explained above, according to the present invention, the information processing apparatus and the like can self-organize an internal state to create an environment model.
Exemplary embodiments of the present invention are explained in detail below with reference to the accompanying drawings.
First, an overview of processing according to an embodiment of the present invention is explained with reference to
According to the embodiment, a target system or agent automatically constructs a model of an external environment on the basis of a sensor signal for observation (hereinafter referred to as observation signal) and an action signal of an action taken by the system or the agent. The system or the agent freely generates an intellectual behavior for realizing the automatic construction of the model and realizing an arbitrary state on an internally-perceived model.
In general, the “agent” indicates an autonomous entity that can perceives (e.g., senses) a state of an environment and select a behavior on the basis of a perceived content. However, in the following explanation, rather than the agent, the system is used as an operation entity.
The modeling of an external environment is not specifically limited. However, in this embodiment, a Hidden Markov Model (hereinafter referred to as HMM) is adopted as the modeling of an external environment.
For example, as shown in
In this case, as shown in
Subsequently, as shown in
As shown in
The overview of the processing according to this embodiment is explained above. An information processing system according to this embodiment is explained below.
The system shown in
The sensor unit 21 observes a predetermined physical amount of an environment in which an agent is placed and provides the modeling unit 22 with a result of the observation as an observation signal.
The modeling unit 22 includes a learning unit 31, an HMM storing unit 32, a recognizing unit 33, and a planning unit 34.
The learning unit 31 constructs an HMM using the observation signal of the sensor unit 21 (see
The recognizing unit 33 estimates, when the action unit 25 explained later behaves, respective transitions up to a present state (a present situation) using the HMM stored in the HMM storing unit 32 and an observation signal sequence of the sensor unit 21. An estimation result of the recognizing unit 33 is provided to the behavior control unit 24.
The planning unit 34 plans (calculates) an optimum path from the present state toward a target state using the HMM stored in the HMM storing unit 32 and provides the behavior control unit 24 with the optimum path (see
The innateness controller 23 issues, on the basis of a predetermined innateness rule, various commands for learning of a learning unit 41 of the behavior control unit 24 explained later and provides the learning unit 41 and the action unit 25 with the commands.
The behavior control unit 24 includes the learning unit 41, a controller-table storing unit 42, a controller storing unit 43, and an execution managing unit 44.
The learning unit 41 learns a controller for each of transitions using respective transitions, which are recognized by the recognizing unit 33 on the basis of a behavior result of the action unit 25 conforming to a command from the innateness controller 23, and the command from the innateness controller 23 (see
The execution managing unit 44 generates a command for the action unit 25 such that the action unit 25 behaves along a path provided from the planning unit 34, i.e., realizes respective transitions in the path. The execution managing unit 44 provides the action unit 25 with the command. This command is inversely generated on the basis of information stored in the controller-table storing unit 42 and the controller storing unit 43. Details of processing by the execution managing unit 44 are explained later.
Further details of the system shown in
As a target of the system shown in
In the problem setting in the simple pendulum task in the past, angular velocity ω is given in addition to the angle θ (i.e., two variables are given) as an observation signal. As a target, a target that the simple pendulum 51 swings up, i.e., the angle θ reaches 180° is given. Further, a target function for attaining the target, for example, a target function that a return is given when the angle θ reaches 180° or a higher value is outputted as the angle θ is closer to 180° is designed and given.
On the other hand, one of targets of the system shown in
In step S1, the system shown in
In step S2, the system shown in
In step S3, the system shown in
In step S4, the system shown in
In step S5, the system shown in
Details of the respective kinds of processing in steps S1 to S5 are individually explained below in the order of the steps.
First, the learning processing for an HMM in step S1 is explained.
In an initial state, the action unit 25 outputs a control signal τ generated at random or a control signal τ obtained by adding a proper perturbation to a pattern innately embedded in advance. Such a control signal τ is generated by the action unit 25 on the basis of, for example, a command given by the innateness controller 23.
A time series of the observation signal θ outputted from the sensor unit 21 during this period (hereinafter referred to as time series observation signal) is stored on a not-shown memory of the learning unit 31. A signal 52 shown in
The series of processing explained above is the learning processing for an HMM.
In the learning processing for an HMM, in general, a Baum-Welch algorithm is used. Examples of HMMs applicable to such an algorithm are shown in
However, the system is caused to learn a totally-connected HMM shown in
Therefore, in this embodiment, a hypothesis that almost all phenomena in the natural world can be represented by a sparse connection such as a small world network is adopted. In this embodiment, the Baum-Welch algorithm limited to sparse connection is adopted. Specifically, in this embodiment any one of HMMs shown in
A display example of a result obtained by giving, in the simple pendulum task, a two-dimensional neighborhood restricted HMM with 484 nodes to the system as an initial structure and causing the system to learn a time series observation signal is shown in
The abscissa of FIG. H indicates the angle θ of the simple pendulum 51 as an observation signal. The ordinate of FIG. H indicates the angular velocity ω of the simple pendulum 51. In
It is seen from the display example shown in
As an only exception, a change from θ=−π to θ=π is discontinuous when the simple pendulum 51 makes a full turn. Because nodes at the left end and the right end in
It is also seen from the display example shown in
The learning processing for an HMM in step S1 is explained above. Subsequently, the recognition processing in step S2 is explained below.
The recognition processing is processing for estimating a present state of the system shown in
A result of the recognition processing is used for the controller learning processing in step S3 explained later. The recognition processing is executed as one kind of processing of the behavior control processing in step S5 explained later separately from the processing in step S2 (see step S61 in
Key information in the recognition processing is an observation signal sequence from the past to the present. The Viterbi Algorithm is widely used for state estimation for an HMM. Therefore, in this embodiment, it is assumed that the recognition processing is executed as follows: a state fifty steps before the present is set undefined, i.e., probabilities of respective nodes are set equal, the state fifty steps before the present is set as an initial state, observation results for fifty steps are given, and states in the respective steps are decided by the Viterbi Algorithm to estimate a state of the last fiftieth step, i.e., a present state.
Specifically, for example, in this embodiment, the recognition processing is executed according to a flowchart of
In the following explanation, a transition probability from a node “i” to a node “j” is described as aij or Aij. An initial state probability is described as πi. An observation value (a level of an observation signal) at time t is described as o(t). Likelihood of the observation value o(t) at the node “i” is referred to as observation likelihood and described as bi(o(t)). Present time is described as T.
In step S21, the recognizing unit 33 sets the time t to 0.
In step S22, the recognizing unit 33 multiplies the initial state probability πi with observation likelihood b(0(0)) and sets the initial state probability πi in the respective nodes.
In step S23, the recognizing unit 33 multiplies a state probability at time t with the transition probability Aij and observation likelihood b(0(t+1)) and updates a maximum probability in the node “j” at a transition destination to a state probability of the node “j”.
In step S24, the recognizing unit 33 stores the node “i” at a transition source at that point in a storage table. A constructing location for the storage table is not specifically limited. In this embodiment, for example, it is assumed that the storage table is constructed in the inside of the recognizing unit 33.
In step S25, the recognizing unit 33 increments the time t by 1 (t=t+1).
In step S26, the recognizing unit 33 determines whether the time t has reached the present time T.
When the time t is time before the present time T, the recognizing unit 33 determines in step S26 that the time t has not reached the present time T (NO in step S26). The processing is returned to step S23 and the processing in step S23 and subsequent steps is repeated.
Loop processing from step S23 to step S26 is repeated at every time t=0 to T. When the time t reaches the present time T, the recognizing unit 33 determines in step S26 that the time t has reached the present time T (YES in step S26). The processing proceeds to step S27.
In step S27, the recognizing unit 33 selects a node having a maximum state probability among state probabilities at time t and sets the node as a decided node at time t. In the processing in step S27 immediately after it is determined as YES in the processing in step S26 that the time t has reached the present time T, since the time t is the present time T, a decided node at the present time T is obtained.
In step S28, the recognizing unit 33 extracts the node “i” at the transition source of the node “j” selected in the processing in step S27 from the storage table and sets the node “i” as a node at time t−1.
In step S29, the recognizing unit 33 decrements the time t by 1 (time t=t−1).
In step S30, the recognizing unit 33 determines whether the time t is 0.
When the time t is time after 0, the recognizing unit 33 determines in step S30 that the time t is not 0 (NO in step S30). The processing is returned to step S27 and the processing in step S27 and subsequent steps is repeated.
Loop processing from step S27 to step S30 is repeated at every time t=T to 0. When the time t reaches 0, the recognizing unit 33 determines in step S30 that the time t is 0 (YES in step S30). The recognition processing is finished.
The recognition processing in step S2 is explained above. Subsequently, the controller learning processing in step S3 is explained below.
When the recognition processing in step S2 is executed, the node “i” indicating a state at every time is determined. The transition probability Aij from the node “i” to the node “j” indicating a state at the next time is also determined. In the following explanation, the transition probability Aij is referred to as transition edge Aij as appropriate. It should be noted that, while the lower-case letter “a” is used in such a manner as transition probability aij in the explanation of the recognition processing, the upper-case letter “A” is used in such a manner as transition probability Aij (transition edge Aij) in the explanation of the controller learning processing. This is for the purpose of preventing confusion with the lower-case letter “a” in an action a(t) explained later.
During the learning processing for an HMM in step S1, the system shown in
The learning unit 41 of the behavior control unit 24 samples, with respect to each of caused transition edges Aij, the observation value o(t) (hereinafter abbreviated as observation value “o”) and the action “a” at the point when the transition edge Aij is caused. In this case, if a time series observation signal is a signal for long time, the transition edges Aij are caused many times during the long time. Therefore, the learning unit 41 learns mapping a=Fij(o) with respect to one transition edge Aij using the sampled observation value “o” and action “a”. As a learning method for this function mapping Fij( ), for example, a method like a neural network can be adopted. As a simplest example, a learning method for the function mapping Fij( ) for outputting an average of the action “a” regardless of the observation value “o” can be adopted.
Such a function mapping Fij( ) is stored in the controller storing unit 43 as a controller to be executed by the action unit 25.
A learning result of the controller, i.e., information indicating, for each of the transition edges Aij, which controller (the function mapping Fij( )) corresponds to the transition edge Aij is stored in the controller-table storing unit 42 in a table format. Such a table is referred to as controller table.
In this embodiment, it is assumed that an identifier (ID) for uniquely specifying each of the controllers (function mappings Fij( )) is given to the controller. In this case, as information indicating what the controller (the function mapping Fij( )) associated with a predetermined transition edge Aij is, an ID of the controller can be adopted. Therefore, in this embodiment, for each of the transition edges Aij, an ID of the controller (the function mapping Fij( )) corresponding to the transition edge Aij is stored in the controller table 42. In the controller storing unit 43, each of the controllers (the function mappings Fij( )) is stored to be tied to an ID thereof. An example of a method of using the ID is referred to in explanation of step S70 in
As the controller learning processing in step S3, the processing for performing learning for allocating a controller for outputting an action to each of transitions of a state is explained as an example. However, as the controller learning processing according to this embodiment, besides the example explained above, for example, processing for performing learning for allocating a controller for outputting an action to each of transition destination states can also be adopted.
The planning processing in step S4 is explained below.
At a stage when the controller learning processing in step S3 ends, the system shown in
Therefore, the planning unit 34 sets up a plan (planning) for realizing the attainment of the target. Processing for setting up such a plan is the planning processing in step S4.
The planning unit 34 sets, as a goal, a target designated from the outside or endogenously obtained in the system. In the system shown in
There are various algorithms for the path search. Any one of the algorithms may be adopted. However, in this embodiment, as indicated by a flowchart of
In step S41, the planning unit 34 sets a state probability of the present state node “i” to 1.0 and sets a state probability of the other nodes to 0. The planning unit 34 sets the time t to 0.
In step S42, the planning unit 34 sets the transition probabilities Aij equal to or higher than a threshold (0.01) to 0.9 and sets the other transition probabilities Aij to 0.
In step S43, the planning unit 34 multiplies the state probability at time t with the transition probability Aij and updates a maximum probability in the node “j” at the transition destination to a state probability of the node “j”.
In step S44, the planning unit 34 stores the node “i” at the transition source at that point in the storage table. A constructing location for the storage table is not specifically limited. In this embodiment, for example, it is assumed that the storage table is constructed in the planning unit 34.
In step S45, the planning unit 34 determines whether a state probability of the goal node “g” as the target has exceeded 0.
When the state probability of the goal node “g” as the target is 0, assuming that the state probability has not reached the target, the planning unit 34 determines in step S45 that the state probability of the goal node “g” has not exceeded 0 (NO in step S45). The processing proceeds to step S46.
In step S46, the planning unit 34 determines whether loop processing from step S43 to step S47 has been repeated N times.
The N times repetition means that the state probability has not reached the target yet even if the steps are repeated N times. Therefore, in such a case, i.e., when the planning unit 34 determines in step S46 that the loop processing has been repeated N times (YES in step S46), the planning processing is finished on the assumption that the planning unit 34 has given up the planning.
On the other hand, when the loop processing has not been repeated N times yet, the planning unit 34 determines in step S46 that the loop processing has not been repeated N times (NO in step S46). The processing proceeds to step S47. In step S47, the planning unit 34 increments the time t by 1 (time t=t+1). Thereafter, the processing is returned to step S43 and the processing in step S43 and subsequent steps is repeated.
The loop processing from step S43 to step S47 is repeated several times in this way. As a result, when the state probability of the goal node “g” as the target exceeds 0, assuming that the state probability has reached the target, the planning unit 34 determines in step S45 that the state probability of the goal node “g” has exceeded 0 (YES in step S45). The processing proceeds to step S48.
In step S48, the planning unit 34 selects the goal node “g”. In step S49, the planning unit 34 sets the goal node “g” equal to the node “j”.
In step S50, the planning unit 34 extracts the node “i” at the transition source of the selected node “j” from the storage table and sets the node “i” as a node at time t−1.
In step S51, the planning unit 34 decrements the time t by 1.
In step S52, the planning unit 34 determines whether the time t is 0.
When the time t is time after 0, the planning unit 34 determines in step S52 that the time t is not 0 (NO in step S52). The processing proceeds to step S53. In step S53, the planning unit 34 sets the node “j” equal to the node “i”. Thereafter, the processing is returned to step S50 and the processing in step S50 and subsequent steps is repeated.
Loop processing from step S50 to step S53 is repeated until the time t reaches 0. When the time t reaches 0, the planning unit 34 determines in step S52 that the time t is 0 (YES in step S52). The planning processing is finished. A node sequence formed at this point, i.e., a node sequence from the present state node “i” to the goal node “g” is decided as a path.
The planning processing in step S4 is explained above. Subsequently, the behavior control processing in step S5 is explained below.
In step S61, the execution managing unit 44 of the behavior control unit 24 performs recognition processing for an HMM and selects a node having a highest state probability among all the nodes as a node i_max.
In this embodiment, it is assumed that processing conforming to the flowchart of the example shown in
In step S62, the execution managing unit 44 selects, as the present node i_pathmax, a node having a highest state probability between the last node i_pathmax and the goal node among the nodes on the path.
In step S63, the execution managing unit 44 determines whether a ratio of state probabilities P(i_max) and P(i_pathmax) is equal to or smaller than a threshold (e.g., equal to or smaller than 0.7). The state probability P (i_max) indicates a state probability of the node i_max. The state probability P(i_pathmax) indicates a state probability of the node i_pathmax.
When the ratio of the state probabilities P(i_max) and P(i_pathmax) is equal to or smaller than the threshold, assuming that the present behavior of the system shown in
On the other hand, when the ratio of the state probabilities P(i_max) and P(i_pathmax) exceeds the threshold, assuming that the present behavior of the system shown in
In step S64, the execution managing unit 44 determines whether the system stays in the same node i_pathmax, i.e., whether the node i_pathmax selected in the present processing in S62 and the node i_pathmax selected in the last processing in step S62 are the same.
When the system does not stay in the same node, the system is regarded as moving along the path. Therefore, the execution managing unit 44 determines in step S64 that the system does not stay in the same node i_pathmax (NO in step S64). The processing proceeds to step S68. Processing in step S68 and subsequent steps is explained later.
On the other hand, when the system stays in the same node i_pathmax, it is likely that the system is not moving along the path. Therefore, the execution managing unit 44 determines in step S64 that the system stays in the same node i_pathmax (YES in step S64). The processing proceeds to step S65.
In step S65, the execution managing unit 44 determines whether a state probability of the next node i_next on the path rises to be higher than the last state probability.
When the state probability of the next node i_next does not rise, assuming that the system is not transitioning along the path, the execution managing unit 44 determines in step S65 that the state probability of the next node i_next does not rise to be higher than the last state probability (NO in step S65). In step S66, the execution managing unit 44 sets the node i_pathmax as the node i_next. Thereafter, the processing proceeds to step S68. Processing in step S68 and subsequent steps is explained later.
On the other hand, when the state probability of the next node i_next rises, the execution managing unit 44 determines in step S65 that the state probability of the next node i_next rises to be higher than the last state probability (YES in step S65). The processing proceeds to step S67.
In step S67, the execution managing unit 44 determines whether the system stays in the same node the number of times equal to or larger than N (e.g., fifty).
When the system does not stay in the same node the number of times equal to or larger than N, the execution managing unit 44 determines in step S67 that the system does not stay in the same node the number of times equal to or larger than N (NO in step S67). The processing proceeds to step S68. Processing in step S68 and subsequent steps is explained later.
On the other hand, when the system stays in the same node the number of times equal to or larger than N, the execution managing unit 44 determines in step S67 that the system stays in the same node the number of times equal to or larger than N (YES in step S67). In step S66, the execution managing unit 44 sets the node i_pathmax as the node i_next. In other words, when the system stays in the same node the number of times equal to or larger than N, the execution managing unit 44 regards that the path is advanced by force. Thereafter, the processing proceeds to step S68.
In step S68, the execution managing unit 44 determines whether the system is already on the goal node.
When it is recognized that the system is already on the goal node, the execution managing unit 44 determines in step S68 that the system is already on the goal node (YES in step S68). Assuming that the system has reached the target, the execution managing unit 44 finishes the behavior control processing.
On the other hand, when it is recognized the system is not already on the gold node, the execution managing unit 44 determines in step S68 that the system is not on the gold node (NO in step S68). The processing proceeds in step S69.
In step S69, the execution managing unit 44 decides the transition edge Aij for transitioning to the next node on the past.
In step S70, the execution managing unit 44 invokes the controller (the function mapping Fij( )) allocated to the transition edge Aij. The action unit 25 gives the present observation value “o” to the controller to calculate the action “a” that should be performed.
More accurately, in this embodiment, an ID of the controller (the function aping Fij( )) allocated to the transition edge Aij is read out from the controller-table storing unit 42. The controller (the function mapping Fij( )) specified by the ID is read out from the controller storing unit 43. An output obtained as a result of inputting the present observation value “o” to the function mapping Fij( ) as the controller is the action “a”.
The action “a” is provided to the action unit 25 as a command. Therefore, in step S71, the action unit 25 executes the command “a”.
Thereafter, the processing is returned to step S61. The processing in step S61 and subsequent steps is repeated.
When the execution managing unit 44 determines in step S68 that system is already on the goal node (YES in step S68) and the behavior control processing ends, the execution managing unit 44 may determine again whether the node i_max at that point is truly the goal node. When a result of re-determination is a result indicating that the node i_max is the goal node, the entire control processing for the simple pendulum shown in
The system shown in
The system shown in
The sensor unit 61 is configured as a so-called multi-modal sensor.
The multi-modal sensor is briefly explained below.
As one of concepts obtained by expanding a human interface in the past is present as a concept of the multi-modal interface. As a synonym of the multi-modal interface, for example, there is a word called multi-media interface. Whereas the multi-media interface represents an interface that simply uses plural media (sound, video, tactile sense, etc.), the multi-media interface is referred to as multi-modal interface when the respective media are used in various forms and perform information transmission.
As an example of the multi-modal interface, there is an interface that sets events such as utterance, action, and line of sight as modals, causes these modals to cooperate with one another, simultaneously uses the modals, and combines plural kinds of messages to cause a human to understand a message that the human originally intends to communicate or is naturally transmitted.
In other words, the multi-modal sensor is a sensor for realizing such a multi-modal interface and is a sensor that can detect a physical amount corresponding to each of the plural modals (events).
For example, in the example shown in
Each of the modeling units 62A, 62B, and 62C has a function and a configuration basically same as those of the modeling unit 22 shown in
Therefore, respective HMMs constructed as a result of learning performed by using observation signals for respective three modals of the sensor 61, i.e., HMMs of the three modals are stored in the HMM storing units 72A to 72C. The modals to be modeled by modeling units 62A to 62C are referred to as modals A to C. In this case, respective HMMs of the modals A to C are stored in the HMM storing units 72A to 72C.
Naturally, the number of modals is not limited to three and only has to be equal to or larger than two. However, in that case, modeling units corresponding to the modeling unit 62A equivalent to the number of modals are present.
The causality unit 63 includes a causality learning unit 75, a causality-table storing unit 76, and a causality estimating unit 77.
The causality learning unit 75 learns node transition, which is recognized by a recognizing unit 73AK on the basis of the structure of an HMM of a modal K (K is any one of A to C), and a relation of a state of an HMM of another modal L (L is any one of A to C other than K). A result of the learning is stored in the causality-table storing unit 76. Details of processing by the causality learning unit 75 are explained later.
The behavior control unit 64 includes an execution managing unit 78 and a controller unit 79. The controller unit 79 includes a controller-table storing unit 80 and a controller storing unit 81. The controller-table storing unit 80 and the controller storing unit 81 have functions and configurations basically same as those of the controller-table storing unit 42 and the controller storing unit 43 shown in
When a target is given, the execution managing unit 78 determines the modal K corresponding to the target and provides the modeling unit 62K with the modal K. A planning unit 74K of the modeling unit 62K plans a path according to the target and provides the execution managing unit 78 with the path. The execution managing unit 78 controls the action unit 65 such that the system (the agent) shown in
The action unit 65 performs a predetermined behavior according to a command from the behavior control unit 64.
The system shown in
Specifically, it is assumed that a multi-modal task explained below is given. The multi-modal task has a purpose of allowing a round mobile robot 85 to freely move within an area surrounded by a wall 86 as shown in
The applicant performed, as an experiment by a simulator, movement of the mobile robot 85 shown in
The prototype is referred to above because the simulator adopted this time is not the simulator per se disclosed in the document but a simulator incorporating observation signals and actions shown in
As shown in
It should be note that arrangement positions in
As shown in
On the assumption that the optical sensor 61B is attached in twenty-four directions (same as the directions of the distance sensor 61A) around the robot 85, the optical sensor 61B outputs values corresponding to brightness of light in the respective twenty-four directions as an observation signal. However, taking into account the fact that light has a diffusing characteristic, the values of the observation signal are adapted to be not only values in one direction but also values affected by the sensors around the robot 85. In
The energy sensor 61C observes energy defined as explained below and outputs an observation value of the energy as an observation signal. The energy is consumed in proportion to a movement amount and supplied in proportion to an amount of irradiated light. In
As an action (a behavior), i.e., as a command given to the action unit 65, a command of a movement amount is adopted. Specifically, a command (Δx, Δy) for movement along the abscissa and the ordinate on the simulator shown in
In summary, the robot 85 has a detection function realized by using the twenty-four-dimension distance sensor 61A, the twenty-four-dimension optical sensor 61B, and the one-dimensional energy sensor 61C. The robot 85 also has input and output functions for two-dimensional movement command. The robot 85 is an agent controlled by the system shown in
An overview of a flow of the processing of the system shown in
First, the system shown in
The system shown in
In the control processing for the simple pendulum task shown in
On the other hand, in the system shown in
In the example shown in
A display example of a learning processing result for an HMM by the modeling unit 62A, i.e., a display example of a result obtained by giving, as an initial structure, a two-dimensional neighborhood structure HMM with 400 nodes to the system and causing the system to learn a time series of the observation signal (distance) of the distance sensor 61A is shown in
A display example of a learning processing result for an HMM by the modeling unit 62B, i.e., a display example of a result obtained by giving, as an initial structure, a two-dimensional neighborhood structure HMM with 100 nodes to the system and causing the system to learn a time series of the observation signal (light) of the distance sensor 61B is shown in
A display example of a learning result of the modeling unit 62C, i.e., a display example of a result obtained by giving, as an initial structure, a two-dimensional neighborhood structure HMM with 100 nodes to the system and causing the system to learn a time series of the observation signal (energy) of the distance sensor 61C is shown in
In
In
Concerning
Since the distance sensor 61A senses the wall 86, it is seen that the HMM for distance shown in
Concerning the HMM for light shown in
Concerning the HMM for energy shown in
When the multi-modal task as the target is considered only in terms of an HMM for distance and an action (a command) and controlled to be in an arbitrary state, a state shown in
However, in the problem setting for the multi-modal task, a direct correlation is not always present between the state transitions of the respective HMMs shown in
For example, transition of the HMM for energy shown in
Therefore, in the system shown in
The causality unit 63 can execute processing explained below instead of steps S2 and S3 in
In respective time steps, presently-recognized one node is decided according to recognition results in the respective HMMs shown in
In addition to the decided node, an action (a behavior) performed at that time can be treated as one modal by being discretized. In the following explanation, such a modal is referred to as action modal. A state of the action modal is referred to as action state.
A state of the HMM at time t including the action state is described as Sk,i(t). “k” indicates a modal number. k=0 indicates the action modal. “i” indicates an index representing a state in the modal.
A probabilistic causality mode indicated by Formula (1) is assumed.
Formula (1) indicates that the next state of a certain modal depends on the present state and a state Sm,l of certain another modal.
This “certain modal” is referred to as cause modal. A present station node in the cause modal is referred to as cause node. Then, when the cause modal is the action modal itself, Formula (1) indicates a simple behavior result model in which a node transitioned from the present state node (the cause node) changes according to a behavior (an action) performed at time t.
In the following explanation, finding of a cause modal and a cause node concerning node transition of respective modals is referred to causality estimation. Since the causality estimation is explained in detail later, only an overview of the causality estimation is explained below.
The causality estimation means, when transition occurs in a certain modal, counting states of other modals recognized at that point and deducing a state that occurs simultaneously with the transition at a high frequency. This makes it possible to find cause modals and cause nodes corresponding to respective transitions. The causality learning unit 75 finds a cause modal and a cause node corresponding to each of the transitions by performing such causality estimation for each of the transitions. The cause modal and the cause node for each of the transitions are stored in the causality-table storing unit 76 as a table. In the following explanation, such a table is referred to as causality table.
The overview of the causality estimation is further explained with reference to
In
In this case, the system shown in
On the other hand, in
When the causality table is stored in the causality-table storing unit 76 in this way, the system shown in
Therefore, the system shown in
The system shown in
The following explanation indicates that, concerning an HMM for multi-modal, an arbitrary state can be controlled as a target, i.e., multi-stage behavior control processing can be performed.
In the multi-stage planning processing, as in the case of the simple pendulum task, the planning unit 74K of the modeling unit 62K (K is any one of A to C) sets, as a goal, a target designated from the outside or endogenously obtained in the system. However, a predetermined state (node) in a predetermined modal is set as the goal. In other words, a goal modal and a goal state are set.
Thereafter, the modeling unit 62K executes, for example, planning processing conforming to the flowchart of
For example, the planning unit 74C executes the planning processing for the modal C for energy, a path shown on the right side of
Consequently, the behavior control unit 64 can execute multi-stage behavior control processing explained below.
The execution managing unit 78 of the behavior control unit 64 acquires, from the causality estimating unit 77 of the causality unit 63, cause modals and cause nodes allocated to respective transitions on the path from the start node to the goal node. When the causality estimating unit 77 receives notification of predetermined transition from the execution managing unit 78, the causality estimating unit 77 finds and extracts a cause modal and a cause node allocated to the predetermined transition from the causality-table storing unit 76 and provides the execution managing unit 78 with the cause modal and the cause node.
When the acquired cause modal is an action modal, the execution managing unit 78 can acquire a command corresponding to the cause node from the controller unit 79 and provide the action unit 65 with the command. Therefore, in this case, the execution managing unit 78 only has to execute behavior control processing conforming to the flowchart of
On the other hand, when the cause modal is not an action modal, it is necessary to transition a present state of the cause modal to the cause node. For example, in an example shown in
The execution managing unit 78 acquires, from the causality estimating unit 77 of the causality unit 63, cause modals and cause nodes allocated to respective transitions on the notified path.
The execution managing unit 78 recursively invokes the cause modals and the cause nodes in this way. At a stage when the execution managing unit 78 reaches an action modal that the agent can directly output, the execution managing unit 78 determines an action (a command) at that time and provides the action unit 65 with the action.
Thereafter, at a stage when the execution managing unit 78 reaches the cause node through such a procedure of the behavior control processing, the execution managing unit 78 returns to the original modal and executes the behavior control processing in the modal. In the example shown in
In some case, rather than one cause modal and one cause node, plural cause modals and cause nodes are present as in a large number of problems in the real world. For example, in the example shown in
The causal estimation is explained in detail below with reference to a specific task example
First, as shown in
As shown in
It is assumed that, when a path is created on the HMM for light, a path approaching light from an S (south) side of the light is formed. Since the three light sources (the points 87) are placed on the simulator shown in
Consequently, first, the robot 85 as the agent can calculate path to an outer edge of light on the HMM for distance and move to a nearest light source (in the example shown in
Arbitrary control of a state of the HMM for energy is discussed below. The HMM for energy is not directly related to an action concerning any transition. If the causality estimation is performed well, transition in a direction in which energy rises has high causality with nodes near the light sources (in the example shown in
Therefore, for example, when the causality estimation shown in
The system shown in
The system shown in
As explained above, the system shown in
The series of processing explained above can be executed by hardware or can be executed by software.
In this case, for example, a personal computer shown in
In
The CPU 91, the ROM 92, and the RAM 93 are connected to one another via a bus 94. An input and output interface 95 is also connected to the bus 94.
An input unit 96 including a keyboard and a mouse, an output unit 97 including a display, the storing unit 98 including a hard disk, and a communication unit 99 including a modem and a terminal adapter are connected to the input and output interface 95. The communication unit 99 controls communication performed with other apparatuses (not shown) via a network including the Internet.
A drive 100 is also connected to the input and output interface 95 according to necessity. A removable medium 101 including a magnetic disk, an optical disk, a magneto-optical disk, or a semiconductor memory is inserted in the drive 100 as appropriate. A computer program read out from the removable medium 101 is installed in the storing unit 98 according to necessity.
When the series of processing is executed by software, a program forming the software is installed in, from a network or a recording medium, a computer incorporated in dedicated software, a general-purpose personal computer that can execute various functions by installing various programs, or the like.
As shown in
In
A sensor signal obtained by sensing the modeling target is observed from the modeling target, for example, in time series.
The data processing apparatus performs, using a sensor signal observed from the modeling target, learning of a state transition model, i.e., estimation of parameters of the state transition model and determination of structure of the state transition model.
As the state transition model, for example, an HMM, a Bayesian network, a POMDP (Partially Observable Markov Decision Process) and the like can be adopted. In the following explanation, for example, an HMM is adopted as the state transition model.
The HMM is a state transition model having states and an inter-state transition.
In
In
In
As the output probability density function bj(x), for example, a mixed normal probability distribution is used.
An HMM (a continuous HMM) is defined by the state transition probability aij, the output probability density function bj(x), and the initial probability πi. The state transition probability aij, the output probability density function bj(x), and the initial probability πi are referred to as parameters λ for the HMM={aij,bj(x),πi, i=1, 2, . . . , N, j=1, 2, . . . , N}. N represents the number of states of the HMM.
As a method of estimating the parameters λ of the HMM, as explained above, a Baum-Welch re-estimation method is widely used. The Baum-Welch re-estimation method is a method of estimating parameters based on an EM (Expectation-Maximization) algorithm.
With the Baum-Welch re-estimation method, estimation of the parameters λ of the HMM is performed on the basis of observed time series data x=x1, x2, . . . , xT such that likelihood calculated from an occurrence probability, which is a probability that the time series data is observed (caused), is maximized.
xt represents a signal (a sample value) observed at time t. T represents the length of the time series data (the number of samples).
The Baum-Welch re-estimation method is a parameter estimating method based on likelihood maximization. However, optimality is not guaranteed. An HMM may converge into a local solution depending on the structure of the HMM or initial values of the parameters λ. Details of the HMM and the Baum-Welch re-estimation method are described in, for example, Laurence Rabiner and Biing-Hwang Juang, “Basics of Sound Recognition (two volumes)”, NTT Advanced Technology Corporation (hereinafter also referred to as document A).
HMMs are widely used in sound recognition. However, in the HMMs used for sound recognition, in general, the number of states, a method of state transition, and the like are determined in advance.
The HMM shown in
In
As opposed to an HMM with limitation on state transition like the HMM shown in
The Ergodic HMM is an HMM with a highest degree of freedom in terms of structure. However, when the number of states is large, it is difficult to estimate the parameters λ.
For example, when the number of states of the Ergodic HMM is 1000, the number of state transitions is one million (=1000×1000).
Therefore, in this case, for example, concerning the state transition probability aij of the parameters λ, it is necessary to estimate one million state transition probabilities aij.
Limited state transitions may be sufficient as necessary state transitions depending on a modeling target. However, when it is not known in advance how state transitions should be limited, it is extremely difficult to appropriately estimate such an enormous number of parameters λ. When an appropriate number of states is not known in advance and information for determining the structure of an HMM is not known in advance either, it is more difficult to calculate appropriate parameters λ.
The data processing apparatus shown in
In
A sensor signal observed from a modeling target is inputted to the time-series-data input unit 111. The time-series-data input unit 111 outputs, on the basis of the sensor signal observed from the modeling target, time series data observed from the modeling target (hereinafter also referred to as observed time series data) x=x1, x2, . . . , xT to the data adjusting unit 112.
The time-series-data input unit 111 directly supplies, for example, a sensor signal in time series observed from the modeling target to the data adjusting unit 112 as observed time series data x.
The time-series-data input unit 111 supplies the observed time series data x to the data adjusting unit 112 in response to a request from the evaluating unit 114.
The data adjusting unit 112 adjusts, according to progress of learning of an HMM stored in the model storing unit 115 explained later, time series data used for the learning of the HMM, i.e., the observed time series data x supplied from the time-series-data input unit 111 and outputs time series data after the adjustment (hereinafter also referred to as adjusted time series data) x′=x1′, x2′, . . . , xT′.
The data adjusting unit 112 applies, for example, processing for down-sampling (down-sampling processing) to the observed time series data x=x1, x2, . . . , xT supplied from the time-series-data input unit 111 and outputs the adjusted time series data x′=x1′, x2′, . . . , xT′ obtained by the down-sampling processing.
For example, when the observed time series data x is time series data sampled at 1000 Hz, the data adjusting time 112 converts, with the down-sampling processing, the observed time series data x sampled at 1000 Hz into the adjusted time series data x′ sampled at 100 Hz.
With the down-sampling processing, high-frequency components included in the observed time series data x are removed. The adjusted time series data x′ is changed to time series data including only macro characteristics of the observed time series data x, i.e., low frequency components of the observed time series data x.
Learning of the HMM as the state transition model stored in the model storing unit 115 is performed by using such adjusted time series data x′ including only the macro characteristics of the observed time series data x. Consequently, in the HMM, the macro characteristics of the observed time series data x are acquired.
In the data adjusting unit 112, it is an important problem in performing learning for appropriately acquire characteristics of the observed time series data x to determine how the observed time series data x should be adjusted, i.e., in the present case, into the adjusted time series data x′ of which sampling frequency the observed time series data x should be converted.
As explained above, the data adjusting unit 112 adjusts the observed time series data x according to progress of learning of an HMM. The adjustment is performed such that, as the learning of the HMM progresses, the adjusted time series data x′ changes from time series data including only macro characteristics of the observed time series data x to time series data including micro characteristics, i.e., high-frequency components of the observed time series data x as well.
For example, the data adjusting unit 112 gradually changes a sampling frequency of the adjusted time series data x′ from a small value to a large value as the learning of the HMM progresses.
Specifically, for example, in an initial period of learning, the data adjusting unit 112 sets the sampling frequency of the adjusted time series data x′ to 10 Hz. Thereafter, as the learning progresses, the data adjusting unit 112 sequentially changes the sampling frequency of the adjusted time series data x′ to 50 Hz, 100 Hz, 500 Hz, and 1000 Hz.
In this case, the HMM acquires the macro characteristics of the observed time series data x in the initial period of the learning and acquires the micro characteristics of the observed time series data x as the learning progresses.
Progress state information indicating a state of the progress of the learning is supplied to the data adjusting unit 112 from the evaluating unit 114. The data adjusting unit 112 recognizes the state of the progress of the learning on the basis of the progress state information from the evaluating unit 114 and changes the sampling frequency of the adjusted time series data x′.
In the data adjusting unit 112, as processing for adjusting the observed time series data x from the adjusted time series data x′ including the macro characteristics to the adjusted time series data x′ including the micro characteristics according to the progress of the learning, filter bank processing can be adopted besides the down-sampling processing (processing for curtailing the observed time series data x in a time direction).
In the data adjusting unit 112, when the filter bank processing is adopted, the observed time series data x is filtered by using a predetermined division number of filter banks. Consequently, the observed time series data x is divided into the predetermined division number of frequency components. The predetermined division number of frequency components are outputted as the adjusted time series data x′.
In this case, in the data adjusting unit 112, the number of divisions of the filter banks is gradually changed to a larger number as the learning progresses.
As explained above, the data adjusting unit 112 adjusts the observed time series data x and outputs the adjusted time series data x′=x1′, x2′, . . . , xT′′. T′ represents the length of the adjusted time series data x′.
The adjusted time series data x′ outputted by the data adjusting unit 112 is supplied to the parameter estimating unit 113 and the structure adjusting unit 117.
The parameter estimating unit 113 estimates the parameters k of the HMM stored in the model storing unit 115 using the adjusted time series data x′ supplied from the data adjusting unit 112.
The parameter estimating unit 113 estimates, for example, with the Baum-Welch re-estimation method, the parameters λ of the HMM stored in the model storing unit 115 using the adjusted time series data x′ from the data adjusting unit 112.
The parameter estimating unit 113 supplies new parameters λ obtained by the estimation of the parameters k of the HMM to the model storing unit 115 and causes the model storing unit 115 to store the new parameters λ in a form of overwriting.
In estimating the parameters λ of the HMM, the parameter estimating unit 113 uses values stored in the model storing unit 115 as initial values of the parameters λ.
In the parameter estimating unit 113, when processing for estimating the new parameters λ is performed, the number of times of learning is counted once.
The parameter estimating unit 113 increments the number of times of learning by 1 every time the processing for estimating the new parameters λ is performed and supplies the number of times of learning to the evaluating unit 114.
Further, the parameter estimating unit 113 calculates, from the HMM defined by the new parameters λ, likelihood that the adjusted time series data x′ supplied from the data adjusting unit 112 is observed and supplies the likelihood to the evaluating unit 114.
The likelihood supplied to the evaluating unit 114 by the parameter estimating unit 113 can be calculated by using the observed time series data x rather than the adjusted time series data x′.
The evaluating unit 114 evaluates, on the basis of the likelihood and the number of times of learning supplied from the parameter estimating unit 113, the learned HMM, i.e., the HMM, the parameters λ of which are estimated by the parameter estimating unit 113. The evaluating unit 114 determines, on the basis of a result of the evaluation of the HMM, whether the learning of the HMM should be finished.
For example, until the number of times of learning supplied from the parameter estimating unit 113 reaches a predetermined number of times, the evaluating unit 114 evaluates that the acquisition of characteristics (time series patterns) of the observed time series data x by the HMM is insufficient and determines to continue the learning of the HMM.
When the number of times of learning supplied from the parameter estimating unit 113 reaches the predetermined number of times, the evaluating unit 114 evaluates that the acquisition of characteristics of the observed time series data x by the HMM is sufficient and determines to finish the learning of the HMM.
Until the likelihood supplied from the parameter estimating unit 113 reaches a predetermined value, the evaluating unit 114 evaluates that the acquisition of characteristics (time series patterns) of the observed time series data x by the HMM is insufficient and determines to continue the learning of the HMM.
When the likelihood supplied from the parameter estimating unit 113 reaches the predetermined value, the evaluating unit 114 evaluates that the acquisition of characteristics of the observed time series data x by the HMM is sufficient and determines to finish the learning of the HMM.
When the evaluating unit 114 determines to continue the learning of the HMM, the evaluating unit 114 requests the time-series-data input unit 111, the data adjusting unit 112, and the structure adjusting unit 117 to perform predetermined processing.
The evaluating unit 114 requests the time-series-data input unit 111 to supply observed time series data.
Further, the evaluating unit 114 supplies the number of times of learning and the likelihood to the data adjusting unit 112 as progress state information representing a state of progress of the learning to request the data adjusting unit 112 to perform down-sampling processing corresponding to the progress of the learning.
The evaluating unit 114 requests, according to the progress of the learning, the structure adjusting unit 117 to adjust the structure of the HMM stored in the model storing unit 115.
The model storing unit 115 stores, for example, the HMM as the state transition model.
When new parameters of the HMM are supplied from the parameter estimating unit 113, the model storing unit 115 updates (overwrites) stored values (stored parameters of the HMM) with the new parameters.
Before the learning of the HMM is started, the model storing unit 115 stores the structure of the HMM initialized by the initial-structure setting unit 116 (initial structure), i.e., initial values of parameters of the HMM determined on the basis of limitation concerning the number of states and state transitions of the HMM.
The parameters of the HMM by the parameter estimating unit 113 are estimated from the initial values determined by the initial-structure setting unit 116.
The structure of the HMM stored in the model storing unit 115 is adjusted by the structure adjusting unit 117 according to the progress of the learning. The update of the stored values in the model storing unit 115 is also performed according to parameters of the HMM obtained by the adjustment of the structure of the HMM by the structure adjusting unit 117.
The initial-structure setting unit 116 initializes the structure of the HMM before the learning of the HMM is started and sets parameters of the HMM having the initialized structure (initial structure) (initial parameters).
The initial-structure setting unit 116 sets the initial structure of the HMM, i.e., the number of states and state transitions of the HMM.
Predetermined limitation can be applied to the number of states and the state transitions of the HMM as the initial structure.
For example, the initial-structure setting unit 116 sets the number of states of the HMM to be equal to or smaller than a predetermined number set as the predetermined limitation.
Specifically, for example, the initial-structure setting unit 116 sets the number of states of the HMM to a relatively small number such as sixteen or one hundred.
Further, the initial-structure setting unit 116 appropriately arranges states in the number of states set as the initial structure in an L-dimensional space (L is a positive integer) equal to or larger than one dimension.
For example, when the number of states set as the initial structure is sixteen and the sixteen states are arranged in a two-dimensional space, the initial-structure setting unit 116 arranges the sixteen states in the two-dimensional space in, for example, a lattice shape.
Thereafter, the initial-structure setting unit 116 sets, with respect to the sixteen states arranged in the two-dimensional space, state transitions, i.e., self-transition and state transition to other states.
Predetermined limitations such as limitation that the structure should be sparse structure can be applied to the state transitions set with respect to the sixteen states.
The sparse structure is, rather than dense state transition structure like the Ergodic HMM in which state transition from an arbitrary state to an arbitrary state is possible, structure in which states to which state transition is possible from a certain state are extremely limited.
Even in the sparse structure, at least one state transition to other states is present and self-transition is present.
The initial-structure setting unit 116 obtains the initial structure by, for example, as explained above, applying the predetermined limitation to initialize the structure of the HMM into the sparse structure. Then, the initial-structure setting unit 116 sets initial parameters, i.e., initial values of the state transition probability aij, the output probability density function bj(x), and the initial probability πi in the HMM having the initial structure.
The initial-structure setting unit 116 sets, for example, with respect to each of the states, the state transition probability aij of (valid) state transition, which is possible from the state, to a uniform value (when the number of possible state transitions is M, 1/M). The initial-structure setting unit 116 sets the state transition probability aij of difficult state transition state, i.e., transition other than the state transition set as the sparse state transition to 0.
When, for example, a normal distribution is used as the output probability density function bj(x), the initial-structure setting unit 116 calculates, according to the following formula, an average μ and dispersion σ2 of the observed time series data x=x1, x2, . . . , XT obtained by the time-series-data input unit 111. The initial-structure setting unit 116 sets the normal distribution defined by the average μ and the dispersion σ2 in the output probability density function bj(x) of the respective states sj.
μ=(1/T)Σxt
σ2=(1/T)Σ(xt−μ)2
In the above formula, Σ means summation with time t changed from 1 to length T of the observed time series data x.
Further, the initial-structure setting unit 116 sets initial probabilities πi of the respective states si to a uniform value. When the number of states of the HMM having the initial structure is set to N, the initial-structure setting unit 116 sets the initial probabilities πi of the respective N states si to 1/N.
The HMM with the initial structure and the initial parameters X={aij, bj(x), πi, i=1, 2, . . . , n, j=1, 2, . . . , N} set by the initial-structure setting unit 116 is supplied to and stored in the model storing unit 115. The (initial) structure and the (initial) parameters stored in the model storing unit 115 are updated by learning.
The structure-adjusting unit 117 adjusts, in response to a request from the evaluating unit 114, the structure of the HMM stored in the model storing unit 115 using the adjusted time series data x′ supplied from the data adjusting unit 112. The adjustment of the structure of the HMM performed by the structure adjusting unit 117 includes adjustment of parameters of the HMM necessary for the adjustment of the structure.
As types of the adjustment of the structure of the HMM Performed by the structure adjusting unit 117, there are six types, i.e., division of a state, merging of a state, addition of a state, addition of state transition, deletion of a state, and deletion of state transition.
The processing by the initial-structure setting unit 116 shown in
The initial-structure setting unit 116 can set Ergodic structure as the initial structure of the HMM or can set sparse structure by applying predetermined limitation to the initial structure.
In
In
It is assumed that a distance between states adjacent to each other in the horizontal direction and a distance between states adjacent to each other in the vertical direction are 1. In
In
A method of setting the sparse initial structure is not limited to a method of applying limitation to states arranged on the L-dimensional space to allow only state transitions (including self-transition) to states located in the neighborhood according to a distance between states.
AS the method of setting the sparse initial state, besides the methods explained above, for example, it is possible to adopt a method of selecting, when attention is directed to a certain state, a random number of the states out of all the states and limiting state transitions to only state transitions (including self-transition) to the states selected at random.
As the method of setting the sparse initial structure, it is also possible to adopt the structures of the HMMs shown in
Division of a state performed by the structure adjusting unit 117 as adjustment of the structure of an HMM is explained with reference to
As explained above, in the figure, circles represent states. In the following explanation, circles affixed with a number “i” are described as state si.
In
The division of a state is performed in order to increase the size of the HMM.
In
The division of the state S5 is performed by adding a new state s7, in which state transitions same as those for the state s5 as a division target can be performed and bidirectional state transition to and from the state s5 can be performed.
In
Further, the structure adjusting unit 117 sets, concerning the new state s7, state transition between the state s7 and the state s5 as valid state transition as well.
In the division of a state, for example, the structure adjusting unit 117 state sets parameters of the new state s7 to, so to speak, succeed parameters of the division target state s5.
The structure adjusting unit 117 sets initial probability π7 state of the new state s7 as initial probability π5 of the division target state s5(π7=π5) and sets an output probability density function b7(x) of the new state s7 as an output probability density function b5(x) of the division target state s5 (b7(x)=b5 (x)).
Further, the structure-adjusting unit 117 sets a state transition probability ai7 of state transition from the state si (in
The structure adjusting unit 117 sets a state transition probability a7i of state transition from the new state s7 to the state si as a state transition probability a5i from the division target state s5 to the state si (a7i=a5i).
The structure adjusting unit 117 sets a state transition probability a57 of state transition from the state s5 as the division target to the new state s7 and a state transition probability a75 of state transition from the new state s7 to the division target state s5 to an appropriate value, for example, a57=a75=0.5.
Further, the structure adjusting unit 117 applies normalization processing to necessary parameters of the HMM after the division of a state and finishes processing for dividing a state.
The structure adjusting unit 117 applies normalization processing, which satisfies the following formula, to the initial probability πi and the state transition probability aij of the HMM after the division of a state.
Σπj=1
Σaij=1(i=1, 2, . . . N)
In the above formula, Σ means summation with the variable “j” representing a state changed from 1 to the number of states N of the HMM after the division of a state. In
As a method of the normalization processing, which satisfies the above formula, for example, there is a method of dividing the initial probability πj before the normalization processing by a sum Σπj=π1+π2+ . . . +πN of the initial probability πj before the normalization processing. The same holds true for normalization processing for the state transition probability aij.
In the division of a state, a state as a division target is not limited to one state.
As the state as the division target, a predetermined number “n” of states (n is equal to or larger than 1 and equal to or smaller than N) can be selected, for example, at random out of N states s1 to SN of the HMM before the division of a state.
As the division target state, for example, higher order “n” states having large dispersion σ2 defining the output probability density function bj(x), i.e., higher order “n” states with relatively large fluctuation in an observation value observed from the states can be selected out of the N states s1 to SN of the HMM before the division of a state.
The number “n” of states as division targets can be set at random or can be set to a fixed value. In both the cases, by the division of a state, the structure of the HMM is updated to a structure in which the number of states increase by “n” from the number of states before the division.
Merging of a state performed by the structure adjusting unit 117 as adjustment of the structure of the HMM is explained with reference to
The merging of a state is performed in order to degenerate redundantly-allocated states.
In
The merging of the state s5 into the state s6 is performed by deleting state transition between the merging target state s5 and the merged target state s6 and deleting the merging target state s5 such that the merged target state s6, so to speak, succeeds state transitions (hereinafter also referred to as peculiar state transitions) between the merging target state s5 and other states excluding the merging target state s5 and the merged target state s6.
Therefore, the structure adjusting unit 117 deletes (invalidates) state transition between the merging target state s5 and the merged target state s6.
Further, in
The structure adjusting unit 117 deletes the merging target state s5.
Further, in the merging of a state, for example concerning state transitions succeeded from the merging target state s5 among state transitions of the merged target state s6, the structure adjusting unit 117 sets state transition probabilities ai6 and a6i to succeed state transition probabilities ai5 and a5j of the merging target state s5.
In
Further, the structure adjusting unit 117 sets a state transition probability a26 of state transition from the state s2 to the merged target state s6 as a state transition probability a25 of state transition from the state s2 to the merging target state s5 (a26=a25).
Similarly, the structure adjusting unit 117 sets a state transition probability a64 of state transition from the merged target state s6 to the state s4 as a state transition probability a54 of state transition from the merging target state s5 to the state s4 (a64=a54).
Further, the state adjusting unit 117 sets a state transition probability a46 of state transition from the state s4 to the merged target state s6 as a state transition probability a45 of state transition from the state s4 to the merging target state s5 (a46=a45).
The structure adjusting unit 117 applies normalization processing to necessary parameters of the HMM after the merging of a state and finishes the merging of a state.
The structure adjusting unit 117 applies normalization processing same as that in the case of the division of a state to the initial probability πi of the HMM after the merging of a state and the state transition probability aij.
In the merging of a state, a set of a state to be set as a merging target and a state to be set as a merged target (hereinafter also referred to as merge set) is not limited to one set.
As a pair of states to be set as the merge set, for example, a pair of higher order “n” (n is a value equal to or larger than 1) states with larger correlation among the states out of pairs of states, in which bidirectional state transitions can be performed, in the N states s1 to sN of the HMM before the merging of a state.
The number “n” of pairs of state to be set as the merge set can be set at random or can be set to a fixed value. In both the cases, by the merging of a state, the structure of the HMM is updated to structure in which the number of states decreases by “n” from the number of states before the merging.
A correlation among states used in selecting pairs of states to be set as a merge set is explained.
The correlation among states represents a degree of similarity of state transitions (including self-transition) to the other states, state transitions from the other states, and observation values observed from states. For example, the correlation among states is calculated as explained below.
As explained with reference to
The structure adjusting unit 117 calculates a correlation among states of the HMM stored in the model storing unit 115 using the adjusted time series data x′ supplied from the data adjusting unit 112.
The structure adjusting unit 117 applies a forward-backward algorithm to the HMM stored in the model storing unit 115 and calculates a forward-backward probability pi(t), which is a probability of being present in the state si at every time t (a probability that a state at time t is the state si) of the adjusted time series data x′=x1′, x2′, . . . , xT′′ supplied from the data adjusting unit 112.
The forward-backward algorithm is an algorithm for calculating a probability value as an integrated value of a forward probability αi(t) calculated by propagating a probability of reaching the respective states si forward in a time direction and a backward probability βi(t) calculated by propagating the probability of reaching the respective states si backward in the time direction.
The forward-backward algorithm is described in Document explained above.
The structure adjusting unit 117 calculates, concerning the HMM stored in the model storing unit 115, the forward probability αi(t) of observing the data x1′, x2′, . . . , xt′ of the adjusted time series data x′ and being present in the state si at time t. Further, the structure adjusting unit 117 calculates, concerning the HMM stored in the model storing unit 115, the backward probability βi(t) of being present in the state si at time t and thereafter observing data xt′, xt+1′, . . . , xT′′ of the adjusted time series data x′(being present in the state si at time t when the data xt′, xt+1′, . . . , xT′′ is observed after time t).
The structure adjusting unit 117 calculates the forward-backward probability pi(t) of being present in the state si at time t using the forward probability αi(t) and the backward probability βi(t).
The structure adjusting unit 117 calculates the forward-backward probability pi(t) of being present in the respective states si at every time t=1, 2, . . . , T′ of the adjusted time series data x′.
Forward-backward probabilities pi(1), pi(2), . . . , pi(T′) of a certain state si are time series data with the length T′. This time series data is represented as pi(=pi(1), pi(2), . . . , pi(T′)) as well.
A correlation between the certain state si and the other states sj is represented as pi*pj. The structure adjusting unit 117 calculates, according to the following formula, the correlation Pi*pj between the certain state si and the other states sj using, for example, forward-backward probabilities pi=pi(1), Pi(2), . . . , pi(T′) of the state si and forward-backward probabilities pj=pj(1), pj(2), . . . , pj(T′).
pi*pj=Σpi(t)pj(t)
In the above formula, Σ means summation with the time t changed from 1 to the length T′ of the adjusted time series data x′.
The correlation pi*pj between the states si and sj is high when time change patterns of the forward-backward probability pi of the state si and the forward-backward probability pj of the state si are similar, i.e., when, besides one of the states si and sj, the other is redundantly present.
In this case, if bidirectional state transition is present between the states si and sj, a pair of the states si and sj is selected as a merge set. With one of the redundant states si and sj set as a merging target and the other set as a merged target, a state of the merging target is merged into a state of the merged target.
The structure adjusting unit 117 can also calculate a correlation between states of the HMM stored in the model storing unit 115 using the observed time series data x rather than the adjusted time series data x′.
It is possible to select, for example, at random which of two states selected as a merge set should be set as a merging target or a merged target.
Addition of a state performed by the structure adjusting unit 117 as adjustment of the structure of an HMM is explained with reference to
Like the division of a state explained with reference to
In
The addition of a state is performed by adding the new state s7, in which bidirectional state transition to and from the state s5 as the target to which a state is added can be performed.
Therefore, the structure adjusting unit 117 sets, concerning the new state s7, the self-transition and the state transition to and from the state s5 as valid state transitions.
For example, the structure adjusting unit 117 sets, in the addition of a state, parameters of the new state s7 to, so to speak, succeed parameters of the state s5 as the target to which a state is added.
The structure adjusting unit 117 sets an initial probability π7 of the new state s7 as an initial probability π5 of the state s5 as the target to which a state is added (π7=π5). The structure adjusting unit 117 sets an output probability density function b7(x) of the new state s7 as an output probability density function b5(x) of the state s5 to which a state is added (b7(x)=b5(x)).
The structure adjusting unit 117 sets a state transition probability a57 of state transition from the state s5 as the target to which a state is added to the new state s7 and a state transition probability a75 of state transition from the new state s7 to the state s5 as the target to which a state is added to an appropriate value, for example, a57=a75=0.5.
Further, the structure adjusting unit 117 applies normalization processing to necessary parameters of the HMM after the addition of a state and finishes processing for adding a state.
The structure adjusting unit 117 applies normalization processing same as that in the case of the division of a state to the initial probability πi and the state transition probability aij of the HMM after the addition of a state.
In the addition of a state, a state as the target to which a state is added is not limited to one state.
As the state as a target to which a state is added, an arbitrary number “n” of states (n is equal to or larger than 1 and equal to or smaller than N) can be selected, for example, at random, out of the N states s1 to sN of the HMM before the addition of a state.
As the state as the target to which a state is added, for example, higher order “n” states having large dispersion σ2 defining the output probability density function bj(x), i.e., higher order “n” states with relatively large fluctuation in an observation value observed from the states can be selected out of the N states s1 to sN of the HMM before the addition of a state.
The number “n” of states as targets to which a state is added can be set at random or can be set to a fixed value. In both the cases, by the addition of a state, the structure of the HMM is updated to a structure in which the number of states increases by “n” from the number of states before the addition.
The addition of a state and the division of a state explained with reference to
Therefore, concerning state transition, in the HMM after the division of a state shown in
As a result, in the addition of a state, independency of the new state is high compared with that in the case of the division of a state.
Addition of state transition performed by the structure adjusting unit 117 as adjustment of the structure of an HMM is explained with reference to
The addition of state transition is performed to solve a problem in that state transitions are insufficient for appropriately representing a modeling target in the structure of the HMM stored in the model storing unit 115. In particular, when sparse state transition is set as an initial structure of the HMM by the initial-structure setting unit 116, it is important to add state transition necessary for appropriate representation of the modeling target.
In
In the addition of state transition, the structure adjusting unit 117 sets valid state transition between the states s4 and s6 as the targets of addition of state transition. Further, the structure adjusting unit 117 sets a state transition probability a46 from one of the states s4 and s6 as the targets of addition of state transition, for example, the state s4 to the other, for example, the state s6 and a state transition probability from the other state s6 to one state s4 to an appropriate value, for example, a46=a64=0.5.
The structure adjusting unit 117 applies normalization processing to necessary parameters of the HMM after the addition of state transition and finishes processing for the addition of state transition.
The structure adjusting unit 117 applies normalization processing same as that in the case of the division of a state to a state transition probability aij after the addition of state transition.
In the addition of state transition, a set of two states as targets of addition of state transition (hereinafter also referred to as addition target set) is not limited to one set.
As a pair of states to be set as the addition target set, for example, a pair of higher order “n” (n is a value equal to or larger than 1) states having a large correlation between the states can be selected out of pairs of states, in which bidirectional state transition is not possible, in N states s1 to sN of the HMM before the addition of state transition.
As explained above, a pair of states having a large correlation are selected as the addition target set among the states in which bidirectional state transition is not possible. In this case, for example, when one of two states without state transition (in which it is difficult to perform direct state transition) is a redundant presence for the other, the two states are mechanically connected by state transition.
The number “n” of pairs of states to be set as the addition target set can be set at random or can be set to a fixed value. In both the cases, the structure of the HMM is updated to a slightly complicated structure in which the number of states does not change but state transitions increases by “n”.
Deletion of a state performed by the structure adjusting unit 117 as adjustment of the structure of an HMM is explained with reference to
In
The deletion of a state is performed to delete a state unnecessary for appropriately representing a modeling target.
In
The deletion of a state is performed by deleting the state s5 as the target of deletion and state transitions possible from the state s5 (including state transitions to the state s5).
In
Further, the state adjusting unit 117 applies normalization processing to necessary parameters of the HMM after the deletion of a state and finishes processing for the deletion of a state.
The structure adjusting unit 117 applies normalization processing same as that in the case of the division of a state to an initial probability πi after the deletion of a state and a state transition probability aij.
The structure adjusting unit 117 selects a state to be set as a target of deletion, for example, as explained below.
As explained with reference to
The structure adjusting unit 117 applies a Viterbi method to the HMM stored in the model storing unit 115 and calculates a process of state transition (a sequence of states) (a path) that maximizes likelihood that the adjusted time series data x′=x1′, x2′, . . . , xT′′ from the data adjusting unit 112 is observed (hereinafter also referred to as maximum likelihood path).
The Viterbi method is an algorithm for determining, among paths of state transitions starting from the respective states si, a path (a maximum likelihood path) for maximizing a value (an occurrence probability) obtained by accumulating, over the length T′ of the adjusted time series data x′, a state transition probability a of state-transitioning from the state si to the state sj at time t and a probability of a sample value x′t at time t of the adjusted time series data x′=x1′, x2′, . . . , xT′′ being observed in the state transition (an output probability calculated from an output probability density function bj(x)).
The Viterbi method is described in Document A explained above.
The structure adjusting unit 117 determines, concerning the adjusted time series data x′, a sequence of states s1′, s2′, . . . , sT′′ and then detects a state not forming the maximum likelihood path (a state not included in the maximum likelihood path) among the states of the HMM.
The state not forming the maximum likelihood path can be regarded as a state not always necessary for representing characteristics (time series patterns) of the adjusted time series data x′=x1′, x2′, . . . , xT′′. Therefore, the structure adjusting unit 117 selects the state not forming the maximum likelihood path as a state to be set as a target of deletion.
For example, in the HMM having the states s1 to s9 shown in
As explained above, the structure adjusting unit 117 deletes the state s5 selected as the target of deletion. Consequently, adjustment of the structure for changing the HMM shown in
The structure adjusting unit 117 performs, as the adjustment of the structure of the HMM, deletion of state transition besides the division of a state, the merging of a state, the addition of a state, the addition of state transition, and the deletion of a state explained with reference to
The deletion of state transition is performed in the same manner as the deletion of a state.
As explained above, the structure adjusting unit 117 determines, concerning the adjusted time series data x′, a sequence of states s1′, s2′, . . . , sT′′ as a maximum likelihood path and selects state transition not forming the maximum likelihood path as state transition to be set as a target of deletion.
Further, the structure adjusting unit 117 deletes the state transition selected as the state transition to be set as a target of deletion, applies normalization processing same as that in the case of the division of a state to the state transition probability aij of the HMM after the deletion of state transition, and finishes processing for the deletion of state transition.
A sensor signal from a modeling target is supplied to the time-series-data input unit 111. The time-series-data input unit 111 directly sets, for example, the sensor signal observed from the modeling target as observed time series data x.
As explained above, the observed time series data x is supplied from the time-series-data input unit 111 to the data adjusting unit 112. Besides, the observed time series data x is supplied to the initial-structure setting unit 116 as well and, as explained above, used for setting of the output probability density function bj(x) in the initial-structure setting unit 116.
In step S111, the initial-structure setting unit 116 performs initialization of an HMM.
The initial-structure setting unit 116 initializes the structure of the HMM to an initial structure and sets parameters of the HMM having the initial structure (initial parameters).
Specifically, the initial-structure setting unit 116 sets, as the initial structure of the HMM, the number of states of the HMM and sets sparse state transition in the HMM having the number of states.
Further, the initial-structure setting unit 116 sets, in the HMM having the initial structure, initial values of the state transition probability aij, the output probability density function bj(x), and the initial probability πi as the initial parameters.
As explained above, the HMM in which the initial structure and initial parameters λ={aij, bj(x), πi, i=1, 2, . . . , N, j=1, 2, . . . , N} are set by the initial-structure setting unit 116 is supplied to and stored in the model storing unit 115.
Thereafter, the processing proceeds from step S111 to step S112. The time-series-data input unit 111 supplies the observed time series data x to the data adjusting unit 112. The processing proceeds to step S113.
In step S113, the data adjusting unit 112 performs, as explained with reference to
Besides being supplied from the data adjusting unit 112 to the parameter estimating unit 113, the adjusted time series data x′ is supplied to the structure adjusting unit 117 as well.
In step S114, the parameter estimating unit 113 estimates, with the parameters of the HMM stored in the model storing unit 115 set as initial values, new parameters of the HMM with the Baum-Welch re-estimation method using the adjusted time series data x′ supplied from the data adjusting unit 112.
The parameter estimating unit 113 supplies the new parameters of the HMM to the model storing unit 115 and causes the model storing unit 115 to store the new parameters in a form of overwriting.
The parameter estimating unit 113 increments the number of times of learning, which is reset to 0 during the start of the learning processing in
Further, the parameter estimating unit 113 calculates likelihood of observation of the adjusted time series data x′ from the HMM defined by the new parameters λ and supplies the likelihood to the evaluating unit 114. The processing proceeds from step S114 to step S115.
In step S115, the evaluating unit 114 evaluates, on the basis of the likelihood and the number of times of learning supplied from the parameter estimating unit 113, the HMM for which learning is performed i.e., the HMM for which the parameters λ are estimated by the parameter estimating unit 113 and determines, on the basis of a result of the evaluation of the HMM, whether the learning of the HMM should be finished.
When it is determined in step S115 that the learning of the HMM is not finished, the evaluating unit 114 requests the time-series-data input unit 111, the data adjusting unit 112, and the structure adjusting unit 117 to perform predetermined processing. The processing proceeds to step S116.
In step S116, the structure adjusting unit 117 performs, in response to a request from the evaluating unit 114, processing for adjusting the structure of the HMM stored in the model storing unit 115 using the adjusted time series data x′ supplied from the data adjusting unit 112. The processing returns to step S112.
In step S112, the time-series-data input unit 111 supplies the observed time series data x to the data adjusting unit 112 in response to a request from the evaluating unit 114. The processing proceeds to step S113.
In step S113, the data adjusting unit 112 performs, in response to a request from the evaluating unit 114, adjustment of the observed time series data x supplied from the time-series-data input unit 111 as explained with reference to
The parameter estimating unit 113 estimates parameters of the HMM. The structure adjusting unit 117 adjusts the structure of the HMM defined by the parameters after the estimation. This processing is repeated.
The data adjusting unit 112 performs down-sampling processing with the observed time series data x set as a target, for example, as explained with reference to
On the other hand, when it is determined in step S115 that the learning of the HMM is finished, the learning processing is finished.
As explained above, in the data processing apparatus shown in
As a result, even if a modeling target is complicated, an HMM that appropriately models the modeling target can be obtained.
In general, an HMM having a large number of states and a large number of state transitions is necessary for modeling of a complicated modeling target. However, it is difficult to use, from the beginning, the HMM having a large number of states and a large number of state transitions and correctly estimate parameters of the HMM.
In the data processing apparatus shown in
Further, in the data processing apparatus shown in
In step S121, the structure adjusting unit 117 applies the division of a state explained with reference to
In step S122, the structure adjusting unit 117 calculates a correlation among states forming the HMM after the division of a state using the adjusted time series data x′ supplied from the data adjusting unit 112. The processing proceeds to step S123.
In step S123, the structure adjusting unit 117 applies the merging of a state explained with reference to
In step S124, the structure adjusting unit 117 applies the addition of state transition explained with reference to
In step S125, the structure adjusting unit 117 applies the addition of a state explained with reference to
In step S126, the structure adjusting unit 117 calculates, concerning the adjusted time series data x′ supplied from the data adjusting unit 112, a maximum likelihood path using the HMM after the addition of a state. The processing proceeds to step S127.
In step S127, the structure adjusting unit 117 detects a state and state transition not forming the maximum likelihood path. Further, in step S127, the structure adjusting unit 117 deletes the state and the state transition not forming the maximum likelihood path as explained with reference to
The structure adjusting unit 117 updates the stored values in the model storing unit 115 with parameters of the HMM after the deletion of the state and the state transition. The processing returns to step S121.
As explained above, the structure adjusting unit 117 performs, concerning the HMM stored in the model storing unit 115, six kinds of adjustment of the structure, i.e., the division of a state, the merging of a state, the addition of a state, the addition of state transition, the deletion of a state, and the deletion of state transition.
In
Therefore, the structure adjusting unit 117 performs the adjustment of the structure of the HMM every time the number of times of learning increases by one. However, the adjustment of the structure of the HMM can be performed according to progress of the learning other than the increase in the number of times of learning by one.
The evaluating unit 114 supplies the number of times of learning and the likelihood to the data adjusting unit 112 as progress state information representing a state of programs of the learning. The progress state information can be supplied to the structure adjusting unit 117 as well.
In this case, the structure adjusting unit 117 performs the adjustment of the structure of the HMM according to the progress state information supplied from the evaluating unit 114.
For example, it is possible to cause the structure adjusting unit 117 to perform the adjustment of the structure when the number of times of learning as the progress state information increases by a predetermined number of times from the number of times at the time of the last adjustment of the structure.
It is also possible to cause the structure adjusting unit 117 to perform the adjustment of the structure, for example, when the likelihood as the progress state information falls from a value at the time of the last adjustment of the structure or when a ratio of an increase in the likelihood falls to be equal to or lower than a predetermined value.
The adjustment of the structure of the HMM by the structure adjusting unit 117 does not guarantee that the structure of the HMM converges to an optimum structure that represents the modeling target.
However, according to the adjustment of the structure of the HMM by the structure adjusting unit 117, a state and state transition estimated as being appropriate for representing the modeling target are added and, on the other hand, a state and state transition estimated as being unnecessary for representing the modeling target are deleted. Therefore, even if a modeling target is a complicated modeling target, it is possible to obtain a large HMM that appropriately models the modeling target.
In
Simulation performed for the data processing apparatus shown in
In the simulation, a sequence of coordinates (x,y) of a moving locus of a robot that moves at random in a two-dimensional space was used as the observed time series data x.
A range of the coordinates (x,y) of the two-dimensional space in which the robot could move was set in a range excluding areas of four blocks #1, #2, #3, and #4 indicated by areas of the following formula in a range represented by
−100<x<+100 and −100<y<+100.
Block #1: −70<x<−20, −70<y<−20
Block #2: −70<x<−20, +20<y<+70
Block #3: +20<x<+70, −70<y<−20
Block #4: +20<x<+70, +20<y<+70
The robot was moved 10000 steps (times) in a movable range with an origin (0,0) set as a start position while a very small moving amount (Δx,Δy) was sequentially determined at random.
In
According to
In the simulation, the sequence of the coordinates (x,y) for 10000 steps was used as the observed time series data x. However, it was assumed unknown that the movable range of the robot and the observed time series data x were the coordinates (x,y) in the two-dimensional space.
In the simulation, it was assumed that information concerning the robot, the coordinates (x,y) (moving locus) as the observed time series data x of which were observed, was not given in advance and it was known in advance that the two-dimensional observed time series data x was measured.
In the simulation, the HMM having the sixteen states shown in
In the simulation, the learning of the HMM was finished at a stage when the number of times of learning reached thirty-six.
In the simulation, as explained above, the observed time series data as the sequence of the coordinates (x,y) for 10000 steps, i.e., the observed time series data including 10000 samples was used for the learning. In the first learning of the thirty-six times of learning, the down-sampling processing was applied to the observed time series data including 10000 samples such that a sampling frequency fell to 1/10 of an original sampling frequency. Adjusted time series data including 1000 samples obtained as a result of the down-sampling processing was used for, for example, estimation of parameters of the HMM.
Thereafter, the sampling frequency of the adjusted time series data was gradually increased such that the sampling frequency fell to 1/9, ⅛, 1/7, . . . and 1/1 of the original sampling frequency every time the number of times of learning increased by three. In this case, when the number of times of learning reached twenty-eight or more, the adjusted time series data was the observed time series data itself.
In
In
In
In
According to
In other words, according to
As explained above, in the data processing apparatus shown in
As explained above, functions of the setting of the sparse initial structure of the HMM by the initial-structure setting unit 116, the adjustment of the structure of the HMM by the structure adjusting unit 117, and the adjustment of the observed time series data by the data adjusting unit 112 act. Therefore, it is possible to perform determination of the structure of a large HMM and estimation of parameters, which are difficult to treat in the past.
The data processing apparatus shown in
In the case explained above, the initial-structure setting unit 116 initializes the structure of the HMM to the sparse structure. However, besides, the initial-structure setting unit 116 can initialize the structure of the HMM to, for example, Ergodic structure.
In the case explained above, the data adjusting unit 112 adjusts the observed time series data according to the progress of the learning. However, the adjustment of the observed time series data does not have to be performed. In this case, in the data processing apparatus shown in
It is possible to determine, for example, according to likelihood itself or a degree of a change in the likelihood, whether the adjustment of the structure by the structure adjusting unit 117 should be performed.
The series of processing explained above can be performed by hardware or can be performed by software. When the series of processing is performed by software, a program configuring the software is installed in a general-purpose computer or the like.
The program can be recorded in advance in a hard disk 155 or a ROM 153 as a recording medium incorporated in the computer.
Alternatively, the program can be temporarily or permanently stored (recorded) in a removable recording medium 161 such as a flexible disk, a CD-ROM (Compact Disc Read Only Memory), an MO (Magneto Optical) disk, a DVD (Digital Versatile Disc), a magnetic disk, or a semiconductor memory. Such a removable recording medium 161 can be provided as so-called package software.
Besides being installed in the computer from the removable recording medium 161, the program can be transferred by radio from a download site to the computer via an artificial satellite for digital satellite broadcast or transferred by wire from the download site to the computer via a network such as a LAN (Local Area Network) or the Internet. The computer can receive the program transferred in that way in a communication unit 158 and install the program in the hard disk 155 incorporated therein.
The computer incorporates a CPU (Central Processing Unit) 152. An input and output interface 160 is connected to the CPU 152 via a bus 151. A command is inputted to the CPU 152 by a user via the input and output interface 160 according to, for example, operation of an input unit 157 including a keyboard, a mouse, and a microphone. The CPU 152 executes the program stored in the ROM (Read Only Memory) 153 according to the command. Alternatively, the CPU 152 loads the program stored in the hard disk 155, the program transferred from the satellite or the network, received by the communication unit 158, and installed in the hard disk 155, or the program read out from the removable recording medium 161 inserted in a drive 159 and installed in the hard disk 155 onto a RAM (Random Access Memory) 154 and executes the program. Consequently, the CPU 152 performs processing conforming to the flowcharts explained above or processing performed by the configurations shown in the block diagrams explained above. For example, the CPU 152 outputs a result of the processing from an output unit 156 including an LCD (Liquid Crystal Display) and a speaker, transmits the processing result from the communication unit 158, or records the processing result in the hard disk 155 via the input and output interface 160 according to necessity.
The information processing apparatus shown in
As shown in
The causality-learning processing unit 201 acquires HMMs of plural modals such as the distance HMM, the light HMM, and the energy HMM generated as explained above and performs causality learning. The causality-learning processing unit 201 outputs data obtained by performing the causality learning to the causality-estimation processing unit 202.
Only one node of nodes (states) of an HMM is typically ignited in the same HMM. Therefore, events of the nodes are a set of events that are exclusive to one another and mutually exclusive and collectively exhaustive (MECE). Therefore, it is possible to associate node transition in a certain HMM with node ignition in single or plural other HMMs. For example, changes with time of respective ignition nodes of N types of HMMs are recorded and used for causality learning.
The causality-estimation processing unit 202 performs causality estimation using data supplied from the causality-learning processing unit 201. The causality-estimation processing unit 202 causes the causality-candidate-list storing unit 203 to store a list representing causality obtained by performing the causality estimation.
The causality among the events is represented by a conditional probability as explained later. Acquiring data used for calculating the conditional probability is referred to as causality learning. Calculating a conditional probability using the data acquired by the causality learning and estimating causality is referred to as causality estimation. Causality perception represents a state in which the causality among the events is perceived by the causality estimation.
The causality-candidate-list arrangement processing unit 204 appropriately arranges a causality candidate list stored in the causality-candidate-list storing unit 203.
When a target value is given, the behavior determining unit 205 determines a behavior with reference to the causality candidate list stored in the causality-candidate-list storing unit 203. A behavior of the robot is controlled on the basis of a command representing the behavior determined by the behavior determining unit 205.
In the causality-estimation processing unit 202, basically, causality estimation is performed as explained below. Details of the causality estimation are explained later.
In performing causality estimation for a certain event a1, events that could occur are classified into, at least in a range of experiences of the robot, a set A including events a1, a2, a3, and the like including the event a1 exclusive to one another and mutually exclusive and collectively exhaustive (MECE) and a set B as a set of the other events.
The causality is represented by a conditional probability P (T: ak→al|ak,b) calculated for all events b that have occurred simultaneously with an event ak in the past. T: ak→al represents transition from the event ak to the event al. In the following explanation, T: ak→al is simply represented as T.
Since the conditional probability P(T|ak,b) is represented by the following formula, a conditional probability is calculated from values of N(T,ak,b) and N(ak,b).
P(T\ak,b)=P(T,ak,b)/P(ak,b)=≡N(T,ak,b)/N(ak,b)
N(T,ak,b) represents the number of times the event ak and the event b simultaneously occur and the event al occurs at the next time. N(ak,b) represents the number of times the event ak and the event b simultaneously occur.
The conditional probability calculated in this way has an error. The magnitude of the error is expected to be inversely proportional to √N(T,ak,b). Therefore, if the event b is controlled by, for example, changing granularity to keep N(T,ak, b) at a value in an appropriate range, the error can be reduced.
Specifically, when the number of experiences of the robot is small, rough representation such as “when it is bright” is used as a method of representing the event b. As the number of experiences increase, representation of finer granularity such as “when illuminance is 700 to 800 x”, “when bright light comes in from the behind”, or “when the robot is picked out by bright light simultaneously with warning sound” included in “when it is bright” is used. This makes it possible to perform finer control and causality estimation while keeping an error corresponding to the number of experiences small.
When the causality temporally fluctuates, it is possible to cause the conditional probability to follow the temporal fluctuation in the causality by attenuating the numbers of times of simultaneous occurrence of events N(T,ak,b) and N(ak,b) with time.
For example, it is assumed that trials were performed 1000 times in the past (N(ak,b)=1000) and successful 500 times (N(T,ak,b)=500) and P(T|ak,b)=500/1000=0.5. It is now assumed that trials are performed ten times and successful eight times now. When there is no attenuation with time, if the 10 times of the trials are included, P(T|ak,b)=508/1010=0.503. Therefore, even if the trials are successful at a high probability of eight times in ten times compared with the trials in the past, there is almost no influence on a value of P(T|ak,b).
On the other hand, when there is attenuation with time, the conditional probability is calculated by multiplying the numerator and the denominator with an attenuation ratio and evaluated. Therefore, when the attenuation ratio is set to 0.1, P(T|ak,b) calculated according to the trials in the past is in a state of P(T|ak,b)=50/100=0.5. When a latest trial result is reflected on this state, P(T|ak,b)=58/110=0.527 is obtained. If the trials in the past are much older and further attenuated, for example, P(T|ak,b) calculated according to the trials in the past is in a state of P(T|ak,b)=5/10=0.5. When the latest trial result is reflected on this state, P(T|ak,b)=13/20=0.65 is obtained.
In other words, it is possible to impart followability to the latest trial result to P(T|ak,b) (cause a value of P(T|ak,b) to be substantially affected by the latest trial result). Even when the latest trial result is not present, it is possible to directly use a probability estimation value based on experiences in the past by attenuating the numbers of times simultaneous occurrence N(T,ak,b) and N(ak,b) at the same attenuation ratio. Since a value of N(T,ak,b) affecting the calculation of the estimation error decreases according to the elapse of time, the estimation error gradually increases. It is possible to automatically represent a characteristic that the experiences in the past are indefinite compared with recent experiences.
From the viewpoint of model application, the conditional probability P(T|ak,b) is a formula that gives likelihood of a model that “if (ak,b), T:ak→al occurs” under observation T:ak→al (transition T from the event ak to the event al).
On the other hand, when the conditional probability P(T|ak,b) is set as a posterior probability, the conditional probability P(T|ak,b) can also be regarded as a probability that transition T:ak→al occurs when (ak,b) simultaneously occurs. This makes it possible to use the conditional probability P(T|ak,b) for control for causing the transition T. Specifically, if an event b with maximum P(T|ak,b) among all events b can be attained together with the event ak, a probability of occurrence of the transition T is the highest. Therefore, in order to cause the transition T, such an event b only has to be found to determine a behavior.
A disadvantage of this method of determining a behavior is that the method tends to be affected by the estimation error due to the fluctuation in experiences. For example, it is examined which of an event that has been successful five times out of ten times and an event that has been successful 501 times out of 1000 times should be executed. It is possible that, although an actual rate of success of the former event is 0.55, the rate of success looks 0.5 because the event has been successful only five times in the trials performed ten times. If the event is tried once more, the rate of success may rise to 6/11=0.545 or may fall to 5/11=0.455. Under such circumstances, the behavior determination is not successful if the event b with maximum P(T|ak,b) is simply selected. It can be said that there is a disadvantage in this point. This is because a choice unfortunate only in the beginning is totally disadvantageous and has no chance of recovery.
Therefore, in order to solve such disadvantages, an expected value of an error is calculated by using the number of trials in the past and a conditional probability at the present point and the behavior determination is performed optimistically because of the expected value, i.e., a conditional probability is increased by the expected value and used for the behavior determination. In the example explained above, this is close to the idea that a rate of success obtained when the former event is successful in one trial is 6/11=0.545 and, since this is higher than a rate of success of 502/1001=0.501 obtained when the latter event is successful in one trial, the former event is selected. As a result, since the number of trials of the former event increases and the estimation error decreases, accuracy of the behavior determination is improved.
Processing concerning causality perception performed by the information processing apparatus shown in
In step S201, the causality-learning processing unit 201 acquires HMMs of plural modals and performs causality learning. The causality-learning processing unit 201 outputs data obtained by performing the causality learning to the causality-estimation processing unit 202.
In step S202, the causality-estimation processing unit 202 performs causality estimation using the data supplied from the causality-learning processing unit 201. The causality-estimation processing unit 202 causes the causality-candidate list storing unit 203 to store a causality candidate list representing causality obtained by performing the causality estimation.
In step S203, the causality-candidate-list arrangement processing unit 204 arranges the causality candidate list stored in the causality-candidate-list storing unit 203 and finishes the processing.
The respective kinds of processing are explained below in order.
In the following explanation, it is assumed that M modals in total are present as modals to be set as targets and ith (i=1, 2, . . . , M) modal has nm states. A state j (j=1, 2, . . . , nm) of the ith modal is represented as sij as appropriate. For example, S25 represents that a second modal is in a state 5. A state of the entire system at time t is represented by M-dimensional state vector St=(S1j1,S2j2, . . . , SMjM).
When it is evident which elements of the respective dimensions indicate which state numbers of the modals, to simplify the explanation, a state of the system is represented by a state vector having the state numbers as elements. For example, when states of the modals 1, 2, and 3 are 5, 7, and 11, respectively, a state vector of the entire system including the modals 1, 2, and 3 is represented by S=(S15,S27,S311)=(5,7, 11). For example, when attention is directed to the modals 2 and 3, the state vector is represented by S(2,3)=(S27,S311)=(7, 11).
In the example shown in
First, causality learning performed by the causality-learning processing unit 201 is explained.
At time t=0, 0 is set as values of all counters and the counters are initialized. In the causality learning, two counters, i.e., an event occurrence counter and a transition occurrence counter are used. In the following explanation, t is equal to or larger than 1 (t≧1).
The values of all the counters are attenuated according to an attenuation ratio γ decided in advance such as γ=0.999 every time predetermined time elapses.
A state St of the entire system at time t and a state st−1, at the immediately preceding time are compared. Modals, states of which change, are listed.
Processing explained below is performed with attention directed to values of L in L=1, 2, . . . , min(M−1,MaxCombi). MaxCombi is a parameter for specifying complication of combinations of modals to be taken into account. An arbitrary natural number can be set as MaxCombi min(M−1,MaxCombi) represents a smaller one of values of M−1 and MaxCombi.
Arbitrary one combination among MCL+1 combinations of modals at the time when L+1 modals are selected out of M modals is represented by cM(L+1;). State vectors representing states of respective modals of the arbitrary one combination at time t−1 are represented by ScM(L+1;)t−1.
Concerning cM(L+1;) of the respective combinations, the event occurrence counter corresponding to ScM(L+1;)t−1 is counted up by 1. The event occurrence counter is a counter for counting the number of times of occurrence of an event represented by a state vector corresponding thereto.
When modals, states of which change, are represented as modals “i”, processing explained below is performed with attention directed to the respective modals “i”.
Arbitrary one combination among M−1CL combinations of modals at the time when L modals are selected out of M−1 modals other than the modals “i” is represented by cM(L;i). State vectors representing states of respective modals of the arbitrary one combination at time t−1 are represented by ScM(L;i)t−1.
Concerning cM(L+1;) of the respective combinations, the event occurrence counter corresponding to (ScM(L;i)t−1|Ti), which is a pair of SCM(L;i)t−1 and state transition Tit−1=(Sik(t−1)→Sik(t)) of the modals “i”, is counted up by 1. The transition occurrence counter is a counter for counting the number of times of occurrence of an event represented by a state vector corresponding thereto at timing immediately before occurrence of state transition for calculating causality.
A specific example of the causality learning is explained.
As shown in
A state of the system changes with time as shown in
In
When a value of L is set to 1 or 2, MCL+1, combinations of modals at the time when L+1 modals are selected out of the three modals are {1, 2}, {1, 3}, {2, 3}, and {1, 2, 3} as shown on the left side. The combinations of modals {1, 2}, {1, 3}, {2, 3}, {1, 2, 3} correspond to cM(L+1;) explained above.
When attention is directed to the combinations of modals {1, 2}, if elements of the modal 3 not included in the combination of attention are represented by “−” as shown in
Similarly, the number of state vectors that could be taken is a number obtained by multiplying the number of elements with the number of states of the modals included in the combination of attention. Therefore, when attention is directed to modals {1, 3}, the number of state vectors that could be taken is six as shown in
Event occurrence counters are prepared in association with the respective state vectors. Therefore, in the case of this example, fifty event occurrence counters are prepared in total.
As shown in
When a value of L is set to 1 or 2, arbitrary one combination among M−1CL combinations of modals at the time when L modals are selected out of modals other than the modal 1 is each of {2}, {3}, and {2, 3} as shown on the left side of
When attention is directed to the combination of the modal {2}, if elements of the modal 1 excluded from the combination of the modals are represented by “*” and elements of the modal 3 not included in the combination of attention are represented by “−”, as shown in
Similarly, when attention is directed to the combination of the modal {3}, as shown in
In the case of this example, nineteen transition occurrence counters are prepared in total in association with respective state transitions of the modal 1.
As shown in
When a value of L is set to 1 or 2, arbitrary one combination among M−1CL combinations of modals at the time when L modals are selected out of modals other than the modal 2 is each of {1}, {3}, and {1, 3} as shown on the left side of
When attention is directed to the combination of the modal {1}, if elements of the modal 2 excluded from the combination of the modals are represented by “*” and elements of the modal 3 not included in the combination of attention are represented by “−”, as shown in
Similarly, when attention is directed to the combination of the modal {3}, as shown in
In the case of this example, eleven transition occurrence counters are prepared in total in association with respective state transitions of the modal 2.
As shown in
When a value of L is set to 1 or 2, arbitrary one combination among M−1CL combinations of modals at the time when L modals are selected out of modals other than the modal 3 is each of {1}, {2}, and {1, 2} as shown on the left side of
When attention is directed to the combination of the modal {1}, if elements of the modal 3 excluded from the combination of the modals are represented by “*” and elements of the modal 2 not included in the combination of attention are represented by “−”, as shown in
Similarly, when attention is directed to the combination of the modal {2}, as shown in
In the case of this example, fourteen transition occurrence counters are prepared in total in association with respective state transitions of the modal 3.
In this way, the transition occurrence counter is prepared to associate respective state transitions of a certain modal and all combinations of states of the other modals.
In a state in which such event occurrence counters and transition occurrence counters are prepared, when the time t reaches 1 and the state of the system transitions from [1 1 1] to [1 2 1] as shown in
The event occurrence counters are counted up.
Attention is directed to each of {1, 2}, {1, 3}, {2, 3}, and {1, 2, 3}, which is arbitrary one combination, among MCL+1 combinations of modals. Even occurrence counters corresponding to state vectors representing states at the immediately preceding time of modals included in the combination of attention are counted up by 1. The state vectors representing the states at the immediately preceding time of the modals included in the combination of attention correspond to ScM(L+1;)t−1 explained above.
When attention is directed to the combination of the modals {1, 2}, states of the modals 1 and 2 at the immediately preceding time t=0 are 1. Therefore, the event occurrence counter corresponding to [1 1 −] among the eight state vectors shown in
When attention is directed to the combination of the modals {1, 3}, states of the modals 1 and 3 at the immediately preceding time t=0 are 1. Therefore, the event occurrence counter corresponding to [1 − 1] among the six state vectors shown in
When attention is directed to the combination of the modals {2, 3}, states of the modals 2 and 3 at the immediately preceding time t=0 are 1. Therefore, the event occurrence counter corresponding to [−1 1] among the twelve state vectors shown in
When attention is directed to the combination of the modals {1, 2, 3}, states of the modals 1, 2, and 3 at the immediately preceding time t=0 are 1. Therefore, the event occurrence counter corresponding to [1 1 1] among the twenty-four state vectors shown in
In this way, when the state of the system transitions from [1 1 1] to [1 2 1], as shown in
Further, the transition occurrence counters are counted up.
First, attention is directed to each of {1}, {3}, and {1, 3}, which is arbitrary one combination, among M−1CL combinations of modals at the time when L modals are selected out of the modals other than the modal 2, a state of which changes. Pairs of state vectors representing states at the immediately preceding time of the modals included in the combination of attention and state transition (1→2) of the modal 2 are calculated.
The calculated pairs represent state vectors representing states at the immediately preceding time of the modals included in the combination of attention associated with the state transition (1→2) of the modal 2. As explained with reference to
The transition occurrence counters (
When attention is directed to the combination of the modal {1}, a state of the modal 1 at the immediately preceding time t=0 is 1. Therefore, the transition occurrence counter corresponding to [1 * −] of the two transition occurrence counters associated with the state transition (1→2) of the modal 2 shown in
When attention is directed to the combination of the modal {3}, a state of the modal 3 at the immediately preceding time t=0 is 1. Therefore, the transition occurrence counter corresponding to [− * 1] among the three transition occurrence counters associated with the state transition (1→2) of the modal 2 shown in
When attention is directed to the combination of the modals {1, 3}, states of the modals 1 and 3 at the immediately preceding time t=0 are 1. Therefore, the transition occurrence counter corresponding to [1 * 1] of the six transition occurrence counter associated with the state transition (1→2) of the modal 2 shown in
In this way, when the state of the system transitions from [1 1 1] to [1 2 1], as shown in
Similarly, when the time t reaches 2, a state at t=2 and a state at the immediately preceding time t=1 are compared. A modal, a state of which changes, is determined as not present. As shown in
The event occurrence counters are counted up.
Attention is directed to each of {1, 2}, {1, 3}, {2, 3}, and {1, 2, 3}, which is arbitrary one combination, among MCL+1 combinations of modals at the time when L+1 modals are selected out of three modals. The event occurrence counters corresponding to state vectors representing states at the immediately preceding time of the modals included in the combination of attention are counted up by 1.
When attention is directed to the combination of the modals {1, 2}, a state of the modal 1 at the immediately preceding time t=1 is 1 and a state of the modal 2 at t=1 is 2. Therefore, the event occurrence counter corresponding to [1 2 −] among the eight state vectors shown in
When attention is directed to the combination of the modals {1, 3}, states of the modals 1 and 3 at the immediately preceding time t=1 are 1. Therefore, the event occurrence counter corresponding to [1 − 1] among the six state vectors shown in
When attention is directed to the combination of the modals {2, 3}, a state of the modal 2 at the immediately preceding time t=1 is 2 and a state of the modal 3 at t=1 is 1. Therefore, the event occurrence counter corresponding to [− 2 1] among the twelve state vectors shown in
When attention is directed to the combination of the modals {1, 2, 3}, a state of the modal 1 at the immediately preceding time t=1 is 1, a state of the modal 2 at t=1 is 2, and a state of the modal 3 at t=1 is 3. Therefore, the event occurrence counter corresponding to [1 2 1] among the twenty-four state vectors shown in
In this way, when the state of the system stays in [1 2 1], as shown in
When the time t reaches 3 and the state of the system transitions from [1 2 1] to [2 2 1] as shown in
The event occurrence counters are counted up.
Attention is directed to each of {1, 2}, {1, 3}, {2, 3}, and {1, 2, 3}, which is arbitrary one combination, among MCL+1 combinations of modals at the time when L+1 modals are selected out of the three modals. Even occurrence counters corresponding to state vectors representing states at the immediately preceding time of modals included in the combination of attention are counted up by 1.
When attention is directed to the combination of the modals {1, 2}, a state of the modal 1 at the immediately preceding time t=2 are 1 and a state of the modal 2 at t=2 is 2. Therefore, the event occurrence counter corresponding to [1 2 −] among the eight state vectors shown in
When attention is directed to the combination of the modals {1, 3}, states of the modals 1 and 3 at the immediately preceding time t=2 are 1. Therefore, the event occurrence counter corresponding to [1 − 1] among the six state vectors shown in
When attention is directed to the combination of the modals {2, 3}, a state of the modal 2 at the immediately preceding time t=2 is 2 and a state of the modal 3 at t=2 is 1. Therefore, the event occurrence counter corresponding to [− 2 1] among the twelve state vectors shown in
When attention is directed to the combination of the modals {1, 2, 3}, a state of the modal 1 at the immediately preceding time t=2 is 1, a state of the modal 2 at t=2 is 2, and a state of the modal 3 at t=2 is 1. Therefore, the event occurrence counter corresponding to [1 2 1] among the twenty-four state vectors shown in
In this way, when the state of the system transitions from [1 2 1] to [2 2 1], as shown in
Further, the transition occurrence counters are counted up.
First, attention is directed to each of {2}, {3}, and {2, 3}, which is arbitrary one combination, among M−1CL combinations of modals at the time when L modals are selected out of the modals other than the modal 1, a state of which changes. Pairs of state vectors representing states at the immediately preceding time of the modals included in the combination of attention and state transition (1→2) of the modal 1 are calculated.
The transition occurrence counters (
When attention is directed to the combination of the modal {2}, a state of the modal 2 at the immediately preceding time t=2 is 2. Therefore, the transition occurrence counter corresponding to [* 2 −] among the four transition occurrence counters associated with the state transition (1→2) of the modal 1 shown in
When attention is directed to the combination of the modal {3}, a state of the modal 3 at the immediately preceding time t=2 is 1. Therefore, the transition occurrence counter corresponding to [*− 1] among the three transition occurrence counters associated with the state transition (1→2) of the modal 1 shown in
When attention is directed to the combination of the modals {2, 3}, a state of the modal 2 at the immediately preceding time t=2 is 2 and a state of the modal 3 at t=2 is 1. Therefore, the transition occurrence counter corresponding to [*2 1] among the twelve transition occurrence counters associated with the state transition (1→2) of the modal 1 shown in
In this way, when the state of the system transitions from [1 2 1] to [2 2 1], as shown in
When the time t reaches 4 and the state of the system transition from [2 2 1] to [2 4 3] as shown in
The event occurrence counters are counted up.
Attention is directed to each of {1, 2}, {1, 3}, {2, 3}, and {1, 2, 3}, which is arbitrary one combination, among MCL+1 combinations of modals at the time when L+1 modals are selected out of the three modals. Even occurrence counters corresponding to state vectors representing states at the immediately preceding time of modals included in the combination of attention are counted up by 1.
When attention is directed to the combination of the modals {1, 2}, states of the modals 1 and 2 at the immediately preceding time t=3 are 2. Therefore, the event occurrence counter corresponding to [2 2 −] among the eight state vectors shown in
When attention is directed to the combination of the modals {1, 3}, a state of the modal 1 at the immediately preceding time t=3 is 2 and a state of the modal 3 at t=3 is 1. Therefore, the event occurrence counter corresponding to [2 − 1] among the six state vectors shown in
When attention is directed to the combination of the modals {2, 3}, a state of the modal 2 at the immediately preceding time t=3 is 2 and a state of the modal 3 at t=3 is 1. Therefore, the event occurrence counter corresponding to [−2 1] among the twelve state vectors shown in
When attention is directed to the combination of the modals {1, 2, 3}, a state of the modal 1 at the immediately preceding time t=3 is 2, a state of the modal 2 at t=3 is 2, and a state of the modal 3 at t=3 is 1. Therefore, the event occurrence counter corresponding to [2 2 1] among the twenty-four state vectors shown in
In this way, when the state of the system transitions from [2 2 1] to [2 4 3], as shown in
Further, the transition occurrence counters are counted up. When two modals are listed, the same processing is repeated for the respective modals.
First, attention is directed to each of {1}, {3}, and {1, 3}, which is arbitrary one combination, among M−1CL combinations of modals at the time when L modals are selected out of the modals other than the modal 2, a state of which changes. Pairs of state vectors representing states at the immediately preceding time of the modals included in the combination of attention and state transition (2→4) of the modal 2 are calculated.
The transition occurrence counters (
When attention is directed to the combination of the modal {1}, a state of the modal 1 at the immediately preceding time t=3 is 2. Therefore, the transition occurrence counter corresponding to [2 * −] of the two transition occurrence counters associated with the state transition (2→4) of the modal 2 shown in
When attention is directed to the combination of the modal {3}, a state of the modal 3 at the immediately preceding time t=3 is 1. Therefore, the transition occurrence counter corresponding to [− * 1] among the three transition occurrence counters associated with the state transition (2→4) of the modal 2 shown in
When attention is directed to the combination of the modals {1, 3}, a state of the modal 1 at the immediately preceding time t=3 is 2 and a state of the modal 3 at t=3 is 1. Therefore, the transition occurrence counter corresponding to [2 * 1] of the six transition occurrence counters associated with the state transition (2→4) of the modal 2 shown in
Subsequently, attention is directed to each of {1}, {2}, and {1, 2}, which is arbitrary one combination, among M−1CL combinations of modals at the time when L modals are selected out of the modals other than the modal 3, a state of which changes. Pairs of state vectors representing states at the immediately preceding time of the modals included in the combination of attention and state transition (1=3) of the modal 3 are calculated.
The transition occurrence counters (
When attention is directed to the combination of the modal {1}, a state of the modal 1 at the immediately preceding time t=3 is 2. Therefore, the transition occurrence counter corresponding to [2 * −] of the two transition occurrence counters associated with the state transition (1→3) of the modal 3 shown in
When attention is directed to the combination of the modal {2}, a state of the modal 2 at the immediately preceding time t=3 is 2. Therefore, the transition occurrence counter corresponding to [*2 −] among the four transition occurrence counters associated with the state transition (1→3) of the modal 3 shown in
When attention is directed to the combination of the modals {1, 2}, states of the modals 1 and 2 at the immediately preceding time t=3 are 2. Therefore, the transition occurrence counter corresponding to [2 2 −] among the eight transition occurrence counters associated with the state transition (1→3) of the modal 3 shown in
In this way, when the state of the system transitions from [2 2 1] to [2 4 3], as shown on the left side of in
When the time t reaches 5 and the state of the system transition from [2 4 3] to [3 4 3] as shown in
The event occurrence counters are counted up.
Attention is directed to each of {1, 2}, {1, 3}, {2, 3}, and {1, 2, 3}, which is arbitrary one combination, among MCL+1 combinations of modals at the time when L+1 modals are selected out of the three modals. Even occurrence counters corresponding to state vectors representing states at the immediately preceding time of modals included in the combination of attention are counted up by 1.
When attention is directed to the combination of the modals {1, 2}, a state of the modal 1 at the immediately preceding time t=4 is 2 and a state of the modal 2 at t=4 is 4. Therefore, the event occurrence counter corresponding to [2 4 −] among the eight state vectors shown in
When attention is directed to the combination of the modals {1, 3}, a state of the modal 1 at the immediately preceding time t=4 is 2 and a state of the modal 3 at t=4 is 3. Therefore, the event occurrence counter corresponding to [2 − 3] among the six state vectors shown in
When attention is directed to the combination of the modals {2, 3}, a state of the modal 2 at the immediately preceding time t=4 is 4 and a state of the modal 3 at t=4 is 3. Therefore, the event occurrence counter corresponding to [−4 3] among the twelve state vectors shown in
When attention is directed to the combination of the modals {1, 2, 3}, a state of the modal 1 at the immediately preceding time t=4 is 2, a state of the modal 2 at t=4 is 4, and a state of the modal 3 at t=4 is 3. Therefore, the event occurrence counter corresponding to [2 4 3] among the twenty-four state vectors shown in
In this way, when the state of the system transitions from [2 4 3] to [3 4 3], as shown in
Further, the transition occurrence counters are counted up.
First, attention is directed to each of {2}, {3}, and {2, 3}, which is arbitrary one combination, among M−1CL combinations of modals at the time when L modals are selected out of the modals other than the modal 1, a state of which changes. Pairs of state vectors representing states at the immediately preceding time of the modals included in the combination of attention and state transition (2→3) of the modal 1 are calculated.
The transition occurrence counters (
When attention is directed to the combination of the modal {2}, a state of the modal 2 at the immediately preceding time t=4 is 4. Therefore, the transition occurrence counter corresponding to [* 4 −] among the four transition occurrence counters associated with the state transition (2→3) of the modal 1 shown in
When attention is directed to the combination of the modal {3}, a state of the modal 3 at the immediately preceding time t=4 is 3. Therefore, the transition occurrence counter corresponding to [* − 3] among the three transition occurrence counters associated with the state transition (2→3) of the modal 1 shown in
When attention is directed to the combination of the modals {2, 3}, a state of the modal 2 at the immediately preceding time t=4 is 4 and a state of the modal 3 at t=4 is 3. Therefore, the transition occurrence counter corresponding to [* 4 3] among the twelve transition occurrence counters associated with the state transition (2→3) of the modal 1 shown in
In this way, when the state of the system transitions from [2 4 3] to [3 4 3], as shown in
The causality learning is advanced by repeating the processing explained above. Information representing values of the event occurrence counters and values of the transition occurrence counters obtained by the causality learning is supplied from the causality-learning processing unit 201 to the causality-estimation processing unit 202.
The causality estimation performed by the causality-estimation processing unit 202 is explained.
When causality of state transition Ti=(Sik→Sik′) in the modal “i” is estimated, processing explained below is performed with attention directed to values of L in L=1, 2, . . . , min(M−1,MaxCombi).
Arbitrary one combination among M−1CL combinations of modals at the time when L modals are selected out of M−1 modals other than the modal “i” is represented as cM(L;i). A state vector pattern corresponding to cM(L;i) is represented as ScM(L;i).
When respective state vectors of the state vector pattern ScM(L;i) is represented as ScM(L;i)j, a conditional probability P(Ti|ScM(L;i)j) of the state transition Ti is calculated for the respective state vectors ScM(L;i)j.
Specifically, Specifically, Sik as a state of a transition source of the modal “i” is added as an element of the state vector ScM(L;i)j, whereby a state vector (Sik,ScM(L;i)j) is generated and a value NS of an event occurrence counter corresponding to the generated state vector (Sik,ScM(L;i)j) is acquired.
When the value NS of the event occurrence counter corresponding to the state vector (Sik,ScM(L;i)j) is 0, a conditional probability P(Ti|ScM(L;i)j) of a state transition Ti corresponding to the state vector ScM(L;i)j is set as σ0. σ0 is a fixed value equal to or larger than 0 and equal to or smaller than 1 that gives a minimum probability.
On the other hand, when the value NS of the event occurrence counter corresponding to the state vector (Sik,ScM(L;i)j) is equal to or larger than 1, a value NT of a transition occurrence counter corresponding to (Ti|ScM(L;i)j) is acquired. (Ti|ScM(L;i)j) represents a state vector ScM(L;i)j associated with the state transition Ti.
A provisional probability value p0=NT/NS is calculated by dividing the value NT of the transition occurrence counter by the value NS of the event occurrence counter.
When a value σ is represented by the following Formula (2) by using the provisional probability value p0 and the value NS of the event occurrence counter, the conditional probability P(Ti|ScM(L;i)j) of the state transition Ti with respect to the state vector ScM(L;l)j is represented by Formula (3) below. In Formula (3), min represents that a smaller value of 1 and p0+σ is set as a conditional probability P.
Adding the value a to the provisional probability value p0 represents setting a value obtained by optimistically considering an estimation error of a probability based on experiences as a final conditional probability P.
Besides, it is also possible to correct the conditional probability P to 0.5 according to Formula (4) below. Correction for setting p+ασ as the conditional probability P is also possible.
p=max(0.5,p−σ) . . . p>0.5
p=min(0.5,p+σ) . . . otherwise (4)
An event of the state transition as the target of estimation of causality is an event having two values, i.e., whether the event occurs or not. Therefore, the event can be modeled by the Beroulli trial of the occurrence probability p. For example, the occurrence probability p=N(X,T)/N(X) obtained from N(X) times of sampling is considered to have an estimation error of an expected value √p(1−p)/N(X). Therefore, a probability increased by the value σ calculated in the same manner is calculated as the final conditional probability P. However, since an estimation error is 0 at P=0 or P=1, in practice, the value σ as an estimation error is calculated by using an appropriate parameter σ0.
When the conditional probability P(Ti|ScM(L;i)j) equal to or larger than a threshold is calculated, the state vector ScM(L;i)j as the target is registered in the causality candidate list together with the conditional probability P(Ti|ScM(L;i)j). The causality candidate list is a list of the state vector ScM(L;i)j having causality with the state transitions Ti. The state vector ScM(L;i)j is associated for each of state transitions Ti in order from one having a highest conditional probability P(Ti|ScM(L;i)j). A state represented by the state vector ScM(L;i)j is a causality candidate of the state transition T′.
A specific example of the causality estimation is explained.
In the following explanation, a state vector having causality with the state transition (1→2) of the modal 2 is estimated.
The state vector having the causality with the state transition (1→2) of the modal 2 is a state vector representing states of both the modal 1 and the modal 3 or a state vector representing one of the modal 1 and the modal 3. Therefore, when a value of L is set to 1 or 2 and arbitrary one combination of M−1CL combinations of modals at the time when L modals are selected out of the modals other than the modal 2 are considered, the combinations are {1}, {3}, and {1, 3}. The respective combinations of the modals correspond to cM (L;i).
As shown in
For example, [1 * −] or [2 * −] of patterns of two state vectors [1 * −] and [2 * −] corresponding to {1} corresponds to the state vector ScM(L;i)j. The following processing is performed with respective eleven state vectors ScM(L;i)j shown in
For example, in the following explanation, a conditional probability of the state transition (1→2) of the modal 2 with respect to the state vectors [1 * −] and [1 * 1] shown in
When attention is directed to [1 * −], as shown on the left side of
The value NS (
After the value NS of the even occurrence counter corresponding to [1 1 −] is acquired, the value NT (the left side of
A conditional probability of the state transition (1→2) of the modal 2 with respect to [1 * −] is calculated on the basis of the value NS of the event occurrence counter and the value NT of the transition occurrence counter. When the value NS of the event occurrence counter is 0, σ0 is calculated as the conditional probability. When the value NS of the event occurrence counter is equal to or larger than 1, the conditional probability is calculated according to Formula (3).
Similarly, when attention is directed to [1 * 1], as shown on the right side of
The value NS of an event occurrence counter corresponding to [1 1 1] and the value NT of a transition occurrence counter corresponding to [1 * 1] (the right side of
A conditional probability of the state transition (1→2) of the modal 2 with respect to [1 * 1] is calculated on the basis of the value NS of the event occurrence counter and the value NT of the transition occurrence counter.
The conditional probability calculated as explained above is registered in the causality candidate list in association with the state vectors as appropriate and stored in the causality-candidate-list storing unit 203.
Arrangement of the causality candidate list performed by the causality-candidate-list arrangement processing unit 204 is explained.
The arrangement of the causality candidate list is merging of the state vectors registered in the causality candidate list. The arrangement of the causality candidate list corresponds to controlling the event b by, for example, changing granularity to keep N(T,ak,b) at a value in an appropriate range. The arrangement of the causality candidate list is performed at predetermined timing.
A state vector ScM(L;)k defined as a pair of specific states in L modals is discussed below.
Possibility of merging is determined between the state vector ScM(L;)k and a state vector (ScM(L;)k,Sij) obtained by adding a specific state Sij of the modal “i”, which is one modal not included in the L modals, to the state vector ScM(L;)k. The state vector ScM(L;)k and the state vector (ScM(L;)k,Sij) are state vectors registered in the causality candidate list in association with a conditional probability of the same state transition.
The state vector (ScM(L;)k,Sij) is a state vector obtained by adding Sij to the state vector ScM(L;)k. Therefore, it can be said that, conceptually, the state vector ScM(L;)k is a state vector higher in order than the state vector (ScM(L;)k,Sij). The determination of possibility of merging is determination concerning whether the low order state vector is included in the high order state vector and considered the same.
A conditional probability P of target state transition with respect to the state vector ScM(L;)k is represented by Formula (5). A conditional probability P′ of the same state transition with respect to the state vector (ScM(L;)k,Sij) is represented by Formula (6).
In this case, determination of possibility of merging of both the state vectors is performed according to Formula (7). α is an appropriate merging coefficient.
|p0−p′0|>α(σ+σ′) . . . Merging is difficult Otherwise . . . Merging is possible (7)
Such determination of possibility of merging is performed between the state vector ScM(L;)k and all ni state vectors (ScM(L;)k,Sij) of the modal “i” obtained by adding the specific state Sij to the state vector (ScM(L;)k.
If at least one of the ni state vectors (ScM(L;)k,Sij) can be determined as difficult to be merged, the state vector ScM(L;)k is deleted from the causality candidate list. State vectors conceptually low in order remain in the causality candidate list.
On the other hand, when a state vector that can be determined as difficult to be merged is not present in the ni state vectors (ScM(L;)k,Sij), the ni state vectors (ScM(L;)k,Sij) and all state vectors including the state vectors (ScM(L;)k,Sij) in a part thereof (higher-order state vectors) are deleted. Low-order state vectors are collectively treated by conceptually high-order state vectors.
A specific example of the arrangement of the causality candidate list is explained.
A state vector [1 * −] shown in
In this case, possibility of merging is determined between [1 * −] and state vectors [1 * 1], [1 * 2], and [1 * 3] shown in
A conditional probability of certain state transition of the modal 2 with respect to [1 * −] is calculated according to Formula (5). A conditional probability of the same state transition of the modal 2 with respect to each of [1 * 1], [1 * 2], and [1 * 3] is calculated according to Formula (6).
It is determined according to Formula (7) whether [1 * −] and [1 * 1], [1 * 2], and [1 * 3] can be merged.
When it is determined that all of [1 * 1], [1 * 2], and [1 * 3] can be merged with [1 * −], [1 * 1], [1 * 2], and [1 * 3] are deleted from the causality list and only [1 * −] is left.
On the other hand, it is determined that it is difficult to merge at least on of [1 * 1], [1 * 2], and [1 * 3], [1 * −] is deleted from the causality list and [1 * 1], [1 * 2], and [1 * 3] are left.
An idea of the arrangement of the causality list is explained with the Simpson's paradox as an example.
The Simpson's paradox is explained below.
There were eighty patients of a certain disease. When a doctor treated a half of the patient (forty of the eighty patients), twenty of the forty patients were cured. When the doctor did not treat another half of the patients (remaining forty of the eighty patients), sixteen of the forty patients were cured. A curing ratio 50 attained when the patients were treated is higher than a natural curing ratio 40 attained when no patients were treated. Therefore, the treatment is considered to be effective.
Among the eighty patients, there were forty males and forty females. Since the number of treated male patients and the number of treated female patients were different, a result is analyzed according to sex.
Thirty males were treated and eighteen of the thirty males were cured. The remaining ten males were not treated. Seven of the ten males were naturally cured. Concerning only the males, a curing ratio 60 attained when the males were treated is lower than a natural curing ratio 70 attained when the males were not treated. It would be considered better not to treat the males.
Only ten females participated in the treatment. Only two of the ten females were cured. Nine of the remaining thirty females were naturally cured. In the case of the females, a curing ratio 20 attained when the females were treated is lower than a natural curing ratio 30 attained when the females were not treated. As in the case of the males, it would be considered better not to treat the females.
It is a question whether this treatment has an effect or rather has a side effect.
As modals that should be taken into account concerning transition, i.e., curing=(disease→health), there are “treatment” and “male and female”. A relation among these modals is summarized as follows:
P(curing|treat) = 0.5
σ = 0.079
P(curing|not treat) = 0.4
σ = 0.078
P(curing|treat, male) = 0.6
σ = 0.089
P(curing|treat, female) = 0.2
σ = 0.13
P(curing|not treat, male) = 0.7
σ = 0.14
P(curing|not treat, female) = 0.3
σ = 0.084
P(curing|male) = 0.63
σ = 0.077
P(curing|female) = 0.28
σ = 0.071
A question is what causality that should truly be taken into account is. The method of arranging causality (arrangement of the state vectors of the causality candidate list) is applied. In the following explanation, a merging coefficient α is set to 1.
Causality concerning P(curing|treat) is arranged as follows:
|P(curing|treat)−P(curing|treat,male)|=0.1<(0.079+0.089)=0.17
|P(curing|treat)−P(curing|treat,female)|=0.3>(0.079+0.13)=0.21
Therefore, it is difficult to merge P(curing|treat, male) and P(curing|treat, female). P(curing|treat) is deleted.
Similarly, concerning P(curing|not treat), causality is as follows:
|P(curing|not treat)−P(curing|not treat,male)|=0.3>(0.078+0.14)=0.22
|P(curing|not treat)−P(curing|not treat,female)|=0.1<(0.078+0.084)=0.16
Therefore, it is difficult to merge P(curing|not treat, male) and P(curing|not treat, female). P(curing|treat) is deleted.
On the other hand, when P(curing|male) is checked, both P(curing|treat, male) and P(curing|not treat, male) can be merged. Therefore, both P(curing|treat, male) and P(curing|not treat, male) are deleted and only P(curing|male) is left.
|P(curing|male)−P(curing|treat,male)|=0.03<(0.077+0.089)=0.17
|P(curing|male)−P(curing|not treat,male)|=0.07<(0.077+0.14)=0.22
Similarly, when P(curing|female) is checked, P(curing|treat, female) and P(curing|not treat, female) can be merged. Therefore, P(curing|treat, female) and P(curing|not treat, female) are deleted and only P(curing|female) is left.
|P(curing|female)−P(curing|treat,female)|=0.08<(0.071+0.13)=0.20
|P(curing|female)−P(curing|not treat,female)|=0.02<(0.071+0.084)=0.15
Consequently, causality effective at present is only “difference between male and female →presence or absence of curing”. It is difficult to conclude whether the curing is effective or has a side effect. In other words, a causality analysis of L=2 that simultaneously takes into account both “difference between male and female” and “presence or absence of curing” is unnecessary. It is considered sufficient to analyze causality within a range of L=1.
Thereafter, experiments were repeated and the number of patients was increased to one hundred times as large as the number of the patients in the experiment explained above. However, as explained below, a paradoxical situation did not change.
There were eight thousand patients of a certain disease. When a doctor treated a half of the patient (four thousand of the eight thousand patients), two thousand of the four thousand patients were cured. When the doctor did not treat another half of the patients (remaining four thousand of the eight thousand patients), one thousand six hundred of the four thousand patients were cured.
Among the eight thousand patients, there were four thousand males and four thousand females. The number of treated male patients and the number of treated female patients were different.
Three thousand males were treated and one thousand and eight hundred of the three thousand males were cured. The remaining one thousand males were not treated. Seven hundred of the one thousand males were naturally cured.
Only one thousand females participated in the treatment. Only two hundred of the one thousand females were cured. Nine hundred of the remaining three thousand females were naturally cured.
Taking into account such a situation, it is examined what is derived from the above experiment result.
Under such a situation, probabilities are the same as those explained above. Since the number of samples N is increased to one hundred times as large as the number in the experiment explained above, only the expected value a of a magnitude of an error is reduced to 1/10 of the in the experiment explained above.
Causality candidates concerning P(curing|treat) are arranged as explained below according to the same calculation. Therefore, it is difficult to merge P(curing|treat, male) and P(curing|treat, female). P(curing|treat) is deleted.
|P(curing|treat)−P(curing|treat,male)|=0.1<(0.0079+0.0089)=0.017
|P(curing|treat)−P(curing|treat,female)|=0.3>(0.0079+0.013)=0.021
Similarly, concerning P(curing|not treat), causality is as follows:
|P(curing|not treat)−P(curing|not treat,male)|=0.3>(0.0078+0.014)=0.0022
|P(curing|not treat)−P(curing|not treat,female)|=0.1<(0.0078+0.0084)=0.016
Therefore, it is difficult to merge P(curing|not treat, male) and P(curing|not treat, female). P(curing|treat) is deleted.
When P(curing|male) is checked, it is difficult to merge P(curing|treat, male) and P(curing|not treat, male). Therefore, P(curing|male) is deleted.
|P(curing|male)−P(curing|treat,male)|=0.03<(0.0077+0.0089)=0.017
|P(curing|male)−P(curing|not treat,male)|=0.07<(0.0077+0.014)=0.022
Similarly, concerning P(curing|female), it is difficult to merge P(curing|treat, female) and P(curing|not treat, female). Therefore, P(curing|female) is deleted.
|P(curing|female)−P(curing|treat,female)|=0.08>(0.0071+0.013)=0.020
|P(curing|female)−P(curing|not treat,female)|=0.02>(0.0071+0.0084)=0.015
Consequently, in this case, the causality analysis of L=2 that simultaneously takes into account “difference between male and female” and “presence or absence of treatment”. In other words, causality that should be taken into account is summarized as follows:
P(curing|treat, male) = 0.6
σ = 0.0089
P(curing|treat, female) = 0.2
σ = 0.013
P(curing|not treat, male) = 0.7
σ = 0.014
P(curing|not treat, female) = 0.3
σ = 0.0084
If attention is directed to only one factor of “difference between male and female” and “presence or absence of treatment”, wrong determination is made. As it is evident from the above, in this case, it can be concluded that curing performance is better when the patients are not treated regardless of male or female. In other words, in this treatment, a side effect is larger than an effect.
The causality estimation of L=2 is necessary rather than the causality estimation of L=1, which directs attention only to the presence of absence of treatment, because the difference between male and female concerning specific numerical values of the curing performance is obviously significant and it is necessary to take into account the difference between male and female in addition to the presence or absence of treatment. Actually, it is this necessity of stratification that solves the paradox.
Since the merging of the state vectors is performed by the causality-candidate-list arrangement processing unit 204 according to necessity, it is possible to leave state vectors that are necessary in terms of representing causality with certain state transition.
Behavior determination processing performed by the behavior determining unit 205 by using the causality candidate list appropriately arranged as explained above and stored in the causality-candidate-list storing unit 203 is explained with reference to a flowchart of
In step S211, the behavior determining unit 205 acquires a target value. The target value is, for example, a value representing one state of a certain modal to be set as a target.
In step S212, the behavior determining unit 205 reads out the causality candidate list stored in the causality-candidate-list storing unit 203 and determines a behavior for transitioning a state of the modal to the state represented by the target value. For example, the behavior determining unit 205 determines transitions from a present state of the modal to the state of the target value and acquires a predetermined number of causality candidates of the respective transitions out of causality candidates registered in the causality candidate list in order from one with a highest conditional probability. The behavior determining unit 205 causes a robot to perform a behavior for transitioning states of other modals to a state represented by a state vector that is one causality candidate having the highest conditional probability or a conditional probability equal to or higher than a fixed level selected out of the acquired causality candidates.
When the causality estimation is appropriately performed, as shown in
As shown in
It is also possible to perform the causality estimation in determining a behavior.
As explained above, it is possible to narrow down candidates of events having causality and stabilize learning by formulating causality such that “occurrence of a certain event b” “causes state transition from an event ak to an event al (the events al and ak are exclusive to each other)” of another modal ai. Since the events al and ak are exclusive to each other, an event causing state transition ak→al is not present in the modal ai. Therefore, it is possible to exclude events in the modal ai.
Plural mechanisms of occurrence of the event al could be present. However, it is possible to ease complexity of a problem by treating the problem as a causality estimation problem of the state transition ak→al .
It is possible to cause a state transition probability representing causality having fluctuation with time to dynamically follow the causality by gently attenuating with time the number of times of simultaneously occurrence of events counted by a counter. Further, it is possible to balance the user and the search of causality taking into account possibility of fluctuation with time.
It is possible to solve tradeoff between the use and the search of causality by formulating causality between the state transition T: ak→al and the event b in a form of a conditional probability P(T|ak,b) such that behavior determination is performed by taking into account the number of times of simultaneous occurrence of the events and the expected value a of estimated error estimated from the conditional probability. In other words, it is possible to realize appropriate behavior regardless of whether the number of data samples is large or small.
In
Whereas parameter tuning is necessary in the other methods in the past, parameter tuning is unnecessary in the method of using the conditional probability with a probability increased by the expected value σ. Therefore, it can be said that the method of using the conditional probability with a probability increased by the expected value σ is practical.
The series of processing explained above can be performed by hardware or can be performed by software. When the series of processing is performed by software, a program configuring the software is installed, from a program recording medium, in a computer incorporated in dedicated hardware, a general-purpose personal computer that can execute various functions by installing various programs, or the like.
A CPU (Central Processing Unit) 211, a ROM (Read Only Memory) 212, and a RAM (Random Access Memory) 213 are connected to one another by a bus 214.
An input and output interface 215 is connected to the bus 214. An input unit 216 including a keyboard, a mouse, and a microphone, an output unit 217 including a display and a speaker, a storing unit 218 including a hard disk and a non-volatile memory, a communication unit 219 including a network interface, and a drive 220 that drives a removable medium 221 such as an optical disk or a semiconductor memory are connected to the input and output interface 215.
In the computer configured as explained above, the CPU 211 loads, for example, a program stored in the storing unit 218 onto the RAM 213 via the input and output interface 215 and the bus 214 and executes the program, whereby the series of processing is performed.
The program executed by the CPU 211 is provided by, for example, being recorded in the removable medium 221 or transmitted via a wired or wireless transmission medium such as a local area network, the Internet, or a digital broadcast and is installed in the storing unit 218.
The program executed by the computer may be a program for performing processing in time series according to the order explained in this specification or may be a program for performing processing in parallel or at necessary timing such as when the program is invoked.
Embodiments of the present invention are not limited to the embodiment explained above. Various modifications are possible without departing from the spirit of the present invention.
It should be understood by those skilled in the art that various modifications, combinations, sub-combinations, and alterations may occur depending on design requirements and other factors insofar as they are within the scope of the appended claims or the equivalents thereof.
Hidai, Kenichi, Sabe, Kohtaro, Kawamoto, Kenta, Minamino, Katsuki, Suzuki, Hirotaka
Patent | Priority | Assignee | Title |
Patent | Priority | Assignee | Title |
5995963, | Jun 27 1996 | Fujitsu Limited | Apparatus and method of multi-string matching based on sparse state transition list |
6212510, | Jan 30 1998 | HANGER SOLUTIONS, LLC | Method for minimizing entropy in hidden Markov models of physical signals |
6801656, | Nov 06 2000 | KONINKLIJKE PHILIPS ELECTRONICS, N V | Method and apparatus for determining a number of states for a hidden Markov model in a signal processing system |
7076102, | Sep 27 2001 | Koninklijke Philips Electronics N.V. | Video monitoring system employing hierarchical hidden markov model (HMM) event learning and classification |
7203635, | Jun 27 2002 | Microsoft Technology Licensing, LLC | Layered models for context awareness |
7260558, | Oct 25 2004 | PACKET INTELLIGENCE LLC | Simultaneously searching for a plurality of patterns definable by complex expressions, and efficiently generating data for such searching |
20020165717, | |||
20050004786, | |||
20050256817, | |||
20060241927, | |||
20060248026, | |||
20080300879, | |||
20090018877, | |||
JP5054068, |
Executed on | Assignor | Assignee | Conveyance | Frame | Reel | Doc |
Jan 21 2009 | SABE, KOHTARO | Sony Corporation | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 034856 | /0021 | |
Jan 21 2009 | MINAMINO, KATSUKI | Sony Corporation | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 034856 | /0021 | |
Jan 22 2009 | KAWAMOTO, KENTA | Sony Corporation | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 034856 | /0021 | |
Jan 22 2009 | SUZUKI, HIROTAKA | Sony Corporation | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 034856 | /0021 | |
Jan 22 2009 | HIDAI, KENICHI | Sony Corporation | ASSIGNMENT OF ASSIGNORS INTEREST SEE DOCUMENT FOR DETAILS | 034856 | /0021 | |
Jun 26 2013 | Sony Corporation | (assignment on the face of the patent) | / |
Date | Maintenance Fee Events |
Jan 31 2017 | ASPN: Payor Number Assigned. |
Apr 08 2020 | M1552: Payment of Maintenance Fee, 8th Year, Large Entity. |
Jun 03 2024 | REM: Maintenance Fee Reminder Mailed. |
Date | Maintenance Schedule |
Oct 25 2019 | 4 years fee payment window open |
Apr 25 2020 | 6 months grace period start (w surcharge) |
Oct 25 2020 | patent expiry (for year 4) |
Oct 25 2022 | 2 years to revive unintentionally abandoned end. (for year 4) |
Oct 25 2023 | 8 years fee payment window open |
Apr 25 2024 | 6 months grace period start (w surcharge) |
Oct 25 2024 | patent expiry (for year 8) |
Oct 25 2026 | 2 years to revive unintentionally abandoned end. (for year 8) |
Oct 25 2027 | 12 years fee payment window open |
Apr 25 2028 | 6 months grace period start (w surcharge) |
Oct 25 2028 | patent expiry (for year 12) |
Oct 25 2030 | 2 years to revive unintentionally abandoned end. (for year 12) |