For multi-agent reinforcement learning in Markov games, knowledge extraction and sharing are key research problems. State list extracting means to calculate the optimal shared state path from state trajectories with c...For multi-agent reinforcement learning in Markov games, knowledge extraction and sharing are key research problems. State list extracting means to calculate the optimal shared state path from state trajectories with cycles. A state list extracting algorithm checks cyclic state lists of a current state in the state trajectory, condensing the optimal action set of the current state. By reinforcing the optimal action selected, the action policy of cyclic states is optimized gradually. The state list extracting is repeatedly learned and used as the experience knowledge which is shared by teams. Agents speed up the rate of convergence by experience sharing. Competition games of preys and predators are used for the experiments. The results of experiments prove that the proposed algorithms overcome the lack of experience in the initial stage, speed up learning and improve the performance.展开更多
Equipment systems-of-systems (SOS) effectiveness evaluation can provide important reference for construction and optimization of the equipment SoS. After discussing the basic theory and methods of parallel experimen...Equipment systems-of-systems (SOS) effectiveness evaluation can provide important reference for construction and optimization of the equipment SoS. After discussing the basic theory and methods of parallel experiments, we depict an SoS effectiveness analysis and evaluation method using parallel expe- riments theory in detail. A case study is carried out which takes the missile defense system as an example. An artificial system of the missile defense system is constructed with the multi-agent modeling method. Then, single factor, multiple factors and defense position deployment computational experiments are carried out and evaluated with the statistical analysis method. Experiment re- sults show that the altitude of the secondary interception missile is not the key factor which affects SoS effectiveness and putting the defense position ahead will increase defense effectiveness. The case study demonstrates the feasibility of the proposed method.展开更多
基金supported by the National Natural Science Foundation of China (61070143 61173088)
文摘For multi-agent reinforcement learning in Markov games, knowledge extraction and sharing are key research problems. State list extracting means to calculate the optimal shared state path from state trajectories with cycles. A state list extracting algorithm checks cyclic state lists of a current state in the state trajectory, condensing the optimal action set of the current state. By reinforcing the optimal action selected, the action policy of cyclic states is optimized gradually. The state list extracting is repeatedly learned and used as the experience knowledge which is shared by teams. Agents speed up the rate of convergence by experience sharing. Competition games of preys and predators are used for the experiments. The results of experiments prove that the proposed algorithms overcome the lack of experience in the initial stage, speed up learning and improve the performance.
文摘Equipment systems-of-systems (SOS) effectiveness evaluation can provide important reference for construction and optimization of the equipment SoS. After discussing the basic theory and methods of parallel experiments, we depict an SoS effectiveness analysis and evaluation method using parallel expe- riments theory in detail. A case study is carried out which takes the missile defense system as an example. An artificial system of the missile defense system is constructed with the multi-agent modeling method. Then, single factor, multiple factors and defense position deployment computational experiments are carried out and evaluated with the statistical analysis method. Experiment re- sults show that the altitude of the secondary interception missile is not the key factor which affects SoS effectiveness and putting the defense position ahead will increase defense effectiveness. The case study demonstrates the feasibility of the proposed method.