Learning sequences of compatible actions among agents

2002-03-01
Action coordination in multiagent systems is a difficult task especially in dynamic environments. If the environment possesses cooperation, least communication, incompatibility and local information constraints, the task becomes even more difficult. Learning compatible action sequences to achieve a designated goal under these constraints is studied in this work. Two new multiagent learning algorithms called QACE and NoCommQACE are developed. To improve the performance of the QACE and NoCommQACE algorithms four heuristics, state iteration, means-ends analysis, decreasing reward and do-nothing, are developed. The proposed algorithms are tested on the blocks world domain and the performance results are reported.
ARTIFICIAL INTELLIGENCE REVIEW

Suggestions

Tracking of ground targets with interacting multiple model estimator
Acar, Duygu; Baykal, Buyurman; Department of Electrical and Electronics Engineering (2011)
Interacting Multiple Model (IMM) estimator is used extensively to estimate trajectories of maneuvering targets in cluttered environment. In the standard tracking methods, it is assumed that movement of target is applicable to a certain model and the target could be monitored via the usage of status predictions of that model. However, targets can make different maneuvering movements. At that time, expression of target dynamic model with only one model can be insufficient. In IMM approach, target dynamic mode...
MESSAGE SCHEDULING ALGORITHMS FOR NETWORKED CONTROL SYSTEMS WITH TIME SLOTTED COMMUNICATION PROTOCOLS
Senol, Sinan; Leblebicioğlu, Mehmet Kemal; Schmidt, Ece Gueran (2008-04-22)
Feedback control systems, of which control loops are closed over a real-time communication network, are called Networked Control Systems (NCSs). In this study the message scheduling problem for NCSs with time slotted communication protocols is tackled First, an algorithmic approach is presented to check the necessary and sufficient conditions for a given message schedule such that the real-time requirements of the system are satisfied Second, starting from an initial schedule, Simulated Annealing is used to...
Voluntary Behavior on Cortical Learning Algorithm Based Agents
Sungur, Ali Kaan; Sürer, Elif (2016-09-23)
Operating autonomous agents inside a 3D workspace is a challenging problem domain in real-time for dynamic environments since it involves online interaction with ever-changing decision constraints. This study proposes a neuroscience inspired architecture to simulate autonomous agents with interaction capabilities inside a 3D virtual world. The environment stimulates the operating agents based on their place and course of action. They are expected to form a life cycle composed of behavior chunks inside this ...
Improving reinforcement learning using distinctive clues of the environment
Demir, Alper; Polat, Faruk; Department of Computer Engineering (2019)
Effective decomposition and abstraction has been shown to improve the performance of Reinforcement Learning. An agent can use the clues from the environment to either partition the problem into sub-problems or get informed about its progress in a given task. In a fully observable environment such clues may come from subgoals while in a partially observable environment they may be provided by unique experiences. The contribution of this thesis is two fold; first improvements over automatic subgoal identifica...
Dynamic constraint satisfaction algorithm for online feature model reconfiguration
Oğuztüzün, Mehmet Halit S.; Entekhabi, Sina (2019-07-01)
Dynamically reconfigurable systems are able to respond to changes in their operational environments by reconfiguring themselves automatically. Dynamic software product lines are dynamically reconfigurable systems with an explicit variability model that guides the reconfiguration. In this work, feature models are used as the variability model. An emerging situation in the environment can lead to some relevant changes to the current configuration: some features must be activated, and some must be deactivated....
Citation Formats
F. Polat, “Learning sequences of compatible actions among agents,” ARTIFICIAL INTELLIGENCE REVIEW, pp. 21–37, 2002, Accessed: 00, 2020. [Online]. Available: https://hdl.handle.net/11511/44700.