Article ID: | iaor20043723 |
Country: | United Kingdom |
Volume: | 21 |
Issue: | 2 |
Start Page Number: | 80 |
End Page Number: | 91 |
Publication Date: | Mar 2004 |
Journal: | Expert Systems |
Authors: | Monekosso N., Remagnino P. |
Keywords: | artificial intelligence: expert systems |
The paper presents the pheromone-Q-learning (Phe-Q) algorithm, a variation of Q-learning. The technique was developed to allow agents to communicate and jointly learn to solve a problem. Phe-Q learning combines the standard Q-learning technique with a synthetic pheromone that acts as a communication medium speeding up the learning process of cooperating agents. The Phe-Q update equation includes a belief factor that reflects the confidence an agent has in the pheromone (the communication medium) deposited in the environment by other agents. With the Phe-Q update equation, the speed of convergence towards an optimal solution depends on a number of parameters including the number of agents solving a problem, the amount of pheromone deposit, the diffusion into neighbouring cells and the evaporation rate. The main objective of this paper is to describe and evaluate the performance of the Phe-Q algorithm. The paper demonstrates the improved performance of cooperating Phe-Q agents over non-cooperating agents. The paper also shows how the Phe-Q learning can be improved by optimizing all the parameters that control the use of the synthetic pheromone.