The analysis and performance evaluation of the pheromone-Q-learning algorithm

The analysis and performance evaluation of the pheromone-Q-learning algorithm

0.00 Avg rating0 Votes
Article ID: iaor20043723
Country: United Kingdom
Volume: 21
Issue: 2
Start Page Number: 80
End Page Number: 91
Publication Date: Mar 2004
Journal: Expert Systems
Authors: ,
Keywords: artificial intelligence: expert systems
Abstract:

The paper presents the pheromone-Q-learning (Phe-Q) algorithm, a variation of Q-learning. The technique was developed to allow agents to communicate and jointly learn to solve a problem. Phe-Q learning combines the standard Q-learning technique with a synthetic pheromone that acts as a communication medium speeding up the learning process of cooperating agents. The Phe-Q update equation includes a belief factor that reflects the confidence an agent has in the pheromone (the communication medium) deposited in the environment by other agents. With the Phe-Q update equation, the speed of convergence towards an optimal solution depends on a number of parameters including the number of agents solving a problem, the amount of pheromone deposit, the diffusion into neighbouring cells and the evaporation rate. The main objective of this paper is to describe and evaluate the performance of the Phe-Q algorithm. The paper demonstrates the improved performance of cooperating Phe-Q agents over non-cooperating agents. The paper also shows how the Phe-Q learning can be improved by optimizing all the parameters that control the use of the synthetic pheromone.

Reviews

Required fields are marked *. Your email address will not be published.