Neuro-Evolution for Multi-Agent Policy Transfer in
Didi, Sabre and Geoff Nitschke (2016) Neuro-Evolution for Multi-Agent Policy Transfer in
RoboCup Keep-Away. In Proceedings International Conference on Autonomous Agents and Multiagent Systems (AAMAS 2016), pages 1281-1282, Singapore.
An objective of transfer learning is to improve and speedup learning on target tasks after training on a different, but related source tasks. This research is a study of comparative Neuro-Evolution (NE) methods for transferring evolved multi-agent policies (behaviors) between multi-agent tasks of varying complexity. The efficacy of five variants of two NE methods are compared for multi-agent policy transfer. The NE method variants include using the original versions (search directed by a fitness function), behavioural and genotypic diversity based search to replace objective based search (fitness functions) as well as hybrid objective and diversity (behavioral and genotypic) maintenance based search approaches. The goal of testing these variants to direct policy search is to ascertain an appropriate method for boosting the task performance of transferred multi-agent behaviours. Results indicate that an indirect encoding NE method using hybridized objective based search and behavioral diversity maintenance yields significantly improved task performance for policy transfer between multi-agent tasks of increasing complexity. Comparatively, NE methods not using behavioral diversity maintenance to direct policy search performed relatively poorly in terms of efficiency (evolution times) and quality of solutions in target tasks.
|EPrint Type:||Conference Poster|
|Subjects:||I Computing Methodologies: I.2 ARTIFICIAL INTELLIGENCE|
|Deposited By:||Nitschke, Geoff|
|Deposited On:||23 November 2017|