Utilizing state-of-art NeuroES and GPGPU to optimize Mario AI
Faculty of Computing, Blekinge Institute of Technology, SE-371 79 Karlskrona, Sweden
Blekinge Institute of Technology, 2014
@article{lovgren2014utilizing,
title={Utilizing state-of-art NeuroES and GPGPU to optimize Mario AI},
author={Lovgren, Hasse},
year={2014}
}
CONTEXT: Reinforcement Learning (RL) is a time consuming effort that requires a lot of computational power as well. There are mainly two approaches to improving RL efficiency, the theoretical mathematics and algorithmic approach or the practical implementation approach. In this study, the approaches are combined in an attempt to reduce time consumption. OBJECTIVES: We investigate whether modern hardware and software, GPGPU, combined with state-of-art Evolution Strategies, CMA-Neuro-ES, can potentially increase the efficiency of solving RL problems. METHODS: In order to do this, both an implementational as well as an experimental research method is used. The implementational research mainly involves developing and setting up an experimental framework in which to measure efficiency through benchmarking. In this framework, the GPGPU/ES solution is later developed. Using this framework, experiments are conducted on a conventional sequential solution as well as our own parallel GPGPU solution. RESULTS: The results indicate that utilizing GPGPU and state-of-art ES when attempting to solve RL problems can be more efficient in terms of time consumption in comparison to a conventional and sequential CPU approach. CONCLUSIONS: We conclude that our proposed solution requires additional work and research but that it shows promise already in this initial study. As the study is focused on primarily generating benchmark performance data from the experiments, the study lacks data on RL efficiency and thus motivation for using our approach.
July 9, 2014 by hgpu