Javascript must be enabled for the correct page display

Actor-Critic Reinforcement Learning with Neural Networks in Continuous Games

Leuenberger, G. (2017) Actor-Critic Reinforcement Learning with Neural Networks in Continuous Games. Bachelor's Thesis, Artificial Intelligence.

[img]
Preview
Text
AI_BA_2017_LEUENBERGER.pdf - Published Version

Download (526kB) | Preview
[img] Text
Toestemming.pdf - Other
Restricted to Backend only

Download (80kB)

Abstract

Reinforcement learning agents with artificial neural networks have previously been shown to acquire human level dexterity in discrete video game environments where only the current state of the game and a reward are given at each time step. A harder problem than discrete environments is posed by continuous environments where the states, observations, and actions are continuous, which is what this thesis focuses on. The algorithm called the Continuous Actor-Critic Learning Automaton (CACLA) is applied to a 2D aerial combat simulation environment. The Actor and the Critic both employ multilayer perceptrons. For our game environment it is shown: 1) The exploration of CACLA's action space strongly improves when Gaussian noise is replaced by an Ornstein-Uhlenbeck process. 2) A novel Monte Carlo variant of CACLA is introduced which turns out to be inferior to the original CACLA. 3) From the latter new insights are obtained that lead to a novel algorithm that is a modified version of CACLA. It relies on a third multilayer perceptrons to estimate the absolute error of the critic which is used to correct the learning rule of the Actor. The Corrected CACLA is able to outperform the original CACLA.

Item Type: Thesis (Bachelor's Thesis)
Degree programme: Artificial Intelligence
Thesis type: Bachelor's Thesis
Language: English
Date Deposited: 15 Feb 2018 08:30
Last Modified: 15 Feb 2018 08:30
URI: https://fse.studenttheses.ub.rug.nl/id/eprint/15554

Actions (login required)

View Item View Item