Abstract
This paper investigates reinforcement learning problems where a stochastic time delay is present in the reinforcement signal, but the delay is unknown to the learning agent. This work posits that the agent may receive individual reinforcements out of order, which is a relaxation of an important assumption in previous works from the literature. To that end, a stochastic time delay is introduced into a mobile robot line-following application. The main contribution of this work is to provide a novel stochastic approximation algorithm, which is an extension of Q-learning, for the time-delayed reinforcement problem. The paper includes a proof of convergence as well as grid world simulation results from MATLAB, results of line-following simulations within the Cyberbotics Webots mobile robot simulator, and finally, experimental results using an e-Puck mobile robot to follow a real track despite the presence of large, stochastic time delays in its reinforcement signal.
from Biomedical Engineering via ola Kala on Inoreader http://ift.tt/1LmDJBX
via IFTTT
from #Med Blogs by Alexandros G.Sfakianakis via Alexandros G.Sfakianakis on Inoreader http://ift.tt/1oeuiQs
via IFTTT
Δεν υπάρχουν σχόλια:
Δημοσίευση σχολίου