International Journal of Computer Applications |
Foundation of Computer Science (FCS), NY, USA |
Volume 76 - Number 1 |
Year of Publication: 2013 |
Authors: Evans Miriti, Peter Waiganjo, andrew Mwaura |
10.5120/13212-0593 |
Evans Miriti, Peter Waiganjo, andrew Mwaura . Comparing Action as Input and Action as Output in a Reinforcement Learning Task. International Journal of Computer Applications. 76, 1 ( August 2013), 24-28. DOI=10.5120/13212-0593
Generalization techniques are useful for enabling an agent to be able to approximate the value of states it has not encountered so far in reinforcement learning. They are also useful as memory use minimization mechanisms in situations where the state space is too large such that is infeasible to represent every state in the state space in the computer memory. Artificial Neural Networks are one generalization technique that is usually employed. Various network structures have been proposed in literature. In this study, two of the structures that have been proposed were implemented in a robot navigation task and their performance compared. The results indicate that having a network structure where there is an output node for each of the possible actions, is superior to the structure in which the selected action is fed as an input to the network and its value output by the single network output node.