Exploring Deep Recurrent Q-Learning for Navigation in a 3D Environment

Journal Title: EAI Endorsed Transactions on Creative Technologies - Year 2018, Vol 5, Issue 14

Abstract

Learning to navigate in 3D environments from raw sensory input is an important step towards bridging the gap between human players and artificial intelligence in digital games. Recent advances in deep reinforcement learning have seen success in teaching agents to play Atari 2600 games from raw pixel information where the environment is always fully observable by the agent. This is not true for first-person 3D navigation tasks. Instead, the agent is limited by its field of view which limits its ability to make optimal decisions in the environment. This paper explores using a Deep Recurrent Q-Network implementation with a long short-term memory layer for dealing with such tasks by allowing an agent to process recent frames and gain a memory of the environment. An agent was trained in a 3D first-person labyrinth-like environment for 2 million frames. Informal observations indicate that the trained agent navigated in the right direction but was unable to find the target of the environment.

Authors and Affiliations

Rasmus Kongsmar Brejl, Henrik Purwins, Henrik Schoenau-Fog

Keywords

Related Articles

Emotional interactive movie: adjusting the scenario according to the emotional response of the viewer

Emotional interactive movie is a kind of film unfolding in different ways according to the emotion the viewer experiences. The movie is made of several sequences; their combination determines the particular scenario expe...

Multi-GPU based framework for real-time motion analysis and tracking in multi-user scenarios

Video processing algorithms present a necessary tool for various domains related to computer vision such as motion tracking, event detection and localization in multi-user scenarios (crowd videos, mobile camera, scenes w...

Eye-tracking Data, Complex Networks and Rough Sets: an Attempt Toward Combining Them

Eye-tracking sequences can be considered in terms of complex networks. On the basis of complex network representation of eye-tracking data, we define a measure, derived from rough set theory, for assessing the cohesion o...

Inducing omnipotence or powerlessness in learners with developmental and attention difficulties through structuring technologies

Schoolwork of learners with developmental and attention difficulties is often characterised by low productivity, many errors due to carelessness or inattention and poor organizing ability. Focus learners have difficultie...

A Mathematical Model of Game Refinement and Its Applications to Sports Games

This paper explores a mathematical model of game progress. We claim that a realistic model of the game progress during the in-game period is not linear but exponential. The second derivative value, i.e., acceleration in...

Download PDF file
  • EP ID EP45886
  • DOI http://dx.doi.org/10.4108/eai.16-1-2018.153641
  • Views 258
  • Downloads 0

How To Cite

Rasmus Kongsmar Brejl, Henrik Purwins, Henrik Schoenau-Fog (2018). Exploring Deep Recurrent Q-Learning for Navigation in a 3D Environment. EAI Endorsed Transactions on Creative Technologies, 5(14), -. https://europub.co.uk/articles/-A-45886