6533b822fe1ef96bd127d903

RESEARCH PRODUCT

The Dreaming Variational Autoencoder for Reinforcement Learning Environments

Per-arne AndersenMorten GoodwinOle-christoffer Granmo

subject

FOS: Computer and information sciencesMaskinlæringComputer Science - Machine LearningVDP::Computer technology: 551Artificial Intelligence (cs.AI)VDP::Datateknologi: 551Computer Science - Artificial IntelligenceMachine learningDeep learningMachine Learning (cs.LG)

description

Reinforcement learning has shown great potential in generalizing over raw sensory data using only a single neural network for value optimization. There are several challenges in the current state-of-the-art reinforcement learning algorithms that prevent them from converging towards the global optima. It is likely that the solution to these problems lies in short- and long-term planning, exploration and memory management for reinforcement learning algorithms. Games are often used to benchmark reinforcement learning algorithms as they provide a flexible, reproducible, and easy to control environment. Regardless, few games feature a state-space where results in exploration, memory, and planning are easily perceived. This paper presents The Dreaming Variational Autoencoder (DVAE), a neural network based generative modeling architecture for exploration in environments with sparse feedback. We further present Deep Maze, a novel and flexible maze engine that challenges DVAE in partial and fully-observable state-spaces, long-horizon tasks, and deterministic and stochastic problems. We show initial findings and encourage further work in reinforcement learning driven by generative exploration.

10.1007/978-3-030-04191-5_11http://hdl.handle.net/11250/2596208