{ "cells": [ { "cell_type": "markdown", "metadata": {}, "source": [ "We presented at IJCNN, 2015 the following paper, which won the [Best Paper Award](http://www.ijcnn.org/assets/docs/ijcnn2015-awards.pdf)\n", "\n", " * Anderson, C., Lee, M., and Elliott, D., \"[Faster Reinforcement Learning After Pretraining Deep Networks to Predict State Dynamics](publications/pretrainijcnn15.pdf)\", Proceedings of the IJCNN, 2015, Killarney, Ireland.\n", "\n", "*Abstract:* Deep learning algorithms have recently appeared that pretrain hidden layers of neural networks in unsupervised ways, leading to state-of-the-art performance on large classification problems. These methods can also pretrain networks used for reinforcement learning. However, this ignores the additional information that exists in a reinforcement learning paradigm via the ongoing sequence of state, action, new state tuples. This paper demonstrates that learning a predictive model of state dynamics can result in a pretrained hidden layer structure that reduces the time needed to solve reinforcement learning problems. " ] }, { "cell_type": "markdown", "metadata": {}, "source": [ "Below is a sequence of animations showing the progress of reinforcement learning. We simulated the dynamics of the cart-pole problem with full 360 pole rotation and collisions of the cart at the ends of the track. Details of this simulation will be provided here soon." ] }, { "cell_type": "markdown", "metadata": { "collapsed": true }, "source": [ "