3 years ago

Model-Based Action Exploration.

Glen Berseth, Michiel van de Panne

Deep reinforcement learning has great stride in solving challenging motion control tasks.

Recently there has been a significant amount of work on methods to exploit the data gathered during training, but less work is done on good methods for generating data to learn from.

For continuous actions domains, the typical method for generating exploratory actions is by sampling from a Gaussian distribution centred around the mean of a policy.

Although these methods can find an optimal policy, in practise, they do not scale well, and solving environments with many actions dimensions becomes impractical.

We consider learning a forward dynamics model to predict the result, ($s_{t+1}$), of taking a particular action, ($a$), given a specific observation of the state, ($s_{t}$).

With a model such as this we, can perform what comes more naturally to biological systems that have already collect experience, we perform internal predictions of outcomes and endeavour to try actions we believe have a reasonable chance of success.

This method greatly reduces the space of exploratory actions, increasing learning speed and enables higher quality solutions to difficult problems, such as robotic locomotion.

Publisher URL: http://arxiv.org/abs/1801.03954

DOI: arXiv:1801.03954v1

You might also like
Never Miss Important Research

Researcher is an app designed by academics, for academics. Create a personalised feed in two minutes.
Choose from over 15,000 academics journals covering ten research areas then let Researcher deliver you papers tailored to your interests each day.

  • Download from Google Play
  • Download from App Store
  • Download from AppInChina

Researcher displays publicly available abstracts and doesn’t host any full article content. If the content is open access, we will direct clicks from the abstracts to the publisher website and display the PDF copy on our platform. Clicks to view the full text will be directed to the publisher website, where only users with subscriptions or access through their institution are able to view the full article.