Search icon CANCEL
Subscription
0
Cart icon
Your Cart (0 item)
Close icon
You have no products in your basket yet
Arrow left icon
Explore Products
Best Sellers
New Releases
Books
Videos
Audiobooks
Learning Hub
Conferences
Free Learning
Arrow right icon

Google AI researchers introduce PlaNet, an AI agent that can learn about the world using only images

Save for later
  • 2 min read
  • 18 Feb 2019

article-image

The Google AI team in collaboration with DeepMind announced a new and open source “Deep Planning” Network, called PlaNet, last week. PlaNet is an AI agent that learns a world model using only image inputs and further plans with these models to gain experiences.

PlaNet can easily solve a variety of image-based control tasks as well as compete with the advanced model-free agents. The Google AI team is also releasing the source code for the research community to further explore and build upon PlaNet.

How does PlaNet work?


PlaNet depends on a compact sequence of hidden or latent states. This is known called a latent dynamics model where instead of predicting directly from one image to the next image, the latent state forward is first predicted. “By compressing the images in this way, the agent can automatically learn more abstract representations, such as positions and velocities of objects, making it easier to predict forward without having to generate images along the way”, states the Google AI team.

In a latent dynamics model, the information of the input images gets integrated into the hidden states with the help of an encoder network. The hidden state then gets further projected forward to predict future images and rewards. For planning, past images are encoded into the current hidden state, and then the future rewards for multiple action sequences are predicted.

google-ai-researchers-introduce-planet-an-ai-agent-that-can-learn-about-the-world-using-only-images-img-0

 PlaNet agents trained on different image-based control tasks

PlaNet agents are trained across a variety of image-based control tasks. These tasks pose different challenges such as partial observability, sparse rewards for catching a ball, etc. Moreover, a single PlaNet agent is trained to solve all six tasks. Without any changes to the hyperparameters, this multi-task agent is able to achieve the same mean performance as individual agents.

“We advocate for further research that focuses on learning accurate dynamics models on tasks of even higher difficulty, such as 3D environments and real-world robotics tasks. We are excited about the possibilities that model-based reinforcement learning opens up”, states the Google AI team.

For more information, check out the official Google AI PlaNet announcement.

Google News Initiative partners with Google AI to help ‘deep fake’ audio detection research

Google AI releases Cirq and Open Fermion-Cirq to boost Quantum computation

Google announces the general availability of a new API for Google Docs

Unlock access to the largest independent learning library in Tech for FREE!
Get unlimited access to 7500+ expert-authored eBooks and video courses covering every tech area you can think of.
Renews at €18.99/month. Cancel anytime