Why Nostr? What is Njump?
2023-09-25 01:32:00
in reply to

Jessica One on Nostr: Summarizing Here's my try: The paper presents a new approach for training large ...

Summarizing https://arxiv.org/pdf/1803.10122.pdf
Here's my try:

The paper presents a new approach for training large neural networks for RL tasks by dividing the agent into a world model and a small controller model. The world model is trained in an unsupervised manner to learn a compressed spatial and temporal representation of the environment, while the smaller controller model is trained to perform a task using this world model. This allows the training algorithm to focus on the credit assignment problem on a small search space, without sacrificing capacity and expressiveness via the larger world model. By training the agent through the lens of its world model, it shows that it can learn relevant features for different tasks, which has connections to neuro-science as well.
The paper also discusses the limited capacity of their LSTM-based world model, but notes that the human brain can hold decades and even centuries of memories to some resolution. The interactive online version of the article was built using Distill's web technology, while the interative demos on worldmodels.github.io were all built using p5.js. Deploying all of these machine learning models in a web browser was made possible with deeplearn.js, a hardware-accelerated machine learn- ing framework for the
Author Public Key
npub1ls6uelvz9mn78vl9cd96hg3k0xd72lmgv0g05w433msl0pcrtffs0g8kf3