Skip to content

Latest commit

 

History

History
114 lines (97 loc) · 6.87 KB

README.md

File metadata and controls

114 lines (97 loc) · 6.87 KB

Awesome-AGI-Paper

Our dream is creating real AI who is suitable for all domains, here are some papers we high recommanded

Representation learning

AutoEncoder

Disentangled-representations

General theory of VAE

GAN

DRL

Task-Agnostic Reinforcement Learning

Successor Representations

OPTIONS SKILL

  • Learning Abstract Options
  • Successor Options: An Option Discovery Framework for Reinforcement Learning

Model base

Model free

Meta learning

Memory

HUBE

NTM

Episode memory

Association memory

Few shot learning

Mutual infomation

General

Key point

Curiosity-driven Exploration by Self-supervised Prediction

Learning a predictor and use predict error as internal reward, and they jointly train an inverse dynamics model for encoder, who project observation to a space that is invariant to parts of the environment that do not affect the agent or task.

EMI

  • Learn the representation of states and the action such that the representation of the corresponding next state following linear dynamics
  • Intrinsic reward augmentation
  • https://github.com/snu-mllab/EMI

THE THERMODYNAMICS OF MACHINE LEARNING

In this work we offer an information-theoretic framework for representation learn- ing that connects with a wide class of existing objectives in machine learning. We develop a formal correspondence between this work and thermodynamics and dis- cuss its implications.

RECURRENT REINFORCEMENT LEARNING:A HYBRID APPROACH

propose a new family of hybrid models that combines the strength of both supervised learning (SL) and reinforcement learning (RL), trained in a joint fashion: The SL component can be a recurrent neural networks (RNN) or its long short-term memory (LSTM) version, which is equipped with the desired property of being able to capture long-term dependency on history, thus providing an effective way of learning the representation of hidden states. The RL component is a deep Q-network (DQN) that learns to optimize the control for maximizing long-term rewards. Extensive experiments in a direct mailing campaign problem demonstrate the effectiveness and advantages of the proposed approach

Dynamics-Aware Unsupervised Discovery of Skills

Graphical models, Information Bottleneck and Unsupervised Skill Learning