Skip to content

createamind/Awesome-AGI-Paper

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

43 Commits
 
 
 
 

Repository files navigation

Awesome-AGI-Paper

Our dream is creating real AI who is suitable for all domains, here are some papers we high recommanded

Representation learning

AutoEncoder

Disentangled-representations

General theory of VAE

GAN

DRL

Task-Agnostic Reinforcement Learning

Successor Representations

OPTIONS SKILL

  • Learning Abstract Options
  • Successor Options: An Option Discovery Framework for Reinforcement Learning

Model base

Model free

Meta learning

Memory

HUBE

NTM

Episode memory

Association memory

Few shot learning

Mutual infomation

General

Key point

Curiosity-driven Exploration by Self-supervised Prediction

Learning a predictor and use predict error as internal reward, and they jointly train an inverse dynamics model for encoder, who project observation to a space that is invariant to parts of the environment that do not affect the agent or task.

EMI

  • Learn the representation of states and the action such that the representation of the corresponding next state following linear dynamics
  • Intrinsic reward augmentation
  • https://github.com/snu-mllab/EMI

THE THERMODYNAMICS OF MACHINE LEARNING

In this work we offer an information-theoretic framework for representation learn- ing that connects with a wide class of existing objectives in machine learning. We develop a formal correspondence between this work and thermodynamics and dis- cuss its implications.

RECURRENT REINFORCEMENT LEARNING:A HYBRID APPROACH

propose a new family of hybrid models that combines the strength of both supervised learning (SL) and reinforcement learning (RL), trained in a joint fashion: The SL component can be a recurrent neural networks (RNN) or its long short-term memory (LSTM) version, which is equipped with the desired property of being able to capture long-term dependency on history, thus providing an effective way of learning the representation of hidden states. The RL component is a deep Q-network (DQN) that learns to optimize the control for maximizing long-term rewards. Extensive experiments in a direct mailing campaign problem demonstrate the effectiveness and advantages of the proposed approach

Dynamics-Aware Unsupervised Discovery of Skills

Graphical models, Information Bottleneck and Unsupervised Skill Learning

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published