Skip to content

Implementation of Scheduled Policy Optimization for task-oriented language grouding

License

Notifications You must be signed in to change notification settings

udacitysimon/walk_the_blocks

 
 

Repository files navigation

Scheduled Policy Optimization for Natural Language Communication with Intelligent Agents

Models and Algorithms

See files under walk_the_blocks/BlockWorldRoboticAgent/srcs/

  • learn_by_ppo.py run this file for training, you can change the schedule mechanism in the function ppo_update(), these are the options:

    • do imitation every 50
    • do imitation based on rules
    • imitation 1 epoch and then RL 1 epoch

    example: python learn_by_ppo.py -lr 0.0001 -max_epochs 2 -entropy_coef 0.05

  • policy_model.py the network achitecture and loss functions:

    • PPO Loss
    • Supervised Loss
    • Advantage Actor-Critic Loss

Instructions

For the usage of the Block-world environment, please refer to https://github.com/clic-lab/blocks

Train the RL agents

  • S-REIN *

About

Implementation of Scheduled Policy Optimization for task-oriented language grouding

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Classic ASP 61.3%
  • C# 22.5%
  • Python 15.8%
  • Other 0.4%