Skip to content

Solving Markov Decision Processes

License

Notifications You must be signed in to change notification settings

GersiD/MDPs.jl

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

30 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

MDPs: Markov Decision Processes

Build Status

This is experimental software which can fail or change in unpredictable ways.

About

A set of tools for solving primarily tabular Markov Decision Processes. The project aims to

  1. Quickly and reliably solve small tabular MDPs with integral states and actions using standard algorithms. See IntMDP.
  2. Formulate and solve generic MDPs. See TabMDP.
  3. Analyze MDP value functions and policies. See value_iteration and policy_iteration.
  4. Framework algorithms that easily extend to new objectives. See InfiniteH and FiniteH.
  5. Simulate tabular MDPs. See Simulate.
  6. Provide a framework for solving general large MDPs. See MDP.

The project is build around the following main data structures:

  • MDP: A structure that supports Bellman updates and value function and qvalue computation. The model assumes that the number of actions is discrete and sufficiently small.
  • TabMDP: A structure that specialized MDPs to tabular state spaces. The states and actions may have arbitrary data types
  • IntMDP: A generic tabular MDP which has states and actions indexed by integers. States 0 and below are assumed to be terminal and allow no actions or transitions

One of the main goals of the project is to support exploration of algorithms for solving various reinforcement learning objectives. To achieve this goal, the project involves a number of abstraction. The benefit of this approach is that one can implement new dynamic programming algorithm with a relatively high degree of code reuse.

Installation

The package is not avaialble from the julia repository. You can install it directly from github:

using Pkg
Pkg.add(url="https://github.com/RiskAverseRL/MDPs.jl")

Similar packages

POMDPs is a comprehensive package that suports both MDPs and POMDPs and offers a variety of solvers.

About

Solving Markov Decision Processes

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Julia 100.0%