Skip to content

blengereau/DynamicProgramming

Repository files navigation

Dynamic Programming & Reinforcement Learning applied to a Neo-Classical Growth Model

This project solves an optimal growth model using Dynamic Programming techniques by approximating iteratively the value function. It presents Reinforcement Learning algorithms to evaluate arbitrary policies on the agent control variable, and update the policy function ofthe agent using a greedy procedure.

Additionally, this project calibrates a Neural Network that approximates the optimal path of the policy function from an arbitrary amount of capital.

1. Model

Consider an agent who allocates her time between producing the consumption good C, and accumulating human capital H. The agent seeks to maximize her discounted utility, as captured by the following objective:

Model

The production function reads in (i). Normalizing the labor supply of the agent to one, the accumulation law for human capital is given by (ii), where 𝛿 is the depreciation rate of human capital, while 𝐿t is the share of the labor supply dedicated to the production of the consumption good.

We assume a CRRA Utility function: U(C) = C^(1-𝜎)/(1-𝜎).

2. Value function and optimal policy

The Bellman Equation of the agent is:

Bellman Equation and Feasability set

We iteratively approximate the value function, reached by the following optimal policy function:

Value Function

Policy function

3. Evaluation of arbitrary policies

4. Neural Network calibration to estimate the optimal path

We aim to calibrate a neural network that approximates the path starting from an arbitrary initial stock of human capital and converging toward the steady state. To calibrate the neural network, we minimize the distance between the simulated and theoretical values of the:

  • Euler equation

Euler Equation

  • Accumulation law of human capital

Accumulation law

  • Initial stock of human capital

Initial condition

We estimate the optimal path using a deep neural network calibration with back propagation of the loss of the residuals associated with the three equations. We observe a convergence to the steady state of the model.

Optimal path

About

Resolution of an optimal growth model using Dynamic Programming and Deep Learning

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published