Skip to content

DiODeProject/Optimal-policy-for-value-based-decision-making

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

18 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Optimal-policy-for-value-based-decision-making

These codes modify the analysis presented in the paper by Tajima, Drugowitsch & Pouget (2016) [1], to examine the implication of maximising geometric-discounted future rewards on the optimal policy, rather than maximising arithmetic mean reward rate.

N.B. reward rate across trials is not maximised, but set to 0; this has the effect of considering only single trial dynamics, since the cost of waiting for the next trial becomes zero.

CITATION:

[1] Satohiro Tajima*, Jan Drugowitsch*, and Alexandre Pouget. Optimal policy for value-based decision-making. Nature Communications, 7:12400, (2016). *Equally contributed.

USAGE: valueDecisionBoundaryRR.m() generates figures used in Figs. 3 or 6 in the paper. You can switch the utility function and reward calculation assumed in the model by removing commented-out lines in the code, and toggling the value of the geometric variable, respectively

About

These are the codes related to a comment on the following publication: Tajima, S., Drugowitsch, J., and Pouget, A. Optimal policy for value-based decision-making. Nature Communications, 7:12400, (2016).

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • MATLAB 100.0%