Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Independent Q-learning for BPD and Congestion Game #51

Merged
merged 11 commits into from
Mar 25, 2024
Merged

Independent Q-learning for BPD and Congestion Game #51

merged 11 commits into from
Mar 25, 2024

Conversation

hiazmani
Copy link
Collaborator

No description provided.

Copy link
Collaborator

@ffelten ffelten left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Mostly LGTM. Just a few comments. We might have to harmonize the plot style for the paper but that is quite simple.

cap_min, cap_max, mix_min, mix_max = self.g_cap_min, self.g_cap_max, self.g_mix_min, self.g_mix_max

# Normalize the rewards
cap_norm = (reward[0] - cap_min) / (cap_max - cap_min)
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I believe we have a NormalizeReward wrapper made for this. Is it different?

Copy link
Collaborator Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

The NormalizeReward wrapper normalizes the rewards that the agents receive directly. In the BDP environment, agents can be rewarded according to two reward schemes (local/global), but the reported results in the graphs are always using the (normalized) 'global' reward scheme regardless of what the agents receive. This method is primarily used to compute these normalized global rewards.

momaland/learning/iql/iql.py Outdated Show resolved Hide resolved
momaland/learning/iql/iql.py Show resolved Hide resolved
@rradules rradules merged commit 361e2e8 into main Mar 25, 2024
10 checks passed
@rradules rradules deleted the iql branch March 25, 2024 17:59
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

None yet

3 participants