Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Bug Report] evals.py ablates all heads when it needs to ablate only one #163

Closed
1 task done
shehper opened this issue May 28, 2024 · 3 comments
Closed
1 task done

Comments

@shehper
Copy link
Contributor

shehper commented May 28, 2024

If you are submitting a bug report, please fill in the following details and use the tag [bug].

Describe the bug
When evaluating SAEs trained on individual attention head outputs, the reconstructed score should be calculated by zero-ablating only that singular head.

Here, zero_abl_loss calls zero_ablate_hook, which zero-ablates all heads.

Additional context
I noticed this while training SAEs on individual attention head outputs. The reconstructed score looked really good (~0.97) but the feature dashboards looked quite mediocre.

Checklist

  • I have checked that there is no similar issue in the repo (required)
@shehper
Copy link
Contributor Author

shehper commented May 28, 2024

I will be happy to submit a PR for this.

@jbloomAus
Copy link
Owner

Hey Shepher,

Thanks for raising this. So obvious in hindsight. I'd love a PR for this.

@shehper
Copy link
Contributor Author

shehper commented May 30, 2024

Sorry, I have just been a little swamped with the research sprint. I'll submit a PR over the weekend.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants