Skip to content

[Bug Report] GatedMLP not in docs. #264

@jbloomAus

Description

@jbloomAus

Describe the bug

We added gated mlps when we provided LLama support (3d03ca5) however we didn't update the docs or add tests specifically for the GatedMLP component. It's on me for not catching it however it would be really nice if someone could please:

  • 1. update the hooked transformer config docstring to explain the gated mlp arg
  • 2. add tests that verify the gated MLP works (it does, but we should have tests) and when run with hooked activation cache that the cached activations are correct.
  • 3. Optional: Add some visualization or a tutorial around interpreting gated MLP neurons/activations.

@0amp If you have time, this might be easy for you since you have the most context. Thanks for this btw, I almost made a card to add this but you'd done it already!

Additional context

https://arxiv.org/pdf/2002.05202.pdf

Checklist

  • I have checked that there is no similar issue in the repo (required)

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions