Skip to content

Conversation

@mattdangerw
Copy link
Member

@mattdangerw mattdangerw commented Apr 10, 2024

Tricky bug, keras.layers.Embedding has started overriding save_own_variables/load_own_variables for quantization. We can no longer rely on absolute indexing for the reverse embedding, e.g. self.reverse_embeddings.assign(store["1"])

I also couldn't figure out a way to continue to support loading tied weights untied. This is because the embedding will now unconditionally error if weight counts are mismatched here: https://github.com/keras-team/keras/blob/v3.2.0/keras/layers/core/embedding.py#L232-L264

I don't think we have a direct need for this functionality today, so I just removed it, but could be worth considering how to bring back in the future.

@mattdangerw mattdangerw added the kokoro:force-run Runs Tests on GPU label Apr 10, 2024
@kokoro-team kokoro-team removed the kokoro:force-run Runs Tests on GPU label Apr 10, 2024
@mattdangerw mattdangerw added the kokoro:force-run Runs Tests on GPU label Apr 10, 2024
@kokoro-team kokoro-team removed kokoro:force-run Runs Tests on GPU labels Apr 10, 2024
Tricky bug, keras.layers.Embedding has start overriding
save_own_variables/load_own_variables for quantization. We can
no longer rely on absolute indexing for the reverse embedding, e.g.
store["1"] = self.reverse_embeddings

I also couldn't figure out a way to continue to support loading
tied weights untied. This is because the embedding will now
unconditionally error if weight counts are mismatched here:
https://github.com/keras-team/keras/blob/v3.2.0/keras/layers/core/embedding.py#L232-L264

I don't think we have a direct need for this functionality today, so I
just removed it, but could be worth considering how to bring back
in the future.
@mattdangerw mattdangerw merged commit 6cdbf6f into keras-team:master Apr 10, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants