Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

The impacts of BatchNorm instead of LayerNorm #1

Closed
Ye-D opened this issue May 31, 2023 · 2 comments
Closed

The impacts of BatchNorm instead of LayerNorm #1

Ye-D opened this issue May 31, 2023 · 2 comments

Comments

@Ye-D
Copy link

Ye-D commented May 31, 2023

In the CrypTen-based MPCFormer, such as MPC-Bert, I noticed that the LayerNorm is replaced by BatchNorm (due to CrypTen does not support LayerNorm now). Will this modification influence Bert's performance (e.g., accuracy)? Due to the source code did not give the model-loading scripts, it might be difficult for me to check it in experiments. Thanks for your help.

@DachengLi1
Copy link
Owner

@Ye-D thanks for the great question! We measure accuracy in plain PyTorch, and only system throughput with Crypten, as reported in the paper. We are implementing modules such as Layernorm in Crypten to measure the accuracy.

By module loading, do you mean something like this ?

@Ye-D
Copy link
Author

Ye-D commented Jun 1, 2023

@DachengLi1 Thanks for the reply!

@Ye-D Ye-D closed this as completed Jun 1, 2023
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants