Skip to content
This repository has been archived by the owner on Oct 31, 2023. It is now read-only.

The input/output feature dimensions of Transformer Encoder and Causal Transformer Decoder? #41

Open
yxgz opened this issue May 10, 2022 · 1 comment

Comments

@yxgz
Copy link

yxgz commented May 10, 2022

Hi, thanks for your great project!
I am wondering the input/output feature dimensions of Transformer Encoder. The description in Section 4.1 of the paper shows the input/output feature dimensions are both 768D, is it right? However, the description in Section 4.4 of the paper shows the
input feature dimension of Causal Transformer Decoder is 2048D, what is the output feature dimension of Causal Transformer Decoder? And is there a dimension conversion (768D->2048D) before using Causal Transformer Decoder?

@rohitgirdhar
Copy link
Contributor

Hi, thanks for your interest and apologies for the delay.
You are correct, there is a linear layer that does this mapping --

self.mapper_to_inter = nn.Linear(backbone_dim,

Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants