Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Questions related to your paper #5

Open
hbell99 opened this issue Nov 22, 2021 · 1 comment
Open

Questions related to your paper #5

hbell99 opened this issue Nov 22, 2021 · 1 comment

Comments

@hbell99
Copy link

hbell99 commented Nov 22, 2021

Hi~ I've read your paper and the code, and I am little confused about some details.

Firstly, the coordinates of bounding boxes are discretized and the vocabulary for layouts contains the 8-bit uniform quantization token (which is token indexes are 0-127 in your code), categorial token, padding token, bos and eos token. But what if during inference stage, the predicted coordinate of the trained model doesn't lie in 0-127 tokens? Will you do any post processing to correct these mis-predicted layouts?

Secondly, in your paper, it is said that 'we minimize KL-Divergence between soft-max predictions and output one-hot distribution with Label Smoothing', but why the code is still a cross entropy loss?

@duzhenjiang113
Copy link

I also have a question. The paper described that the input order of the data is a random sequence, but from the code point of view, it is input in a specific sort of order.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants