Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

I am revising the model to solve QA task.. #32

Closed
0525stone opened this issue Sep 1, 2022 · 1 comment
Closed

I am revising the model to solve QA task.. #32

0525stone opened this issue Sep 1, 2022 · 1 comment

Comments

@0525stone
Copy link

Hi, I am working on your code to solve QA task.
I have a question.

Currently my dataset consists of context, question, answer(each question have pair of answer & context).
The length of question is too short(mostly, after tokenizing with 'bert-base-multilingual', 6 to 7 tokens are generated). so I add to make model run because default model setting requires 'chunk_size = 64'.

here is my question.
I think the role of token is just added to fill out empty space. I think that means logically is nothing.
Once training, I made two chunks and put those in. First chunk is made by "question: blahblahblah" and second chunk is made by "answer: blahblah"(those sentences are too short.. as I mentioned earlier)

Once I put input such as "question: blahblahblah? answer: " into the checkpoint model, the model spit out [CLS] question : blahblahblah? answer: [SEP] [PAD] [PAD] [PAD] [PAD] [PAD] [PAD] [PAD] [PAD] [PAD] [PAD] [PAD] [PAD] [PAD] [PAD] [PAD] [PAD] [PAD] [PAD] [PAD] [PAD] [PAD] [PAD] [PAD] [PAD] [PAD] [PAD] [PAD] [PAD] [PAD] [PAD] [PAD] [PAD] [PAD] [PAD].
I have no clue why my model stay dumb...

@0525stone 0525stone closed this as not planned Won't fix, can't repro, duplicate, stale Sep 1, 2022
@yerinNam
Copy link

I'm having the same issue, did you solve it?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants