New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
QQP test evaluation is extremely slow #209
Comments
(This is why we couldn't do the test set evaluation this week.) |
investigating |
QQP test is quite big, so that's got to be most of it. a 100% slowdown in evaluation is okay, but would be pretty conspicuous here. See also: #145 |
I can confirm that QQP test eval works, though, so it may be safe to ignore. |
ok |
Wow this is excruciatingly slow, 1hr+ for me on P100. |
This is pretty bad—it seems to be a problem even on dev, and it's almost certainly a result of #121. |
Just occurred to me: we should use a smarter batcher / iterator during eval (and validation). IIRC we got pretty decent speed ups. |
How easy would it be? Would it break streaming? If *very* and *no*, give it
a try!
…On Wed, Jul 25, 2018 at 5:04 PM Alex Wang ***@***.***> wrote:
Just occurred to me: we should use a smarter batcher / iterator during
eval (and validation). IIRC we got pretty decent speed ups.
—
You are receiving this because you authored the thread.
Reply to this email directly, view it on GitHub
<#209 (comment)>,
or mute the thread
<https://github.com/notifications/unsubscribe-auth/ABOZWYhK_oRn6XMlsnGuRjbvl4w2ewbXks5uKN04gaJpZM4VYyTy>
.
|
I think it would not be that bad; we already use it during training. Wouldn't break anything. |
Smart batching doesn't seem to help because the batch utilization was already pretty high, I guess because the QQP are pretty similar in length. What we could do is jack up the batch size during evaluation only? |
Worth a try, if it's easy to set up.
…On Wed, Jul 25, 2018 at 9:40 PM Alex Wang ***@***.***> wrote:
Smart batching doesn't seem to help because the batch utilization was
already pretty high, I guess because the QQP are pretty similar in length.
What we could do is jack up the batch size during evaluation only?
—
You are receiving this because you authored the thread.
Reply to this email directly, view it on GitHub
<#209 (comment)>,
or mute the thread
<https://github.com/notifications/unsubscribe-auth/ABOZWSM-dzC-R1Ck0HWxQcIEFMnsoSmKks5uKR36gaJpZM4VYyTy>
.
|
What GPUs were you running on previously? For some reason I'm getting pretty relatively fast eval times (~10m) for QQP test on 1080s... |
Odd—this was also on NYU 1080s. I don't recall seeing any evaluation speed
fixes since Friday...
Much of the time was spent on sorting—is that process cached somehow?
…On Wed, Jul 25, 2018 at 10:17 PM Alex Wang ***@***.***> wrote:
What GPUs were you running on previously?
For some reason I'm getting pretty relatively fast eval times (~10m) for
QQP test on 1080s...
—
You are receiving this because you authored the thread.
Reply to this email directly, view it on GitHub
<#209 (comment)>,
or mute the thread
<https://github.com/notifications/unsubscribe-auth/ABOZWTi9QZv2r1Cypk-aH23hKF7j6819ks5uKSaxgaJpZM4VYyTy>
.
|
Sorting is not cached...I was also waiting 2,5 hours on p100....I'm trying to debug it on CPU |
Alex—was that definitely test? QQP dev is much smaller than test.
…On Wed, Jul 25, 2018 at 10:36 PM Jan Hůla ***@***.***> wrote:
Sorting is not cached...I was also waiting 2,5 hours on p100....I'm trying
to debug it on CPU
—
You are receiving this because you authored the thread.
Reply to this email directly, view it on GitHub
<#209 (comment)>,
or mute the thread
<https://github.com/notifications/unsubscribe-auth/ABOZWYuG2OhsOeQ3nO6KzoaRcrU0BzGmks5uKSsvgaJpZM4VYyTy>
.
|
Yes, on dev it's also very fast (~1m) and test is ~12m Profiling now. Maybe it's due to having an untrained encoder? But I'm running the same script on the p100 and it's much slower (though less than an hour). |
Is this really slow? it seems to go through a rate of ~60 batches/30 seconds, which is the same rate as qnli or mnli. qqp is just a big dataset, (test is ~390K. the rest of the test sets are ~20K or less; val is 40k, the rest of the val sets are also <20K). The logging helps a lot. EDIT: Not going to close because others might also wonder why qqp takes for ever |
Fair enough. Feel free to close if there's nothing more to be done.
…On Thu, Jul 26, 2018 at 4:10 PM Patrick Xia ***@***.***> wrote:
Is this really slow? it seems to go through a rate of ~60 batches/30
seconds, which is the same rate as qnli or mnli.
qqp is just a big dataset, (test is ~390K. the rest of the test sets are
~20K or less; val is 40k, the rest of the val sets are also <20K).
The logging helps a lot.
—
You are receiving this because you authored the thread.
Reply to this email directly, view it on GitHub
<#209 (comment)>,
or mute the thread
<https://github.com/notifications/unsubscribe-auth/ABOZWeF3rFy0yLE0MVvIA8Woo10AJxI_ks5uKiIngaJpZM4VYyTy>
.
|
One run has been on QQP Test for 2.5 hours without signs of progress. GPU usage is non-zero but low. This seems to have changed since #121.
@Jan21 @iftenney, any guesses? Did you verify that QQP test works?
Dev also appears to be quite slow, but I don't have numbers yet.
The text was updated successfully, but these errors were encountered: