Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

How many milliseconds of delay does different chunk correspond to? #17

Closed
TeaPoly opened this issue Dec 9, 2020 · 6 comments
Closed

Comments

@TeaPoly
Copy link
Contributor

TeaPoly commented Dec 9, 2020

I notice chunk 1 degradation 30% (from 5.51 to 7.83) in section "Unified Dynamic chunk" ctc greedy search.

Please let me know if my understanding is incorrect:
Chunk 1 delay conversion is 12layers*1chunk*40ms(conv2d) = 480 ms.

@robin1001
Copy link
Collaborator

The latency is only related to chunk size and sampling rate(conv2d, 40ms), so for chunk 1, the delay is 1 * 40=40ms, and for chunk 16, the delay is 16 * 40 = 640ms.

We are preparing the paper, I will provide the paper link in a few days once it is ready.

@TeaPoly
Copy link
Contributor Author

TeaPoly commented Dec 10, 2020

Thanks for your reply.

@TeaPoly TeaPoly closed this as completed Dec 10, 2020
@robin1001
Copy link
Collaborator

I notice chunk 1 degradation 30% (from 5.51 to 7.83) in section "Unified Dynamic chunk" ctc greedy search.

Please let me know if my understanding is incorrect:
Chunk 1 delay conversion is 12layers1chunk40ms(conv2d) = 480 ms.

Hi TeaPloy, please see more details at https://arxiv.org/abs/2012.05481, the paper is the first draft, we are still refining it. Hope it helps.

@robin1001 robin1001 pinned this issue Dec 11, 2020
@TeaPoly
Copy link
Contributor Author

TeaPoly commented Dec 11, 2020

I notice chunk 1 degradation 30% (from 5.51 to 7.83) in section "Unified Dynamic chunk" ctc greedy search.
Please let me know if my understanding is incorrect:
Chunk 1 delay conversion is 12layers_1chunk_40ms(conv2d) = 480 ms.

Hi TeaPloy, please see more details at https://arxiv.org/abs/2012.05481, the paper is the first draft, we are still refining it. Hope it helps.

Thanks.

@TeaPoly
Copy link
Contributor Author

TeaPoly commented Dec 11, 2020

The latency is only related to chunk size and sampling rate(conv2d, 40ms), so for chunk 1, the delay is 1 * 40=40ms, and for chunk 16, the delay is 16 * 40 = 640ms.

We are preparing the paper, I will provide the paper link in a few days once it is ready.

Bother you again, whether the history context is infinite?

截屏2020-12-11 下午7 06 10

@robin1001
Copy link
Collaborator

The latency is only related to chunk size and sampling rate(conv2d, 40ms), so for chunk 1, the delay is 1 * 40=40ms, and for chunk 16, the delay is 16 * 40 = 640ms.
We are preparing the paper, I will provide the paper link in a few days once it is ready.

Bother you again, whether the history context is infinite?

截屏2020-12-11 下午7 06 10

Yes,history context is the whole history.

@mikelei mikelei unpinned this issue Jan 26, 2021
wantt referenced this issue Apr 13, 2022
* [runtime] refactor decoder, asr_model to support more platforms

* fix topo error when rescoring

* support Copy for asr_model

* fix wrong check

* refine copy

* set inter-op thread
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants