-
Notifications
You must be signed in to change notification settings - Fork 7
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Testing in a higher frequency #3
Comments
I try this way but It seems do not work/(ㄒoㄒ)/~~ |
@Hatins I also had problems with that, didn't work, and then I realized I didn't change the frequency of event data meaning re-processing it. Did you do that? @NikolaZubic @magehrig Can Mamba from Gu & Dao be used here to also evaluate at different frequencies, and will Mamba give even better results than S5? Btw, thanks for sharing S5 as zip in PyTorch, it is really valuable. |
Hi @looper99 Actually, I haven't quite grasped your question. Are you suggesting evaluating high-frequency event data without recreating it? |
I think what he means is, can your frequency generalization technique also be applied in the s6 (ssm) model, also known as mamba? Does this model perform well even at high frequencies? |
Have you ever encountered this problem? When I adjust the training: precision: 32 to 16, an error will be reported. Do you know how to solve it? thanks |
Yes, the S5 model seems to need a higher precision, that 32. |
I did it in the same way as in the pendulum task outlined in the following link: GitHub - lindermanlab/S5 - Issue #13. It worked just fine for event data. I experimented and saw that there is a way to train with precision 16, but you have to add But, when I tried to train this way, I can't get the mAP from the paper, which I get when using precision 32. I can only do GEN1 since I have one A100. |
Hi @looper99 |
@Hatins : Mamba requires less memory, but its not a continuous model anymore. |
@looper99 Regarding the Mamba, it is a purely discrete model in the sense that the selectivity of B, C, and delta on the input enables it to learn what to remember and what to forget. All the continuous-time theory breaks on it. So, you cannot use Mamba for evaluation at different frequencies. What is good about Mamba is that the VRAM requirements are much less, so it is more practical in that sense. No attention, no MLPs, but expressive. Mamba is for discrete data. Why the authors decided to move from this continuous setting with Mamba, I don't know (probably to compete with Transformers on purely discrete data problems), but the ideal model would be the one that can do both discrete and continuous very well and can be adapted by re-scaling the delta. This itself is an open research question. |
I tried your method, but still encountered errors in the modifications made.I don't know why. |
Hi @NikolaZubic
Thanks for your nice work and opening source!
I have a question when testing in a higher frequency, which parameters we should change?
I have found two places about the 'step_scale':
and
in S5SSM moudle.
If I want two testing in frequecy 200Hz (training in frequency 20), how should I adjust the parameter?
Should I change the above step_scale from 1.0 to 0.1 both?
The text was updated successfully, but these errors were encountered: