Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

How do you set hyperparameters? #3

Open
jinmu-wxy opened this issue Mar 21, 2024 · 3 comments
Open

How do you set hyperparameters? #3

jinmu-wxy opened this issue Mar 21, 2024 · 3 comments

Comments

@jinmu-wxy
Copy link

Hello, I wonder how you set hyperparameters, such as learning rate, batch size, and the number of epochs. We just obtained a 68 score on the Crame-D dataset using the command in the readme.

python main.py --train --ckpt_path ckpt --gpu_ids 0 --batch_size 64 --lorb base --modulation Normal --epochs 100 --dataset CREMAD --gs_flag

@Cecile-hi
Copy link
Owner

Hi, I think maybe you can try --batch_size 16 and specify --av_alpha 0.55, I used this hyper and get a 77.69 acc, like shown below:
thumbnail_0A9AC1C2@BE2C3A2B 8103F96500000000
thumbnail_B13CC1BB@146D5A30 8103F96500000000

@shicaiwei123
Copy link

hello, with the following command,

python main.py --train --ckpt_path ckpt --gpu_ids 0 --batch_size 64 --lorb base --modulation Normal --epochs 100 --dataset CREMAD --gs_flag

I also got the result of 68.8.

When I set the batch size as 16 and av_alpha as 0.55, I got the result of 73.6, which is still far from the result in the paper.

Are there any else parameters?

@hubaak
Copy link

hubaak commented May 7, 2024

@Cecile-hi Hi, I set batch_size=16, epoch=200, lr_decay_step=150, and other hyperparameters to default. I got a score of 0.800 on the CREMA-D dataset, which is as good as expected. Here is the figure of the accuracy curve:
image
However, when I used the same parameter for uni-modal learning(named late fusion in your paper), I got an even higher accuracy of 0.812, much higher than 0.663 in Table 1.
image
And the corresponding curve is here:
image
I tried other parameters but found that for the same set of parameters, uni-modal learning always seemed to have a competitive performance to your method. So I am a little bit confused by the results now. I'd appreciate it if you could tell me the detailed hyperparameter setting of Table 1 :).

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants