Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Some questions about the traininig time and the finally results #4

Closed
Henry0528 opened this issue Sep 22, 2023 · 11 comments
Closed

Some questions about the traininig time and the finally results #4

Henry0528 opened this issue Sep 22, 2023 · 11 comments

Comments

@Henry0528
Copy link

First thank you for your excellent work, while I met some questions when reproducing the results:

  1. I find that the training time is extremely long. I used 3 3090 to train the model on the MVTec dataset and it took me almost 2 days to finish the whole class training process.
  2. There is some large gap between the report results and the one I reproduced, like class carpet
    AUROC: 0.6737560033798218
    AUROC pixel level: 0.8384530544281006
    threshold: 0.59561723
    I kept almost all of the parameters unchanged in the config.yaml except for the batchsize, I wonder whether the setting is the same as the one you used and if not , how can I make the proper change to get the similar results.
@arimousa
Copy link
Owner

Training the model on 15 categories of MVTec for 2 days is realistic since diffusion models are expensive to train.
Some hyperparameter tuning is necessary to achieve optimal results as mentioned in the config file and readme. We will soon publish the best settings for each category.

@Jay-zzcoder
Copy link

I also met this question about the results of carprt, there is a huge gap between the report results and the one I reproduced, I try epochs 1000, 1500, 2000 , AD epoch 1,2 3 and many hyperparameter

@arimousa
Copy link
Owner

For carpet w = 0 and DA_chp =0 would be the best. However, I will publish checkpoints and settings very soon.

@Jay-zzcoder
Copy link

For carpet w = 0 and DA_chp =0 would be the best. However, I will publish checkpoints and settings very soon.

There is likely no AD_chp in config.yaml, you mean DA_epochs or others?

@arimousa
Copy link
Owner

arimousa commented Oct 30, 2023

The DA_epochs indicates the number of iterations for fine-tuning the feature extractor. And DA_chp indicates the checkpoint you load the checkpoint. For carpet setting them to zero would results in the best.

@Jay-zzcoder
Copy link

The DA_epochs indicates the number of iterations for fine-tuning the feature extractor. And DA_chp indicates the checkpoint you load the checkpoint. For carpet setting them to zero would results in the best.

AD_chp=0 means I onlt need to train one iteration to fine-tuning the feature extractor for carpet?

@arimousa
Copy link
Owner

It means a pretrained feature extractor outperforms a fine-tuned one.

@Jay-zzcoder
Copy link

It means a pretrained feature extractor outperforms a fine-tuned one.

No fine tuning? directly use pretrained feature extractor?

@arimousa
Copy link
Owner

It means a pretrained feature extractor outperforms a fine-tuned one.

No fine tuning? directly use pretrained feature extractor?

Exactly.

@Jay-zzcoder
Copy link

It means a pretrained feature extractor outperforms a fine-tuned one.

No fine tuning? directly use pretrained feature extractor?

Exactly.

I directly use pretrained feature extractor and set w=0, I reproduced the best results is:
Image AUROC: 91.7
Pixel AUROC: 93.9
PRO: 81.2

@arimousa
Copy link
Owner

arimousa commented Dec 5, 2023

checkpoints are published

@arimousa arimousa closed this as completed Dec 5, 2023
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants