Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Changing model size doesn't seem to improve detection, actually makes it worse? #37

Closed
mellerbeck opened this issue Aug 4, 2024 · 1 comment

Comments

@mellerbeck
Copy link

model_size: str. The size of the model to use. It can be 'n' (nano), 's' (small), 'm' (medium) or 'l' (large). Larger models are more accurate but slower. Default: 's'.

I haven't tested with a large sample size, but at first pass if I change the model size to anything besides s it stops detecting.

@Eric-Canas
Copy link
Owner

HI! Yes, you are true, it may actually happen.

Larger models are usually more capable for difficult task, but also more prone to overfit data if it is not varied enough.

I manually tag the dataset I use for detection, so it size grows in every tagging-training round, but actually, its current size may be not large enough to give larger models the data they need to squeeze their capabilities.

So it is posible that currently, while the dataset is not large enough, larger models have suffered from overfit, actually decreasing their capabilities on real world problems.

Sorry for the inconvinience, I'll modify that part of the documentation. While that claim should become true in the long future, it is likely not happening now.

PD: Thanks for the sponsoring <3

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants