-
Notifications
You must be signed in to change notification settings - Fork 22
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
training set, validation set, and test set #475
Comments
Hello Chenyong 👋 Actually, we do have calibration, validation and evaluation datasets 🙂
Datasets are given to the BIOMOD_FormatingData :
Cut of the original dataset into calibration and validation is done within BIOMOD_Modeling function through ℹ️ Note that there are several possibilities to build your calibration and validation datasets. You can find more details within the Cross-validation vignette. When calling to get_evaluations function to retrieve your evaluation values, you will see 3 columns in your output table : calibration, validation and evaluation. They will only be filled if the corresponding dataset was provided, so evaluation will be empty if you did not provide evaluation dataset to BIOMOD_FormatingData, and validation column will be empty if for example you ask to build a model with all the data ( ➡️ Note also that we have plenty of tutorial and documentation onto our website, and you can have an overview of the package functions within this presentation. So do not hesitate to have a look 👀 Hope it helps, |
Hello Chenyongpeng,
No. As the
In your picture :
Hope it helps, |
Hello, biomod2 team!
Species records are randomly divided into a training set (75% of the data) for model calibration and a test set (25% of the data) for validation.
--This is somewhat misleading. Typically there should be three dataset parts: : a training set, a validation set, and a test set. The validation set is used to fine-tune the model's hyperparameters, while the test set is used to assess the model accuracy to unseen data.
Whether biomod2 contains validation sets. If not, can you add a validation set and use techniques such as cross-validation to mitigate the overfitting problem?
The text was updated successfully, but these errors were encountered: