Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

GaussianProcess score function #29

Merged
4 commits merged into from Dec 13, 2010
Merged

GaussianProcess score function #29

4 commits merged into from Dec 13, 2010

Conversation

dubourg
Copy link
Contributor

@dubourg dubourg commented Dec 9, 2010

Hi list,

After a discussion with Alex (Gramfort) who reviewed the gaussian_process module, we decided to re-implement the score function so that it natively performs a leave-one-out estimate of the determination coefficient (as usually done with this kind of model) without exposing the internals to the user.
(See agramfort@f068df4)

There is an example of usage of this new score function in the plot_gp_diabetes_dataset.py example.

Cheers,
Vincent

agramfort and others added 2 commits November 28, 2010 22:31
- a few cosmits in the examples
- implementation of real built-in score function for the GaussianProcess which evaluates a leave-one-out estimate of the determination coefficient (as usually done w/ this kind of model).
@ogrisel
Copy link
Member

ogrisel commented Dec 11, 2010

@dubourg what is the status of this pull request?

Conflicts:
	examples/gaussian_process/plot_gp_diabetes_dataset.py
	examples/gaussian_process/plot_gp_probabilistic_classification_after_regression.py
@agramfort
Copy link
Member

merged !

thanks

@dubourg
Copy link
Contributor Author

dubourg commented Dec 13, 2010

Thanks for the mege.
I can see that you renamed the plot_gp_diabetes_dataset example as it does not produce any plot anymore.

But I'd like to review the implementation of the cross_val_score function to address the problem discussed last night: one cannot perform a leave-one-out estimate on the default Regressor's score function as inhereted from base.RegressorMixin. Indeed, the r2_score(y_true, y_pred) would return a -Inf array as it computes the variance on a single sample, thus y_true -y_true.mean() = 0 and finally R2 = -Inf...

I think that a cross_val_predict function could do the trick for all regressors in the scikits, and it would allow to make adequation plots: y_pred_on_folds vs. y_true in addition to the cross_val score estimates.

GaelVaroquaux pushed a commit that referenced this pull request Feb 4, 2013
FIX: some of Gael comments
jnothman pushed a commit to jnothman/scikit-learn that referenced this pull request Nov 20, 2016
ENH : NumpyDocString subclass collections.Mapping
This pull request was closed.
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

None yet

3 participants