New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Model Selection via BIC and EBIC #13
Comments
There is another sample splitting based model selection approach that is an alternative to cross-validation, which I will write up as well. These BIC approaches, StARS and cross-validation together pretty exhaustively cover all model selection approaches that try to pick out one penalty parameter from a grid of penalty parameters. There is another school of model averaging, where you don't try to pick out a single optimal lambda at all, but instead try to combine multiple models. We will cover this in the weighted graphical lasso issue (#8) |
Ok, will see if I can finish the other branch and start this this week. On Sun, Jul 10, 2016 at 6:15 PM mnarayan notifications@github.com wrote:
|
Wow this is awesome, so simple. On Sun, Jul 10, 2016 at 6:09 PM mnarayan notifications@github.com wrote:
|
So when I threshold the values in the precision
I start to get reasonable results with ebic, but without some thresholding, I find that all values in the resulting matrix are nonzero (clearly). |
Are all the matrices in the path dense? What is the largest lambda value and largest covariance value? On Tue, Jul 19, 2016, 8:29 PM Jason Laska notifications@github.com wrote:
|
Ohh, are you making sure lambda path values are going from largest to On Wed, Jul 20, 2016, 5:29 PM Manjari Narayan manjari.narayan@gmail.com
|
Given sparse regularized inverse covariance estimates over a grid of regularization parameters, a popular criteria to choose the optimal penalty and corresponding estimate is to apply the Bayesian (or Swartz) Information Criterion or the BIC and the Extended BIC (EBIC) in high dimensional regimes.
The BIC criterion is defined as
BIC(lam) = -2 * Loglikelihood(Sigma, Theta(lam)) + (log n) * (# of non-zeros in Theta(lam))
The EBIC criterion is defined as
EBIC(lam) = - 2 * Loglikelihood(Sigma, Theta(lam)) + (log n) * (# of non-zeros in Theta(lam)) + 4 * (# of non-zeros in Theta(lam)) * (log p) * gam
Here
The goal is to implement model selection using the above criteria as an alternative to cross-validation.
References:
BIC in sparse inverse covariance estimation
EBIC in sparse inverse covariance estimation
The text was updated successfully, but these errors were encountered: