Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Amplitudes of the GMM #4

Open
erlebach opened this issue Sep 23, 2017 · 3 comments
Open

Amplitudes of the GMM #4

erlebach opened this issue Sep 23, 2017 · 3 comments

Comments

@erlebach
Copy link

Hi,

How do you ensure that the amplitudes of the GMM (pi[i]) remain positive and sum to unity, when minimizing the cost function with respect to these amplitudes? Thanks.

Gordon
@slim1017
Copy link
Owner

Hi,

  1. pi remain positive:
    Empirically,we found that pi always remain positive during training (note that there is a log pi in loss function, maybe this is the reason), so we don't need to worry about the constraint pi >= 0;

  2. sum to unity:
    we could normalize the pi vector to make its sum equal one after each batch or epoch or entire training. The three ways achieve similar accuracy.

The parameter pi has much less impact on model training and clustering accuracy than the mean and variance parameters of GMM.

@daib13
Copy link

daib13 commented Sep 25, 2017

Do you think using a softmax layer to produce pi is a better way or a worse way? Because sofmax layer ensures that pi is always positive and sum to 1. But it may cause problems like gradient vanishing or other problems in SGD.

@colobas
Copy link

colobas commented May 17, 2019

Hi

The parameter pi has much less impact on model training and clustering accuracy than the mean and variance parameters of GMM.

Does this hold even for datasets where the clusters aren't balanced? It seems odd.

Also, if this is the case, couldn't one fix the the gmm_pis?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants