Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Can‘t achieve the scores in paper #3

Closed
PaffxAroma opened this issue Jul 22, 2021 · 9 comments
Closed

Can‘t achieve the scores in paper #3

PaffxAroma opened this issue Jul 22, 2021 · 9 comments

Comments

@PaffxAroma
Copy link

PaffxAroma commented Jul 22, 2021

I'm trying to reproduce the paper, but cann't reach 0.85 ACC runing this code on GoogleNews-S. All hyper-parameters are set to same values in paper, and the data is enhanced with contextual argumentation. The running result shows, not the model but the representation with K-Means performances better with 0.75 acc , and the model just reaches 0.62 acc. When I increase the clustering head lr, the result with model still remains 0.62 level. What should I do to improve this?

@PaffxAroma PaffxAroma changed the title Can‘t reach the results in paper Can‘t achieve the scores in paper Jul 22, 2021
@yanhan19940405
Copy link

I also did not achieve good results. After visualizing the data embedding space, I found that the sentence embedding matrix generated by SCCL does not have obvious discrimination. What happened to your follow-up?
image

@Dejiao2018
Copy link

Thanks for your interests in our work @PaffxAroma. To your questions,

  1. The reported accuracy of google-s is 83.1 instead of 85. 2) As we claimed in the paper, ACC is reported as the KMeans clustering results. 3) Also you should check how the clustering accuracy changes along the learning process. Arbitrary long learning process will result in degenerated performance.

@yanhan19940405 , can you provide more context about your plot? Is it a TSNE visualization? If so, why only one color here. Please refer to my answer to your original question. thanks

@yanhan19940405
Copy link

yanhan19940405 commented Jul 30, 2021 via email

@yanhan19940405
Copy link

yanhan19940405 commented Jul 30, 2021 via email

@Dejiao2018
Copy link

Dejiao2018 commented Jul 30, 2021

Please refer to Table 3 in our paper. Back translation does not perform well in our experiment, which we do not recommend for contrastive learning based short text clustering.

Also for your problem, in addition to the data augmentation, I do encourage you check 1) and 2) possible causes in my response to your original question #4 , which should more likely cause the problems you encounter.

As for the sentence embedding matrix, should it be (m, 768) instead, m indicates the batch size? It seems BERT-flow focuses on pairwise semantic similarity only, I'm not sure whether the statement there can generalize to categorical data. I may encourage using the TSNE plot on the (distil)bert embeddings instead.

@yanhan19940405
Copy link

yanhan19940405 commented Jul 30, 2021 via email

@yanhan19940405
Copy link

Sorry, I just noticed your last reply. Yes, 128 means embedding size, which is obtained by linear transformation from 768 dimensions. m represents the overall sample number (not clearly stated here,feel sorry)

@rajat-tech-002
Copy link

rajat-tech-002 commented Aug 16, 2021

Thanks for your interests in our work @PaffxAroma. To your questions,

  1. The reported accuracy of google-s is 83.1 instead of 85. 2) As we claimed in the paper, ACC is reported as the KMeans clustering results. 3) Also you should check how the clustering accuracy changes along the learning process. Arbitrary long learning process will result in degenerated performance.

@yanhan19940405 , can you provide more context about your plot? Is it a TSNE visualization? If so, why only one color here. Please refer to my answer to your original question. thanks

@Dejiao2018 . The idea in the paper in quite good. I like the approach.
So, all the results reported in the paper are with Bert Embeddings and Kmeans? Rather than with Clustering Head? What was the reason for not reporting results with Cluster Head? Was the ACC with Clustering Head always less than with Kmeans? Thanks

@1085737319
Copy link

I'm trying to reproduce the paper, but cann't reach 0.85 ACC runing this code on GoogleNews-S. All hyper-parameters are set to same values in paper, and the data is enhanced with contextual argumentation. The running result shows, not the model but the representation with K-Means performances better with 0.75 acc , and the model just reaches 0.62 acc. When I increase the clustering head lr, the result with model still remains 0.62 level. What should I do to improve this?

What is the parameter setting of the data set SearchSnippets?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

5 participants