-
Notifications
You must be signed in to change notification settings - Fork 18
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Questions regarding the dataset and best hyper-paramters #1
Comments
Hi Felix, I should have added the hyperparameters to the readme file (which I will very soon). CMU: NA: WORLD: Why Twitter-WORLD is smaller in size than Twitter-US? Also note that the random seeds are changed so you might not get the exact results (unfortunately) but in several runs it might be a little better or worse, but in general comparable. Don't hesitate to contact me if there were more issues. Afshin |
Hi Afshin, Thank you for your instant response. I ran the first command: I ran 10 other runs with different seeds (from 1 to 10), but I still can't get an acc@161 higher than 55. I wonder if I made some mistakes. Here is how I create the environment. conda create --name geo --file requirements.txt However, I got an error when it tried to import lasagne (version 0.1). pip install --upgrade https://github.com/Theano/Theano/archive/master.zip
pip install --upgrade https://github.com/Lasagne/Lasagne/archive/master.zip After that I can run the gcnmain.py script without errors. -Felix |
Hi Felix, Regarding the Lasagne and Theano update, you're right, they should be upgraded. Regarding running with default hyperparameters, we shouldn't do that because the default hyperparameters are not suitable for all the three datasets, they're just there so that the code runs (e.g. the default hidden layer is only 100 and the bucket is 300 which should be 300 300 300 and 50, respectively for cmu). Regarding when you run with the command I had this mistake of putting -silent in the previous command that I put here, If you remove it you'll see the correct results like this: https://gist.github.com/felixgwu/8ae4c6e7a887092ae30c82fea6d6db40 Don't hesitate to send me feedback if something was still wrong, I'd love to fix errors, help. Thanks Felix. Afshin |
Hi Afshin, Thank you so much! Here is the log in case someone else also wants to reproduce it. I'll try the other two larger datasets and hopefully, I can reproduce them. -Felix |
Hi Felix, Great News. After running for the first time, the code saves the preprocessed dataset in dump.pkl in the dataset dir. Next time it loads that by default. If that file is made with incorrect hyperparameters, it'll still load it even if the new hyperparameters (e.g. bucket size) is correct. To stop it from doing that we can use -builddata option to force it to reproduce dump.pkl. Thanks a lot Felix for the typo fixes and all the other help (I'll add them to the repo asap). It made the code easier for everyone else to reproduce. Afshin |
Hi,
Thank you for sharing the code of this impressive work.
I have two questions regarding how to reproduce the results in the paper.
Based on the paper, it seems that the TWITTER-WORLD dataset is larger than the TWITTER-US; however, when I downloaded the data from this link, I found that the files in the na folder is larger than the ones in the world folder, which confuses me. I wonder if there is a naming typo here.
I tried the following command to get the GCN results with the default hyper-paramters on GEOTEXT:
python gcnmain.py -save -i cmu -d data/cmu -enc latin1
Unfortunately, I only get:
PM dev results:
PM Mean: 565 Median: 103 Acc@161: 54
PM test results:
PM Mean: 578 Median: 99 Acc@161: 53
This is a lot worse than the Mean: 546 Median: 45 Acc@161: 60 reported in the paper.
Could you please share the commands you used to produce the amazing GCN results on all three datasets in Table 1 of the paper?
The text was updated successfully, but these errors were encountered: