Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Cuda gpu #3

Closed
mariembenslama opened this issue Apr 27, 2019 · 4 comments
Closed

Cuda gpu #3

mariembenslama opened this issue Apr 27, 2019 · 4 comments

Comments

@mariembenslama
Copy link

mariembenslama commented Apr 27, 2019

Hello,

I have a RAM with 13GB, I activated cuda in the params.py file but the training and test are still slow comparing to the capacity of my machine.

It's supposed to run quickly I mean.

I'm wondering if cuda is actually working or not in reality?

@Holmeyoung
Copy link
Owner

@mariembenslama since our displayInterval is set to 100 and batchSize is set to 64, it means the training loss will be printed every 6400 samples. You can set the displayInterval to 10 then it may seems ten times quickly. Of course, you can use the command watch -n 1 nvidia-smi to checkout the using case of GPU, it will show something like

Sun Apr 28 17:44:06 2019
+-----------------------------------------------------------------------------+
| NVIDIA-SMI 410.104	  Driver Version: 410.104      CUDA Version: 10.0     |
|-------------------------------+----------------------+----------------------+
| GPU  Name        Persistence-M| Bus-Id        Disp.A | Volatile Uncorr. ECC |
| Fan  Temp  Perf  Pwr:Usage/Cap|         Memory-Usage | GPU-Util  Compute M. |
|===============================+======================+======================|
|   0  Tesla P40           Off  | 00000000:3B:00.0 Off |                    0 |
| N/A   36C    P0   128W / 250W |   1063MiB / 22919MiB |     66%      Default |
+-------------------------------+----------------------+----------------------+

+-----------------------------------------------------------------------------+
| Processes:                                                       GPU Memory |
|  GPU       PID   Type   Process name                             Usage      |
|=============================================================================|
|    0    249645      C   python                                      1053MiB |
+-----------------------------------------------------------------------------+

If there are still problems, please let me know, thank you~

@mariembenslama
Copy link
Author

mariembenslama commented Apr 28, 2019

Thanks! I guess it's good now :) it jumped quickly to a good intervalle of accuracy and loss ^-^

Thank youuu~~~
Is the model (DL in general) always taking too long to grasp all the alphabets?
I have about 3340 alphabets, at epoch 9/1000 Loss is 0.5456 and Accuracy is 0.004 but it's growing up.
Right now I have 100000 train and 10000 test, I will train this sample, add a checkpoint, then add more data and train again.

Do you think by completing the train it'll give a good accuracy in the end?

Sans titre

@Holmeyoung
Copy link
Owner

@mariembenslama I believe you can do it! But you can add more training data to it to let it know the right decline direction. If you have more training data, after 9/1000 epoch, the loss can be lower! Good luck!

@mariembenslama
Copy link
Author

Thank you! I will do as you say 😊

Thanks a lot again 😀😀😀

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants