-
Notifications
You must be signed in to change notification settings - Fork 1.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
How to shuffle data for training? #29
Comments
The data is automatically shuffled for you. I was constantly forgetting to do that with CLI caffe, so I went ahead and made it the default. Are you saying that converging quickly is a problem? I'm not sure I understand your issue. |
I created database, by parsing the directory with the next structure:
Where pos - images with pedestrian, and neg - without. For training I choosed simple LeNet architecture. As you can see accuracy is very good. But when I try to test on some samples, CNN all time classifies as neg class (see figure below). |
Ok so in your Accuracy/Loss graph, you can see that the validation loss is much higher than your training loss. That usually means one of two things:
In general, those "standard networks" are not expected to solve all of your problems. They are simply a starting point to give you a few examples of the types of neural network architectures that have been used before. Once you find that neither of them solves your problem perfectly (and they usually won't), then you'll have to step into the world of designing a neural network yourself. That's the whole point of DIGITS, really - making it easier to try different model architectures and get a feel for if they're working quickly, while helping you avoid making some of the more mundane mistakes like syntax errors. |
For posterity, here is what it looks like when a model (AlexNet, in this case) fits a dataset (VOC 2007, in this case) relatively well. This dataset is really hard (widely varying image scales, huge category skew), so the accuracy doesn't go very high, but you can see that the loss on the training set and the validation set are both decreasing (so learning is happening), and they are staying close together (so the function you are learning is a good predictor for images in the validation set as well as the training set). |
Hi there! I took LeNet architecture only as base, I've changed it for my data. Input size is 64x128, two outputs. Pedestrian data - 20k I ask only because it look like all labels in created database are equal 0. |
@drozdvadym, can you open a new issue? Thanks for reporting it - that is indeed a bug and needs to be fixed, but it doesn't belong in this thread. |
Ok |
I am not sure, but all networks that was trained on two classes, converge very fast. And for the several first epochs, network recognizes only negative classes. Maybe it due to not shuffled data?
The text was updated successfully, but these errors were encountered: