-
Notifications
You must be signed in to change notification settings - Fork 147
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Procedure for training a new dataset #13
Comments
Lets go through this step-by-step:
|
Thanks a lot! I am close to training it on a new dataset. However, I have one question. I have been trying to duplicate your results on the Also, can I see what happens internally in the network while training? Does the |
Your plots look quite good... I suggest letting the training run until loss/accuracy level out. Every image in the |
So, I looked at the video created by
|
By the way: which SVHN dataset are you using? |
I am using the SVHN dataset that you linked to in this comment. |
BTW, I tried running the text recognition model on an image in my dataset. I used the following commmand:
I got the following error: Is there any reason for this? I remember successfully running this once. |
Yeah, but which of the three datasets in this archive are you using? You should not use |
I am using the
|
Oh okay, Your prediction is |
Oh! I get it. I used the model that you provided with on your site. I thought it could give good results on non-FSNS datasets. I know that models aren't really interchangeable across datasets but I thought I could expect a bit more accuracy. What are the remaining dictionaries for? |
You tried to use the FSNS model for your image? Yeah that won't work -.- deep learning systems are unfortunately not good enough for such kind of transfer (at least, yet). If you try with the text recognition model, it should work better 😄 It is actually a dictionary of lists. Each key in the dictionary is the predicted text and the value to the key is a list of bboxes, with |
Ah! I thought so! I hoped that the FSNS model could at least recognize latin characters! :D I guess the only thing remaining is to train it on my own dataset. |
The FSNS model is able to recognize latin characters, but the smple you are using, is totally different from everything the model has seen before... |
Hi @rohit12 I got the same problem in your closed issues, |
you forgot to add metainformation to the label_file. This is described here. |
@Bartzi Thank you for your reply!!! Do you mean this item? |
@lmolhw5252 If you take a look at the #6, the second or the third comment from Christian will be pointing towards a download. That is the SVHN dataset that he has trained on. Add This tells the network that there are 4 text regions with a maximum of 4 characters in there. |
Hi Christian, I am getting the following error while training on a new dataset. Do you have any hints for me?
|
Yes, the input size is defined by the variable It could be that you are experiencing the same bug, as somebody else. I fixed this in the repository. The problem seems to be that the shape of the supplied labels does not fit to the shape of your predictons. |
Yes, you'll need to create the |
I am facing the same issue. I have pulled the updated repository . Kindly help.
|
The shape of the arrays does not fit, as we can see from the error. |
Thanks for the response. Yes , my data has some punctuations. I have added them in character map . |
Are there any punctuations that are not allowed in the data labels? @Bartzi |
No, there should be no restriction of punctuations. It might be difficult for the network to learn to predict the punctuation, but this might be circumvented with extra training data containing such punctuation |
Hello Christian,
I am starting a new issue here so that others who want to train your code on a new dataset in the future are able to do so.
According to me, the steps to train a new model are as follows:
char_map
for the model.How do I actually create a
char_map
? I want the model to predict words rather than individual characters. I guess I am sort of confused about the nature of thechar_map
.How should the GT be? I have an image with 4-5 text regions and each text region has some number of words in it. So should my GT file start with
5 40
indicating that there are 5 text regions with 40 characters each? What if the number of text regions that I indicate is far more than the actual number of text regions in some image?When I create the GT file, do I just mention write down each character separated by a tab? Can I include words in the GT file?
Thanks!
The text was updated successfully, but these errors were encountered: