-
Notifications
You must be signed in to change notification settings - Fork 1.1k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
problem in the output #31
Comments
You are welcome @argman , Great effort btw. As i said before i had a x1,y1 and x2,y2 so for example the annotations of the last image is 550,1667,833,1667,833,1737,550,1737,text what could be wrong ? suggestions ? |
setting text-scale 1024 , and input size 512 could make such a problem ? |
@rmmal , I think your problem is different. The definition of text maybe confusing for the network to learn. In the paper, text is defined as a single line. From your annotation, i cannot find what's the standard of a text line. |
I edited the dataset and started now training again , I will see what will happen and tell you. Thanks for your help @argman |
@rmmal any updates regarding your training & results? |
I have trained a model with this command:
python multigpu_train.py --gpu_list=0,1,2 --input_size=512 --batch_size_per_gpu=14 --checkpoint_path=/backup/EAST/
--text_scale=1024 --training_data_path=/DATA/EAST/data/ --geometry=RBOX --learning_rate=0.0001 --num_readers=12
and i've waited till:
Step 007130, model loss 0.0316, total loss 0.0827, 7.33 seconds/step, 5.73 examples/second
first Question should i make him , do more iterations or this is enough ???
second Question:
The output of all the images seems to be 1 size , why this is happening ?
i couldn't see many variations in the output dimensions
examples:
![screenshot from 2017-09-11 09-42-23](https://user-images.githubusercontent.com/6699411/30263652-1f2f0344-96d6-11e7-83b4-44ebc74edcad.png)
![screenshot from 2017-09-11 09-42-37](https://user-images.githubusercontent.com/6699411/30263654-1f372e7a-96d6-11e7-8721-2c79046c5c40.png)
![screenshot from 2017-09-11 09-42-58](https://user-images.githubusercontent.com/6699411/30263653-1f30a528-96d6-11e7-83d5-cc78907b7e25.png)
![screenshot from 2017-09-11 09-43-21](https://user-images.githubusercontent.com/6699411/30263656-1f5632b6-96d6-11e7-9310-bb8f92bf2a15.png)
so what's missing to be able to detect blocks of text ?
The text was updated successfully, but these errors were encountered: