-
Notifications
You must be signed in to change notification settings - Fork 255
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
I have few questions. #5
Comments
@GodOfSmallThings |
@BowieHsu |
@GodOfSmallThings It may help, you can try to do some experiment. |
Pretraining on SynthText will help. @BowieHsu: when SynthText is added for pretraining, an fmean of 85% can be achieved. http://rrc.cvc.uab.es/?ch=4&com=evaluation&task=1&e=1&f=1&d=0&p=0&s=1 However, quite a lot more iterations are needed. |
Thank you very much ! |
dude,1e-2 learning rate will improve F-score |
1e-2 learning rate caused loss = Nan, so I change it to 1e-3.... |
@lizzyYL so you need train in 1e-3 learning rate for 100 iter and then just train it in 1e-2. |
@BowieHsu Yes, that's what I did the first time. Running train.sh, but loss=nan at 3w iters. |
@lizzyYL I also tried at batch 8 for first time, and got Nan. When batch 24, Nan didn't occurred and performance went up to 82.4%. |
@GodOfSmallThings Understand! Thank you for your reply~ |
@lizzyYL @dengdan @BowieHsu @GodOfSmallThings @comzyh |
@tsing-cv so,have you print the pixellink results in test_any_image.py? |
@tsing-cv |
@dengdan |
@GodOfSmallThings I meet the same problem, After training on SynthText for about 10W iter using 1e-2 learning rate and then finetune on ICDAR2015, the F-score stay 0 forever. |
@dengdan @GodOfSmallThings another thing, I tried add BN ops after every conv layer which leads F-score drop about 5~6%. |
@tsing-cv
You'll probably see at some point training is working. Find the point where training fails, checking the result every 10000 times. @BowieHsu |
@GodOfSmallThings have you tried to use staircase learning rate instead of fix it at 1e-2, I also tried Adam, perform bad, I think the reason should be upscale 11 convolution, the channel of those 11 convolution are too few which leads network unstable. |
@BowieHsu |
Anybody encountered this problem? Traceback (most recent call last):
File "/home/a??/anaconda2/envs/caffe_tf/lib/python3.6/site-packages/tensorflow/python/framework/op_def_library.py", line 510, in _apply_op_helper
preferred_dtype=default_dtype)
File "/home/a??/anaconda2/envs/caffe_tf/lib/python3.6/site-packages/tensorflow/python/framework/ops.py", line 926, in internal_convert_to_tensor
ret = conversion_func(value, dtype=dtype, name=name, as_ref=as_ref)
File "/home/a??/anaconda2/envs/caffe_tf/lib/python3.6/site-packages/tensorflow/python/framework/constant_op.py", line 229, in _constant_tensor_conversion_function
return constant(v, dtype=dtype, name=name)
File "/home/a??/anaconda2/envs/caffe_tf/lib/python3.6/site-packages/tensorflow/python/framework/constant_op.py", line 208, in constant
value, dtype=dtype, shape=shape, verify_shape=verify_shape))
File "/home/a??/anaconda2/envs/caffe_tf/lib/python3.6/site-packages/tensorflow/python/framework/tensor_util.py", line 383, in make_tensor_proto
_AssertCompatible(values, dtype)
File "/home/a??/anaconda2/envs/caffe_tf/lib/python3.6/site-packages/tensorflow/python/framework/tensor_util.py", line 303, in _AssertCompatible
(dtype.name, repr(mismatch), type(mismatch).__name__))
TypeError: Expected int32, got 8.0 of type 'float' instead.
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "train_pixel_link.py", line 294, in <module>
tf.app.run()
File "/home/a??/anaconda2/envs/caffe_tf/lib/python3.6/site-packages/tensorflow/python/platform/app.py", line 48, in run
_sys.exit(main(_sys.argv[:1] + flags_passthrough))
File "train_pixel_link.py", line 287, in main
batch_queue = create_dataset_batch_queue(dataset)
File "train_pixel_link.py", line 153, in create_dataset_batch_queue
capacity = 500)
File "/home/a??/anaconda2/envs/caffe_tf/lib/python3.6/site-packages/tensorflow/python/training/input.py", line 927, in batch
name=name)
File "/home/a??/anaconda2/envs/caffe_tf/lib/python3.6/site-packages/tensorflow/python/training/input.py", line 722, in _batch
dequeued = queue.dequeue_many(batch_size, name=name)
File "/home/a??/anaconda2/envs/caffe_tf/lib/python3.6/site-packages/tensorflow/python/ops/data_flow_ops.py", line 464, in dequeue_many
self._queue_ref, n=n, component_types=self._dtypes, name=name)
File "/home/a??/anaconda2/envs/caffe_tf/lib/python3.6/site-packages/tensorflow/python/ops/gen_data_flow_ops.py", line 2418, in _queue_dequeue_many_v2
component_types=component_types, timeout_ms=timeout_ms, name=name)
File "/home/a??/anaconda2/envs/caffe_tf/lib/python3.6/site-packages/tensorflow/python/framework/op_def_library.py", line 519, in _apply_op_helper
repr(values), type(values).__name__))
TypeError: Expected int32 passed to parameter 'n' of op 'QueueDequeueManyV2', got 8.0 of type 'float' instead. |
@tsing-cv |
Why it always give these notes, while training? All Pred BBoxes cannot drawn in the picture? @GodOfSmallThings, @dengdan, @BowieHsu |
@tsing-cv Sorry, I don't know. I didn't dig to that problem yet. |
@small-wong thank you |
Hi, can anyone obtain the result of 83.7%? |
i am training the model to detect English text in documents, how can i make sure model has converged? i try to finetune the model author provided by my images, what can loss reach? it converges so slow... @GodOfSmallThings @BowieHsu @dengdan @lizzyYL @tsing-cv |
@cjt222 you can download icdar accuracy calculate bash to test whether your model has converged. |
@small-wong Thank you ! |
how can i train model by vgg pretrain model?i download vgg model from |
@cjt222 If you want to use pretrained model, you should write a loading model code file. |
@small-wong can you share the model again? the model to detect Chinese text |
@tsing-cv Bounding box (-19,538,78,592) is completely outside the image and will not be drawn. |
@GodOfSmallThings At the time you reaching the result of 83.7%, Can you share the loss? I still get around 0.4-0.5 and when i used model to predict, it predict empty box. |
config.batch_size_per_gpu = int(config.batch_size_per_gpu) |
I have few questions noted below.
Thank you!
The text was updated successfully, but these errors were encountered: