Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Support output predict #47

Open
thachkysanh1996 opened this issue Jan 8, 2019 · 11 comments
Open

Support output predict #47

thachkysanh1996 opened this issue Jan 8, 2019 · 11 comments

Comments

@thachkysanh1996
Copy link

After trained model with your tutorial. (Using your dic.txt and fsns in data/train/tfexample.record). Then i use the model-checkpoint to infer 2 image. But the output is not right. Please help me!

screenshot from 2019-01-08 14-56-51

@A-bone1
Copy link
Owner

A-bone1 commented Jan 8, 2019

@thachkysanh1996 This should be a problem with coding and decoding. You can check the relevant information. You can also use python2 to try the correctness of the result through this command
"print( "b********".decode('utf-8'))""

@thachkysanh1996
Copy link
Author

thachkysanh1996 commented Jan 8, 2019

@A-bone1 thanks you !!!

@thachkysanh1996
Copy link
Author

@A-bone1 Does your project work on photo ocr ? or just a image cropped with text

@A-bone1
Copy link
Owner

A-bone1 commented Jan 9, 2019

@thachkysanh1996 The type of task depends on your training data. If your training data is photo ocr, the model can work on photo ocr.

@thachkysanh1996
Copy link
Author

@A-bone1 Thanks! I have another question. If training data is photo ocr with multiple text in image. How can i format the label ?

@A-bone1
Copy link
Owner

A-bone1 commented Jan 10, 2019

@thachkysanh1996 I have not tried this situation, but I think it is possible to add a text detection network in front of the text recognition network. It may be better to crop the detected text from the original image and then perform text recognition separately.

@thachkysanh1996
Copy link
Author

@A-bone1 Thanks you !!!

@thachkysanh1996
Copy link
Author

@A-bone1 I wrote the code to generate syntext data for fsns (Background white and put text on it). After training model 120k step, it will recognize the data like i syn.

ask
asdasda

But when test with cut image (in real). It cannot recognize and accuracy is very poor.
How can i make my model good prediction in actual image? How can you generate synthtext image for training? Please help me!

@A-bone1
Copy link
Owner

A-bone1 commented Jan 15, 2019

@thachkysanh1996 Is the prediction result of the test set similar to the training set accurate? If the training set differs greatly from the test set style, the accuracy is really poor. I think the generated training set should be as similar as possible to the test set. In addition, the cut image should not be distorted after reshaping.

@thachkysanh1996
Copy link
Author

@A-bone1 Yes, the test set is text image from the real world. (Different with training set which i generated like image above). The accuracy is really poor. I think that my generate is very simple, not inconsistent light condition like text image in the wild.

In addition, the cut image like this. How can i know about distortion after the model reshape image?

fsns_train_08

@tuongtranngoc
Copy link

hi ! @thachkysanh1996 how to predict a single image (ramdom size)?? what file do you run?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants