-
Notifications
You must be signed in to change notification settings - Fork 622
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Some errors in the 800K datasets: oversized word/char box, missing labels #101
Comments
Hi, |
@ankush-me For (1), it's hard to check whether a box is oversized. I think automatic check is an AI task itself :) but as you replied in #13 #15, it may have been caused by the fonts applied in those samples. the percentage is low. (we didn't manually count it) For (3), the percentage is a bit high, but is easy to check and correct. We print an error message every time our pre-processing program encounters this problem, and it seems like, nearly around half the samples contain such errors. ( We shuffle the sample list every time the program starts). We just ignore and discard those invalid boxes. For (2), wrong annotation includes: (a) the GT characters are different from what are actually on the image(e.g. GT chars are 'the' while chars on the image are 'HHH'). (b) there are no corresponding chars on the image while GT char list indicates there should be. p.s. the (p1,p1,p2,p2) mentioned above has no corresponding chars on the image either |
Hi, Apologies for the delay in replying. Regarding incorrect GT:
|
Hi, thanks for your reply. We checked the source code, and found out the possible reason why some chars may missing in the graph. The way your code prints chars onto the img is: (1) Compute a 'surface' according to the depth information and semantic segmentation result. When the interruption happens, chars that have already been printed are not eliminated, and the text is still kept in the set of contours that would be the ground truth. This is where the problem is. If a misunderstood it, please correct me. |
I do not really understand your interpretation of the "interruption" --- Can you point to the code where you think the text is still kept after "interruption"? |
Hi, I think the problem is here in this function.
(1) It's possible that the first Please correct me if I'm wrong. |
@Jyouhou Thank you -- yes it does seem like a bug! |
That seems like a bug, but not really a bug now! |
Hi, I am having similar problems as those discussed in #13 #15.
I am using the pre-generated 800K dataset to train a model, and found that there exist the following issues:
(1) Some word/char boxes are oversized, as discussed in #13, #15.
(2) Some word recognition annotations are wrong.
(3) There are some confusing bounding box coordinate values, e.g. negative value, coordinates that cross over the image boundary, char box coordinates that actually consist of 2 pairs of vertexes(e.g. p1,p1,p2,p2, while 4 different points are expected).
The text was updated successfully, but these errors were encountered: