-
Hello, I want to fine-tune detection and recognition models on my custom dataset. |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment 7 replies
-
Hi @hkhanr 👋 About the recognition format, I'm not sure I understand your question, but the README does provide examples about the format. But text recognition doesn't have any localization information. Text detection does. So if you have the OCR annotations (localization + text) of some documents. For the text detection training, only remove the text information. For the text recognition, you will have to crop each localized word. And then, imagine you have a folder called "images" and put the cropped images in it:
Otherwise, end-to-end OCR training is not available in docTR :) I hope that helps! |
Beta Was this translation helpful? Give feedback.
Hi @hkhanr 👋
About the recognition format, I'm not sure I understand your question, but the README does provide examples about the format. But text recognition doesn't have any localization information. Text detection does.
So if you have the OCR annotations (localization + text) of some documents. For the text detection training, only remove the text information. For the text recognition, you will have to crop each localized word.
And then, imagine you have a folder called "images" and put the cropped images in it:
crop1.jpg, crop2.jpg, crop3.jpg
. Then your labels should be in json format in a "labels.json" next to the "images" folder, and should have this structure: