-
Notifications
You must be signed in to change notification settings - Fork 127
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Any work around to retain original form of words ? #5
Comments
Hello @PSanni, For your first problem, namely, retaining original form of words, I do not know how to adress it. Though, for your second question, I was able to use another dataset of my own (actually being trained). Hereby the solution I came up with. I hope it can be applied to your usecase. This project makes use of the datasets from this other project https://github.com/ku21fan/STR-Fewer-Labels, as mention in Datasets.md, with few workarounds. I thouroughly followed the instructions and was able to start a training with parseq on my own dataset. Edit: the training terminates but the test shows really inconsistent results. Maybe the .mdb file is still problematic. I am exploring this issue |
@PSanni for now, you can just directly edit and comment out Line 85 in 98959c9
Note that some preprocessed datasets have had the spaces within labels removed. For the datasets which I preprocessed (COCO, OpenVINO, TextOCR), the spaces within the labels should be intact. For fine-tuning on other datasets, you have two options:
|
- Expose normalize_unicode parameter of LmdbDataset - Add remove_whitespace flag for disabling whitespace removal in labels
I think its a good idea to include an annotation samples and required input format to the model. |
The LMDB format used is unchanged from prior work. The conversion from text labels to token IDs is handled by |
In addition to disabling whitespace (space, tabs, new line, etc.) removal, make sure you add the space character ' ' to Closing this now since all issues have been addressed already. Feel free to reopen if I missed anything. |
The model currently cannot retain the original form of words. For example, in image if words are "sunflower oil", it returns "sunfloweroil" without space. Is there any work around to address it?
Also, is it possible to fine-tune this model on other dataset such as XFUND (https://github.com/doc-analysis/XFUND) ?
The text was updated successfully, but these errors were encountered: