-
Notifications
You must be signed in to change notification settings - Fork 202
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
make lists -j32
doesn't seem to be honoring the thread count. (Also happens when calling make training -j32
)
#382
Comments
Worked around this with the below bash scripting to speed things up:
|
I always used |
Hi - not necessarily the answer you were looking for, but Tesstrain is essentially a wrapper to help you run a sequence of Tesseract binaries with hopefully the correct parameters. Here is my way to significantly speed up the development process -
|
@stweil do you have an example makefile for training without lstmf files? |
Hi team,
I'm training a model on some font with
START_MODEL=eng
and while the resulting .traineddata can correctly guess a lot of things with the font there are some which trip it up. It was only trained on a couple thousand lines.To try and lazily solve this problem I'm trying again with far more training lines than previously. (160k; very overkill and likely pointless cycles to train).
During
make training
I've noticed that many preparation steps take place in parallel but it seems thelists
step is callingtesseract data/font-ground-truth/abc_00001.tif data/font-ground-truth/abc_00001 --psm 13 lstm.train
one .tif at a time.In my limited experience with this software this seems like a step that could be run concurrently and would help speed up the initial data preparation step in getting to the actual training part of the process without having to resort to scripting.
Is it possible to make this training preparation step run in parallel with multiple
-jxx
jobs?The text was updated successfully, but these errors were encountered: