all repos — videocr @ 94e2ca5b8e015762efb529c2ffc1f917a8db7497

Extract hardcoded subtitles from videos using machine learning

update README
Yi Ge me@yige.ch
Mon, 29 Apr 2019 17:02:35 +0200
commit

94e2ca5b8e015762efb529c2ffc1f917a8db7497

parent

3f6af992a507da137945d1d0a0864a2678635356

1 files changed, 4 insertions(+), 4 deletions(-)

jump to
M README.mdREADME.md

@@ -67,11 +67,11 @@ time_end='', conf_threshold=65, sim_threshold=90, use_fullframe=False)

``` Write subtitles to `file_path`. If the file does not exist, it will be created automatically. -## Adjustable Parameters +### Parameters - `lang` - Language of the subtitles in the video. + Language of the subtitles in the video. Besides `eng` for English, all language codes on [this page](https://github.com/tesseract-ocr/tessdata_best/tree/master/script) are supported. - `time_start` and `time_end`

@@ -87,8 +87,8 @@ - `sim_threshold`

Similarity threshold for subtitle lines. Neighbouring subtitles with larger [Levenshtein](https://en.wikipedia.org/wiki/Levenshtein_distance) ratios than this threshold will be merged together. The default value is fine for most cases. - Make it closer to 0 if you get too few subtitle lines, or make it closer to 100 if you get too many duplicated subtitles. + Make it closer to 0 if you get too many duplicated subtitle lines, or make it closer to 100 if you get too few subtitle lines. - `use_fullframe` - By default, only the bottom half of each frame is used for OCR in order to reduce errors. You can explicitly make the algorithm handle the full frame. + By default, only the bottom half of each frame is used for OCR. You can explicitly use the full frame if your subtitles are not within the bottom half of each frame.