tesseract

How to read Words from Identity Card using Tesseract OCR.?

岁酱吖の 提交于 2019-12-03 13:25:51
I am working information reading from Identity Card information using Tesseract Library.I got Confidence score of each word or each line. Box[0]: x=13, y=12, w=1134, h=57, confidence: 40, text: REPUYBLIQUE FRANCAISE Box[1]: x=21, y=75, w=1119, h=50, confidence: 42, text: 7 NN99 3W F 59W Box[2]: x=17, y=137, w=539, h=52, confidence: 30, text: V7 7 D5 NOM1BOHEL Box[3]: x=6, y=189, w=954, h=46, confidence: 0, text: Box[4]: x=12, y=239, w=1016, h=34, confidence: 40, text: 5 Q HV2 H CHRISTIANL NICBLE HBNIOIJE Box[5]: x=21, y=310, w=975, h=53, confidence: 67, text: 2 E 20 06 1329 Box[6]: x=28, y=372

Open-CV - Not loading correctly

ぃ、小莉子 提交于 2019-12-03 13:17:58
I'm using Ubuntu 14.04 and I'm trying to compile this code, but I get these errors no matter what, I believe it has something to do with including the OpenCV library, but I'm not sure. Could anyone help me out? Errors: main.cc:66:37: error: ‘CV_RETR_EXTERNAL’ was not declared in this scope main.cc:66:55: error: ‘CV_CHAIN_APPROX_NONE’ was not declared in this scope main.cc:81:28: error: ‘CV_BGR2GRAY’ was not declared in this scope The Code(sorry for the formatting, I just can't get this right): #include <opencv2/imgproc.hpp> #include <opencv2/highgui/highgui.hpp> #include <tesseract/baseapi.h>

training tesseract for handwritten text

人盡茶涼 提交于 2019-12-03 12:50:00
I need to identify handwritten text (icr). No need to understand arbitrary text - I am able to instruct my users to write very clearly, with separate letters and etc. However still there will be some amount of difference between any training set and the real letters. I am hoping to train tesseract for this purpose. Has anyone tried this? Any hope in this path? You must have fonts similar to those handwriting letters. You may create them with any font designing tool(a sample is here ). Then you can follow the training process as described here . 来源: https://stackoverflow.com/questions/10763017

configure: error: leptonica library missing (when building tesseract-ocr-3.01 on MinGW)

余生颓废 提交于 2019-12-03 12:46:27
When running configure it fails with checking for leptonica... yes checking for pixCreate in -llept... no configure: error: leptonica library missing But I have leptonica 1.69 built (downloaded source and ran ./configure && make install ) Edit I think configure: error: leptonica library missing is a bit misleading, please note that it first says checking for leptonica... yes , and then fails on checking for pixCreate in -llept... no . So maybe the problem is not that the library is missing, but something else. I finally managed to make it compile, after reading this and this thread. The proper

Alternative to Tesseract OCR Training?

拟墨画扇 提交于 2019-12-03 11:33:24
问题 For the past 3 months I've been trying to train the Tesseract With identifying a collection of images I've had, due a real lack of proper documentation, and very high level of complexity I'm starting to give up on Tesseract as a solution. I'm looking for an alternative, which would be relatively pain free for training, I'm not looking to rediscover the wheel here. If there isn't anything free, I guess paid solutions would have to do (nothing above 200$) 回答1: Based on your comment, all you

chinese character recognition using Tesseract OCR

梦想的初衷 提交于 2019-12-03 10:19:54
问题 I have been using Tesseract 3.0.2 OCR SDK for image text extraction. But if I use Chinese text images and pass through OCR then Tesseract doesn't provide me the Chinese characters instead of that I am getting numeric and english characters. But I need Chinese characters as displayed in the image I am using. How can I achieve this? Is there any way I can obtain Chinese characters rather than any other characters? 回答1: You need to download chinese trained data (it will be a file like chi_sim

How can I train my Python based OCR with Tesseract to train with different National Identity Cards?

≡放荡痞女 提交于 2019-12-03 10:17:53
问题 I am working with python to make an OCR system that reads from the ID Cards and give the exact results from the image but it is not giving me the righteous answers as there are so many wrong characters that the tesseract reads. How can I train tesseract in a way that it reads the ID card perfectly and gives us the right and exact details, furthermore how can I get myself to the .tiff file and to make tesseract work for my project. 回答1: Steps to improve Pytesseract recognition: 1) Clean your

OCR match frame´s position to field in credit card

被刻印的时光 ゝ 提交于 2019-12-03 10:17:43
问题 I am developing an OCR to detect credit card. After scanning the image I get a list of words with it´s positions. Any tips/suggestions about the best approach to detect which words correspond to each field of credit card (number, date, name)? For example: position = 96.00 491.00 text = CARDHOLDER Thanks in advance 回答1: Your first problem is that most OCRs are not optimised for small amounts of text that take up most of the "page" (or card image, in your case) in spatially separated chunks.

Tesseract-OCR 3.02 with libc++

∥☆過路亽.° 提交于 2019-12-03 10:17:30
问题 Xcode 4.6, iOS SDK 6.1, tesseract-ocr 3.02 Since the last OpenCV versions are built using libc++ , and tesseract-ocr is built using libstdc++ , they can't be used together in one xcode project. So, I'm trying to build tesseract using libc++. Using the script here (updating the base sdk and deploy target to 6.1), tesseract is being built just fine, and works in my xcode project once the C++ standard library is set to the compiler default. Than, I tried altering the script to build it with libc

Floor Plan Text Recognition & OCR

本秂侑毒 提交于 2019-12-03 09:43:01
The objective is to create bounding boxes using text recognition methods (eg: OpenCV) for US floor plan images, which can then be fed into a text reader (eg: LSTM or tesseract). Several methods which have been tried cv2.findContours and cv2.boundingRect methods have been attempted but have largely failed to generalise to different types of floor plans (there is a wide deviation in how the floor plans look). For example, cv2.findContours using grayscale, adaptive thresholds, erosion and dilation (with various iterations) before applying the cv2.findContours function results in the bellow. Note