Oct 17, 2023

Exploring Optical Character Recognition (OCR): An Experiment with OpenCV and PyTesseract

This blog delves into grayscale and color image processing, shedding light on accuracy and challenges and uncovering OCR's vast potential.
Priyamvada
PriyamvadaSoftware Engineer - III
lines

The utility of OCR extends beyond its applications in various domains, from document digitization to text extraction in images.

In this blog, we are exploring the performance and reliability of OCR using OpenCV and PyTesseract on a diverse set of images.

The blog comprises two key phases:

  1. The first focuses on extracting text from grayscale images
  2. The second is dedicated to detecting and extracting text from color images

The results will shed light on the accuracy and challenges associated with OCR, providing insights into its potential applications and limitations.

Experimental Setup

Step 1: Library Installation

Before commencing the experiment, we installed two crucial libraries:

  • OpenCV (Open Source Computer Vision Library): An open-source library specializing in computer vision and machine learning tasks, including image processing and object detection.
  • Python-tesseract (Pytesseract): An optical character recognition (OCR) tool in Python, known for its ability to extract text from images.

Here’s how we go about it:

Step 2: Extracting Text from a Grayscale Image

Our journey commences with the extraction of text from a grayscale image. We'll begin by loading an input image from which we intend to extract text.

Screenshot 2023-10-17 at 4.31.26 PM.png

Import Libraries

Reading and Resizing the Image

We read the image and resize it to the desired dimensions.

Note: If you wish to save the resized image, you can use the Image.save() method.

Extracting Text

We employ the image_to_string method from the Pytesseract class to extract text from the image.

Here is what the output looks like :

Untitled - 2023-10-17T163522.140.png

Step 3: Detecting and Extracting Text from Color Images

Now, we venture into the process of extracting text from color images. Take a look at the example below, showcasing the color image from which we'll be extracting text:

Screenshot 2023-10-17 at 4.37.01 PM.png

Here, we will draw rectangular bounding boxes around the text using OpenCV.

Importing Libraries

Image Preprocessing

We read the image and converted it to grayscale using cv2.cvtColor.

We then convert the grayscale image into a binary image. Binary images have only two possible pixel values, often 0 for black and 1 (or 255) for white. This simplifies the information and is typically achieved through thresholding, a technique for distinguishing the foreground from the background.

Step 4: Bounding Boxes and Text Extraction

We define a rectangular kernel using cv2.getStructuringElement in OpenCV.

In this function, the first argument is the grayscale image, and the second argument is our threshold value T, which we've set to 0. That's because Otsu's method automatically calculates our optimal threshold value. The third argument is the output value when a pixel passes the threshold test.

The fourth argument is the thresholding type, which is logically combined with two methods.

The cv2.threshold function returns a tuple of two values: the threshold value T and the thresholded image itself.

Screenshot 2023-10-17 at 4.42.04 PM.png

We the create a rectangular kernel with OpenCV's cv2.getStructuringElement function. In OpenCV, you have the option to use either the cv2.getStructuringElement function or NumPy to define your structuring element.

Dilation acts like a magnifying glass for important parts of the image, making them larger. This helps connect broken text together, especially in challenging cases. We achieve this using the cv2.dilate function, which helps define text boundaries.

Screenshot 2023-10-17 at 4.45.57 PM.png

Step 5: Text Detection and Cropping

We use the cv2.findContours method to identify the areas covered by white pixels in the image.

We then draw bounding boxes around each of these areas, helping us isolate and focus on each block of text. With the bounding boxes in place, we crop out these rectangular sections, making text extraction using Pytesseract more manageable.

Here's the output image after drawing bounding boxes around text blocks.

Screenshot 2023-10-17 at 4.48.36 PM.png

And here's a snapshot of the extracted results saved in a text file.

Screenshot 2023-10-17 at 4.49.13 PM.png

Final Words

After extensive testing on various image types and formats, it becomes evident that OCR, while powerful, may only sometimes match the precision of certain commercial solutions at our disposal. Tesseract, however, shines when it encounters document images that exhibit:

  • The crisp separation between foreground text and background.
  • Proper horizontal alignment and suitable scaling.
  • High-quality image resolution.

The game-changer lies in harnessing the synergy of deep learning with OCR, which can profoundly enhance OCR accuracy, even when dealing with diverse fonts. The latest release of Tesseract introduces deep learning-based OCR, a significant leap in accuracy driven by LSTM and RNNs.

Hire our Development experts.