Every business is subject to know their customer (KYC), In order to detect and avoid any fraud in financial transactions. When this KYC is done electronically then this process is known as e-KYC. Now a day from banking to telecom organization to other business e-KYC has become an inevitable feature to have. At ValueFirst we also understand this and our Chabot offering (Surbo) has taken care of it. Our offering based upon the hybrid solution, which can extract textual information out of provided identity card images. This solution is an amalgamation of standard image processing techniques and deep learning models.Our solution based on 5 sequential steps, which start from the reshaping the image to standard size after that, in second step solution detects the edges in the images, in third step houghines are identified on detected edges and image get rotated if it is tilted and in the fourth and fifth step text detection and text recognition is done on Identity card respectively. We got around 85% of accuracy by our implemented solution.
The implemented approach for automatic text extraction from identity cards is consisting of five steps as discussed below.
Text extraction steps in implemented solutions
Following is the description of the steps in our implemented solution.
As discussed in the step descriptions that an identity card has to go through all the five phases (Reshape, Edge Detection, Hough Lines and Angle correction, Text detection, and Text recognition) to extract text out of the image. Following is the heuristic for the implemented solution.Image→(Reshaping| Edge detection |Hough Lines|Angel Correction|Textdetection|Text Recoginition)→Text
Equation 2. 1 Heuristic of the implemented solution
In order to give better intuition of what is happening behind the scene, we have described our implementation in the form of heuristic. Following is the heuristic of our implemented solution.
Equation 2. 2 Heuristic for our solution.
Description of each step mentioned in this above heuristic already discussed in the step description section.
As identity cards hold sensitive and personal information so there no public dataset available for Identity cards. It is very difficult to collect a reasonable amount of personal identity cards within an organization. In order to overcome this challenge, we have collected a few identity cards and then generated a reasonable amount of testing identity cards using image augmentation techniques. This image augmentation technique involves brightness changes in the images, height, and width shift in the images and horizontally flipping of the images.
We have evaluated our implementation solution on 56 different identity cards rotated at a different angle, with different lighting conditions and with different zoom sizes, etc. When our implemented solution is fed with an image it produces an image with corrected orientation, an identified text area with a red color box on the image, and with the text recognized text. The input and output images sample is shown in figure 3.1.
Sample demonstration of our solution
On the testing dataset, our solution touches 85% accuracy on a diverse set of identity cards (these identity cards involve Aadhar cards, Pan Cards, and driving license). Following is the table which shows the accuracy of our implemented solution on various fields (these specific fields are chosen as they common in most of the identity card on which testing has run). Text ColumnAccuracy#ID0.830233DOB0.847619FatherName0.850000Name0.837209
Text recognition accuracy
Following the bar, the plot shows the accuracy we get on different fields with a threshold line at 80%.
Barchart for accuracy on different identity card text field.
As demonstrated by bar chart for all of the field bars crosses over the red threshold line. It means the solution is able to identify mentioned identity cards fields correctly with an accuracy of more than 80%.
Traditionally the Identity card recognition was done using the image processing the technique, which is good for reading the text from the straight identity card images or the identity card which are not having varying resolutions, rotation tilts. Moreover, different lighting conditions tend to provide different outputs in traditional solutions. But in our implemented solution different resolutions of Identity cards and orientation of the images are automatically taken care of. For correcting the resolutions our system converts provided identity card to standard resolution and for correcting the orientation our system has used CNN based model and Hough lines techniques. The amalgamation of different phases in the implemented model has provided an accuracy of around 85%.
However, the implemented solution is giving an accuracy of around 85%. Yet there are the areas where we can make further improvements in the implemented solution. These areas are as follows:
Reach out to us on marketing@vfirst.com for any assistance and information.