Fig 1 - uploaded by Sunil Kumar Kopparapu
Content may be subject to copyright.
Offline signature (bitmap image) and the corresponding online signature. Observe that the online signature has information on how the signature was written (shown by arrow direction). 

Offline signature (bitmap image) and the corresponding online signature. Observe that the online signature has information on how the signature was written (shown by arrow direction). 

Source publication
Article
Full-text available
The problem of offline to online script conversion is a challenging and an ill-posed problem. The interest in offline to online conversion exists because there are a plethora of robust algorithms in online script literature which can not be used on offline scripts. In this paper, we propose a method, based on heuristics, to extract online script in...

Context in source publication

Context 1
... B ACKGROUND Offline script recognition by computers is well-addressed in literature and still attracts attention from several researchers around the globe [1], [2]. There is rich literature in both online and offline script research. The main difference between online 1 and offline 2 signatures lies in the fact that online signature captures the manner in which the signature was written while the offline signature has no information on how the signature was generated or written. In this paper, we give a brief survey of the offline script literature. Recent surveys by Steinherz et. al [2] for offline script and Lorigo and Govindraju [1] specifically for Arabic offline script show the interest offline script recognition research carries. Rigoli [3] et al, compare the use of Hidden Markov models (HMMs) for both on-line and off-line signature verification. While HMM is a good model to capture second order statistics [3]; the use of HMM for signature verification is questionable. In practice for a given person we can have limited number of signature samples (at maximum three or four) to model the signature through its statistics. Sabourin and Drouhard [5] proposed the use of artificial neural networks (ANN) to model signatures. Neural networks, like, HMMs need a large number of samples to model the signature. Coetzer et. al [7] use discrete radon transform as the parameters to model the signature as HMMs. Leung et al [8] track features and the position of the stroke to verify signatures. Peter et al [12] use wavelet parameters to verify signatures. where Qi et. al [15] set up the problem of signature verification in a multiresolution framework. They process the signature at different resolution and then use the output at different resolutions to verify signatures. Xiao et al [16] use a modified Bayesian network approach to verify signatures. While the approach suggested in each of these references is based either on offline or on online signatures, there have been approaches suggested [17] which tend to take multi-modal cues to verify signatures 3 . In this paper, we propose a method for deriving online information from offline script. To the best of our knowledge, there is no reported work in literature that deals with procedures to convert an offline script into an online script. The closest work is the work reported by Zimmer and Ling [17] . They propose a hybrid handwritten signature verification system where the online reference data acquired through a digitizing tablet serves as the basis for the segmentation process of the corresponding scanned off-line data. Local foci of attention over the image are determined through a self-adjustable learning process in order to pinpoint the feature extraction process. Both local and global primitives are processed and the decision about the authenticity of the specimen is defined through similarity measurements. In Section II we formulate the problem of deriving online information from offline script, in Section III we give details of the procedure adopted to derive online information followed by experimental results in Section IV and conclusions in Section V. II. P ROBLEM F ORMULATION The problem that we address in this paper is one of extracting online information from an offline script in the form of a bitmap image. Essentially, we need to derive the way the script was written by looking at the final shape of the script. Clearly, this is an ill-posed problem 4 . Observe that, shown only the bitmap image and in the absence of knowledge about English script, there are several ways in which the signature could have been written. While there is no such ambiguity in determining how the signature was written when online information is available. The problem of offline to online signature conversion is to identify the way the signature was written from the bitmap image 5 . Fig. 1 shows an offline signature and its equivalent online information. The direction of the arrow shows the way the online signature was written. It also captures the number of strokes (four in this case) in the signature (pen-lifts). How ever what is not depicted is the order in which the four segments were written. In this paper we do not address the problem of identifying the order in which the segments were written. For script belonging to the same set 6 , it is possible to traverse the signature using rules based on heuristics. These derived rules would primarily be based on the knowledge of how people write that script (left to right, top to bottom, etc) in that particular language. A large portion of such rules would be based primarily on the language in which the signature is written in addition to heuristics. III. O FFLINE TO O NLINE C ONVERSION The bitmap offline signature image consists of the actual signature (the written part) and the background (the paper on which the signature was signed). The primary idea of our methodology is to intelligently traverse the written part in the bitmap image signature. We assume the script/signature to form a path/road and the traversal scheme being a truck driver who is trying to stay on the path. The driver of the vehicle steers the truck along the signature path so as to stay on the path all the time. Block diagram of the complete offline to online conversion process is shown in Fig. 2. The offline signature is obtained by scanning a signature on a paper with the help of a scanner. The original offline bitmap signature images are normally 8 bit which results in any of the pixel in the image taking a value between 0 and 255 , namely, 256 ( 2 8 ) gray pixel intensity levels. The first step in the offline to online signature conversion is to preprocess the bitmap image using some basic image processing to remove any noise that might have cropped up in the scanning process. We use a 5 × 5 median filter to remove noise. This gray level image is binarized using dynamic thresholding method. The threshold is determined by observing that the histogram plot of the gray level image would largely be a two hump plot. We take the two highest peaks 7 in the pixel- intensity histogram of the bitmap and record the corresponding intensities. The binarizing threshold is set at the intensity that lies at the middle of the two intensities. The actual signature traversal 8 is carried out on the binary image. The binarised image is then traversed using the truck driver steering his truck on the signature. The traversal process that we have adopted is influenced by the way a truck driver steers the truck when driving the truck on a road. The start of the road or the signature is determined by scanning the bitmap image from top to bottom and from left to right (probably one would adopt another strategy if one were to look at script in a different language). The first road pixel becomes the start point of the truck. Now the strategy adopted by truck driver is to steering the truck such that the truck stays on and in the middle of the road. We assume the written part of the signature to be the road to be traversed. We construct a virtual truck with two wheels, which sense if the truck is going off the road (see Fig. 3) by determining the ration of the road pixels under each of the wheels. When the truck is tending to get off the road (namely the number of road pixels under the left and right truck are not approximately same) it steers itself so that the number of pixels under both the left and the right wheel are same and hence stays on the middle of the road. This allows the virtual truck to traverse the signature. The places of intersection in the signature, where two paths cross each other, both the paths become available for the truck to steer itself. In such situations, we direct the truck to proceed in the direction in which it has been moving. In order that the truck is able to traverse a signature accurately, the two wheels of the truck should be as close to the corresponding two edges of the signature. The thickness of different signatures is not uniform because of different writing material used. This requires that the size of the truck be a function of the thickness of the signature. The size of the truck is dependent on the average width of the signature. The average signature width is calculated (see Fig. 4) as the normalized average of highest three sectional widths with respect to the x -axis. Suppose there are n sectional widths estimated in a signature image. and let X i be the sectional width of the i th section. Without loss of generality we can assume that X i is arranged in the increasing order of widths for increasing i . The average width of the signature is calculated ...

Similar publications

Article
Full-text available
Over the last decades, expression classification and face recognition have received substantial attention in computer vision and pattern recognition with more recent efforts focusing on understanding and modelling expression variations. In this paper, we present an expression classification and expression-invariant face recognition method by synthe...

Citations

... In offline handwriting, the sample is recorded in the form of a twodimensional image which takes account only the spatial information, while in online handwriting, the sample is recorded as a temporal sequence of two-dimensional coordinate points (x,y) representing the pen tip trajectory movement [1]. The conversion of the input signal from representation to another one is possible using several techniques like in [2]. ...
Conference Paper
Full-text available
The deep learning-based approaches have proven highly successful in handwriting recognition which represents a challenging task that satisfies its increasingly broad application in mobile devices. Recently, several research initiatives in the area of pattern recognition studies have been introduced. The challenge is more earnest for Arabic scripts due to the inherent cursiveness of their characters, the existence of several groups of similar shape characters, large sizes of respective alphabets, etc. In this paper, we propose an online Arabic character recognition system based on hybrid Beta-Elliptic model (BEM) and convolutional neural network (CNN) feature extractor models and combining deep bidirectional long short-term memory (DBLSTM) and support vector machine (SVM) classifiers. First, we use the extracted online and offline features to make the classification and compare the performance of single classifiers. Second, we proceed by combining the two types of feature-based systems using different combination methods to enhance the global system discriminating power. We have evaluated our system using LMCA and Online-KHATT databases. The obtained recognition rate is in a maximum of 95.48% and 91.55% for the individual systems using the two databases respectively. The combination of the on-line and off-line systems allows improving the accuracy rate to 99.11% and 93.98% using the same databases which exceed the best result for other state-of-the-art systems.