A Comprehensive Survey on Different Techniques and Applications of Digital Image Processing

13
75 Proc. of Int. Conf. on Computing, Communication & Manufacturing 2014 A Comprehensive Survey on Different Techniques and Applications of Digital Image Processing Satadal Saha 1 , Subhadip Basu 2 , and Mita Nasipuri 2 1 ECE Dept., MCKV Institute of Engg., Liluah, Howhah, India Email: [email protected] 2 CSE Dept., Jadavpur University, Kolkata, India Email: {subhadip, mnasipuri}@cse.jdvu.ac.in Abstract—Digital image found its first application almost hundred years back. But the evolution of digital image processing is intimately tied to the development of the digital computer. Because of its inherent nature, digital image requires so much storage space and their processing requires so much computational power that progress in the field of digital image processing had been highly dependent on the development of modern digital computer which came only in 1940s. Till then digital image processing has become a key area of research to a group of people. Today it has enormous applications in the fields of medical science, biotechnology, transportation management, document analysis and retrieval, outdoor surveillance, remote sensing etc. This paper presents a brief introduction to digital image processing, followed by a comprehensive survey of its various techniques and applications. Index Terms—digital image processing, computer vision, object detection, text localization, human motion detection, image segmentation Introduction Digital Image Processing refers to the field of processing digital images with the help of a digital computer. A digital image is composed of a finite number of elements commonly referred to as picture elements, image elements, pels or pixels. Pixel is the term most widely used to denote the basic element of a digital image. Like human imaging system, digital imaging devices can acquire digital images and store them for processing. But unlike human who can capture images only for visible band of electromagnetic spectrum, imaging devices can capture images covering whole range of electromagnetic spectrum. Thus, digital image processing encompasses a wide variety of application areas in the domains of medical science, remote sensing, traffic monitoring, document analysis and retrieval etc. On the other hand, Computer Vision is another domain of research area, where the ultimate objective is to use computer system to emulate human vision by learning from the environment and being able to make inferences on specific situations and take necessary actions on the basis of the inferences. This area includes another branch of expertise known as Artificial Intelligence (AI) whose objective is to emulate human intelligence. The area of Image Analysis or Image Understanding is in between image processing and computer vision. There is no general agreement among research community regarding the borderline between image processing, image analysis and computer vision. Sometimes a distinction is made by defining image processing as a discipline in which both the input and output of the process are images. It is a process that involves primitive operations such as reduction of noise, contrast enhancement, and image sharpening etc. Image analysis is a process that is characterized by the fact that its inputs are generally images, but its outputs are attributes extracted from those images (e.g., edges, contours, and the identity of individual objects). And finally, computer vision may be defined as a process that involves “making sense” of an ensemble of

Transcript of A Comprehensive Survey on Different Techniques and Applications of Digital Image Processing

  75  

Proc. of Int. Conf. on

Computing, Communication & Manufacturing 2014

A Comprehensive Survey on Different Techniques

and Applications of Digital Image Processing Satadal Saha1, Subhadip Basu2, and Mita Nasipuri2

1ECE Dept., MCKV Institute of Engg., Liluah, Howhah, India Email: [email protected]

2CSE Dept., Jadavpur University, Kolkata, India Email: {subhadip, mnasipuri}@cse.jdvu.ac.in

Abstract—Digital image found its first application almost hundred years back. But the evolution of digital image processing is intimately tied to the development of the digital computer. Because of its inherent nature, digital image requires so much storage space and their processing requires so much computational power that progress in the field of digital image processing had been highly dependent on the development of modern digital computer which came only in 1940s. Till then digital image processing has become a key area of research to a group of people. Today it has enormous applications in the fields of medical science, biotechnology, transportation management, document analysis and retrieval, outdoor surveillance, remote sensing etc. This paper presents a brief introduction to digital image processing, followed by a comprehensive survey of its various techniques and applications. Index Terms—digital image processing, computer vision, object detection, text localization, human motion detection, image segmentation

Introduction Digital Image Processing refers to the field of processing digital images with the help of a digital computer. A digital image is composed of a finite number of elements commonly referred to as picture elements, image elements, pels or pixels. Pixel is the term most widely used to denote the basic element of a digital image. Like human imaging system, digital imaging devices can acquire digital images and store them for processing. But unlike human who can capture images only for visible band of electromagnetic spectrum, imaging devices can capture images covering whole range of electromagnetic spectrum. Thus, digital image processing encompasses a wide variety of application areas in the domains of medical science, remote sensing, traffic monitoring, document analysis and retrieval etc.

On the other hand, Computer Vision is another domain of research area, where the ultimate objective is to use computer system to emulate human vision by learning from the environment and being able to make inferences on specific situations and take necessary actions on the basis of the inferences. This area includes another branch of expertise known as Artificial Intelligence (AI) whose objective is to emulate human intelligence. The area of Image Analysis or Image Understanding is in between image processing and computer vision. There is no general agreement among research community regarding the borderline between image processing, image analysis and computer vision. Sometimes a distinction is made by defining image processing as a discipline in which both the input and output of the process are images. It is a process that involves primitive operations such as reduction of noise, contrast enhancement, and image sharpening etc. Image analysis is a process that is characterized by the fact that its inputs are generally images, but its outputs are attributes extracted from those images (e.g., edges, contours, and the identity of individual objects). And finally, computer vision may be defined as a process that involves “making sense” of an ensemble of

  76  

recognized objects, as in image analysis and performing the cognitive functions normally associated with vision.

Based on the preceding discussions, it is seen that a logical place of overlap between image processing and image analysis is the area of recognition of individual regions or objects within an image. Thus, in a broader sense, digital image processing encompasses processes whose inputs and outputs are images and also the processes that extract attributes from images, up to and including the recognition of individual objects. To clarify the concept further, let us consider the problem of automated recognition of text within a general scene image. The processes of acquiring an image containing the text, pre-processing that image, segmenting the individual characters, describing the characters in the form of feature values suitable for computer processing and recognizing those individual characters can well be considered within the scope of digital image processing. Making sense of the content of the image may be viewed as being in the domain of image analysis and even computer vision, depending on the level of complexity of the problem corresponding to the level of expected solution implied by the statement “making sense”.

Applications of Digital Image Processing There are numerous applications of digital image processing staring from early 20th century till now. One of the oldest applications of digital images, as reported in literatures, was sending pictures by submarine cable between London and New York. In 1920, introduction of the Bartlane cable picture transmission system improved the time required to transport a picture across the Atlantic ocean [1]. Specialized printing equipment coded the pictures for cable transmission and then the images were reconstructed at the receiving end. Visual quality of these early digital pictures was improved towards the end of 1921 by introducing a technique based on photographic reproduction. The early Bartlane systems were capable of coding images in five distinct grey levels, which was increased to 15 grey levels in 1929. Although the examples just cited involve digital images, the evolution of digital image processing is intimately tied to the development of the digital computer. Because of its inherent nature, digital images require so much storage space and their processing requires so much computational power that progress in the field of digital image processing had been highly dependent on the development of digital computers. Though the idea of a computer was conceived long back, what is called a modern digital computer came only in 1940s with the introduction of two key concepts by John von Neumann: (1) a memory to hold a stored program and data, and (2) conditional branching. These two ideas are the foundation of a central processing unit (CPU), which is the heart of computers today. Starting with the von Neumann architecture, there were a series of key advances from the invention of the transistor by Bell Laboratories in 1948 to the present use of ultra large scale integration (ULSI) in 1980s that led to computers powerful enough to be used for digital image processing. The first computers powerful enough to carry out meaningful image processing tasks appeared in the early 1960s and the first potential digital image processing using those computers for improving images from a space probe began at the Jet Propulsion Laboratory, California in 1964 when the pictures of the moon, transmitted by Ranger 7, were processed by a computer to correct various types of image distortions [2]. The invention of computed tomography (CT) in the early 1970s is one of the most important events in the application of image processing in medical science. Another emerging area of application is image registration, which is the process of aligning two or more images (the reference and sensed images) of the same scene taken at different time, from different viewpoints, and/or by different sensors. This has enormous applications in medical science field and others [3–5]. From the 1960s until the present days, the field of image processing has grown rapidly [6–8]. In addition to applications in medicine and in the space programs, digital image processing techniques are now used in a broad range of applications. Study of the aerial and satellite imagery is another area of the application of image processing. In physics and allied fields, image processing techniques are used to enhance resulting images of the experiments in areas like high-energy plasmas and electron microscopy. Archaeological use of image processing includes restoration of blurred images that were the only available records of rare artefacts lost or damaged after being photographed. Other successful applications of image processing techniques can be found in the field of bioinformatics, astronomy, chemistry, nuclear medicine, law enforcement, defence, industrial applications etc. The aforesaid examples illustrate the cases where the results of processing are intended for human interpretation.

Another major area of application of digital image processing techniques is in solving problems dealing with machine perception which may have little resemblance to visual features that humans use in interpreting the content of an image. Biometric recognition is an important and wide application area where digital image processing plays an important role. Automatic face recognition [9], fingerprint recognition, palm

  77  

recognition, retina recognition have been the well researched areas in this domain. Content based image retrieval [10] is another domain of research in which images are classified and indexed according their content such that the desired class of images can be retrieved as and when needed in future. Sign language [11] and attitude analysis [12] are the areas where lexical meaning of arm/body movement and facial expression are analyzed and high level semantics are inferred. Other typical areas of machine perception that utilize image processing techniques are optical character recognition (OCR), robotics, military signal processing, meteorology and environmental assessment.

Human motion tracking With more advancement in image processing techniques, evolution in electronic devices and development of highly efficient sensors and imaging devices, automatic surveillance has become a growing demand from different administrative authorities over the globe. Human motion tracking is one of the needs in this domain. Many survey reports [13–15] have been published on computer vision based human motion tracking. Real time tracking of human motion has been reported in [16–19]. A 3D-model based human motion tracking has been published in [20]. An analysis and synthesis of human movement is reported in [21], [22]. Surveillance in virtual reality is also reported in different literatures [23–25]. Tracking human motion in a scene and counting number of interactive people has been reported in [26], [27]. Other related applications like gesture recognition, prediction of motion trajectories and robotics are also important research areas of digital image processing [28–30].

Pedestrian behaviour analysis Another application area related to human motion tracking is pedestrian behaviour analysis and collision avoidance. A report on Pedestrian Collision Avoidance Systems is reported in [31]. A generalized way of object motion detection has been reported in [32], [33]. Technical report on pedestrian behaviour analysis has been published in a number of literatures [34–42]. Along with pedestrian behaviour and motion tracking, prediction of trajectory of pedestrian leading to collision with on-road vehicles has also been an area of research in vehicular technology [43], [44]. Tracking the pedestrian [45], [46] is not only important in day light but also at night at very low light condition, in which case normal surveillance cameras may not capture quality images. Night vision cameras are used for this purpose and digital image processing techniques are used on the images captured from night vision devices [47–50].

Document image processing Document image processing has become a growing area of research because of the requirement of automatic processing of scanned document images in the field of content-based information indexing and retrieval, automatic annotation and structuring of images. Text extraction involves detection, localization, tracking, enhancement and recognition of the text from a given image. However, variations of text due to differences in size, style, orientation, and alignment, as well as low image contrast and complex background make the problem of automatic text extraction extremely difficult and challenging job [51]. Lot of research work has already been done in this field during last couple of decades. Poignant et al. [52] presented a video OCR system that detects and recognizes overlaid texts in video as well as its application to person identification in video documents. An end-to-end real-time scene text localization and recognition method is proposed in [53].

Texts contain textural features, which are normally classified using Support Vector Machine (SVM) or Multilayer Perceptron (MLP) for the detection of regions as text or non-text. Random forest is another classifier which has the ability of maintaining accuracy in small number of labelled dataset and providing good performance in unbalanced dataset as in the case of unbalanced text and non-text distribution. In [54], a method of merging different random forests trained with different kinds of features is proposed to improve the accuracy of classification.

Gupta et al. [55] applied discrete wavelet transform (DWT) for extracting text region from complex background. In another work [56], Haar Discrete Wavelet Transform is employed. Liang et al. [57] also used

  78  

DWT to detect edges of candidate text regions. The resulted detail component sub-bands contain both text and non-text edges. Then thresholding is used on the intensity of the edges to distinguish text and non-text edges. Text regions are composed of vertical, horizontal and diagonal edges. Morphological dilation operators are applied to connect isolated text edges of each detail component sub-band in a transformed binary image. According to the experimental results, real text regions are the overlapped portion of three kinds of dilated edges. Hence, logical AND operator is applied to three kinds of dilated edges and final text regions are obtained.

Stroke width and local contrast are used by Subramanian et al. [58] to develop a fast, simple, and effective algorithm to detect character strokes. Strokes of a pen or brush are the marks created when writing or painting. Stroke filter is a spatial filter which can detect the strokes in images. By nature, at the stroke points, (1) the image has intensities different from its lateral regions, (2) its lateral regions have similar intensities to each other, and (3) it is nearly homogenous. Jung et al. [59], [60] used stroke filter to localize text within scene and video images. The candidature of the text regions thus identified is verified using SVM classifier. The text regions are then refined by boundary shrinking, character recombination and character extension.

Video image processing Anthimopoulos et al. [61] proposed a two-stage system for text detection in video images. In the first stage, text lines are detected based on the edge map of the image leading to a high recall rate with low computational time expenses. In the second stage, the result is refined using a sliding window and an SVM classifier trained on features obtained by a new Local Binary Pattern-based operator that describes the local edge distribution. The whole algorithm is used in a multi-resolution fashion enabling detection of texts with larger variation in size. A survey on text information extraction in images and videos is reported in [62].

Based on multilingual analysis, Lyu et al. [63] proposed a comprehensive, efficient video text detection, localization, and extraction method, which emphasizes the multilingual capability over the whole processing. The text detection is carried out by edge detection, local thresholding and hysteresis edge recovery. The coarse-to-fine localization scheme is then performed to identify text regions accurately. The text extraction consists of adaptive thresholding, dam point labelling, and inward filling. The proposed method is also robust to various background complexities and text appearances.

Very recently, due to the ease availability of camera mounted hand held devices, document images are also generated by camera captured devices rather than traditional scanning devices. Moreover, in commercial surveillance applications, processing the text within the scene has become a growing need. Yang et al. proposed an approach [64] for automated text detection in video data combining two stages: firstly, an edge-based multi-scale text detector to identify potential text candidates with high recall rate and small computational time expenses and secondly, candidate text lines are refined by an image entropy based improvement algorithm and a Stroke Width Transform (SWT) based verification procedure.

Vehicle license plate localization Most of the existing automatic traffic monitoring systems [65], [66] concentrate on localization of license plate regions and subsequent interpretation of the characters therein. In general, most of the works on license plate recognition system apply edge based features for localizing standardized license plate regions. Some of these works [65], [66] capture the image of a vehicle carefully placed in front of a camera occupying the full view of it and taking a clear image of the license plate. Spanish license plate is localized and recognized in [67] using connected component labeling (CCL) algorithm of different binarized images generated from different thresholds. It also uses the aspect ratio and distance of the plate from the center of the image as a characteristic for localization of the plate. Kwasnicka et al. proposed a hybrid method [66] to localize the license plate. During the localization phase, the position of the characters is used in [65]. It assumes that no significant edge lies near the license plate and characters are disjoint. An FPGA based license plate recognition system is reported in [68] using real time video processing. It uses low memory and is relatively faster than software based solutions. A work on localization of Iranian license plate is done in [69]. The system designed by Jia et al. [70] uses mean shift algorithm for localization of license plate giving

  79  

satisfactory result for license plates having color different from the body color of the vehicle. An application of generalized text localization and segmentation scheme is reported in [71], [72], where text color and spatial information like edge, texture and spatial distribution are used to localize the text within the image. A morphology based method is proposed in [73] giving high accuracy of localization with slightly lower recognition accuracy.

Most of the Indian commercial vehicles have license plates with yellow background and black foreground (for texts). A technique based on the method of region growing with color based segmentation has been developed in [74]. RGB and HSI color models are used for color segmentation. An artificial neural network based technique has been reported in [75], in which the network is trained with color features of the characters of the license plate. The trained network is then used to localize the license plate in the video snapshots captured from the surveillance cameras. It is seen that when the characters of the license number are written horizontally, the vertical edges of each character appear at regular interval and they have a specific height. The pattern and concentration of the vertical edges also remain in conformity with the pattern of the license number. This phenomenon have been explored to develop technique [76], [77] to find the license plate region within the image. The edges created by the characters within the license plate are extracted using Sobel’s vertical edge operator [78]. A license plate localization technique based on Hough transform has been reported in [79]. A hybrid method is developed and reported in [80], [81] that uses Hough transform inside the regions found by vertical edge based multi-stage technique to identify the desired regions where the vertical edges are very close to each other and select the license plate accordingly. An effective edge based technique [82] has been developed for localization of license plate region from the captured video snapshots taken by the surveillance cameras in unconstrained outdoor environment. The overall technique is applied on pre-processed grey scale images. Different edge map images are subsequently generated using different heuristically chosen thresholds. A rule based segment analysis engine then identifies relevant connected regions from the binarized image for true localization of the license plate regions.

Image Processing Techniques

A. Image binarization Binarization has been the area of research for the last forty years or so [83], mainly to find a single threshold value or a set of threshold values for converting a grey scale image into a binary image. Most of the algorithms till developed are of generic type with or without using local information or special content within the image. The major contribution of research for binarization is to recover or extract information from a degraded document images. A nonparametric and unsupervised method of automatic threshold selection is presented by Nobuyuki Otsu [84] which uses a method based on grey level histogram and it maximizes the intra-class variance to total variance. Bernsen’s binarization method [85] computes the local maximum and minimum intensities for each pixel around the neighborhood of that pixel. Niblack proposed a method [86] which determines the threshold for each pixel by shifting a rectangular window across the image. Sauvola et al. developed an algorithm [87] for text and picture segmentation within an image and binarized the image using local threshold. In this method, neighbors of each pixel are considered to determine whether the central pixel will become black or white. In [88], adaptive Wiener filter [89] is used as a pre-processing step based on the statistics of the neighbouring pixels. It then performs an initial estimation of foreground pixels and the result is then refined using Sauvola’s adaptive thresholding method [87] with 𝑘=0.2, to get the binarized image 𝑆(𝑥, 𝑦). It then performs approximate estimation of background surface 𝐵(𝑥, 𝑦) of the pre-processed image 𝐼(𝑥, 𝑦) using the method proposed in [90]. Valverde et al. proposed a method [91] that binarizes the image using Niblack’s technique, with 𝑘=-0.2, within a neighbourhood of 15×15 pixels. It then performs morphological opening with a structuring element favouring 4-connectivity. A slight modification of Niblack’s method is applied in [92] by Zhang et al. to binarize the document image of bound volumes. It first detects the shadow boundary. It then binarizes the image using slight modification of Niblack’s method with 𝑘=0.1 and 𝑅=100. Milewski et al. [93] uses a binarization based method for separating handwritten text from carbon copy medical form images. A comparative study of different document image binarization methods have been discussed in different literatures [94–96]. Saha et al. proposed three methods for giving the membership value to each pixel at each level of histogram equalization. The net membership value of each

  80  

pixel is then calculated by merging the membership values obtained at each level of hierarchy linearly [97], [98] or non-linearly [99]. The image is then binarized depending on the net membership value of each pixel.

B. Image segmentation Segmentation is a procedure that is used to partition an image into its constituent parts or objects. Autonomous segmentation is one of the most difficult tasks in digital image processing. Lot of works in the area of image segmentation has already been reported in the form of research papers during last couple of decades or so. Three types of algorithms are discussed in a research report [100]. In one earlier work [101], a region based representation of image and video useful to multimedia services has been represented. It then proposes a partition tree representation of image. Bouman et al. [102] uses multi-scale Bayesian image segmentation on multispectral remotely sensed images. A work is reported in [103] in which text is localized and segmented from complex images and videos. The method applies multi-resolution approach and uses multimedia text detector and segmentor. It also estimates text color using vector quantization method. A new discrimination function is proposed in [104] for segmenting touching characters in printed document image based on both pixel and profile projections. A dynamic recursive segmentation algorithm is developed for effectively segmenting touching characters. In [105], a grey scale based character segmentation and recognition methodology is proposed in which the candidate segmentation points could be found efficiently by using topographic features and projection profiles of grey scale images. Basu et al. [106] has reported text line extraction from multi-skewed handwritten document images using water flow technique. In another work [107], text lines are segmented from unconstrained handwritten document images with four types of scripts by comparing neighborhood of connected components. A character segmentation algorithm for Arabic text is proposed in [108]. An iterative multi-model sub-image binarization of handwritten character segmentation is proposed in [109]. A lexicon-driven handwritten Japanese character string segmentation is proposed in [110].

In [111], a license plate localization and character segmentation method is proposed. The proposed method employed hybrid binarization technique to reduce the segmentation problem caused by dirt. Hegt et al. [112] proposed a projection based method for license plate character segmentation. A vertical projection based character segmentation of license plate characters is proposed in [113]. Haris et al. [114] proposed a hybrid multidimensional image segmentation algorithm which combines edge and region-based techniques through the morphological algorithm of watersheds. Saha et al. proposed a method [115], which uses hierarchical segmentation of touching characters from camera captured images is based on the method of binarization at the different levels of object hierarchy.

C. Optical character recognition Recognition is the process that assigns a label to an object based on its descriptors. Different objects or patterns can be classified by means of finding discriminating feature set exhibiting different properties of the objects or patterns. More specifically, features are the quantitative measures of a pattern, that as a whole impose a membership value of the pattern to a particular class. The purpose of feature extraction is to reduce the amount of data required for representation of the properties of different patterns. An exhaustive survey on different feature extraction techniques for recognition of isolated/segmented characters has been published in [116], [117]. Surveys on feature descriptors from the perspective of shape analysis and object recognition has been published in different literatures [118–120]. Several features descriptors, viz. Fourier descriptors, graph descriptions, projection profile, contour profiles, geometric moment, Zernike moments, spline curve approximations have been investigated and subsequently published in different literatures [121–123]. With the progress of research in feature extraction techniques, different effective feature descriptors, viz. gradient-based features [124], [125], contour based features [126], concavity and curvature features [127], [128], pixel distance features [129] , directional distance features [130], wavelet based features [131], chain code features [132], histogram of gradient (HOG) based features [133–136] are being reported in the domain of character recognition. A topological feature combination has been proposed in [137–139] for recognition of handwritten characters/numerals. Different types of textural features are extracted in [140]. In [141], a work is reported for feature extraction for different type of scripts. An efficient fuzzy based system is reported in a work [142] for localization of license plate and Kohonen’s self organizing neural model is used for

  81  

recognition of characters. A two stage hybrid recognition system combining statistical and structural features is proposed in [143]. Llorens et al. [67] used HMM to recognize the characters. Quad-tree based longest run (QTLR) features are extracted from each binarized character segment using a technique described in [141], [144].

Conclusions Digital image processing has found its utility not only in the field of computer science but also in other domains of engineering and technology. This paper presents a comprehensive survey of different applications of it. It has other applications as well, which has not been reported here. Similarly, regarding image processing techniques, only image binarization, image segmentation and character recognition have been discussed here. Binarization plays very crucial role in pre-processing stage for most of the image processing applications. Though lot of research work has already been done in this field, still there is huge scope of research in this area. There is still scope of research in the field of image segmentation because a good segmentation result always leads to a good recognition result. There may not be enough scope of research for optical character recognition for English script. But there is growing scope of research for recognition of regional scripts across different parts of India. Development of new feature set and optimization of it may be a new direction of research in this regard.

  82  

REFERENCES

[1] M. D. McFarlane, “Digital pictures fifty years ago,” Proceedings of the IEEE, vol. 60, no. 7, pp. 768–770, 1972. [2] H. C. Andrews, A. G. Tescher, and R. P. Kruger, “Image processing by digital computer,” Spectrum, IEEE, vol.

9, no. 7, pp. 20–32, 1972. [3] T. McInerney and D. Terzopoulos, “Deformable models in medical image analysis,” in Mathematical Methods

in Biomedical Image Analysis, 1996., Proceedings of the Workshop on, 1996, pp. 171–180. [4] B. Zitova and J. Flusser, “Image registration methods: a survey,” Image and vision computing, vol. 21, no. 11,

pp. 977–1000, 2003. [5] I. Sluimer, A. Schilham, M. Prokop, and B. van Ginneken, “Computer analysis of computed tomography scans

of the lung: a survey,” Medical Imaging, IEEE Transactions on, vol. 25, no. 4, pp. 385–405, 2006. [6] K. S. Fu and A. Rosenfeld, “Pattern recognition and image processing,” Computers, IEEE Transactions on, vol.

100, no. 12, pp. 1336–1346, 1976. [7] A. Rosenfeld, “Image pattern recognition,” Proceedings of the IEEE, vol. 69, no. 5, pp. 596–605, 1981. [8] A. Rosenfeld, “Image analysis: Problems, progress and prospects,” Pattern Recognition, vol. 17, no. 1, pp. 3–12,

1984. [9] B. Fasel and J. Luettin, “Automatic facial expression analysis: a survey,” Pattern Recognition, vol. 36, no. 1, pp.

259–275, 2003. [10] Y. Liu, D. Zhang, G. Lu, and W. Y. Ma, “A survey of content-based image retrieval with high-level semantics,”

Pattern Recognition, vol. 40, no. 1, pp. 262–282, 2007. [11] S. C. W. Ong and S. Ranganath, “Automatic sign language analysis: A survey and the future beyond lexical

meaning,” Pattern Analysis and Machine Intelligence, IEEE Transactions on, vol. 27, no. 6, pp. 873–891, 2005. [12] R. L. Teach and E. H. Shortliffe, “An analysis of physician attitudes regarding computer-based clinical

consultation systems,” Computers and Biomedical Research, vol. 14, no. 6, pp. 542–558, 1981. [13] D. M. Gavrila, “The visual analysis of human movement: A survey,” Computer vision and image understanding,

vol. 73, no. 1, pp. 82–98, 1999. [14] T. B. Moeslund and E. Granum, “A survey of computer vision-based human motion capture,” Computer Vision

and Image Understanding, vol. 81, no. 3, pp. 231–268, 2001. [15] J. K. Aggarwal and Q. Cai, “Human motion analysis: A review,” in Nonrigid and Articulated Motion Workshop,

1997. Proceedings., IEEE, 1997, pp. 90–102. [16] K. Akita, “Image sequence analysis of real world human motion,” Pattern Recognition, vol. 17, no. 1, pp. 73–

83, 1984. [17] C. R. Wren, A. Azarbayejani, T. Darrell, and A. P. Pentland, “Pfinder: Real-time tracking of the human body,”

Pattern Analysis and Machine Intelligence, IEEE Transactions on, vol. 19, no. 7, pp. 780–785, 1997. [18] Q. Cai and J. K. Aggarwal, “Tracking human motion using multiple cameras,” in Pattern Recognition, 1996.,

Proceedings of the 13th International Conference on, 1996, vol. 3, pp. 68–72. [19] A. Senior, “Tracking People with Probabilistic Appearance Models,” in IEEE Int. Workshop on PETS, 2002, pp.

48–55. [20] D. M. Gavrila and L. S. Davis, “3-D model-based tracking of humans in action: a multi-view approach,” in

IEEE Computer Society Conference on CVPR, 1996, pp. 73–80. [21] T. W. Calvert and A. E. Chapman, “Analysis and synthesis of human movement,” in Handbook of pattern

recognition and image processing (vol. 2), Academic Press, Inc., 1994, pp. 431–474. [22] Z. Jiang, Z. Lin, and L. S. Davis, “Recognizing human actions by learning and matching shape-motion prototype

trees,” Pattern Analysis and Machine Intelligence, IEEE Transactions on, vol. 34, no. 3, pp. 533–547, 2012. [23] F. Z. Qureshi and D. Terzopoulos, “Surveillance camera scheduling: A virtual vision approach,” Multimedia

systems, vol. 12, no. 3, pp. 269–283, 2006. [24] F. Z. Qureshi and D. Terzopoulos, “Surveillance in virtual reality: System design and multi-camera control,” in

Computer Vision and Pattern Recognition, 2007. CVPR’07. IEEE Conference on, 2007, pp. 1–8. [25] F. Qureshi and D. Terzopoulos, “Distributed coalition formation in visual sensor networks: A virtual vision

approach,” in Distributed Computing in Sensor Systems, 2007, pp. 1–20. [26] A. E. C. Pece, “From cluster tracking to people counting,” in IEEE Workshop on Performance Evaluation of

Tracking and Surveillance, 2002, pp. 9–17. [27] L. Marcenaro, L. Marchesotti, and C. S. Regazzoni, “Tracking and counting multiple interacting people in

indoor scenes,” in Third IEEE international workshop on performance evaluation of tracking and surveillance, Copenhagen, 2002, pp. 56–61.

[28] K. Rangarajan, W. Allen, and M. Shah, “Matching motion trajectories using scale-space,” Pattern recognition, vol. 26, no. 4, pp. 595–610, 1993.

[29] F. Qureshi and D. Terzopoulos, “Intelligent perception and control for space robotics,” Machine Vision and Applications, vol. 19, no. 3, pp. 141–161, 2008.

[30] R. E. Kahn, M. J. Swain, P. N. Prokopowicz, and R. J. Firby, “Gesture recognition using the perseus architecture,” in Computer Vision and Pattern Recognition, 1996. Proceedings CVPR’96, 1996 IEEE Computer Society Conference on, 1996, pp. 734–741.

  83  

[31] T. Gandhi and M. M. Trivedi, “Pedestrian collision avoidance systems: A survey of computer vision based recent studies,” in Intelligent Transportation Systems Conference, 2006. ITSC’06. IEEE, 2006, pp. 976–981.

[32] C. Papageorgiou and T. Poggio, “A trainable system for object detection,” International Journal of Computer Vision, vol. 38, no. 1, pp. 15–33, 2000.

[33] O. Barinova, V. Lempitsky, and P. Kholi, “On detection of multiple object instances using hough transforms,” Pattern Analysis and Machine Intelligence, IEEE Transactions on, vol. 34, no. 9, pp. 1773–1784, 2012.

[34] L. Zhao and C. E. Thorpe, “Stereo-and neural network-based pedestrian detection,” Intelligent Transportation Systems, IEEE Transactions on, vol. 1, no. 3, pp. 148–154, 2000.

[35] F. Bu and C. Y. Chan, “Pedestrian detection in transit bus application: sensing technologies and safety solutions,” in Intelligent Vehicles Symposium, 2005. Proceedings. IEEE, 2005, pp. 100–105.

[36] P. Lombardi and B. Zavidovique, “A context-dependent vision system for pedestrian detection,” in Intelligent Vehicles Symposium, 2004 IEEE, 2004, pp. 578–583.

[37] C. F. Wakim, S. Capperon, and J. Oksman, “A Markovian model of pedestrian behavior,” in Systems, Man and Cybernetics, 2004 IEEE International Conference on, 2004, vol. 4, pp. 4028–4033.

[38] L. Havasi, Z. Szlávik, and T. Szirányi, “PEDESTRIAN DETECTION USING DERIVED THIRD-ORDER SYMMETRY OF LEGS A novel method of motion-based information extraction from video image-sequences,” Computer Vision and Graphics, pp. 733–739, 2006.

[39] G. Antonini and M. Bierlaire, “Capturing interactions in pedestrian walking behavior in a discrete choice framework,” Transportation Research Part B, 2005.

[40] M. Oren, C. Papageorgiou, P. Sinha, E. Osuna, and T. Poggio, “Pedestrian detection using wavelet templates,” in Computer Vision and Pattern Recognition, 1997. Proceedings., 1997 IEEE Computer Society Conference on, 1997, pp. 193–199.

[41] P. Viola, M. J. Jones, and D. Snow, “Detecting pedestrians using patterns of motion and appearance,” in Computer Vision, 2003. Proceedings. Ninth IEEE International Conference on, 2003, pp. 734–741.

[42] T. Hashiyama, D. Mochizuki, Y. Yano, and S. Okuma, “Active frame subtraction for pedestrian detection from images of moving camera,” in Systems, Man and Cybernetics, 2003. IEEE International Conference on, 2003, vol. 1, pp. 480–485.

[43] Y. Abramson and B. Steux, “Hardware-friendly pedestrian detection and impact prediction,” in Intelligent Vehicles Symposium, 2004 IEEE, 2004, pp. 590–595.

[44] A. Shashua, Y. Gdalyahu, and G. Hayun, “Pedestrian detection for driving assistance systems: Single-frame classification and system level performance,” in Intelligent Vehicles Symposium, 2004 IEEE, 2004, pp. 1–6.

[45] L. Kratz and K. Nishino, “Tracking pedestrians using local spatio-temporal motion patterns in extremely crowded scenes,” Pattern Analysis and Machine Intelligence, IEEE Transactions on, vol. 34, no. 5, pp. 987–1002, 2012.

[46] P. Dollar, C. Wojek, B. Schiele, and P. Perona, “Pedestrian detection: An evaluation of the state of the art,” Pattern Analysis and Machine Intelligence, IEEE Transactions on, vol. 34, no. 4, pp. 743–761, 2012.

[47] T. Tsuji, H. Hattori, M. Watanabe, and N. Nagaoka, “Development of night-vision system,” Intelligent Transportation Systems, IEEE Transactions on, vol. 3, no. 3, pp. 203–209, 2002.

[48] Y. Fang, K. Yamada, Y. Ninomiya, B. K. P. Horn, and I. Masaki, “A shape-independent method for pedestrian detection with far-infrared images,” Vehicular Technology, IEEE Transactions on, vol. 53, no. 6, pp. 1679–1697, 2004.

[49] X. Liu and K. Fujimura, “Pedestrian detection using stereo night vision,” Vehicular Technology, IEEE Transactions on, vol. 53, no. 6, pp. 1657–1665, 2004.

[50] F. Xu, X. Liu, and K. Fujimura, “Pedestrian detection and tracking with night vision,” Intelligent Transportation Systems, IEEE Transactions on, vol. 6, no. 1, pp. 63–71, 2005.

[51] C. P. SUMATHI, T. Santhanam, and N. Priya, “Techniques and Challenges of automatic Text Extraction in Complex Images: A Survey,” Journal of Theoretical and Applied Information Technology, vol. 35, no. 2, 2012.

[52] J. Poignant, L. Besacier, G. Quénot, and F. Thollard, “From text detection in videos to person identification,” in Multimedia and Expo (ICME), 2012 IEEE International Conference on, 2012, pp. 854–859.

[53] L. Neumann and J. Matas, “Real-time scene text localization and recognition,” in Computer Vision and Pattern Recognition (CVPR), 2012 IEEE Conference on, 2012, pp. 3538–3545.

[54] Y. Zhang, C. Wang, B. Xiao, and C. Shi, “A New Method for Text Verification based on Random Forests,” in International Conference on Frontiers in Handwriting Recognition, 2012, pp. 109–113.

[55] N. Gupta and V. K. Banga, “Image Segmentation for Text Extraction,” in 2nd International Conference on Electrical, Electronics and Civil Engineering (ICEECE’2012), 2012, pp. 182–185.

[56] N. Gupta and V. K. Banga, “Localization of Text in Complex Images Using Haar Wavelet Transform,” International Journal of Innovative Technology and Exploring Engineering (IJITEE), vol. 1, no. 6, pp. 111–115, 2012.

[57] C. W. Liang and P. Y. Chen, “DWT based text localization,” International Journal of Applied Science and Engineering, vol. 2, no. 1, pp. 105–116, 2004.

[58] K. Subramanian, P. Natarajan, M. Decerbo, and D. Castañòn, “Character-stroke detection for text-localization and extraction,” in Document Analysis and Recognition, 2007. ICDAR 2007. Ninth International Conference on, 2007, vol. 1, pp. 33–37.

  84  

[59] C. Jung, Q. Liu, and J. Kim, “A stroke filter and its application to text localization,” Pattern Recognition Letters, vol. 30, no. 2, pp. 114–122, 2009.

[60] C. Jung, Q. Liu, and J. Kim, “A new approach for text segmentation using a stroke filter,” Signal Processing, vol. 88, no. 7, pp. 1907–1916, 2008.

[61] M. Anthimopoulos, B. Gatos, and I. Pratikakis, “A two-stage scheme for text detection in video images,” Image and Vision Computing, vol. 28, no. 9, pp. 1413–1426, 2010.

[62] K. Jung, K. In Kim, and A. K Jain, “Text information extraction in images and video: a survey,” Pattern recognition, vol. 37, no. 5, pp. 977–997, 2004.

[63] M. R. Lyu, J. Song, and M. Cai, “A comprehensive method for multilingual video text detection, localization, and extraction,” Circuits and Systems for Video Technology, IEEE Transactions on, vol. 15, no. 2, pp. 243–255, 2005.

[64] H. Yang, B. Quehl, and H. Sack, “Text detection in video images using adaptive edge detection and stroke width verification,” in Systems, Signals and Image Processing (IWSSIP), 2012 19th International Conference on, 2012, pp. 9–12.

[65] C. N. E. Anagnostopoulos, I. E. Anagnostopoulos, V. Loumos, and E. Kayafas, “A license plate-recognition algorithm for intelligent transportation system applications,” Intelligent Transportation Systems, IEEE Transactions on, vol. 7, no. 3, pp. 377–392, 2006.

[66] H. Kwaśnicka and B. Wawrzyniak, “License plate localization and recognition in camera pictures,” in 3rd Symposium on Methods of Artificial Intelligence, 2002, pp. 243–246.

[67] D. Llorens, A. Marzal, V. Palazón, and J. Vilar, “Car license plates extraction and recognition based on connected components analysis and HMM decoding,” Pattern Recognition and Image Analysis, pp. 357–362, 2005.

[68] H. Caner, H. S. Gecim, and A. Z. Alkar, “Efficient embedded neural-network-based license plate recognition system,” Vehicular Technology, IEEE Transactions on, vol. 57, no. 5, pp. 2675–2683, 2008.

[69] H. Mahini, S. Kasaei, and F. Dorri, “An efficient features-based license plate localization method,” in Pattern Recognition, 2006. ICPR 2006. 18th International Conference on, 2006, vol. 2, pp. 841–844.

[70] W. Jia, H. Zhang, X. He, and M. Piccardi, “Mean shift for accurate license plate localization,” in Intelligent Transportation Systems, 2005. Proceedings. 2005 IEEE, 2005, pp. 566–571.

[71] S. P. Chowdhury, S. Dhar, K. Rafferty, A. K. Das, and B. Chanda, “Robust Extraction of Text from Camera Images using Colour and Spatial Information Simultaneously,” Journal of Universal Computer Science, vol. 15, no. 18, pp. 3325–3342, 2009.

[72] S. P. Chowdhury, S. Dhar, A. K. Das, B. Chanda, and K. McMenemy, “Robust extraction of text from camera images,” in Document Analysis and Recognition, 2009. ICDAR’09. 10th International Conference on, 2009, pp. 1280–1284.

[73] S. H. Kasaei, S. M. Kasaei, and S. A. Kasaei, “New Morphology Based Method for Robust Iranian Car Plate Detection and Recognition,” International Journal of Computer Theory and Engineering, vol. 2, no. 2, pp. 264–268, 2010.

[74] S. Saha, S. Basu, M. Nasipuri, and D. K. Basu, “An Offline Technique for Localization of License Plates for Indian Commercial Vehicles,” in Proc. of National Conference on Computing and Communication Systems (COCOSYS-09), 2009, pp. 206–211.

[75] S. Saha, S. Basu, M. Nasipuri, and D. K. Basu, “Localization of license plates from surveillance camera images: A color feature based ann approach,” International Journal of Computer Applications IJCA, vol. 1, no. 23, pp. 27–31, 2010.

[76] S. Saha, S. Basu, M. Nasipuri, and D. K. Basu, “License Plate localization from vehicle images: An edge based multi-stage approach,” International Journal on Recent Trends in Engineering (Computer Science), vol. 1, no. 1, pp. 284–288, 2009.

[77] S. Saha, S. Basu, and M. Nasipuri, “Automatic Localization and Recognition of License Plate Characters for Indian Vehicles,” International Journal of Computer Science & Emerging Technologies, vol. 2, no. 4, pp. 520–533, 2011.

[78] I. Sobel and G. Feldman, “A 3 × 3 isotropic gradient operator for image processing,” in Presented at a talk at the Stanford Artificial Project, 1968.

[79] S. Saha, S. Basu, M. Nasipuri, and D. K. Basu, “A Hough Transform based Technique for Text Segmentation,” Journal of Computing, vol. 2, no. 2, pp. 134–141, 2010.

[80] S. Saha, S. Basu, and M. Nasipuri, “Development of a Stop-Line Violation Detection System for Indian Vehicles,” in Handbook of Research on Computational Intelligence for Engineering, Science and Business, S. Bhattacharyya and P. Dutta, Eds. IGI Global, 2013, pp. 200–227.

[81] S. Saha, S. Basu, and M. Nasipuri, “License Plate Localization Using Vertical Edge Map and Hough Transform Based Technique,” in Proceedings of the International Conference on Information Systems Design and Intelligent Applications (INDIA 2012) held in Visakhapatnam, India, January 2012, pp. 649–656.

[82] S. Saha, S. Basu, M. Nasipuri, and D. K. Basu, “Localization of License Plates from Indian Vehicle Images Using Iterative Edge Map Generation Technique,” Journal of Computing, vol. 3, no. 6, pp. 48–57, 2011.

[83] J. S. Weszka, R. N. Nagel, and A. Rosenfeld, “A threshold selection technique,” Computers, IEEE Transactions on, vol. 100, no. 12, pp. 1322–1326, 1974.

  85  

[84] N. Otsu, “A threshold selection method from gray-level histograms,” Automatica, vol. 11, no. 285–296, pp. 23–27, 1975.

[85] J. Bernsen, “Dynamic thresholding of grey-level images,” in International Conference on Pattern Recognition, 1986, pp. 1251–1255.

[86] W. Niblack, An introduction to digital image processing. Strandberg Publishing Company, 1985. [87] J. Sauvola and M. Pietikäinen, “Adaptive document image binarization,” Pattern Recognition, vol. 33, no. 2, pp.

225–236, 2000. [88] B. Gatos, I. Pratikakis, and S. J. Perantonis, “Adaptive degraded document image binarization,” Pattern

Recognition, vol. 39, no. 3, pp. 317–327, 2006. [89] R. C. Gonzales and R. E. Woods, Digital Image Processing, vol. 6. Prentice Hall, 2002. [90] M. Seeger and C. Dance, “Binarising camera images for OCR,” in Document Analysis and Recognition, 2001.

Proceedings. Sixth International Conference on, 2001, pp. 54–58. [91] J. S. Valverde and R.-R. Grigat, “Optimum binarization of technical document images,” in Image Processing,

2000. Proceedings. 2000 International Conference on, 2000, vol. 3, pp. 985–988. [92] Z. Zhang and C. L. Tan, “Recovery of distorted document images from bound volumes,” in Document Analysis

and Recognition, 2001. Proceedings. Sixth International Conference on, 2001, pp. 429–433. [93] R. Milewski and V. Govindaraju, “Binarization and cleanup of handwritten text from carbon copy medical form

images,” Pattern recognition, vol. 41, no. 4, pp. 1308–1315, 2008. [94] M. Nandy and S. Saha, “An Analytical Study of different Document Image Binarization Methods,” in

Proceedings of IEEE National Conference on Computing and Communication Systems (COCOSYS-09), UIT, Burdwan, 2009, pp. 71–76.

[95] M. Sezgin, “Survey over image thresholding techniques and quantitative performance evaluation,” Journal of Electronic imaging, vol. 13, no. 1, pp. 146–168, 2004.

[96] P. K. Sahoo, S. Soltani, and A. K. C. Wong, “A survey of thresholding techniques,” Computer vision, graphics, and image processing, vol. 41, no. 2, pp. 233–260, 1988.

[97] S. Saha, S. Basu, and M. Nasipuri, “Binarization of Document Images Using Hierarchical Histogram Equalization Technique with Linearly Merged Membership Function,” in Proceedings of the International Conference on Information Systems Design and Intelligent Applications (INDIA 2012) held in Visakhapatnam, India, January 2012, pp. 639–647.

[98] S. Saha, S. Basu, M. Nasipuri, and D. K. Basu, “Image binarization based on grey membership parameters of pixels,” 26-Mar-2013.

[99] S. Saha, S. Basu, M. Nasipuri, and D. K. Basu, “A Novel Scheme for Binarization of Vehicle Images Using Hierarchical Histogram Equalization Technique,” in International Conference on Computer, Communication, Control and Information Technology (C3IT 2009), 2009, pp. 270–275.

[100] R. G. Casey and E. Lecolinet, “A survey of methods and strategies in character segmentation,” Pattern Analysis and Machine Intelligence, IEEE Transactions on, vol. 18, no. 7, pp. 690–706, 1996.

[101] P. Salembier and F. Marqués, “Region-based representations of image and video: segmentation tools for multimedia services,” Circuits and Systems for Video Technology, IEEE Transactions on, vol. 9, no. 8, pp. 1147–1169, 1999.

[102] C. A. Bouman and M. Shapiro, “A multiscale random field model for Bayesian image segmentation,” Image Processing, IEEE Transactions on, vol. 3, no. 2, pp. 162–177, 1994.

[103] R. Lienhart and A. Wernicke, “Localizing and segmenting text in images and videos,” Circuits and Systems for Video Technology, IEEE Transactions on, vol. 12, no. 4, pp. 256–268, 2002.

[104] S. Liang, M. Shridhar, and M. Ahmadi, “Segmentation of touching characters in printed document recognition,” Pattern Recognition, vol. 27, no. 6, pp. 825–840, 1994.

[105] S.-W. Lee, D.-J. Lee, and H.-S. Park, “A new methodology for gray-scale character segmentation and recognition,” Pattern Analysis and Machine Intelligence, IEEE Transactions on, vol. 18, no. 10, pp. 1045–1050, 1996.

[106] S. Basu, C. Chaudhuri, M. Kundu, M. Nasipuri, and D. Basu, “Text line extraction from multi-skewed handwritten documents,” Pattern Recognition, vol. 40, no. 6, pp. 1825–1839, 2007.

[107] A. Khandelwal, P. Choudhury, R. Sarkar, S. Basu, M. Nasipuri, and N. Das, “Text line segmentation for unconstrained handwritten document images using neighborhood connected component analysis,” Pattern Recognition and Machine Intelligence, pp. 369–374, 2009.

[108] T. Sari, L. Souici, and M. Sellami, “Off-line handwritten arabic character segmentation algorithm: Acsa,” in Frontiers in Handwriting Recognition, 2002. Proceedings. Eighth International Workshop on, 2002, pp. 452–457.

[109] A. Dawoud and M. S. Kamel, “Iterative multimodel subimage binarization for handwritten character segmentation,” Image Processing, IEEE Transactions on, vol. 13, no. 9, pp. 1223–1230, 2004.

[110] C.-L. Liu, M. Koga, and H. Fujisawa, “Lexicon-driven segmentation and recognition of handwritten character strings for Japanese address reading,” Pattern Analysis and Machine Intelligence, IEEE Transactions on, vol. 24, no. 11, pp. 1425–1437, 2002.

  86  

[111] J.-M. Guo and Y.-F. Liu, “License plate localization and character segmentation with feedback self-learning and hybrid binarization techniques,” Vehicular Technology, IEEE Transactions on, vol. 57, no. 3, pp. 1417–1424, 2008.

[112] H. A. Hegt, R. J. De La Haye, and N. A. Khan, “A high performance license plate recognition system,” in Systems, Man, and Cybernetics, 1998. 1998 IEEE International Conference on, 1998, vol. 5, pp. 4357–4362.

[113] X. Jia, X. Wang, W. Li, and H. Wang, “A novel algorithm for character segmentation of degraded license plate based on prior knowledge,” in Automation and Logistics, 2007 IEEE International Conference on, 2007, pp. 249–253.

[114] K. Haris, S. N. Efstratiadis, N. Maglaveras, and A. K. Katsaggelos, “Hybrid image segmentation using watersheds and fast region merging,” Image Processing, IEEE Transactions on, vol. 7, no. 12, pp. 1684–1699, 1998.

[115] S. Saha, S. Basu, and M. Nasipuri, “Hierarchical Segmentation of Falsely Touching Characters from Camera Captured Degraded Document Images,” International Journal of Computer Science Issues, vol. 8, no. 4, pp. 90–98, 2011.

[116] Ø. Due Trier, A. K. Jain, and T. Taxt, “Feature extraction methods for character recognition-a survey,” Pattern recognition, vol. 29, no. 4, pp. 641–662, 1996.

[117] M. Yang, K. Kpalma, and J. Ronsin, “A survey of shape feature extraction techniques,” Pattern recognition, pp. 43–90, 2008.

[118] S. Loncaric, “A survey of shape analysis techniques,” Pattern recognition, vol. 31, no. 8, pp. 983–1001, 1998. [119] J. Sokolowski and A. Zochowski, “On the topological derivative in shape optimization,” SIAM journal on

control and optimization, vol. 37, no. 4, pp. 1251–1272, 1999. [120] P. Heider, A. Pierre-Pierre, R. Li, and C. Grimm, “Local shape descriptors, a survey and evaluation,” in

Proceedings of the 4th Eurographics conference on 3D Object Retrieval, 2011, pp. 49–56. [121] M. Cheriet, N. Kharma, C.-L. Liu, and C. Suen, Character recognition systems: a guide for students and

practitioners. Wiley-Interscience, 2007. [122] M. Cheriet, M. El Yacoubi, H. Fujisawa, D. Lopresti, and G. Lorette, “ICFHR 2008 Panel Discussion:

Handwriting recognition research: Twenty years of achievement... and beyond,” Pattern recognition, vol. 42, no. 12, pp. 3131–3135, 2009.

[123] D. Zhang and G. Lu, “Review of shape representation and description techniques,” Pattern recognition, vol. 37, no. 1, pp. 1–19, 2004.

[124] C.-L. Liu and C. Y. Suen, “A new benchmark on the recognition of handwritten Bangla and Farsi numeral characters,” Pattern Recognition, vol. 42, no. 12, pp. 3287–3295, 2009.

[125] C.-L. Liu, K. Nakashima, H. Sako, and H. Fujisawa, “Handwritten digit recognition: benchmarking of state-of-the-art techniques,” Pattern Recognition, vol. 36, no. 10, pp. 2271–2285, 2003.

[126] G. Y. Chen, T. D. Bui, and A. Krzyzak, “Contour-based handwritten numeral recognition using multiwavelets and neural networks,” Pattern Recognition, vol. 36, no. 7, pp. 1597–1604, 2003.

[127] K. T. Miura, R. Sato, and S. Mori, “A method of extracting curvature features and its application to handwritten character recognition,” in Document Analysis and Recognition, Proceedings of the Fourth International Conference on, 1997, vol. 2, pp. 450–454.

[128] D. Zhang and G. Lu, “A comparative study of curvature scale space and Fourier descriptors for shape-based image retrieval,” Journal of Visual Communication and Image Representation, vol. 14, no. 1, pp. 39–57, 2003.

[129] I.-S. Oh and C. Y. Suen, “A Feature for character recognition based on Directional Distance Distributions,” in Document Analysis and Recognition, Proceedings of the Fourth International Conference on, 1997, vol. 1, pp. 288–292.

[130] H. Fujisawa and C.-L. Liu, “Directional pattern matching for character recognition revisited,” algorithms, vol. 13, p. 14, 2003.

[131] S. E. N. Correia, J. M. de Carvalho, and R. Sabourin, “On the performance of wavelets for handwritten numerals recognition,” in Pattern Recognition, Proceedings. 16th International Conference on, 2002, vol. 3, pp. 127–130.

[132] S. Hoque, K. Sirlantzis, and M. C. Fairhurst, “A new chain-code quantization approach enabling high performance handwriting recognition based on multi-classifier schemes,” in Proceedings of International Conference on Document Analysis and Recognition (ICDAR), 2003, pp. 834–838.

[133] K. Terasawa and Y. Tanaka, “Slit style HOG feature for document image word spotting,” in Document Analysis and Recognition, 2009. ICDAR’09. 10th International Conference on, 2009, pp. 116–120.

[134] K. Pramod Sankar, C. V. Jawahar, and R. Manmatha, “Nearest Neighbor based Collection OCR,” in Proc. DAS, 2010.

[135] P. Sankar K, C. V. Jawahar, and R. Manmatha, “Nearest neighbor based collection OCR,” in Proceedings of the 9th IAPR International Workshop on Document Analysis Systems, 2010, pp. 207–214.

[136] V. Chandrasekhar, G. Takacs, D. Chen, S. Tsai, R. Grzeszczuk, and B. Girod, “CHoG: Compressed histogram of gradients a low bit-rate feature descriptor,” in Computer Vision and Pattern Recognition, 2009. CVPR 2009. IEEE Conference on, 2009, pp. 2504–2511.

[137] N. Das, J. M. Reddy, R. Sarkar, S. Basu, M. Kundu, M. Nasipuri, and D. K. Basu, “A statistical-topological feature combination for recognition of handwritten numerals,” Applied Soft Computing, vol. 12, no. 8, pp. 2486–2495, 2012.

  87  

[138] N. Das, K. Acharya, R. Sarkar, S. Basu, M. Kundu, and M. Nasipuri, “A Novel GA-SVM Based Multistage Approach for Recognition of Handwritten Bangla Compound Characters,” in Proceedings of the International Conference on Information Systems Design and Intelligent Applications (INDIA 2012) held in Visakhapatnam, India, January 2012, pp. 145–152.

[139] S. Bag, G. Harit, and P. Bhowmick, “Topological features for recognizing printed and handwritten Bangla characters,” in Proceedings of the 2011 Joint Workshop on Multilingual OCR and Analytics for Noisy Unstructured Text Data, 2011, p. 10.

[140] R. M. Haralick, K. Shanmugam, and I. H. Dinstein, “Textural features for image classification,” Systems, Man and Cybernetics, IEEE Transactions on, no. 6, pp. 610–621, 1973.

[141] S. Basu, N. Das, R. Sarkar, M. Kundu, M. Nasipuri, and D. Kumar Basu, “A novel framework for automatic sorting of postal documents with multi-script address blocks,” Pattern Recognition, vol. 43, no. 10, pp. 3507–3521, 2010.

[142] S. L. Chang, L. S. Chen, Y. C. Chung, and S. W. Chen, “Automatic license plate recognition,” Intelligent Transportation Systems, IEEE Transactions on, vol. 5, no. 1, pp. 42–53, 2004.

[143] X. Pan, X. Ye, and S. Zhang, “A hybrid method for robust car plate character recognition,” Engineering Applications of Artificial Intelligence, vol. 18, no. 8, pp. 963–972, 2005.

[144] S. Basu, N. Das, R. Sarkar, M. Kundu, M. Nasipuri, and D. K. Basu, “A hierarchical approach to recognition of handwritten Bangla characters,” Pattern Recognition, vol. 42, no. 7, pp. 1467–1484, 2009.