


Vol 28, No 1 (2018)
- Year: 2018
- Articles: 20
- URL: https://journals.rcsi.science/1054-6618/issue/view/12258
Mathematical Method in Pattern Recognition
A Probabilistic Model of Fuzzy Clustering Ensemble
Abstract
A probabilistic model of clustering ensemble based on a collection of fuzzy clustering algorithms and a weighted co-association matrix is proposed. An expression for the upper bound of the misclassification probability of an arbitrary pair of objects is obtained depending on the characteristics of the ensemble. This expression is used to determine the optimal weights of the algorithms.



Regression Estimate of a Multidimensional Decision Function in the Two-Alternative Pattern-Recognition Problem
Abstract
The approximation properties of the regression estimate of the decision function in the two-alternative pattern-recognition problem are studied. This is used to find the dependence of the approximation quality on the discretization methods of the values of the random variable and its dimension.



Approximation Scheme for a Quadratic Euclidean Weighted 2-Clustering Problem
Abstract
We consider the strongly NP-hard problem of partitioning a finite set of Euclidean points into two clusters so as to minimize the sum (over both clusters) of the weighted sums of the squared intra-cluster distances from the elements of the cluster to its center. The weights of the sums are equal to the cardinalities of the clusters. The center of one of the clusters is given as input, while the center of the other cluster is unknown and is determined as the mean value over all points in this cluster, i.e., as the geometric center (centroid). The version of the problem with constrained cardinalities of the clusters is analyzed. We construct an approximation algorithm for the problem and show that it is a fully polynomial-time approximation scheme (FPTAS) if the space dimension is bounded by a constant.



Convolutional Neural Network Structure Transformations for Complexity Reduction and Speed Improvement
Abstract
Two methods of convolution-complexity reduction, and therefore acceleration of convolutional neural network processing, are introduced. Convolutional neural networks (CNNs) are widely used in computer vision problems. In the first method, we propose to change the structure of the convolutional layer of the neural network into a separable one, which is more computationally simple. It is shown experimentally that the proposed structure makes it possible to achieve up to a 5.6-fold increase in the operating speed of the convolutional layer for 11 × 11-sized convolutional filters without loss in recognition accuracy. The second method uses 1 × 1 fusing convolutions to increase the number of convolution outputs along with decreasing the number of filters. It decreases the computational complexity of convolution and provides an experimental processing speed increase of 11% in the case of large convolutional filters. It is shown that both proposed methods preserve accuracy when tested with the recognition of Russian letters, CIFAR-10, and MNIST images.



Representation, Processing, Analysis, and Understanding of Images
An Iterative Thinning Algorithm for Binary Images Based on Sequential and Parallel Approaches
Abstract
Thinning is an important process in several applications of computer vision. It aims to find the onepixel midline of the pattern in binary image. In spite of different thinning methods that have been proposed, the existing methods are not capable of solving all thinning problems. In this work, a new iterative thinning method for binary images was proposed based on a hybrid technique of sequential and parallel approaches. It consists of three stages. The first pre-processing stage determined and prepared the contour. Next, the peeling stage tested and removed unwanted pixels. When the first two stages did not meet more pixels, the postprocessing stage prepared the skeleton to produce the final one-pixel width skeleton. In this work, the first and last stages adopted the sequential approach, while the second stage was based on the parallel approach. To evaluate the performance of the proposed method, a selected and DIBCO2010-H_DIBCO2010_GT benchmark datasets were used with benchmark measurement techniques for thinning processing. The results were compared with Huang, Zhang, K3M, and Abu-Ain methods. The experiments show that the proposed method is implemented well with all types of thinning problems, better than other methods. It is simple to design, its result skeleton has one-pixel width, and it preserves the topology and connectivity.



Robust Visual Tracking Based on Convex Hull with EMD-L1
Abstract
Factors such as drastic illumination variations, partial occlusion, rotation make robust visual tracking a difficult problem. Some tracking algorithms represent a target appearances based on obtained tracking results from previous frames with a linear combination of target templates. This kind of target representation is not robust to drastic appearance variations. In this paper, we propose a simple and effective tracking algorithm with a novel appearance model. A target candidate is represented by convex combinations of target templates. Measuring the similarity between a target candidate and the target templates is a key problem for a robust likelihood evaluation. The distance between a target candidate and the templates is measured using the earth mover’s distance with L1 ground distance. Comprehensive experiments demonstrate the robustness and effectiveness of the proposed tracking algorithm against state-of-the-art tracking algorithms.



Relative Flow Estimates for Shot Boundary Detection
Abstract
This paper proposes a simple approach based on Relative Flow Estimates (RFE) for shot cut detection. The property of Relative flow estimates can be used for abrupt cut detection and a correction mechanism for gradual camera-shot transition detection (e.g., fade-in and fade-out, dissolves, wipes). The exacted feature vector in each frame can be mapped into a 3-D space along the continuous time axis, and these feature data can be treated as a virtually constructed pipe with fluid flowing in the 3-D axis. Compared with existing approaches, the new RFE-based algorithm can directly detect shot cut. A wide range of test videos are used to evaluate the performance of the proposed method. The experimental results show that the new scheme can produce promising results.



Bi-dimensional Empirical Mode Decomposition and Nonconvex Penalty Minimization Lq (q = 0.5) Regular Sparse Representation-based Classification for Image Recognition
Abstract
This paper reports an innovative pattern recognition technique for fracture microstructure images based on Bi-dimensional empirical mode decomposition (BEMD) and nonconvex penalty minimization Lq (q = 0.5) regular sparse representation-based classification (NPMLq-SRC) algorithm. The detailed procedures of this work can be divided into three steps, i.e., the preprocessing stage, the feature extraction stage and the image classification stage. We test and validate the proposed method through real data from metallic alloy fracture images. The case verification results show that our proposal can obtain a much higher recognition accuracy than the conventional Back Propagation Neural Networks (BPNN for short), the L1-norm minimization sparse representation-based classification (L1-SRC) and the BEMD combined with L1-norm minimization sparse representation-based classification (BEMD+L1-SRC) methods, respectively. Specifically, the proposed BEMD+NPMLq-SRC (q = 0.5) method outperforms the BEMD+L1-SRC method by 3.33% improvement of the average recognition accuracy, and outperforms L1-SRC method by 14.06% improvement of the average recognition accuracy, respectively.



Optimizing the Quantization Parameters of the JPEG Compressor to a High Quality of Fine-Detail Rendition
Abstract
This paper describes a new algorithm for adaptive selection of DCT quantization parameters in the JPEG compressor. The quantization parameters are selected by classification of blocks based on the composition of fine details whose contrast exceeds the threshold visual sensitivity. Fine details are identified by an original search and recognition algorithm in the N-CIELAB normalized color space, which allows us to take visual contrast sensitivity into account. A distortion assessment metric and an optimization criterion for quantization of classified blocks to a high visual quality are proposed. A comparative analysis of test images in terms of compression parameters and quality degradation is presented. The new algorithm is experimentally shown to improve the compression of photorealistic images by 30% on average while preserving their high visual quality.



An Improved Spatiogram Similarity Measure for Object Tracking
Abstract
Spatiogram is a generalization of the histogram. It adds high-order spatial information so that the target can be described more precisely. It is important to choose a suitable method of measuring the similarity between two spatiograms when Spatiogram is applied to the target tracking field. However, the original similarity measure based on spatiogram has the limitation of the insufficient discriminative power. Therefore, in this paper, we propose a new spatiogram similarity measure method called BJSD (Bhattacharyya coefficient and Jensen-Shannon divergence). The similarity of the color feature and the similarity of the spatial distribution are calculated by the Bhattacharyya coefficient (BC) and Jensen-Shannon divergence (JSD). Experiments show that the improved similarity measure has better discriminative than the Conaire’s method and the tracking results are more stable and accurate than the traditional mean shift tracking method.



Wrong Matching Points Elimination after Scale Invariant Feature Transform and Its Application to Image Matching
Abstract
When images are rotated and the scale varies or there are similar objects in the images, wrong matching points appear easily in the scale invariant feature transform (SIFT). To address the problem, this paper proposes a SIFT wrong matching points elimination algorithm. The voting mechanism of Generalized Hough Transform (GHT) is introduced to find the rotation and scaling of the image and locate where the template image appears in the scene in order to completely reject unmatched points. Through a discovery that the neighborhood diameter ratio and direction angle difference of correct matching pairs have a quantitative relationship with the image’s rotation and scaling information, we further remove the mismatching points accurately. In order to improve image matching efficiency, a method for finding the optimal scaling level is proposed. A scaling multiple is obtained through training of sample images and applied to all images to be matched. The experimental results demonstrate that the proposed algorithm can eliminate wrong matching points more effectively than the other three commonly used methods. The image matching tests have been conducted on images from the Inria BelgaLogos database. Performance evaluation results show that the proposed method has a higher correct matching rate and higher matching efficiency.



Software and Hardware for Pattern Recognition and Image Analysis
A Novel Method to Measure Sub-micro Repeatability of the High-Precision Positioning Control System Based on Digital Image Correlation Method
Abstract
To measure and analyze the repeatability of the high-precision positioning control system, an optical technique based on Digital Image Correlation (DIC) method coupling with digital microscope is proposed. In this study, the reliability and feasibility of this method is verified in high-accuracy motorized linear platform, the repeatability of which is calibrated as ±0.3 μm. To determine the in-plane displacement, the DIC requires two speckles images during each hysteresis process to represent the displacement change of platform using digital microscope. In essence, the method can be considered as a “plane to plane” measurement, which can effectively reduce the random error compared to “point to point” measurements. In last, five groups of displacement data under different loadings were obtained and be used to testify the feasibility of this method. In summary, the present study provides a simple, low-cost and efficient non-contact optical technique to detect sub-micro repeatability.



Applied Problems
Normal and Abnormal Tissue Classification in Positron Emission Tomography Oncological Studies
Abstract
Positron Emission Tomography (PET) imaging is increasingly used in radiotherapy environment as well as for staging and assessing treatment response. The ability to classify PET tissues, as normal versus abnormal tissues, is crucial for medical analysis and interpretation. For this reason, a system for classifying PET area is implemented and validated. The proposed classification is carried out using k-nearest neighbor (KNN) method with the stratified K-Fold Cross-Validation strategy to enhance the classifier reliability. A dataset of eighty oncological patients are collected for system training and validation. For every patient, lesion (abnormal tissue) and background (normal tissue around the lesion) are contoured on PET images using a semi-automatic method. Then 160 vectors are obtained to train and validate the KNN. Each vector is composed by thirty Standardized Uptake Values (SUVs) characterizing the area under investigation (lesion or background). In one case, vectors are labeled as normal or abnormal tissues by a nuclear medicine physician using a semi-automatic method; in other cases, Fuzzy C-means (FCM) and k-means are used for labelling vectors in an unsupervised manner. This study aims to evaluate the performance of the proposed classifier comparing it to the Linear Kernel Support Vector Machine (KSVM). The method accuracy is evaluated by comparison with the gold standard in terms of correct classification. Experimental results show that the KNN method achieves the highest classification accuracy using the semi-automatic labelling training (Sensitivity: 86.25%; Specificity: 90.00%; Negative Predictive Value: 88.37%; Precision: 89.81%; Accuracy: 88.12%; Error: 11.87%). In addition, the proposed method shows real-time performance; it could be applied to the field classification of PET images assisting physicians into discrimination of normal and abnormal tissue during radiation treatment planning.



Particular Use of BIG DATA in Medical Diagnostic Tasks
Abstract
The paper presents the main research results in the area of data mining application to medicine. We propose a new information technology of data mining for different classes of biomedical images based on the methodology of diagnostically relevant information selection and creation of informative characteristics. Application of Big Data technology in proposed systems of medical diagnostics has allowed to improve the learning set quality and reduce the classification error. Based on these results, the conclusion is made, that the usage of many heterogeneous sources of diagnostic information made it possible to improve the overall quality of the diagnostics.



Empirical Mode Decomposition for Signal Preprocessing and Classification of Intrinsic Mode Functions
Abstract
Empirical mode decomposition (EMD) is an adaptive, data-driven technique for processing and analyzing various types of non-stationary signals. EMD is a powerful and effective tool for signal preprocessing (denoising, detrending, regularity estimation) and time-frequency analysis. This paper discusses pattern discovery in signals via EMD. New approaches to this problem are introduced, which involve well-known information criteria along with some other proposed ones, which have been investigated and developed for our particular tasks. In addition, the methods expounded in the paper may be considered as a way of denoising and coping with the redundancy problem of EMD. A general classification of intrinsic mode functions (IMFs, empirical modes) in accordance with their physical interpretation is offered and an attempt is made to perform classification on the basis of the regression theory, special classification statistics and some cluster- analysis algorithm. The main advantage of the innovations is their capability of working automatically. Simulation studies have been undertaken on multiharmonic signals. We also cover some aspects of hardware implementation of EMD.



Monitoring Animal Diseases in Remote Area
Abstract
In densely populated countries, medical facilities are not much available in country-side areas. If it is about animal health, it is rare to get a good cure. The owner has to take the animal to the Vet and will have to do it every time if he is living in remote places. Most countries in South Asia have this situation where animal husbandry is a common source of earning bread. Many times either the animal is not getting cure on time or diseases are detected at a later stage which harms them most. This paper describes a cellular communication based mobile system which can be used to monitor the health of animals in the remote country-side area. Also, the owner can get a prescription from the doctor and cure animals according to doctor’s advice. The developed system uses client-server model and can work for multiple patients who are connected with the same doctor. The Vet can access this system on his PDA or desktop using a web application.



A Coarse-to-Fine Strategy for Vehicle Logo Recognition from Frontal-View Car Images
Abstract
This paper proposes a vehicle logo recognition (VLR) system centered on front-view cars, which has been largely neglected by vision community in comparison to other object recognition tasks. The study focuses on local features that describe structural characteristics by locating the logo of a car using a coarse-to-fine (CTF) strategy that first detects the bounding box of a car then the grille and at last, the logo. The detected logo is then used to recognize the make of a car in a reduced time. Our system starts to progress in detecting the bounding box of a car by means of a vocabulary voting and scale-adaptive mean-shift searching strategy. The system continues to process in locating the bounding box of an air-intake grille using a scale-adaptive sliding window searching technique. In the next level, the bounding box of a logo is located by means of cascaded classifiers and circular region detection techniques. The classification of vehicle logos is carried out on the patch-level as occurrences of similar visual words from a visual vocabulary, instead of representing the patchbased descriptors as bag-of-features and classifying them using a standard classifier. The proposed system was tested on 25 distinctive elliptical shapes of vehicle logos with 10 images per class. The system offers the advantage of accurate logo recognition of 86.3% in the presence of significant background clutter. The proposed scheme could be independently used for part recognition of grille detection and logo detection.



Traffic Sign Classification with a Convolutional Network
Abstract
I approach the traffic signs classification problem with a convolutional neural network implemented in TensorFlow reaching 99.33% accuracy. The highlights of this solution would be data pre-processing, data augmentation pipeline, pre-training and skipping connections in the network. I am using Python as programming language and TensorFlow as a fairly low-level machine learning framework.



Total Margin Based Balanced Relative Margin Machine
Abstract
Inspired by the total margin algorithm, we extend balanced relative margin machine (BRMM) by introducing surplus variables, and propose a total margin based balanced relative (TM-BRMM). TMBRMM not only solves the loss of information points involved, but also addresses outliers at the outer boundaries that limit the maximum distance from points to separating hyperplane. Furthermore, by means of kernel function, it is easy to solve nonlinear separable datasets. The experiments on UCI datasets verify the feasibility and superiority of TM-BRMM.



Erratum


