 Original Article
 Open access
 Published:
Analytical study of two feature extraction methods in comparison with deep learning methods for classification of small metal objects
Visual Computing for Industry, Biomedicine, and Art volume 5, Article number: 13 (2022)
Abstract
This paper addresses the efficiency of two feature extraction methods for classifying small metal objects including screws, nuts, keys, and coins: the histogram of oriented gradients (HOG) and local binary pattern (LBP). The desired features for the labeled images are first extracted and saved in the form of a feature matrix. Using three different classification methods (nonparametric Knearest neighbors algorithm, support vector machine, and naïve Bayesian method), the images are classified into four different classes. Then, by examining the resulting confusion matrix, the performances of the HOG and LBP approaches are compared for these four classes. The effectiveness of these two methods is also compared with the “You Only Look Once” and faster regionbased convolutional neural network approaches, which are based on deep learning. The collected image set in this paper includes 800 labeled training images and 180 test images. The results show that the use of the HOG is more efficient than the use of the LBP. Moreover, a combination of the HOG and LBP provides better results than either alone.
Introduction
Object classification is one of the most important problems in the field of image processing and machine vision [1, 2]. In general, classification methods are divided into parametric and nonparametric methods. Parametric methods seek to extract parameters for describing a particular model based on a trainingdata analysis. After creating the desired model, it is possible to classify new samples [3, 4]. The support vector machine (SVM) and naïve Bayesian (NB) classifier are the most important parametric methods [5,6,7,8,9]. In parametric methods, a single model is considered for all of the input data. Although reducing the classification problem to finding a few finite parameters seems logical, this assumption may not be correct, potentially leading to a misdiagnosis of the output.
Nonparametric algorithms involve finding close or similar samples based on a suitable distance function, and then interpolating to find the correct output. The knearest neighbor (KNN) algorithm is one of the most popular nonparametric methods [10,11,12]. In nonparametric methods, there is no need to calculate the parameters in the training phase; however, an algorithm is designed by using training examples to classify new data. Because these methods are built directly on data instead of estimating or predicting parameters, they often achieve higher accuracy than parametric methods. Especially in cases where the distribution of the training data is such that it cannot be modeled with a finite number of parameters, the use of nonparametric methods seems more logical. The main drawback of nonparametric methods (including KNN) is that it is necessary to have all of the training samples to make decisions regarding a new sample. This leads to increased memory and computational costs, particularly for large datasets. If the training samples are expressed in the form of simple and short descriptors, it is possible to significantly avoid increasing the required memory volume and computational complexity.
Deep learning and deep networks have progressed in recent years and have achieved acceptable results in many applications, including object detection and recognition [4, 13,14,15,16,17,18,19]. Unlike traditional methods, these methods do not require a separate step to manually extract the necessary features.
One of the most wellknown architectures proposed for object detection using deep learning [16] is “You Only Look Once” (YOLO) [20]. The importance of deep learning to various applications, such as those in agriculture, medicine, surveillance, and monitoring systems, is increasing daily [21,22,23]. The YOLO algorithm was first introduced in 2016, aiming to detect objects with a high speed and accuracy. This method introduced a new structure for object recognition systems. Owing to the significant attention it has received, different versions of YOLO have been implemented. YOLO stands for “You only look at the image once.” This term refers to the ability of the human visual system to detect objects at a glance. Therefore, the YOLO object recognition system is designed to provide a detection method that is similar to that of the human visual system. The YOLO algorithm consists of a 24layer convolutional neural network (CNN) for feature extraction, and two fully connected layers for predicting the probabilities and coordinates of objects.
The faster regionbased CNN (faster RCCN) [24] is another stateoftheart deeplearningbased technique. It was introduced in 2015 as a region proposal network (RPN) for sharing fullimage convolutional features with a detection network, thereby enabling nearly costfree region proposals. An RPN is a fully convolutional network that simultaneously predicts the bounding boxes and objectness scores at each position. It is the most widely used stateoftheart version of the regionbased CNN (RCNN) [25].
The detection and classification of metal objects are important applications for machine image and vision processing, especially in industry and commerce [26,27,28]. In ref. [26], a deep CNNbased technique was proposed to detect metal screws and microdefects on their surfaces. Images of different types of metal screws were captured using industrial cameras. The proposed deep network architecture was then used to diagnose and check the flawlessness of the screw. The experimental results showed that this proposed technique could achieve a detection accuracy of 98%. However, this method was trained only on screw images, and did not include other categories of metal objects.
An automatic mobile recognition system for separating coins from banknotes was presented in ref. [27]. The proposed method was based on a scaleinvariant feature transform color descriptor extraction method, and could be run on a smartphone. The results from this method were reviewed based on a set of images of banknotes and coins common in Jordan. Although this method had good accuracy in separating coins from banknotes, it seems that the problem (and method) resembled shape detection more than object classification. Because the coins were all circular and the banknotes were rectangular, the proposed method was not generalizable to detect other objects.
The method proposed in ref. [28] focused on a screw identification system for use in various industries, particularly in the automotive industry. The proposed method was based on a backpropagation neural network. The results from the experiments showed that the system could detect moving objects on a production line with appropriate accuracy; however, it was limited to detecting only two categories (screws and nuts).
One of the most critical challenges in object classification is choosing an appropriate featureextraction method [13]. An improper selection at this stage can affect the accuracy of classification, and lead to errors in decisionmaking. Accordingly, it is necessary to conduct comprehensive research on the efficiency of various featureextraction methods. A feature extraction method may work well with a particular classification algorithm, but may not work very well if used in another classifier. Thus, it is necessary to compare the results from several classification algorithms to conduct a detailed analytical study. In view of the above, the main purpose of this study is to comprehensively compare the performances of the histogram of oriented gradients (HOG) and local binary pattern (LBP) feature vectors.
The present study addresses the efficiency of the HOG and LBP approaches in classifying four groups of metal objects (screws, nuts, keys, and coins). To study these features in detail, three different classification methods are employed, i.e., the nonparametric KNN algorithm, SVM, and NB methods. The accuracy of each method is investigated using the HOG and LBP feature vectors, along with a combination of these two feature vectors. Then, the effectiveness of these methods is compared with those of the YOLO and faster RCNN deep learning methods.
The major contributions of this study include (1) providing an analytical study of the use of HOG and LBP methods for the classification of small objects; (2) using three conventional methods (KNN, SVM, and NB) for a comprehensive comparison of the HOG and LBP; (3) using YOLO as a deeplearning method for a more detailed comparison; and (4) providing a diverse image set of metal objects able to be used in future research as a benchmark set for comparing different methods.
The following section discusses the methods in the proposed structure. The results and discussion section analyzes the obtained results from traditional techniques and deep learning methods, e.g., YOLO and faster RCNN. The last section presents the conclusions of this study.
Methods
Figure 1 shows the general structure of the proposed method for image classification. By extracting the corresponding descriptors for each training sample, all of the training data are in the form of a matrix in which each row represents the feature vector of a sample, and each column represents a feature in the feature space. The training data were labeled; therefore, each related class is available at the start of the proposed algorithm. A parametric or nonparametric model is created based on these training samples for predicting the class of a test image. As shown in Fig. 1, feature extraction involves the use of one of the HOG or LBP methods, or of a combination of these two methods. Classification refers to one of the three methods mentioned above: KNN, SVM, and NB. With the arrival of new samples (test images), the desired features are extracted, and the label of the unknown sample is estimated using the created model. After the test samples are determined, a confusion matrix is generated for each classification model. The efficiencies of the different methods are examined using the confusion matrices.
HOG
The HOG was introduced in 2005 for pedestrian detection in static images. Today, this method plays an important role in identifying humans in movies [29, 30], as well as in various other applications such as sketchbased image retrieval [31] and realtime vehicle detection [32].
To extract the HOG, an image is first filtered using the horizontal and vertical operators in the x and ydirections, so that image gradients are obtained for the x and ydirections.
In the above, I is the original image. Dx and Dy are the filter masks in the x and y directions, respectively, and are defined as the vectors of Eq. (3). In ref. [29], more complex masks, including Sobel operators, were used to calculate the image gradients, and their performance was evaluated. This study reveals that the use of the masks of Eq. (3), in addition to providing simplicity, leads to better results in pedestrian detection. In Eq. (3), Gx and Gy denote the gradients of the image in the x and y directions, respectively, and the sign * indicates the convolution operation. In convolution operations, the neighbors of a central pixel are summed with specified weights, and the result is placed as the current pixel value. The weights are determined by a weight matrix or convolution mask. After calculating Gx and Gy, the magnitude and orientation of the gradient in each pixel are obtained as follows:
Here, G is the magnitude of the gradient, θ is the gradient direction, and i and j represent rows and columns in the image, respectively. To calculate the gradient histogram in each cell, the gradient orientation is first limited to a range of 0–180, as follows:
To calculate the histogram of gradients, the distance between 0–180° is divided by n equal distances, representing the number of directions of the gradient or histogram bars. Each of these distances forms a histogram channel. The range from 0–180° is used instead of the 360° range because usually, additional bars are needed for extraction in the range of 0–360°. Thus, the smaller range saves more time for feature extraction. Experimental observations have also shown that using a 360° range has little effect on improving the results relative to a 180° range. As discussed in ref. [29], a ninebar histogram achieved better results in experiments; accordingly, the present study uses the same number of bars to calculate the HOG.
To calculate the histogram, the image is divided into several cells. Each pixel then votes for one of the histogram channels, based on its gradient orientation. These votes are weighted based on the magnitude of the gradient in that pixel. This generates a histogram for each cell for describing the gradient of the pixels. In some cases, the HOG is calculated for a block (consisting of several cells) by connecting the histograms of adjacent cells (Fig. 2).
LBP
The LBP is another type of visual descriptor used in various machine vision applications [33,34,35]. This descriptor can be used for powerful feature extraction methods for face descriptions [36], analyzing wear in carpets [37], etc.
The feature vector in the LBP for grayscale images is calculated as follows.

The desired image is divided into several blocks, and each block is divided into several cells.

The following calculations are performed for each pixel in a cell.

Each pixel is compared to its eight neighboring pixels. The neighboring pixels are examined individually in a particular direction (e.g., clockwise).

When the center pixel is larger than the neighboring pixel, the number ‘0’ is written; otherwise, the number ‘1’ is written. In this manner, an eightbit number is obtained by comparing the central pixel with its eight neighbors. For convenience, this number is usually converted to a decimal number between 0–255 (Fig. 3).

A histogram of the numbers obtained in the previous step is calculated for each cell. This histogram has 256 bars (from 0–255), and each bar shows the number of repetitions of a specific number in that cell.

If necessary, the desired histogram is normalized.

The histogram of the entire block is obtained by connecting the histograms of neighboring cells. Thus, if a block contains four cells, the generated feature vector has a length of 256 × 4.
Figure 3 shows an example of binarypattern calculations in a 3 × 3 neighborhood. The binary pattern 00010011 is assigned to the central pixel, with a gray level of ‘5’ in the image on the left. After completing the calculations for all of the blocks, the generated feature vectors are processed by using an appropriate model to categorize the desired images. These classifiers can be used to categorize objects, recognize faces, analyze textures, and so on.
Different types of LBP algorithms have been proposed, with various changes relative to the original algorithm. One of the most useful and widely used types of LBP is the uniform pattern, which can significantly reduce the length of the feature vector [35]. This idea stems from the fact that the number of occurrences of certain binary patterns, called uniform binary patterns, is particularly important.
A generated binary pattern for a pixel is called a uniform pattern if it has a maximum of two 0–1 or 1–0 transitions. For example, 00010000 is a uniform pattern with two transitions, 0–1 and 1–0, but pattern 01010111, with five transitions, is not uniform. Uniform binary patterns with the highest number of events correspond to the basic features of the image, such as its edges, corners, and important points [34].
Therefore, uniform patterns can be considered as factors in identifying the main features of an image. All nonuniform patterns are assigned to a single bin, and each uniform pattern has a separate bin. As 58 uniform patterns are in the range of 0–255, the uniform LBP feature vector will have a length of 59; this is significantly reduced from the length of 256 in an ordinary LBP.
In this manner, by comparing local neighborhoods and calculating a uniform LBP histogram, an image signature is created to represent the type of texture. The generated signatures are sufficiently distinctive for images falling into different classes. Therefore, the LBP can be used to classify the textures.
KNN algorithm
The KNN algorithm is one of the most common nonparametric classification methods [10,11,12, 38]. In nonparametric methods, there is no need to calculate the parameters in the learning phase. Using the data itself, an algorithm is designed to check whether the new data belong to the training class. The advantage of these methods is that they do not require parameter estimations, and are usually more accurate than parametric methods; however, their main limitation is that they require all of the training samples to classify new samples. This increases the memory and computational costs, especially for large datasets. The KNN classification steps used to categorize images are as follows.

The training datasets and related labels are uploaded, and then a value of K is chosen as the number of neighbors.

The distance between the test image and each training sample is calculated.

The training samples are sorted in ascending order, based on the distances calculated in the previous step.

The first K items are selected from the sorted set.

The label is checked for the selected items in the previous step.

The most frequently labeled class is selected as the predicted class for the test sample.
Figure 4 shows an example of twoclass classification using the KNN algorithm. In this case, if k = 3, the test sample is considered as being in class B, because out of the three close neighbors, two neighbors are labeled B, and one neighbor is labeled A.
If k = 6, the situation will be different because in this case, four neighbors are labeled A and two neighbors are labeled B; therefore, the test sample is placed in class A. The accuracy of the classification algorithm can be verified by comparing the predicted labels with the actual labels of the test samples.
SVMbased classifier
The SVM is one of the currently widely used methods for classification [5, 6]. The current popularity of the SVM method can be compared with the popularity of neural networks over the past decade. The SVM is based on a linear classification of data. Figure 5 shows an example of a dataset that can be categorized linearly. Several lines are drawn to categorize the data. In a linear division of data, an attempt is made to select a line with a more reliable margin. Quadratic programming is used to find the optimal linear separator; this is a known method for solving limited problems.
The basic idea of the SVM is that, assuming that the categories are linearly separable, a line is obtained with a maximum margin to separate the categories. To find such a separator, two boundary lines are drawn parallel to the separator line, and are separated such that they collide with the data. The separation that maximizes the training data margin among the linear separators minimizes the generalization error. The training data closest to the separating line are called the support vectors (Fig. 5). Notably, for dimensions greater than two, the term ‘hyperplane’ is used instead of ‘line.’ A hyperplane is a geometric concept representing a generalization of the concept of a plane in ndimensions. In other words, the hyperplane defines a subsequent k subspace in an ndimensional space such that k < n.
One of the best properties of SVM is that, in cases where the data are not linearly separable, the SVM maps the data to a larger dimension using a nonlinear mapping function Φ. In this way, the data can be linearly separated in this new space.
This implies that samples that are not linearly separable in their original space I move to a new feature space called F to create a hyperplane for separating them. When this hyperplane returns to its original space I, it forms a nonlinear curve. As shown in Fig. 6, the input data are not linearly separable, and no line can accurately represent the boundary between the two classes. However, by mapping them from a twodimensional space to a threedimensional space, it is possible to create a hyperplane for separating the boundaries of these two classes.
NB categories
NB is a probabilitybased machinelearning algorithm that can be used for a wide range of classification problems [7,8,9]. Common applications of the NB algorithm include spam filtering, document classification, and emotion prediction.
This NB algorithm uses the Bayesian theorem to produce results, based on a hypothesis of strong independence between the features. This implies that changing the value of one feature does not directly affect the value of any other feature. Although this assumption is simplistic (as the algorithm’s name implies) for realworld datasets, the NB classifier has nevertheless found a worthy place among classification algorithms.
Suppose X = (x_{1}, x_{2}, .. x_{n}) expresses a data sample as a vector of n independent variables. To calculate the probability P[C_{k} (x_{1}, x_{2}, .. x_{n})], it is sufficient to use the joint probability, and to simplify it using a conditional probability concerning the independence of the variables.
Deep learning methods
YOLO is a stateoftheart realtime object detection system. It uses a single neural network to obtain a full image. This network divides the image into regions, and predicts bounding boxes and probabilities for each region. These bounding boxes are weighted using the predicted probabilities.
To use the YOLO algorithm, it is first necessary to prepare training images.
Preparing a training set for the YOLO algorithm is different from the approach for traditional methods, such as SVM or KNN. It is necessary to draw a bounding box around each object, and the corresponding class is determined by the user. Various programs can be used to draw bounding boxes and tag them. In this study, the MAKESENSE (https://www.makesense.ai/) program is used to tag the training images.
In this manner, a text file is created for each of the training images, in which the specifications of an object are written in each row. The first number is related to the class of the object; then the coordinates of the center of the rectangle and its length and width are written. The number of rows in the file is equal to the number of objects in the image. Figure 7 shows an example of the training data and generated file from the output of the MAKESENSE program. This program is also used to tag images for another deep learning method (faster RCNN).
Faster RCNN is a deep convolutional network used for object detection, and works as a single, endtoend, unified network for the user. The network can predict the locations of multiple objects in a short time. Researchers at UC Berkeley developed RCNN [25] in 2014. The RCNN is a deep convolutional network capable of detecting 80 different types of objects in images. In comparison with the generic pipeline for the object detection methods, the foremost contribution of RCNN is the extraction of features based on a CNN.
The RCNN consists of three principal modules. The first module produces 2000 region proposals using a selective search algorithm. After resizing to a fixed predefined size, the next module extracts a feature vector of length 4096 from each region proposal. The third module utilizes a pretrained SVM algorithm to classify the region proposal into one of the object classes, or as background. The RCNN model has some weaknesses: it is a multistage model, where each stage is an independent part. Consequently, it cannot be trained endtoend. It captures the extracted features from a pretrained CNN on the disk to train the SVMs. This requires a bulk storage on the order of gigabytes. The RCNN depends on a selective search algorithm for creating region proposals, which takes a long time. In addition, this algorithm cannot be customized for detection problems. Each region proposal is fed without dependence on the CNN for the feature extraction, making it inappropriate to run the RCNN in real time. As an extension of the RCNN model, the fast RCNN model was proposed [24] to overcome some of these limitations.
Results and Discussion
In this study, conventional classification methods (SVM, KNN, and NB) were implemented using MATLAB software version 2019. A set of images, including 800 training images and 180 test images, was prepared to analyze the efficiency of the HOG and LBP methods. More specifically, 200 training and 45 test images were considered for each of the four categories (screws, nuts, keys, and coins).
In the case of the HOG, the size of each cell was 64 × 64 pixels, each block contained 2 × 2 cells, the number of bins in the orientation histograms was 10, and the length of the HOG feature vector was 360. For the LBP, the cell size was 64, and a rotationally invariant uniform feature vector with a length of 160 was employed. Thus, the combined feature vector (HOGLBP) had a length of 520. There was an attempt to provide sufficient variety in terms of the types of objects and lighting conditions. Figure 8 shows examples of images in the collected dataset.
Table 1 lists the confusion matrices obtained using the SVM method. This table has three matrices, from top to bottom, corresponding to the states of the HOG, LBP, and a combination of them (HOGLBP). The KNNbased classification and NBbased classification results are shown in Tables 2 and 3, respectively. As shown in Table 1, when using the SVMbased classification, the feature extraction by the LBP method performed worse than that by HOG for detecting keys and coins, but performed better than the HOG for detecting nuts.
In detecting the nuts, using the HOG results in 34 correct detections, but this value is 36 for the LBP. This table also shows that using the HOGLBP feature vector combination, which results from the connection of two separate feature vectors, yields better results than the independent use of either of these features. Figure 9 shows the receiver operating characteristic (ROC) curves of the SVM and area under the curve (AUC) values for the three different feature vectors. This figure shows that the combined feature vector (HOGLBP) achieves better results than the independent use of the HOG and LBP.
Table 2 presents the results for the KNN classifier. In the case of the KNN algorithm, choosing an appropriate value for k may affect the efficiency of the classification. The value of k = 3 achieved a greater accuracy for metal object classification in the experiments. The matrices in this table indicate that using the HOG leads to better results than using the LBP. As shown in this table, using the HOGLBP combination to detect three classes (i.e., screws, nuts, and coins) shows better results than using each of them separately. However, the use of the HOGLBP feature performs worse than the independent use of the HOG in the case of the key class (39 vs 42).
An examination of Table 3 shows that HOG performs better than the LBP in the case of the NB classifier. This table also indicates that the HOGLBP combination does not achieve a higher accuracy than the HOG and LBP separately when using am NB classifier.
Figure 10 depicts the ROC curves for the SVM, KNN, and NB approaches, along with the AUC values. As shown, the SVM classifier achieves better results than the KNN and NB classifiers. Therefore, from Figs. 9 and 10, it can be concluded that using the SVM classifier for the HOGLBP vector provides the best result for the classification of small metal objects when using conventional methods. In the following, the effectiveness of two deep learning methods is examined: YOLO and faster RCNN.
In implementing the deep learning method, 10% of the training images were employed as a validation set. Table 4 presents the confusion matrix obtained using YOLO version 5 and faster RCNN. As shown in Table 4, the classification achieves high accuracy with these two methods. In the case of YOLO, one nut is not placed in any of the classes. In the case of faster RCNN, one nut is placed in an incorrect class (coin) and part of one key is recognized as a nut, as can be seen from Fig. 11. One of the advantages of deep learning methods such as YOLO and faster RCNN is that they are able to identify the locations of the objects as well as classify them in multiobject conditions, i.e., where there are several objects in one frame. In contrast, in traditional methods, this capability does not exist, and it is necessary to use an auxiliary algorithm to detect the location of an object. An example of multiple detections is shown in Fig. 12.
Notably, the YOLO and faster RCNN algorithms do not require a separate step to extract the feature vectors. In this respect, they may be superior to featurebased methods such as KNN and SVM. However, implementing a deep learning algorithm has more hardware requirements than featurebased extraction methods. In the present study, the YOLO and faster RCNN algorithms were implemented using Google Colab. Another drawback of YOLO and faster RCNN is the method for labeling the training data, which makes the training phase of these algorithms more difficult and timeconsuming than conventional methods.
Examination of the confusion matrices in Tables 1, 2, 3 and 4 indicates that generally, feature extraction in the HOG method performs better than that in the LBP method, but their combination feature vector (HOGLBP) is more accurate than their independent use for classifying small metal objects. In addition, the SVMbased method performs better than the KNN and NBbased methods, and the NBbased method achieves lower accuracy than the other methods. Although the SVMbased method is less accurate than YOLO and faster RCNN, it has fewer hardware requirements and an easier training phase than deep learning methods.
Conclusions
In this study, the efficiency of two feature extraction methods (the HOG and LBP) in classifying small metal objects including screws, nuts, keys, and coins was evaluated. Three different classifications, including KNN, SVM, and NBbased methods were used. The experiments indicate that generally, the HOG is better than the LBP, and that using their combination feature vector (HOGLBP) is better than using each separately. The effectiveness of these two methods was also compared with those of YOLO and faster RCNN, which are based on deep learning. Although deep learning methods do not require a separate step for feature extraction, they require more powerful platforms than traditional methods.
The strength of conventional feature extraction methods, such as the HOG and LBP, is that the extracted feature vector can be used in different classifiers to select a more accurate classifier. However, the disadvantage of these methods is that they can only classify detected objects. In other words, they cannot detect the positions of foreground objects in an image. In such a situation, a foreground extraction method must be applied, and then a traditional method (such as the HOG) can be used to produce the feature vector for each foreground region. In contrast, deep learning methods determine object positions in an image, as well as classifying those objects. This is the most positive point of the YOLO and faster RCNN algorithms, but their disadvantage is that they require manual image annotation in the training phase, which is timeconsuming and laborious for a large dataset. Nevertheless, the importance of deep learning methods is increasing in various applications, such as agriculture, medicine, and surveillance systems.
In continuing the research done in this study, other texture feature analyses, such as the Gabor filter, can be examined. Fourclass classification can be developed to classify with greater numbers. In addition, ensemble methods for considering various sizes of objects can be employed. A comparison with other deep learning methods will provide a better understanding of the efficiency of the different feature extraction methods. In general, the results of this study can be applied to factories and industrial workshops. Therefore, obtaining industrial images and providing practical implementations of the proposed method in real environments can be explored in future studies.
Availability of data and materials
The datasets used and/or analyzed during the current study are available from the corresponding author upon reasonable request.
Abbreviations
 HOG:

Histogram of oriented gradients
 LBP:

Local binary pattern
 KNN:

Knearest neighbors
 YOLO:

You Only Look Once
 SVM:

Support vector machines
 CNN:

Convolutional neural network
 RPN:

Region proposal network
 RCNN:

Regionbased convolutional neural network
 NB:

Naïve Bayesian
 ROC:

Receiver operating characteristic
 AUC:

Area under the curve
References
Hussain N, Khan MA, Kadry S, Tariq U, Mostafa RR, Choi JI et al (2021) Intelligent deep learning and improved whale optimization algorithm based framework for object recognition. Hum Cent Comput Inf Sci 11:34.
Walambe R, Marathe A, Kotecha K (2021) Multiscale object detection from drone imagery using ensemble transfer learning. Drones 5(3):66. https://doi.org/10.3390/drones5030066
Masood H, Zafar A, Ali MU, Khan MA, Ahmed S, Tariq U et al (2022) Recognition and tracking of objects in a clustered remote scene environment. Comput Mater Contin 70(1):16991719. https://doi.org/10.32604/cmc.2022.019572
Hussain N, Khan MA, Sharif M, Khan SA, Albesher AA, Saba T et al (2020) A deep neural network and classical features based scheme for objects recognition: an application for machine inspection. Multimedia Tools Appl https://doi.org/10.1007/s11042020088523
Miao YY, Song JX (2014) Abnormal event detection based on SVM in video surveillance. Paper presented at the 2014 IEEE workshop on advanced research and technology in industry applications, IEEE, Ottawa, 2930 September 2014.
Hasan H, Shafri HZM, Habshi M (2019) A comparison between support vector machine (SVM) and convolutional neural network (CNN) models for hyperspectral image classification. IOP Conf Ser Earth Environ Sci 357(1):012035. https://doi.org/10.1088/17551315/357/1/012035
Fang K (2019) Naive bayes image classification based on multiple features. Comp Software Media Appl 2(1):1215.
Bustomi MA, Faricha A, Ramdhan A, Faridawati (2018) Integrated image processing analysis and Naïve Bayes Classifier method for lungs Xray image classification. ARPN J Eng Appl Sci 13(2):718724.
Park DC (2016) Image classification using naïve bayes classifier. Int J Comput Sci Electron Eng 4(3):135139.
Zhou NR, Liu XX, Chen YL, Du NS (2021) Quantum Knearestneighbor image classification algorithm based on KL transform. Int J Theoretical Phys 60(3):12091224. https://doi.org/10.1007/s10773021047477
Garcia C (2016) A nearestneighbor algorithm for targeted interaction design in social outreach campaigns. Kybernetes 45(8):12431256. https://doi.org/10.1108/K0920150236
Wang YX, Wang RJ, Li DF, AduGyamfi D, Tian KB, Zhu YX (2019) Improved handwritten digit recognition using quantum knearest neighbor algorithm. Int J Theoretical Phys 58(7):23312340. https://doi.org/10.1007/s10773019041245
Rashid M, Khan MA, Alhaisoni M, Wang SH, Naqvi SR, Rehman A et al (2020) A sustainable deep learning framework for object recognition using multilayers deep features fusion and selection. Sustainability 12(12):5037. https://doi.org/10.3390/su12125037
Batool A, Nisar MW, Shah JH, Khan MA, ElLatif AAA (2022) iELMNet: integrating novel improved extreme learning machine and convolutional neural network model for traffic sign detection. Big Data https://doi.org/10.1089/big.2021.0279
Khan S, Khan MA, Alhaisoni M, Tariq U, Yong HS, Armghan A et al (2021) Human action recognition: a paradigm of best deep learning features selection and serial based extended fusion. Sensors 21(23):7941. https://doi.org/10.3390/s21237941
Garg D, Goel P, Pandya S, Ganatra A, Kotecha K (2018) A deep learning approach for face detection using YOLO. Paper presented at the 2018 IEEE punecon, IEEE, Pune, 30 November2 December 2018. https://doi.org/10.1109/PUNECON.2018.8745376
Varadarajan V, Garg D, Kotecha K (2021) An efficient deep convolutional neural network approach for object detection and recognition using a multiscale anchor box in realtime. Future Internet 13(12):307. https://doi.org/10.3390/fi13120307
Walambe R, Marathe A, Kotecha K, Ghinea G (2021) Lightweight object detection ensemble framework for autonomous vehicles in challenging weather conditions. Comput Intell Neurosci 2021:5278820. https://doi.org/10.1155/2021/5278820
Rashid M, Khan MA, Sharif M, Raza M, Sarfraz MM, Afza F (2019) Object detection and classification: a joint selection and fusion strategy of deep convolutional neural network and SIFT point features. Multimedia Tools Appl 78(12):1575115777. https://doi.org/10.1007/s1104201870310
Redmon J, Divvala S, Girshick R, Farhadi A (2016) You only look once: Unified, realtime object detection. Paper presented at the 2016 IEEE conference on computer vision and pattern recognition, IEEE, Las Vegas, 2730 June 2016. https://doi.org/10.1109/CVPR.2016.91
Khan MA, Muhammad K, Sharif M, Akram T, Kadry S (2021) Intelligent fusionassisted skin lesion localization and classification for smart healthcare. Neural Comput Appl https://doi.org/10.1007/s0052102106490w
Khan MA, Akram T, Sharif M, Saba T (2020) Fruits diseases classification: exploiting a hierarchical framework for deep features fusion and selection. Multimedia Tools Appl 79(35):2576325783. https://doi.org/10.1007/s11042020092443
Khan MA, Zhang YD, Khan SA, Attique M, Rehman A, Seo S (2021) A resource conscious human action recognition framework using 26layered deep convolutional neural network. Multimed Tools Appl 80(28):3582735849. https://doi.org/10.1007/s11042020094081
Ren SQ, He KM, Girshick R, Sun J (2017) Faster RCNN: Towards realtime object detection with region proposal networks. IEEE Trans Pattern Anal Mach Intell 39(6):11371149. https://doi.org/10.1109/TPAMI.2016.2577031
Girshick R, Donahue J, Darrell T, Malik J (2014) Rich feature hierarchies for accurate object detection and semantic segmentation. Paper presented at the 2014 IEEE conference on computer vision and pattern recognition, IEEE, Columbus, 2328 June 2014. https://doi.org/10.1109/CVPR.2014.81
Song LM, Li XY, Yang YG, Zhu XJ, Guo QH, Yang HD (2018) Detection of microdefects on metal screw surfaces based on deep convolutional neural networks. Sensors 18(11):3709. https://doi.org/10.3390/s18113709
Doush IA, ALBtoush S (2017) Currency recognition using a smartphone: Comparison between color SIFT and gray scale SIFT algorithms. J King Saud UnivComputer Informat Sci 29(4):484492. https://doi.org/10.1016/j.jksuci.2016.06.003
Johan TM, Prabuwono AS (2011) Recognition of bolt and nut using artificial neural network. Paper presented at the 2011 international conference on pattern analysis and intelligence robotics, IEEE, Kuala Lumpur, 2829 June 2011. https://doi.org/10.1109/ICPAIR.2011.5976889
Dalal N, Triggs B (2005) Histograms of oriented gradients for human detection. Paper presented at the 2005 IEEE computer society conference on computer vision and pattern recognition (CVPR'05), IEEE, San Diego, 2025 June 2005. https://doi.org/10.1109/CVPR.2005.177
Vo T, Tran D, Ma WL, Nguyen K (2013) Improved HOG descriptors in image classification with CP decomposition. In: Lee M, Hirose A, Hou ZG, Kil RM (eds) Neural information processing. 20th international conference on neural information processing, Daegu, Korea, November 2013. Lecture notes in computer science, vol 8228. Springer, Heidelberg, 384391. https://doi.org/10.1007/9783642420511_48
Hu R, Collomosse J (2013) A performance evaluation of gradient field hog descriptor for sketch based image retrieval. Comput Vis Image Und 117(7):790806. https://doi.org/10.1016/j.cviu.2013.02.005
Yan G, Yu M, Yu Y, Fan LF (2016) Realtime vehicle detection using histograms of oriented gradients and AdaBoost classification. Optik 127(19):79417951. https://doi.org/10.1016/j.ijleo.2016.05.092
Greeshma KV, Gripsy JV (2020) Image classification using HOG and LBP feature descriptors with SVM and CNN. Int J Eng Res Technol 8(4):14.
Ojala T, Pietikainen M, Maenpaa T (2002) Multiresolution grayscale and rotation invariant texture classification with local binary patterns. IEEE Trans Pattern Anal Mach Intell 24(7):971987. https://doi.org/10.1109/TPAMI.2002.1017623
Barkan O, Weill J, Wolf L, Aronowitz H (2013) Fast high dimensional vector multiplication face recognition. Paper presented at the 2013 IEEE international conference on computer vision, IEEE, Sydney, 18 December 2013. https://doi.org/10.1109/ICCV.2013.246
Ahonen T, Hadid A, Pietikainen M (2006) Face description with local binary patterns: Application to face recognition. IEEE Trans Pattern Anal Mach Intell 28(12):20372041. https://doi.org/10.1109/TPAMI.2006.244
Orjuela SA, Vansteenkiste E, Rooms F, De Meulemeester S, De Keyser R, Philips W (2011) Analysing wear in carpets by detecting varying local binary patterns. Paper presented at the SPIE 7870, Image processing: algorithms and systems IX. SPIE, San Francisco, 3 February 2011. https://doi.org/10.1117/12.871988
Altman NS (1992) An introduction to kernel and nearestneighbor nonparametric regression. Am Stat 6(3):175185. https://doi.org/10.1080/00031305.1992.10475879.
Acknowledgments
Not applicable.
Funding
Not applicable.
Author information
Authors and Affiliations
Contributions
SA, MCharoosaei, and MChinipardaz wrote the code, provided formal analysis and investigation, methodology, and visualization; SA wrote the paper. The authors have read and approved the final manuscript.
Corresponding author
Ethics declarations
Competing interests
The authors declare that they have no competing interests.
Additional information
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Amraee, S., Chinipardaz, M. & Charoosaei, M. Analytical study of two feature extraction methods in comparison with deep learning methods for classification of small metal objects. Vis. Comput. Ind. Biomed. Art 5, 13 (2022). https://doi.org/10.1186/s42492022001116
Received:
Accepted:
Published:
DOI: https://doi.org/10.1186/s42492022001116