Skip to main content
  • Original Article
  • Open access
  • Published:

Defect detection of gear parts in virtual manufacturing

Abstract

Gears play an important role in virtual manufacturing systems for digital twins; however, the image of gear tooth defects is difficult to acquire owing to its non-convex shape. In this study, a deep learning network is proposed to detect gear defects based on their point cloud representation. This approach mainly consists of three steps: (1) Various types of gear defects are classified into four cases (fracture, pitting, glue, and wear); A 3D gear dataset was constructed with 10000 instances following the aforementioned classification. (2) Gear-PCNet+ + introduces a novel Combinational Convolution Block, proposed based on the gear dataset for gear defect detection to effectively extract the local gear information and identify its complex topology; (3) Compared with other methods, experiments show that this method can achieve better recognition results for gear defects with higher efficiency and practicability.

Introduction

Virtual manufacturing is a simulation-based technology for defining, simulating, and visualizing the manufacturing process in the design stage. During manufacturing, product defect detection is closely related to quality assurance. The detection of 3D objects has been widely studied [1,2,3,4]. Mechanical gears are widely used in the power transmission of various industrial machinery, including turbines, motor vehicles, and aircraft [5]. Gear defect detection is crucial in virtual manufacturing to detect faults incurred during the manufacturing simulation. However, gear defects are inevitable in an actual industrial environment with almost 80% of the faults in mechanical transmission systems caused by gear defects [6], resulting in manufacturing and financial losses, in addition to personal safety issues. Thus, defect detection is necessary in mechanical systems.

Traditionally, researchers artificially collected the characteristics of vibration and acoustic emission signals to monitor the condition of rotating machinery [7]. Signal-based methods [1,2,3] are also effective for gears, but they often require accurate physical models and signal processing experience [8, 9], which are insufficient to satisfy the modern industry requirements of intelligence. Sensor data were the basis for detection. Li et al. [10] collected information from different sensors to analyze defect features. However, the defect vibration signals were acquired by running the gear and the defects may be submerged in strong meshing harmonics of various rotary components.

Deep learning has great advantages in image classification [11,12,13] and target detection [14, 15] owing to its feature extraction and nonlinear approximation abilities. Furthermore, intelligent data-driven fault diagnosis technology has been receiving more attention. Li et al. [16] proposed a separation-fusion-based deep learning approach to analyze multi-modal features of gearbox vibration measurements and obtained the results of diagnosis. For traditional methods, modulated signals of gears are impractical in extracting features and detecting defects. On the other hand, image-based computer vision can be used in defect detection [17]. Researchers tried to use 2D images of gears to recognize gear defects [18] besides the gray image transformed by vibration signals [19, 20]. Nonetheless, it is difficult to recognize the defects of gears, especially on the tooth surface owing to its complex concave structure. In addition, the textures of oil stains or rust on gear surfaces with the results from images and cause confusion in defect detection [21].

Compared with the image-based methods, 3D point cloud models with depth data can avoid the misrecognition of gear defects from image texture or oil marks. Charles et al. [22] first proposed a network of point clouds: PointNet. Then, various point cloud-based deep learning networks were successfully used in 3D shape classification, object detection, tracking, and 3D segmentation [23, 24]. Massive, labeled data with defect information is the key to ensuring good detection performance of neural networks. Nevertheless, it is difficult to collect adequate data for the machines, which is a limiting factor for intelligent fault diagnosis. Researchers tackled the issue of lack of labeled data by transfer learning [25, 26] and semi-supervised/unsupervised learning [27] methods. However, a noise-free point cloud can be obtained from the computer aided design (CAD) model of a gear through virtual manufacturing. This makes it significant in checking the defect detection results using point cloud data. Besides, gear model with defects has complex local structures that can be fully represented by point clouds. Therefore, in this study, a new artificial neural network, Gear-PCNet++, is presented based on point clouds extracted from CAD models. In this network, a novel Combinational Convolution Block (CCB) is proposed to replace the convolution layer in Multi-Layer Perception (MLP) networks to extract more gear defect details.

The main contributions of this study are: (1) construction of a data set of 3D gear models, which has 4 typical gear defects: fracture, pitting, glue, and wear; (2) CCB combining multi-level features of gears, which improves the precision rate of defect detection; (3) development of a new network, Gear-PCNet++, based on CCB, enabling gear defects of various types to be recognized with high accuracy.

Methods

Construction of 3D gear sample sets

The data of point clouds can be obtained from 3D scanning, but it is difficult to accurately label the categories for scanned raw point clouds. Based on the geometric properties of gears, an approach for 3D gear data generation is proposed.

In this study, gear defects are classified into four typical types: wear, pitting, glue, and fracture [5]. The gear defects can be represented as a combination of the four typical defects, as illustrated in Fig. 1.

Fig. 1
figure 1

Four typical gear defects

Five basic gears were constructed with different parameters: modulus, tooth number, tooth width, and diameter of center hole (Table 1).

Table 1 Parameters of basic gears

Let W, P, G, and B represent wear, pitting, glue, and fracture, respectively, and S denote normal gear (basic gear). The gear models with defects are generated by combining defects and the basic gear using Eq. 1.

$$Gr_{def,i} = Gr_{bas,j} + Def_{i} = \{ Base,W,P,B\} ,$$
(1)

Where Grdef,i is the i-th generated gear with defects, Grbas,j is the j-th basic gear, and Defi is the defects of the i-th generated gear.

The CAD model is transferred into a point cloud model. Though the CAD model of the gear has a large number of surface elements, effective surfaces are randomly used to discretize points. Finally, a point cloud data set with 10000 gear samples is constructed; some of which are shown in Fig. 2.

Fig. 2
figure 2

Gear data set with defects. The light gray, orange, blue, red, green, and purple spheres represent points of baselines, basic gear, fracture, glue, pitting, and wear, respectively

CCB

The location of gear defects occurred on the tooth surface is frequently similar. Hence, it is difficult to identify gear defects from the local features of point clouds. The boundary information is more critical than other details for gears [22, 28]. The CCB module is proposed (Fig. 3) to improve the ability to identify gear features, especially the boundary lines.

Fig. 3
figure 3

CCB. Channel represents the number of channels of input feature vector. Isbn represents whether to add batch normalization to each convolution layer

A convolution layer significantly improves the efficiency of parameters by sharing weights and is widely used in artificial neural networks. Wu et al. [29] proposed PointConv using Monte Carlo approximation. This architecture is a convolution operation suitable for unstructured point cloud data. It has also been verified that dilated convolution and down sampling were effective ways to expand receptive fields [30]. Dilated Point Convolutions uses K·D nearest neighbors to replace the original k-nearest neighbor partition [31], and extracts the features of each d-th point. With the same parameter, it increases the receptive field of PointConv. This is similar to dilated convolution, but it may lead to loss of details with local features. PointNet++ uses neighborhood-based feature extraction to replace the independent learning of each point [32], notably overcoming the limitations of PointNet [22]. Inspired by Deformable CNN [33], the Deformable KPconv in ref. [34] assigns different convolution kernels to each local geometry.

The receptive fields play an important role in semantic segmentation. Essentially, the size of receptive fields is related to the number of convolution layers and the size of convolution kernels. For deeper networks, larger kernel size corresponds to larger receptive fields but large convolution kernels may cause performance degradation. The sizes of convolution kernels typically used in structured data images are \(3 \times 3\), \(5 \times 5\), or \(7 \times 7\). For unstructured point cloud models, a large convolution kernel will extract a lot of useless inter-point or point-domain information, which may be trivial to the improvement of performance. Multi-scale analysis is another strategy to improve the effect in image semantic segmentation [35,36,37], which can also enrich feature information. In addition, feature pyramid networks [38] is the most commonly used framework. Based on the above multi-scale or multi-level information interaction, this multi-scale synthesis strategy is applied to the convolution and uses a relatively small convolution kernel to obtain feature-rich information. Specifically, convolution kernels with different sizes are used to extract features under different receptive fields, and are then connected to the result of this module. The convolution of \(1 \times 1\) has been widely used in ResNet, GoogLeNet [39], and other architectures. In the aforementioned module, \(1 \times 1\) convolution is also used to achieve dimensional transformation to reduce the number of parameters. Moreover, the selection of convolution kernel size is based on the ideas discussed further.

Points, lines, and faces are the basic geometrical elements of gears. Two and three points can determine the corresponding line and plane, respectively; the point cloud is sparse relative to the original 3D model. It is assumed that a surface contains at least three points, of which two form a boundary line in a point cloud of gears. Then, the relevant geometric element information is extracted using kernel sizes 1, 2, and 3, and the corresponding features can be identified as projection points, pseudo lines, and pseudo surfaces, respectively, to a certain extent.

The point clouds in the input network are usually disordered. As shown in Fig. 4, there are pitting and wear defects in a gear, which are represented by green and blue cuboids, respectively. Ppit-j, Pwear-i, and Pwear-k are the points in pitting and wear, respectively. As for the point Pwear-i, the large convolution kernel can easily extract the feature that makes little contribution to the point.

Fig. 4
figure 4

Ordered/unorganized point cloud and the receptive field corresponding to different convolution kernel sizes. a The ordered point cloud extracted from the gear discretization; b Result of dispersing pitting defects and wear defects into point clouds; c The point cloud, result of dispersing pitting defects and wear defects, input into the neural network after random shuffle; d Representation of the receptive field corresponding to different convolution kernel sizes in the network. In the circular region d, the orange dot represents the convolution of kernel size 1, the green line represents the convolution of kernel size 2, and the red triangle represents the convolution of kernel size 3

To ensure the effectiveness of the extracted feature, defining the neighborhood of a point based on distance is a general strategy, which has been applied in many networks such as PointNet++, SpiderCNN [40], and EdgeConv [41]. Because of the difference between the study herein and the above methods a distance-based optimization strategy (Fig. 5) is proposed to assign corresponding weights to the features extracted by convolution kernels of different sizes.

Fig. 5
figure 5

Feature weights optimization based on distance. \(Conv_{i}\) is the convolution with kernel size i. \(W_{i,j}\) is the weight of the i-th point and \(j\) represents the size of kernel

The input point cloud is set as \({\mathrm{\{ }}p_{0} {,}p_{1} {,} \cdots p_{n} {\mathrm{\} }}\). Taking \(p_{i}\) as an example, the extracted feature is related to three points: \(p_{i}\),\(p_{i + 1}\),\(p_{i + 2}\) whose three-dimensional geometric center is \(p_{i0} = \left( {p_{i} + p_{i + 1} + p_{i + 2} } \right)/3.\).

\(W_{i,j}\) represents the weight of the i-th point whose convolution kernel size is \(j\). When the convolution kernel is 2, the distance between the two points is directly related to the kernel. Thus, \(W_{i,2\;}=k_2\;e^{-\vert p_ip_{i+1}\vert_d}\) can represent the corresponding weight. Similarly, when the convolution kernel is 3, the weight can be evaluated by a girth-related function:\(W_{i,3}=k_3\;e^{-\left(\vert p_ip_{i+1}\vert_d+\vert p_ip_{i+2}\vert_d+\vert p_ip_{i+2}{\vert_d\vert}\right)/3}\).

From the above definition, it is obvious that the proportion of inter point features will decrease with the discretization of points. Therefore, a compensation coefficient \(k\) is added to each weight to extract more local information. Furthermore, an eccentricity coefficient (\(W_{i,1} = k_{1} \cdot e^{{ - \left| {p_{i} p_{io} } \right|_{d} }}\)) is added when the convolution kernel is 1. The distance between pi and pio will decrease the proportion of the projection features of the point.

Through this multi-scale information synthesis, the proposed module can extract richer local features, and the latest extracted feature can be expressed using Eq. 2.

$$F_{i,com\_block} = F_{i,1} + Conv_{i} \left( {F_{i,ori} ,W_{i,1} \cdot F_{i,1} ,W_{i,2} \cdot F_{i,2} ,W_{i,3} \cdot F_{i,3} } \right)$$
(2)

Where Fi,com_block is the output feature of the module; Convi is the dimension transformation; Fi,ori is the input feature; Fi,k and Wi,k are the extracted feature and corresponding weight coefficient, respectively, when the convolution kernel is k. No additional weight calculation operation is required if the neighborhood is defined based on the distance.

Network architecture

First, a gear defect recognition network based on 1D convolution operation is proposed: Gear-PCNet (Fig. 6). The network is composed of feature extraction (CCB-MLP) and final classification modules. Gear-PCNet can learn the representation of gear defects and output their results.

Fig. 6
figure 6

Structure of Gear-PCNet where the number is output channel of the layer

CCB can output features containing both single and inter-point information. The point cloud is the rotation and translation invariance. Projecting the point cloud data into 2D images or expressing it as voxels may lead to information loss. In PointNet, Charles et al. [22] dealt with the above two issues using the maximum value (Eq. 3). In Gear-PCNet, both the maximum and average functions are used (Eq. 4) to extract the features of point clouds and concatenate them.

$${\mathrm F}_{\max}=\mathrm{Max}\left({\mathrm x}_1,\;{\mathrm x}_2,\;...\;,\;{\mathrm x}_{\mathrm n}\right)$$
(3)
$${\mathrm F}_{\mathrm{avg}}=\frac{{\mathrm x}_1+{\mathrm x}_2+...+{\mathrm x}_{\mathrm n}}{\mathrm n}$$
(4)

After pooling and lateral linking the comprehensive features obtained by CCB-MLP, the classification of detected point cloud data can be completed through a fully connected network with three layers.

As verified above, the hierarchical feature learning framework is further applied to Gear-PCNet and Gear-PCNet++ is built based on the 2D convolution operation; the structure of Gear-PCNet++ is shown in Fig. 7. By constructing local region sets, the data set is relatively more concentrated, allowing the radius of local regions to be set small.

Fig. 7
figure 7

Structure of Gear-PCNet+ + 

Unlike PointNet and PointNet++, Gear-PCNet was replaced with CCB to extract feature in Gear-PCNet++ (Fig. 8).

Fig. 8
figure 8

Feature extraction module in Gear-PCNet/Gear-PCNet+ +. a Feature extraction architecture in Gear-PCNet; b Feature extraction architecture in Gear-PCNet+ + 

By using multi-resolution grouping, the two grouped features were propagated to the original points. Then, the two features were concatenated and regarded as the basis for point set segmentation.

Results and Discussion

This approach was evaluated on a set of 10000 samples (gears with defects); their features can be grouped into 5 types: basic gear, fracture, pitting, glue, and wear. The 10000 samples were divided into training, validation, and testing sets in a 8:1:1 ratio, and experiments were run on a PC with a “NVIDIA GeForce RTX 3070” GPU and an “Intel Core i5-10400F @ 2.90GHz” CPU.

Experiment results

The CCB is applied to Gear-PCNet to synthesize the features extracted under different convolution kernels. The number of parameters in Gear-PCNet is given in Table 2.

Table 2 Parameters in Gear-PCNet

PointNet is a classic point cloud classification and segmentation network. The number of parameters in Gear-PCNet (\({7}{\text{.89}} \times {10}^{{5}}\)) is less than that of PointNet (vanilla) (\(2.05 \times 10^{6}\)). The effectiveness of Gear-PCNet was evaluated based on the classification performance of the three networks on gear data set. In addition, the combined CCB in Gear-PCNet was replaced with the structure shown in Fig. 9 to verify the superiority of comprehensive feature information over single feature information.

Fig. 9
figure 9

Single convolution group for replacing CCB

The replaced three networks were denoted: Gear-PCNet-single-1, Gear-PCNet-single-2, and Gear-PCNet-single-3. The addition of a convolution layer of kernel size 4 to the CCB in the Gear-PCNet (Gear-PCNet-4) to validate the effect of bigger kernel size on network performance was tested. The training and testing results of the above structures are presented in Table 3.

Table 3 Classification results of gear data set

Table 3 shows that Gear-PCNet has the best convergence and generalization ability, and can classify and recognize each defect point of a gear with high accuracy. The results of the network with only a single convolution kernel size are inferior to Gear-PCNet verifying that the synthetic feature can more comprehensively express the information of points than a single feature. In Gear-PCNet-4, a lot of information that does not belong to the original point is extracted, and the architecture does not work well.

CCB was replaced with the block (Fig. 10) to verify that the better performance of the network was not due to the addition of the number of convolution layers. The testing accuracy was 78.29%, which shows the effectiveness of the CCB.

Fig. 10
figure 10

Multi convolution with single size kernel for replacing CCB

The CCB had good results by extracting richer features. Gear-PCNet++ and several classical networks were tested on the gear data set. Figure 11 presents the prediction accuracy of the training and validation sets in the training process. It is seen that Gear-PCNet++ and PointNet++ converge faster.mAcc (mean Accuracy) and mIoU (mean Intersection-Over-Union) are the evaluation metrics; the results are listed in Table. 4. KPConv is more accurate in points classification and Gear-PCNet++ is better at object segmentation. Each architecture performs well in gear defect recognition.

Fig. 11
figure 11

Accuracy in training and validation. a Accuracy of training set; b Accuracy of validation set. In both a and b, a marker represents the accuracy of the network at the current epoch of training or validation. Specifically, gray represents PointNet, green represents PointNet+ + , blue represents PointCNN, red represents KPconv and yellow represents Gear-PCNet++ 

Table 4 Segmentation results of gear data set

Discussion of defect recognition

In Experiment results section, the classification and prediction of Gear-PCNet and Gear-PCNet++ is presented, but the types and numbers of defects in different gear models are different. In this section, the identification of defects and points in different models is analyzed based on the performance of Gear-PCNet++ on test samples (1000 gear models). Table 5 presents the recognition results of points in each model in the testing set. The recognition accuracy of 97.90% models is above 95.00%.

Table 5 Recognition accuracy of points in each model

The judgment of defect types was considered correct if the recognition was successful, that is, if there were 3 defects in a model, if and only if the 3 defects were detected, the defect detection is considered correct. A defect existed only if there were more than 10 points labeled with the defect. Under the above settings, 99.90% models were judged correctly. This shows that the recognition results are highly reliable.

Meanwhile, Fig. 12 gives a recognition confusion matrix of each defect type in the testing set. In Fig. 12, the confusion matrix was approximated as a diagonal matrix, which also shows that the approach herein is accurate and effective.

Fig. 12
figure 12

Confusion matrix of defect classification. Each row represents the distribution of predicted labels of points corresponding to each actual label. The depth of the color in the graph represents the predicted percentage

Few recognition results of the network in this study (containing the CAD models of gears and point cloud data) given in Table 6 have the same defect color representation as Fig. 2 and defects representation as Eq. 1.

Table 6 Few recognition results and their original CAD models (point clouds)

Gears also have intersecting defects making it difficult to recognize point category. They can be divided into self-intersection of the same defect features and inter-section of different defect features. Figures. 13a and b show the intersection result of pitting holes and the intersection result of broken tooth and wear, respectively. In Gear-PCNet, these kind of intersection result may require many relevant samples to assist the training of the network; but can be satisfied in Gear-PCNet++.

Fig. 13
figure 13

Gear models with intersecting defects. a Intersection result of pitting holes; b Intersection result of broken tooth and wear

Conclusions

Gear defect recognition plays an important role in mechanical fault diagnosis. In this study, deep learning was used to extract the gear features and determine the gear defects. First, a data set of gear CAD models containing 10000 basic gears with 4 typical defects, was constructed. Second, by setting few strategies a point cloud-based gear set was generated from the gear models. Then, by giving a new CCB with three (1, 2, 3) sizes of convolution kernels, a new network: Gear-PCNet++ which can extract gear features more effectively was proposed. Finally, experimental results showed that the proposed network achieved high recognition accuracy compared to other methods for all types of gear defects.

Availability of data and materials

Not applicable.

Abbreviations

CAD:

Computer aided design

CCB:

Combinational Convolution Block

MLP:

Multi-Layer Perception

mAcc:

Mean Accuracy

mIoU:

Mean Intersection-Over-Union

References

  1. Guo Y, Liu QN, Wu X, Na J (2016) Gear fault diagnosis based on narrowband demodulation with frequency shift and spectrum edit. Int J Eng Technol Innov 6(4):243-254

    Google Scholar 

  2. Guo Y, Zhao L, Wu X, Na J (2019) Vibration separation technique based localized tooth fault detection of planetary gear sets: A tutorial. Mech Syst Signal Process 129:130-147. https://doi.org/10.1016/j.ymssp.2019.04.027

    Article  Google Scholar 

  3. Bansal S, Sahoo S, Tiwari R, Bordoloi DJ (2013) Multiclass fault diagnosis in gears using support vector machine algorithms based on frequency domain data. Measurement 46(9):3469-3481. https://doi.org/10.1016/j.measurement.2013.05.015

    Article  Google Scholar 

  4. Li JL, Li XY, He D, Qu YZ (2020) A domain adaptation model for early gear pitting fault diagnosis based on deep transfer learning network. Proc Inst Mech Eng Part O: J Risk Reliab 234(1):168-182. https://doi.org/10.1177/1748006X19867776

    Article  Google Scholar 

  5. Kumar A, Gandhi CP, Zhou YQ, Kumar R, Xiang JW (2020) Latest developments in gear defect diagnosis and prognosis: A review. Measurement 158:107735. https://doi.org/10.1016/j.measurement.2020.107735

  6. Han B, Yang XH, Ren YF, Lan WG (2019) Comparisons of different deep learning-based methods on fault diagnosis for geared system. Int J Distrib Sens Netw 15(11):1550147719888169. https://doi.org/10.1177/1550147719888169

    Article  Google Scholar 

  7. Samanta B (2004) Gear fault detection using artificial neural networks and support vector machines with genetic algorithms. Mech Syst Signal Process 18(3):625-644. https://doi.org/10.1016/S0888-3270(03)00020-7

    Article  Google Scholar 

  8. Qiao W, Lu DG (2015) A survey on wind turbine condition monitoring and fault diagnosis-Part I: Components and subsystems. IEEE Trans Ind Electron 62(10):6536-6545. https://doi.org/10.1109/TIE.2015.2422112

    Article  Google Scholar 

  9. Qiao W, Lu DG (2015) A survey on wind turbine condition monitoring and fault diagnosis-Part II: Signals and signal processing methods. IEEE Trans Ind Electron 62(10):6546-6557. https://doi.org/10.1109/TIE.2015.2422394

    Article  Google Scholar 

  10. Li X, Xu YX, Li NP, Yang B, Lei YG (2023) Remaining useful life prediction with partial sensor malfunctions using deep adversarial networks. IEEE/CAA J Autom Sinica 10(1):121-134. https://doi.org/10.1109/JAS.2022.105935

    Article  Google Scholar 

  11. He KM, Zhang XY, Ren SQ, Sun J (2016) Deep residual learning for image recognition. Paper presented at the 2016 IEEE conference on computer vision and pattern recognition, IEEE, Las Vegas, 27-30 June 2016. https://doi.org/10.1109/CVPR.2016.90

  12. Krizhevsky A, Sutskever I, Hinton GE (2012) ImageNet classification with deep convolutional neural networks. Paper presented at the 25th international conference on neural information processing systems, APNNS: Lake Tahoe, 3 December 2012

  13. Simonyan K, Zisserman A (2015) Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556

  14. Redmon J, Farhadi A (2018) YOLOv3: An incremental improvement. arXiv preprint arXiv:1804.02767

  15. He KM, Gkioxari G, Dollar P, Girshick R (2017) Mask R-CNN. Paper presented at the 2017 IEEE international conference on computer vision, IEEE, Venice, 22-29 October 2017. https://doi.org/10.1109/ICCV.2017.322

  16. Li C, Sanchez RV, Zurita G, Cerrada M, Cabrera D, Vásquez RE (2015) Multimodal deep support vector classification with homologous features and its application to gearbox fault diagnosis. Neurocomputing 168:119-127. https://doi.org/10.1016/j.neucom.2015.06.008

    Article  Google Scholar 

  17. Yu LY, Wang Z, Duan ZJ (2019) Detecting gear surface defects using background-weakening method and convolutional neural network. J Sensors 2019:3140980. https://doi.org/10.1155/2019/3140980

    Article  Google Scholar 

  18. Su YT, Yan P (2020) A defect detection method of gear end-face based on modified YOLO-V3. Paper presented at the 2020 10th institute of electrical and electronics engineers international conference on cyber technology in automation, control, and intelligent systems (CYBER), IEEE, Xi'an, 10–13 October 2020. https://doi.org/10.1109/CYBER50695.2020.9279161

  19. Li Y, Cheng G, Pang YS, Kuai MS (2018) Planetary gear fault diagnosis via feature image extraction based on multi central frequencies and vibration signal frequency spectrum. Sensors 18(6):1735. https://doi.org/10.3390/s18061735

    Article  Google Scholar 

  20. Kien BH, Iba D, Ishii Y, Tsutsui Y, Miura N, Iizuka T et al (2019) Crack detection of plastic gears using a convolutional neural network pre-learned from images of meshing vibration data with transfer learning. Forsch Ingenieurwes 83(3):645-653. https://doi.org/10.1007/s10010-019-00354-5

    Article  Google Scholar 

  21. Yang J, Li SB, Wang Z, Dong H, Wang J, Tang SH (2020) Using deep learning to detect defects in manufacturing: a comprehensive survey and current challenges. Materials 13(24):5755. https://doi.org/10.3390/ma13245755

    Article  Google Scholar 

  22. Charles RQ, Hao S, Mo KC, Guibas LJ (2017) PointNet: Deep learning on point sets for 3D classification and segmentation. Paper presented at the 2017 IEEE conference on computer vision and pattern recognition, IEEE, Honolulu, 21–26 July 2017. https://doi.org/10.1109/CVPR.2017.16

  23. Ma YL, Zhang YZ, Luo XF (2019) Automatic recognition of machining features based on point cloud data using convolution neural networks. Paper presented at the 2019 international conference on artificial intelligence and computer science, ACM, Wuhan, 12 July 2019. https://doi.org/10.1145/3349341.3349407

  24. Zhang ZB, Jaiswal P, Rai R (2018) FeatureNet: Machining feature recognition based on 3D convolution neural network. Comput-Aided Des 101:12-22. https://doi.org/10.1016/j.cad.2018.03.006

    Article  Google Scholar 

  25. Zhang W, Wang ZW, Li X (2023) Blockchain-based decentralized federated transfer learning methodology for collaborative machinery fault diagnosis. Reliab Eng Syst Saf 229:108885. https://doi.org/10.1016/j.ress.2022.108885

  26. Guo L, Lei YG, Xing SB, Yan T, Li NP (2019) Deep convolutional transfer learning network: A new method for intelligent fault diagnosis of machines with unlabeled data. IEEE Trans Ind Electron 66(9):7316-7325. https://doi.org/10.1109/TIE.2018.2877090

    Article  Google Scholar 

  27. Wu XY, Zhang Y, Cheng CM, Peng ZK (2021) A hybrid classification autoencoder for semi-supervised fault diagnosis in rotating machinery. Mech Syst Signal Process 149:107327. https://doi.org/10.1016/j.ymssp.2020.107327

  28. Sun X, Lian ZH, Xiao JG (2019) SRINet: Learning strictly rotation-invariant representations for point cloud classification and segmentation. Paper presented at the 27th ACM international conference on multimedia, ACM, Nice, 15 October 2019. https://doi.org/10.1145/3343031.3351042

  29. Wu WX, Qi ZG, Li FX (2019) PointConv: Deep convolutional networks on 3D point clouds. Paper presented at the 2019 IEEE/CVF conference on computer vision and pattern recognition, IEEE, Long Beach, 15–20 June 2019. https://doi.org/10.1109/CVPR.2019.00985

  30. Luo WJ, Li YJ, Urtasun R, Zemel R (2016) Understanding the effective receptive field in deep convolutional neural networks. Paper presented at the 30th international conference on neural information processing systems, NIPS, Barcelona, 5 December 2016

  31. Engelmann F, Kontogianni T, Leibe B (2020) Dilated point convolutions: On the receptive field size of point convolutions on 3D point clouds. Paper presented at the 2020 IEEE international conference on robotics and automation, IEEE, Paris, 31 May–31 August 2020. https://doi.org/10.1109/ICRA40945.2020.9197503

  32. Qi CR, Yi L, Su H, Guibas LJ (2017) PointNet++: Deep hierarchical feature learning on point sets in a metric space. Paper presented at the 31st international conference on neural information processing systems, Long Beach, California, 4 December 2017

  33. Dai JF, Qi HZ, Xiong YW, Li Y, Zhang GD, Hu H et al (2017) Deformable convolutional networks. Paper presented at the IEEE international conference on computer vision, IEEE, Venice, 22–29 October 2017. https://doi.org/10.1109/ICCV.2017.89

  34. Thomas H, Qi CR, Deschaud JE, Marcotegui B, Goulette F, Guibas L (2019) KPConv: Flexible and deformable convolution for point clouds. Paper presented at the 2019 IEEE/CVF international conference on computer vision, IEEE, Seoul, 27 October-2 November 2019. https://doi.org/10.1109/ICCV.2019.00651

  35. Liu W, Anguelov D, Erhan D, Szegedy C, Reed S, Fu CY et al (2016) SSD: Single shot multibox detector. Paper presented at the 14th European conference on computer vision, IEEE, Amsterdam, 17 September 2016. https://doi.org/10.1007/978-3-319-46448-0_2

  36. Ronneberger O, Fischer P, Brox T (2015) U-net: Convolutional networks for biomedical image segmentation. Paper presented at the 18th international conference on medical image computing and computer-assisted intervention, MICCAI, Munich, 18 November 2015. https://doi.org/10.1007/978-3-319-24574-4_28

  37. Milletari F, Navab N, Ahmadi SA (2016) V-net: Fully convolutional neural networks for volumetric medical image segmentation. Paper presented at the 2016 fourth international conference on 3D vision, IEEE, Stanford, 25–28 October 2016. https://doi.org/10.1109/3DV.2016.79

  38. Lin TY, Dollar P, Girshick R, He KM, Hariharan B, Belongie S (2017) Feature pyramid networks for object detection. Paper presented at the 2017 IEEE conference on computer vision and pattern recognition, IEEE, Honolulu, 21–26 July 2017. https://doi.org/10.1109/CVPR.2017.106

  39. Szegedy C, Liu W, Jia YQ, Sermanet P, Reed S, Anguelov D et al (2015) Going deeper with convolutions. Paper presented at the 2015 IEEE conference on computer vision and pattern recognition, IEEE, Boston, 7–12 June 2015. https://doi.org/10.1109/CVPR.2015.7298594

  40. Xu YF, Fan TQ, Xu MY, Zeng L, Qiao Y (2018) SpiderCNN: Deep learning on point sets with parameterized convolutional filters. Paper presented at the 15th European conference on computer vision, IEEE, Munich, 8–14 September 2018. https://doi.org/10.1007/978-3-030-01237-3_6

  41. Wang Y, Sun YB, Liu ZW, Sarma SE, Bronstein MM, Solomon JM (2019) Dynamic graph CNN for learning on point clouds. ACM Trans Graph 38(5):146. https://doi.org/10.1145/3326362

    Article  Google Scholar 

  42. Li YY, Bu R, Sun MC, Wu W, Di XH, Chen BQ (2018) PointCNN: Convolution on X-transformed points. Paper presented at the 32nd international conference on neural information processing systems, NIPS, Montréal, 3 December 2018

Download references

Acknowledgements

Not applicable.

Funding

This work was supported by opening fund of State Key Laboratory of Lunar and Planetary Sciences (Macau University of Science and Technology), No. 119/2017/A3; the Natural Science Foundation of China, Nos. 61572056 and 61872347; and the Special Plan for the Development of Distinguished Young Scientists of ISCAS, No. Y8RC535018.

Author information

Authors and Affiliations

Authors

Contributions

AW provided the conceptualization and methodology; ZX, AW, FH and GZ wrote the original draft; AW, ZX, and FH reviewed and edited the paper. All authors have read and agreed to the published version of the manuscript. All authors read and approved the final manuscript.

Corresponding author

Correspondence to Aizeng Wang.

Ethics declarations

Competing interests

The authors have no competing interests in the manuscript.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Xu, Z., Wang, A., Hou, F. et al. Defect detection of gear parts in virtual manufacturing. Vis. Comput. Ind. Biomed. Art 6, 6 (2023). https://doi.org/10.1186/s42492-023-00133-8

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/s42492-023-00133-8

Keywords