Next Article in Journal
Virtual Reality Usability and Accessibility for Cultural Heritage Practices: Challenges Mapping and Recommendations
Next Article in Special Issue
A Novel Bottom-Up/Top-Down Hybrid Strategy-Based Fast Sequential Fault Diagnosis Method
Previous Article in Journal
Communication-Induced Checkpointing with Message Logging beyond the Piecewise Deterministic (PWD) Model for Distributed Systems
Previous Article in Special Issue
Condition-Based Maintenance of HVAC on a High-Speed Train for Fault Detection
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A CNN Prediction Method for Belt Grinding Tool Wear in a Polishing Process Utilizing 3-Axes Force and Vibration Data

by
Wahyu Caesarendra
1,*,
Triwiyanto Triwiyanto
2,
Vigneashwara Pandiyan
3,
Adam Glowacz
4,*,
Silvester Dian Handy Permana
5 and
Tegoeh Tjahjowidodo
6,7,*
1
Faculty of Integrated Technologies, Universiti Brunei Darussalam, Jalan Tungku Link, Gadong BE1410, Brunei
2
Department of Electromedical Engineering, Health Polytechnic Ministry of Health, Pucang Jajar Timur No. 10, Surabaya 60282, Indonesia
3
Laboratory for Advanced Materials Processing, Empa—Swiss Federal Laboratories for Materials Science & Technology, Feuerwerkerstrasse 39, 3602 Thun, Switzerland
4
Department of Automatic Control and Robotics, Faculty of Electrical Engineering, Automatics, Computer Science and Biomedical Engineering, AGH University of Science and Technology, al. A. Mickiewicza 30, 30-059 Kraków, Poland
5
Department of Informatics, Faculty of Creative Industry and Telematics, Universitas Trilogi, Jakarta 12760, Indonesia
6
School of Mechanical and Aerospace Engineering, Nanyang Technological University, Singapore 639815, Singapore
7
Department of Mechanical Engineering, De Nayer Campus, KU Leuven, Jan Pieter de Nayerlaan 5, 2860 Sint-Katelijne-Waver, Belgium
*
Authors to whom correspondence should be addressed.
Submission received: 3 May 2021 / Revised: 29 May 2021 / Accepted: 9 June 2021 / Published: 14 June 2021

Abstract

:
This paper presents a tool wear monitoring methodology on the abrasive belt grinding process using vibration and force signatures on a convolutional neural network (CNN). A belt tool typically has a random orientation of abrasive grains and grit size variation for coarse or fine material removal. Degradation of the belt condition is a critical phenomenon that affects the workpiece quality during grinding. This work focuses on the identifation and the study of force and vibrational signals taken from sensors along an axis or combination of axes that carry important information of the contact conditions, i.e., belt wear. Three axes of the two sensors are aligned and labelled as X-axis (parallel to the direction of the tool during the abrasive process), Y-axis (perpendicular to the direction of the tool during the abrasive process) and Z-axis (parallel to the direction of the tool during the retract movement). The grinding process was performed using a customized abrasive belt grinder attached to a multi-axis robot on a mild-steel workpiece. The vibration and force signals along three axes (X, Y and Z) were acquired for four discrete sequential belt wear conditions: brand-new, 5-min cycle time, 15-min cycle time, and worn-out. The raw signals that correspond to the sensor measurement along the different axes were used to supervisedly train a 10-Layer CNN architecture to distinguish the belt wear states. Different possible combinations within the three axes of the sensors (X, Y, Z, XY, XZ, YZ and XYZ) were fed as inputs to the CNN model to sort the axis (or combination of axes) in the order of distinct representation of the belt wear state. The CNN classification results revealed that the combination of the XZ-axes and YZ-axes of the accelerometer sensor provides more accurate predictions than other combinations, indicating that the information from the Z-axis of the accelerometer is significant compared to the other two axes. In addition, the CNN accuracy of the XY-axes combination of dynamometer outperformed that of other combinations.

1. Introduction

In manufacturing industries, the requirements for high-quality precision parts with complex geometries have been increasing rapidly [1,2,3]. A workpiece generally goes through various stages such as machining, finishing and polishing to attain prescribed design specification and tolerance [4]. The development of science technology, modern control theory, and advanced machining techniques using multi-axis arm robot has enabled automated finishing and polishing for curved surfaces to achieve uniform quality.
The quality of surface finish for a complex geometry part in manufacturing processes, e.g., grinding and polishing, depends primarily on two main variables, i.e., (1) the condition of belt grinding tool and (2) the combination of operating parameters such as cutting speed, force, feed rate, polymer wheel hardness and grit size. Monitoring the condition of the belt grinding tool is necessary because if the tool deterioration is undetected, it will affect the material removal mechanism and ultimately affect the workpiece’s surface quality. In addition, the polishing process parameters are also an important aspect as they must be adaptable during the manufacturing process depending on certain scenarios or condition such as (1) the curvature shape of the workpiece, (2) the changes in abrasive tool wear condition, (3) the hardness of the workpiece, which is generally not fully uniform and (4) the area of the workpiece being manufacture, e.g., at the edges or toward the part center. This study focuses on predicting belt grinding tool condition from brand-new to wear with certain manufacturing parameters.
Several studies on tool condition monitoring found in manufacturing process literature mainly concentrate on the cutting tool and the milling process [5,6,7,8,9], while those on the belt grinding tool condition and prediction are still limited [10]. The abrasive belt grinding is a modification of the traditional rigid grinding process. However, the advantage of belt grinding over traditional grinding lies in its ease to uniformly machine any intricate geometries for workpieces [11,12]. The polymer wheel backing of the belt grinding tool on which the belt embedded with abrasive grains rests enables conformability with the intricate surfaces [13]. Due to the nature of the polymer backing’s compliant, the process is highly nonlinear and dependent on the process parameters [11,14]. The belt grinding tool consists of three primary parts, namely the driving mechanism, the polymer wheel and the belt itself. The driving mechanism generally consists of at least two wheels in which the belt rotates. The compliance depends primarily on the type and contour of the backing material, and finally, the belt grit size and grain composition define the longevity and the material removal characteristics. Apart from the three primary components, the operating style, such as the applied force, the depth of cut, feeding direction, lubrication, the angle of feed etc., also affects the process dynamics [15].
Out of all the process parameters, only the belt degradation over the cycle time cannot be controlled. The degradation of the belt happens with the abrasive grains withering out or by the grains degenerating resulting in loss of performance of the tool [16]. Several approaches have been proposed on monitoring tool states in the grinding process using the sensor data acquired during the process and state of the art machine learning (ML) algorithm [17,18,19,20]. Apart from the tool wear, Deep Learning (DL) based auto-encoder architecture has been used to perform pixel-level classification of weld seam/bead states [20]. Spectrograms computed from the sound signals with DL method have been used to classify wear states of the abrasive belt [10]. A multi-sensor fusion method of vision and sound have been used along with a light gradient boosting machine (LightGBM) algorithm to monitor in-process grinding material removal rate (MRR) [21,22].
The advancement of the manufacturing process equipped with an intelligent method opens the possibilities to answer the existing challenges. An alternate monitoring strategy using CNN and signals input from the accelerometer and force sensor is proposed in this study for monitoring the belt states. To apply ML and DL methods in tool condition monitoring, the inputs representing the dynamic condition between the abrasive grinding tool and the workpiece play an important role. It has been studied that the appearance of vibration, force and torque variable in the polishing process can be determined as the indicator of the tool wear condition [17]. To date, most of tool wear condition monitoring methods in abrasive processes, such as grinding and polishing, rely on human-interference inspection and is typically conducted as an offline measurement exercise. The offline measurements consequently will interrupt the entire grinding and polishing process due to dismounting and re-mounting procedures of the working coupon to its reference point. This interruption leads to disruptions in the production line. This paper aims to open the possibility of using the DL method for belt tool condition prediction and the potential implementation towards online monitoring systems.
CNN is one of the DL algorithms based on the convolution sliding kernel approach [23]. CNN is usually implemented for classification and prediction methods to avoid unnecessary intermediate step of computing sparse representation such as time, frequency and time-frequency features as they can handle raw data with minimum pre-processing. The CNN algorithm has also been applied in 1-dimensional (1-D) signal, e.g., audio and vibration signal, and 2-dimensional (2-D) data, e.g., images. More details of CNN algorithm is described in Section 2. Specific to the application of CNN in the manufacturing area, especially in tool condition monitoring, CNN has been applied in tool condition monitoring such as for end milling [7]. The previous work of the authors in [13] presents the application of analysis of variance (ANOVA) combined with an adaptive neuro-fuzzy inference system (ANFIS) to model material removal. The combined ANOVA and ANFIS method is used to obtain the optimum configuration between the process parameters/variables of abrasive belt grinding such as RPM, feed, force, rubber hardness and grit size and the stock material removal rate. In the present work, the CNN method is applied to the vibration and force signal from a three-axis accelerometer and a three-axis dynamometer collected during the polishing process to monitor the belt degradation. A new belt grinding tool was used in the polishing process of the mild-steel workpiece with certain manufacturing parameters until wear. The workpiece’s polishing process was performed using a customized abrasive belt grinding attached to a multi-axis arm robot. The vibration and force signals are acquired from the 4 discrete sequential conditions of the belt grinding tool from new to wear (brand-new, 5-min cycle time, 15-min cycle time, worn-out) during the polishing process of the mild-steel workpiece. The ‘5-min cycle time’ and ’15-min cycle time’ mean that the belt is prepared by polishing continuously for 5 and 15 minutes, respectively. This study also aims to correlate the vibration and force sensor axes directions with the belt grinding condition.
The paper is organized into 5 Sections. Section 1 briefly reviews the belt degradation, its influence on surface quality, and the real-time monitoring strategies. Section 2 gives a brief overview of the CNN architecture used for this work. Section 3 introduces the robotized belt grinding experimental setup, processing parameters and data acquisition setup. Section 4 presents and discusses the CNN classification results on the sensor data. Finally, Section 5 summarizes this investigation’s findings and future works.

2. Materials and Methods

2.1. Convolutional Neural Network (CNN)

CNN is a feed-forward neural network inspired by the brain’s visual cortex and specializes in processing data with a grid or a sequential structure [24]. CNN is specifically designed for handling multiple arrays as the input. This initial input process of CNN is analogous to imitating the eyes to identify images, followed by a training process for further recognition of the scene [25]. However, to predict well on tricky scenarios, in many cases CNN must be designed with a more complicated architecture. Even though CNN takes higher computational time and hardware resources compared to traditional ML methods (support vector machine (SVM), artificial neural network (ANN) and adaptive neuro-fuzzy inference system (ANFIS) etc.), they are very efficient in processing raw data with minimum pre-processing.
CNN’s are well known as a first applied method for tool wear prediction compared to other DL methods such as recurrent neural network (RNN), gated neural network (GNN) and long short-term memory (LSTM) [10,11,21,26]. CNN can have different architectures, and it depends based on the problem at hand. The model might have different layers such as VGG-16 (16 layers), LeNet, googleNet (22 layers), AlexNet (8 layers), ResNet (152 layers), etc., and also can use different transformation such as ReLu (rectified linear unit), Dropout and batch normalization. The parameters used for the model training are primarily chosen in such a way as to reduce computational cost and hardware utilization. CNN training parameters are chosen based on exhaustive search or on trial and error [27,28].
The first modern CNN used by Tavakkoli et al. [29] has a 7-layer structure (excluding the input layer), namely LeNet-5, which has the following structures C1, S2, C3, S4, C5 and F6 output. In addition, Zhang et al. [27] reported that several layers in a CNN model are to be considered for better performance. The following layers on CNN considered in this work following [27], are as follows:
  • The sub-sampling layer is usually used to reduce the dimensions coming from the input. The sub-sampling layer’s goal is to reduce the number of trainable parameters needed on CNN, which can speed up the performance of CNN. The speed means computational time. By optimizing the CNN architecture, the accuracy and computational time cost could be obtained as fast as possible. The speed also means using fewer resources in identifying images.
  • The convolutional layer works by imitating the properties of the visual cortex of the brain for studying the features from input images. Filters or kernels trained at this layer can identify specific shapes. For example, if it is used to study an image, the filter might learn for edge detection. The following is the equation in the convolution layer:
f g t f τ g t τ d τ
Equation (1) is a formula to calculate the image filter for each pixel. A detailed illustration is presented in Figure 1, Figure 2, Figure 3, Figure 4 and Figure 5, and a summary of all convolution process is shown in Figure 6.
  • The loss layer is the output of CNN, which is an important part of the neural network. This function calculates loss that corresponds to the difference between the model output and the original ground truth. The magnitude of the loss value determines the rate of gradient change and weights updation during backpropagation.
  • The fully connected layer is found in most neural networks. Fully connected layers use matrix multiplication to get the output from that layer.
  • ReLU layer is used for thresholding or similar to the activation function on neural network. They bring in non-linearity to the model. The following is the equation in the ReLU layer max(0,x) function:
g x = ReLU x = { x   ,     x 0 0   ,   x 0
g x = ReLU x = { 1   ,     x 0 0   ,   x 0
At this stage, the map feature results will be fed into the search function for the highest value between x and 0. The architecture of the ReLU layer process is presented in Figure 7.
  • The pooling layer is very similar to the convolutional layer. This operation in the pooling layer is essential for reducing the spatial size of the feature map. To reduce computational cost, a dimension reduction process is necessary. This layer is useful for extracting domain features that are rotational and invariant so that the model training process becomes more effective. The pooling operations are of many types based on model performance, namely maximum, average, minimum pooling, etc. An example of the max-pooling process of feature map 1 to feature map 9 is presented in Figure 8.
  • Figure 9 shows how a 2D input image transforms when it passes through series of a convolutional layer to produce a feature map. The pooling layer is generally used to reduce the feature map size between convolutions. The pooling method that is often used in CNN is the max pooling method. In very rare cases averaging pooling is also used.
  • The flattening layer changes a matrix data form (2D and 3D) into a one-dimensional array to enter the next layer, typically before the output layer or SoftMax activation. This layer places all pixel data in one row and makes a connection with the last layer. An illustration of a 3 × 3 matrix conversion to a one-dimensional matrix is illustrated in Figure 10.
Figure 8. Max pooling: (a) feature map 1; (b) feature map 9.
Figure 8. Max pooling: (a) feature map 1; (b) feature map 9.
Electronics 10 01429 g008
Figure 9. The architecture of the max-pooling layer.
Figure 9. The architecture of the max-pooling layer.
Electronics 10 01429 g009
Figure 10. The architecture of flattening layer.
Figure 10. The architecture of flattening layer.
Electronics 10 01429 g010
The CNN architecture discussed in this section is suitable for solving image classification, object detection, object location and image segmentation problems. Examples of case studies in image recognition using CNN typically have three stages: the input, CNN and output stages. A detailed explanation of the stages of CNN is presented in [30,31,32]. Figure 11 shows the stages carried out by CNN from the input image, convolutional layer, pooling layer and flattening layer.

2.2. CNN Structure Used in the Present Study

The structure of CNN varies depending on the application, for example, the structure presented in [33]. The details of the CNN architecture to classify belt wear states is presented in Figure 12. The CNN architecture consists of 10 layers, namely 6 one dimensional (1D) convolution layer, 1 max-pooling layer, 1 global average pooling layer, 1 drop layer and 1 fully connected layer. Furthermore, each 1D convolution layer constitutes 200 filters, in which the filter weights have been initialized with random values before training. Additionally, the kernel length of size 4 is used in the convolution process between the signal and filter. At the first stage, the CNN receives the raw time-series 1D signals from a three-axis dynamometer sensor (X, Y and Z) and a three-axis accelerometer sensor (X, Y and Z). In the first layer (L1), each raw signal (X, Y and Z) from the dynamometer and accelerometer sensors are convolved with the filter kernels. Subsequently, the results of the convolutions at each layer are used as input to the next layers. In the fourth layer, to reduce the number of inputs, a pooling layer is applied. In this case, the max-pooling is taken. In the next three layers (L5, L6 and L7), the values from the max-pooling layer are convolved with the filter kernels. In the eighth layer (L8), in order to reduce the number of parameters, the global average pooling layer is applied. In this case, all of the values from each filter were averaged, which results in only one value and turns the total number of the parameter to 1200 (6 inputs × 200 filters). In the ninth layer, a dropping layer is applied to prevent overfitting in the training stage. Finally, all of the nodes from the drop layer is fully connected. In layer 4 (L4), the max-pooling layer was chosen because it helps to have a high contrast feature map of the resulting convolution in the middle of the feature engineering process. On the contrary, average pooling was chosen in the last convolution (L8) to smoothen the feature maps.

3. Experimental Setup

This study employs a 3-axis accelerometer and 3-axis dynamometer sensor signature to predict the belt wear that affects the polishing quality. The following subsection presents in detail the experimental polishing setup and data acquisition procedure:

3.1. Polishing Process Equipment, Sensors and Data Acquisition

The experiments were performed using a multi-axis arm robot mounted with an abrasive belt grinding setup, as presented in Figure 13. The experimental setup consisted of ABB 6660-205-193 robot that was primarily used for imparting motion in the grinding direction. The belt grinder used in this work was electrically powered that can run at 11,000 RPM at no-load condition. The belt grinder could polish and grind with belts about 8″ to 3/4″ wide × 18″ long. The belt grinder was coupled to a force control present in the end effector of the ABB robot with the help of a customized bracket. The force control was closed-looped into the robot controller to ensure the force input from the operator and imparted force in the normal direction (Z-axis) are equal. Normal direction is usually preferred to achieve uniform material removal. As far as the experimental trials, a force of 20 N was applied throughout this study. The experimental conditions used in this work are listed in Table 1. Four different belt condition representing four different classes of wear level in this study were prepared beforehand by performing actual grinding processes at different cycle times, as depicted in Table 2. During the polishing experiments, one variable output process parameter was the belt wear condition (four levels). Other process parameters such as feed rate, grinding speed and force control were maintained at a constant value, as presented in Table 1.
Kistler 8763A500 triaxial accelerometer sensor was attached near the tension arm of the electric belt grinder to obtain the vibration signal during polishing (i.e., contact between belt tool and workpiece), as presented in Figure 13. In addition, Kistler 9254 three-component dynamometer was placed below the mild-steel workpiece, as shown in Figure 13. NI data acquisition (DAQ) module and LabVIEW environment were used to acquire vibration signal in a digitized format with a sampling frequency of 2 kHz. In the case of the dynamometer, a DAQ device supported by the DEWESoft platform was used to acquire force signals across normal and tangential directions. The signals corresponding to the force and accelerometer sensors were synchronized offline.
ABB Robot studio was used to configure and was planned for the tool path. The tool path planning for the linear run performed in this work has five different zones, A–E, as presented in Figure 14. The region between A–B was used to align the tool head over the machining region. Zone B–C was used to attain the force required for grinding. C–D was the zone where force control is actively applied. Finally, Zone D–E was used to decelerate the force sensor and robot motion. The whole experiments performed in this study were carried out in a non-lubricated condition on a mild steel workpiece.

3.2. Vibration and Force Data Acquisition

This paper is an extension work of the previous study that applied ML methods to sensor signatures [17,18]. The difference to the work reported by Pandiyan et al. [17] is that they used short-time Fourier transform (STFT) on acoustic emissions signal to detect the change in contact mechanisms caused by tool wear in abrasive belt grinding. However, in the present study, CNN methods are developed to predict the wear level of the belt tool using 3-axes accelerometer and 3-axes dynamometer signatures instead of the acoustic data. All accelerometer and force sparse features from the 3 axes were used as input data in the ML method [18]. However, extracting features and feeding them into the ML algorithms is a two-step procedure and is computationally expensive. In this study, the CNN algorithm evaluates which axis/axes of the sensor signature(s) are more relevant for classifying the abrasive belt wear state using raw sensor signals. An example of raw vibration and force signal from the X-axis sensor direction for different belt tool states, i.e., brand-new, 5-min cycle time, 15-min cycle time and worn-out (from top to bottom) is presented in Figure 15a,b. For interested readers, a complete raw signals plot of Figure 15 is presented in [18].
The raw vibration and force signals dataset were acquired with a sampling frequency of 2 kHz for 3.5 s, resulting in 7000 data points as presented in Figure 15. Prior to the convolution process, the raw datasets were segmented into a smaller data form. The raw datasets were reshaped to 200 data points with 10% overlaps. The new 200 data points were equivalent to a window of 0.1 sec of the sampling frequency of the polishing process. According to previous works [18], 200 data points can still capture the polishing process’s dynamic behavior. The pre-processing of the raw dataset produced 64,000 new rows of datasets for each class and 256,000 new datasets for all four classes combined, i.e., brand-new, 5-min cycle time, 15-min cycle time and worn-out. These datasets were then divided into two sets for training and validation. The CNN model computation used 80% of 256,000 datasets (192,000 datasets) as the training set and another 20% of 256,000 datasets (64,000 datasets) for validation.
A sample of data input signals for the CNN algorithm from three directions (X, Y and Z-axes) of the accelerometer is presented in Figure 16. Based on the visualization of the vibration signals corresponding to four different belt conditions, it is evident that distinguishing them is challenging. An example of force signals from three directions (X, Y and Z-axes) of the dynamometer is presented in Figure 17. Synonymous to the vibration signals, differentiating belt conditions visually is also challenging. A detailed illustration of the data input preparation for the CNN method is presented in Figure 18. As presented in Figure 18, raw vibration and force data were collected from three axes. The raw data was then divided further into a smaller sized window for CNN training.

4. Results and Discussion

4.1. CNN Results of Vibration Data

This sub-section provides the CNN classification results of vibration signals from three categories: (1) single-axis accelerometer (X, Y and Z), (2) double-axes accelerometer (XY, XZ and YZ) and (3) triple-axes (XYZ).
This study is based on the CNN design, as illustrated in Figure 12 on the accelerometer sensor data. To provide a better understanding for the classification labels in Table 3 confusion matrix results in Table 4, Table 5 and Table 6 and Figure 19 and Figure 20, they are renamed as follows (as presented in Table 2):
  • Brand-new as Belt No.1.
  • 5-min cycle time as Belt No.2.
  • 15-min cycle time as Belt No.3.
  • Worn-out as Belt No.4.

4.1.1. Vibration Signal of Single-Axis Accelerometer (X, Y and Z)

As depicted in Table 3a, the CNN classification result in the X-axis indicates that all 796 datasets of the “brand-new” class can be classified correctly without misclassification. In the “5-min cycle time” class, 796 datasets were classified correctly. However, 2 datasets were misclassified as “15-min cycle time”, and other 2 datasets were misclassified as “worn-out”. In the “15-min cycle time” class, all 798 datasets were also successfully classified. For the “worn-out” class, 793 datasets were classified correctly, and only 1 dataset was identified as a “15-min cycle time” class.
Additionally, Table 3b shows that 795 datasets of the “brand-new” class corresponding to the Y-axis can be classified correctly, and 1 dataset was misclassified as a “5-min cycle time” class. In the “5-min cycle time” class, 798 datasets were classified correctly. However, 2 datasets were extremely misclassified as “worn-out” class. In the “15-min cycle time” class, 717 datasets were classified correctly. However, 83 datasets were classified incorrectly as “worn-out” class. The worst classification occurs in the “worn-out” class, where 670 datasets were classified correctly, but 89 datasets were classified incorrectly as “brand-new” and 16 datasets were identified as “5-min cycle time” class, while other 19 datasets were misclassified to the “15-min cycle time” class.
Table 3c shows CNN results on Z-axis vibrational signals. 790 datasets of “brand new” class were classified correctly, 1 dataset was incorrectly classified as “5-min cycle time” class, and 5 datasets were misclassified as “worn-out” class. In the “5-min cycle time” class, 791 datasets were classified correctly, 3 datasets were incorrectly classified as “15-min cycle time” class, and 6 datasets were incorrectly classified as “worn-out” class. For the “15-min cycle time” class, 748 datasets were classified correctly, and there were 52 datasets classified incorrectly as “worn-out” class. Similar to the Table 3b result, the poor classification result occurred in the “worn-out” class, where class 2 datasets were misclassified as “brand-new” class, 47 datasets were incorrectly identified as “5-min cycle time”, and 26 datasets were misclassified as “15-min cycle time” class.
Table 3 shows the accuracy and loss curves during CNN training using the accelerometer sensor data from X, Y or Z axes. In Table 3, it can be seen that processing 7672 data for 50 epoch takes 17 to 20 s. When trained with sensors derived from the X-axis of the accelerometer, the accuracy of the CNN model results in higher accuracy (accuracy: 95.78) than the other axes (Y-axis, accuracy: 87.47%; Z-axis, accuracy: 89.43%).

4.1.2. Vibration Signal of Double-Axis Accelerometer (XY, XZ and YZ)

The sensor data from two axes were combined to examine the dynamic behavior of the abrasive process. This combined signal is inputted into the CNN algorithm. A confusion matrix from the CNN model that used two-axis (XY, XZ and YZ) from the accelerometer sensor is presented in Table 4. Table 4b indicates that the result across XZ-axis shows higher accuracy with the CNN compared to other combination. Table 4a indicates that all 796 datasets of “brand-new” can be classified appropriately. Additionally, in the “5-min cycle time” class, 797 datasets were classified correctly, 1 dataset was misclassified as “brand-new”, and 2 datasets were misclassified as “15-min cycle time”. For the “15-min cycle time” class, all 800 datasets were classified correctly. In the “worn-out” class, 788 datasets were classified correctly, and another 6 datasets were shifted to the “15-min cycle time” class.
Table 4 shows the accuracy and loss curves during CNN training using two inputs from sensors with a combination of XY, XZ or YZ inputs. In Table 4, it can be seen that processing 7672 data with 50 epoch takes 17 to 20 s. When trained with sensors derived from the XZ-axes of the accelerometer, the accuracy of the CNN model results in higher accuracy (accuracy: 96.40%) than the other axes (XY-axes, accuracy: 95.06%; YZ-axes, accuracy: 94.92%).

4.1.3. Vibration Signal of Triple-Axis Accelerometer (XYZ)

Figure 19 shows the confusion matrix of the CNN model that used the data from the three axes of the accelerometer sensor. Figure 19 indicates that 795 datasets of the “brand-new” class can be classified appropriately, and 1 dataset was misclassified as a “5-min cycle time” class. For the “5-min cycle time” class, 794 datasets were classified correctly, 6 datasets were misclassified to the “15-min cycle time” class. In the “15-min cycle time” class, all 800 datasets were classified correctly. In contrast to the “worn-out” class, 765 datasets were classified correctly, and 29 datasets were misclassified as “15-min cycle time” class.
Figure 19 shows the accuracy and loss model during CNN training using all accelerometer sensor inputs. Figure 8 shows that processing 7672 data with the 50 epochs take 18.2 s. When trained with sensors derived from the XYZ-axis accelerometer, the accuracy of the CNN model yields an accuracy of 94.23%.

4.2. CNN Results of Force Signals

This sub-section provides the CNN classification results of force data from three categories: (1) single-axis (X, Y and Z), (2) double-axis (XY, XZ and YZ) and (3) triple-axis (XYZ).

4.2.1. Force Signal of Single-Axis Dynamometer (X, Y and Z)

This study also utilized the CNN method for the three-axis force data, as presented in Table 5. The CNN result in the X-axis (Table 5a) indicates that 793 datasets of the “brand-new” class can be classified appropriately, and 3 datasets were misclassified to the “5-min cycle time” class. In the “5-min cycle time” class, 797 datasets were classified correctly, 3 datasets were identified as “brand-new” class. In the “15-min cycle time” class, 798 datasets were classified correctly, 1 dataset was classified incorrectly to “brand-new”, and 1 dataset was classified incorrectly as “worn-out”. In the “worn-out” class, 747 datasets were classified correctly. However, 47 datasets were classified incorrectly to the “15-min cycle time” class.
Table 5b shows that 794 datasets of the “brand-new” class can be classified, and 2 datasets were incorrectly classified as “5-min cycle time” class. In the “5-min cycle time” class, 796 datasets were classified correctly, and 4 datasets were identified as “worn-out”. In the “15-min cycle time” class, 771 datasets were classified correctly, and 29 datasets were classified incorrectly to the “worn-out” class. Meanwhile, in the “worn-out” class, 764 datasets were classified correctly, and 30 datasets were classified incorrectly to the “15-min cycle time” class.
Table 5c shows that 794 datasets of the “brand-new” class can be classified correctly. However, 2 datasets were misclassified as “5-min cycle time” class. In the “5-min cycle time” class, 798 datasets were classified correctly, and 2 datasets were incorrectly classified as “15-min cycle time”. In the “15-min cycle time” class, 788 datasets were classified correctly. However, 12 datasets were classified incorrectly to the “worn-out” class. In the “worn-out” class, 785 datasets were classified correctly, and 9 datasets were misclassified as “15-min cycle time” class.
Table 5 shows the accuracy and loss curves during CNN training using force data input from X, Y or Z axes. In Table 5, it can be seen that processing 7672 data with 50 epochs took nearly 17 to 20 s. When trained with sensors derived from the X-axis of the dynamometer, the accuracy of the CNN model resulted in higher accuracy (accuracy: 93.07%) than the other axes (Y-axis, accuracy: 92.14%; Z-axis, accuracy: 92.94%).

4.2.2. Force Signal of Double-Axis Dynamometer (XY, XZ and YZ)

The force signature across two directions were combined to examine the dynamic behavior during the abrasive process. This combined signal was input into the CNN algorithm. The confusion matrices from the CNN model that used two-axis (XY, XZ and YZ) is presented in Table 6. According to the CNN results presented in Table 6, Table 6a indicates that the CNN model results show highest than other combinations. In detail, Table 6a indicates that all 796 datasets of the “brand-new” class can be classified perfectly. For the “5-min cycle time” class, 797 datasets were classified correctly, 1 dataset was identified as “brand-new” class, and 2 datasets were incorrectly classified as “15-min cycle time” class. In the “used” class, all 800 datasets were also classified perfectly. Furthermore, in the “worn-out” class, 788 datasets were classified correctly, and 6 datasets were incorrectly identified as “15-min cycle time” class.
Table 6 shows the accuracy and loss curves during CNN training with inputs from axes combinations such as of XY, XZ or YZ. In Table 6, it can be seen that processing 7672 data with the number of 50 epochs takes 17 to 20 s. When trained with sensors derived from the XY-axis of the dynamometer, the accuracy of the CNN model results in higher accuracy (accuracy: 99.80%) than the other axes (XZ-axes, accuracy: 95.53%; YZ-axes, accuracy: 96.47%).

4.2.3. Force Signal of Triple-Axis Dynamometer (XYZ)

Model accuracy and the confusion matrix from the CNN model that used tree axis from dynamometer sensor is presented in Figure 20. Figure 20 indicates that 793 datasets of “brand-new” can be classified appropriately, and three datasets were misclassified to the “5-min cycle time” class. Additionally, in the “5-min cycle time” class, 797 datasets were classified correctly, 2 datasets were misclassified to “brand-new” class and 1 dataset was misclassified to “15-min cycle time” class. Moreover, in the “15-min cycle time” class, 798 datasets were classified correctly and 2 datasets were incorrectly classified to “worn-out” class. For the “worn-out” class, 790 datasets were classified correctly and 4 datasets were misclassified to the “15-min cycle time” class.
Figure 20 shows the accuracy and loss curves during CNN model training using all dynamometer sensor inputs. In Figure 9, it can be seen that processing 7672 data with the number of 50 epochs takes 18.2 s. When trained with sensors derived from the XYZ-axis dynamometer, the accuracy of the CNN model yields an accuracy of 96.34%.

4.3. Statistical Measurement of CNN Prediction and Classification

4.3.1. Accelerometer (Vibration) Data

This section provides a statistical summary of CNN prediction and classification of abrasive belt grinding condition from brand-new to worn-out condition. The statistical summary of metrics such as training time, testing time, accuracy on test data and loss on test data.

Training Time

Table 7 shows the summary of ‘training time’ for the CNN model. Each combination of three-axis accelerometer sensor showed different time required to train the model. Table 7 indicates that a combination of the XY axes of the accelerometer sensor shows less ‘training time’, i.e., 14.71 ± 0.51 s than other axis combination.
A box plot of the ‘training time’ of the CNN model for a different accelerometer axis direction is presented in Figure 21. All of the model’s combination input showed that the standard deviation value is small (Q3-Med), as presented in Table 7. This is indicated that the proposed architecture model was able to train the CNN model. The highest time required to train the model is shown by the XY combination (29.64 s). In training the model, using a single input axis (X, Y or Z), it is shown that the time required is almost similar. However, by using more input (all axis), the time needed to train the model was not reduced.
The boxplot is divided into four quartiles, namely 1st, 2nd, 3rd, and 4th quartiles. The 1st, 2nd, 3rd, and 4th quartiles show a range data distribution of 0–25%, 25–50%, 50–75%, and 75–100%, respectively. The grey and yellow color in the boxplot of Figure 21, Figure 22, Figure 23, Figure 24, Figure 25, Figure 26, Figure 27 and Figure 28. indicates the second, and third quartile, respectively. A wider box in the boxplot shows more distribution of data.

Testing Time

Table 8 shows the summary of ‘testing time’ for the CNN model. Each combination of three-axis accelerometer sensor showed a different time to test the model. Table 8 indicates that a combination of the accelerometer sensor’s YZ-axes shows less time (0.22 ± 0.01 s) than other axis combination.
A box plot of ‘testing time’ of the CNN model for a different combination of accelerometer axis is presented in Figure 22. All of the model’s combination input showed that the standard deviation value is low (Q3-Med), as presented in Table 8. This indicated that the proposed architecture model was able to test the CNN model. The highest time required to test the model is shown by the XZ combination (0.33 s). In training the model, using a single input axis (X, Y or Z), it is shown that the required processing time is almost similar. However, by using more input (all axis), the time needed to train the model was not reduced. This issue was confirmed when we compared between YZ and XYZ combination.

Accuracy of Test Data

Table 9 shows the summary of the performance of the CNN model. Each combination of three-axis accelerometer sensor showed different accuracy. Table 9 indicates that a combination of the accelerometer sensor XY-axes shows better accuracy than that of other combinations (1.00 ± 0.00). The lowest accuracy was found when the model was trained using Y-axis (0.95 ± 0.03).
A box plot of ‘accuracy on test data’ of the CNN model for a different combination of accelerometer axis is presented in Figure 23. The CNN model, which is trained using XY-axis datasets, showed the highest accuracy (1.00 ± 0.00). In contrast, The CNN model based on Y-axis indicated the lowest accuracy (0.89 ± 0.02). Additionally, the CNN which is trained using Y and Z datasets, showed wider deviation among the others axis combination.

Loss on Test Data

Table 10 shows the loss value of the CNN model. The loss value is important to indicate the summation of the error in the evaluation and testing datasets. Each combination of three-axis accelerometer sensor showed different loss value. Table 10 indicates that a combination of the accelerometer sensor’s XY-axes shows a better loss value than other combinations (0.02 ± 0.01). The worst loss value was found when the model was trained using Y-axis (0.20 ± 0.05).
A box plot of ‘loss on test data’ of the CNN model for a different combination of accelerometer axis is presented in Figure 24. The CNN model, which is trained using XY-axis datasets, showed the lowest value (0.02 ± 0.01). In contrast, The CNN model based on the Y and Z axis indicated the higher loss and wider deviation value among others axis combination.

4.3.2. Dynamometer (Force) Data

This section provides a statistical summary of CNN prediction and classification of abrasive belt grinding condition from brand-new to worn-out condition. The statistical summary includes ‘training time’, ‘testing time’, ‘accuracy on test data’ and ‘lost on test data’.

Training Time

Table 11 shows the summary of ‘training time’ for the CNN model. Each combination of three-axis dynamometer sensor showed different time required to train the model. Table 11 indicates that a combination of the XY-axes of the dynamometer sensor shows less training time (14.23 ± 0.54 s) than other axis combinations.
A box plot of the ‘training time’ of the CNN model for a different combination of dynamometer axis is presented in Figure 25. All of the combined input of the model showed that the value of the standard deviation is small, as presented in Table 11. This is indicated that the proposed architecture model was able to train the CNN model. The highest time required to train the model is presented on Z-axis (32.49 s). Using double and triple input axis in training, the model shows that the ‘training time’ required is less than the single-axis, especially XY with 14.23 s on average.

Testing Time

Table 12 shows the summary of ‘testing time’ for the CNN model. Each combination of three-axis accelerometer sensor showed a different time to test the model. Table 12 indicates that a combination of XZ-axes of accelerometer sensor shows less time (0.22 ± 0.01 s) than other axis combination.
A box plot of ‘testing time’ of the CNN model for a different combination of dynamometer axis is presented in Figure 26. All of the model’s combination input showed that the standard deviation value is small, as presented in Table 12. This is indicated that the proposed architecture model was able to test the CNN model. The highest time required to test the model is shown by the XZ combination (0.21 s) compared to the single input axis (X, Y and Z) and double input axis (XY and YZ). In addition, by using more inputs (XYX) the time needed to test the model was not improved even better.

Accuracy on Test Data

Table 13 shows the summary of the performance of the CNN model. Each combination of three-axis dynamometer sensor showed different accuracy. Table 13 indicates that a combination of the XZ-axes of the dynamometer sensor shows better accuracy (1.00 ± 0.00) than other axis combinations. The lowest accuracy was found when the model was trained using Y-axis datasets, i.e., 0.95 ± 0.02.
A box plot of ‘accuracy on test data’ of the CNN model for a different combination of dynamometer axis is presented in Figure 27. The CNN model, which is trained using XZ-axis datasets, showed the highest accuracy (1.00 ± 0.00). In contrast, The CNN model based on Y-axis indicated the lowest accuracy (0.95 ± 0.02). Additionally, the CNN which is trained using Y and Z datasets showed wider deviation among the others axis combination.

Loss on Test Data

Table 14 shows the loss value of the CNN model. The loss value is important to indicate the summation of the error in the evaluation and testing datasets. Each combination of three-axis dynamometer sensor showed different loss value. Table 14 indicates that a combination of the XZ-axes of the accelerometer sensor shows a better loss value than other combinations (0.02 ± 0.01). The worst loss value was found when the model was trained using Y-axis (0.15 ± 0.03).
A box plot of the “loss on test data” of the CNN model for a different combination of dynamometer axis is presented in Figure 28. The CNN model, which is trained using XZ-axis datasets, showed the lowest value (0.02 ± 0.01). In contrast, The CNN model based on the Y and Z axis indicated the higher loss and wider deviation value among others axis combination.

4.3.3. Discussion

Based on the CNN results presented in Section 4.1 for vibration signals and Section 4.2 for force signals, it is found that the sensor data across different axes has significant variation in representing the dynamic behavior of the abrasive process, which is directly related to the belt grinding condition.
In vibration signals, the X-axis direction shows the highest classification accuracy among the single-axis sensors assessment compared to those from the other two single-axis sensors assessment compared to those from the other two (Y-axis and Z-axis). The result of the combining axes showed that the combination of XZ and YZ has higher accuracy than the XY combination. Moreover, the signal combination from the triple-axis (XYZ) does not improve the accuracy. The accuracy summary of the three-axis accelerometer signals assessment is presented in Table 15.
In force signals, all single-axis sensor assessment shows significant accuracy drop due to a lot of misclassified datasets. In addition, the result combining sensor data between two axes shows that higher accuracy is obtained from the combination of XY compared to the other combinations (XZ and YZ). Moreover, the triple-axis (XYZ) signal combination does not show better accuracy compared to the single-axis and double-axis assessment. The accuracy summary of the three-axis accelerometer signals assessment is presented in Table 16.

5. Conclusions

A methodology for abrasive belt grinding condition monitoring and worn state prediction is presented. The methodology is focused on understanding the effect of the directions of vibration and force signatures with respect to the belt wear rate. This proposed methodology also responds to the open question, i.e., opening the possibility of applying DL for tool condition monitoring and its potential application. From the presented result, the methodology revealed that the accelerometer and dynamometer’s direction has a certain relation to the physical behavior of the belt grinding process, especially on the dynamics in the contact between the belt tool and workpiece. In vibration signal analysis, XZ and YZ direction combination have more accuracy than the others. In contrast, in force signal analysis, the combination of XY gives the most accurate results in comparison to the others. Another outcome of this study is also attributed to our future work that the DL model can be embedded into a monitoring measurement system to monitor the belt state during the abrasive process without interrupting the ongoing grinding process. Even though the DL method has potential application in tool wear prediction of mild steel, this study needs further investigation if the proposed DL prediction can be generalized and translated across different materials during grinding, which is the research direction in progress.

Author Contributions

Conceptualization, W.C. and T.T. (Triwiyanto Triwiyanto); methodology, W.C., and T.T. (Triwiyanto Triwiyanto); software, T.T. (Triwiyanto Triwiyanto); validation, W.C. and T.T. (Triwiyanto Triwiyanto); formal analysis, W.C. and T.T. (Triwiyanto Triwiyanto); resources, V.P. and T.T. (Tegoeh Tjahjowidodo); data curation, V.P.; writing—original draft preparation, W.C., T.T. (Triwiyanto Triwiyanto), V.P. and S.D.H.P.; writing—review and editing, W.C., A.G. and T.T. (Tegoeh Tjahjowidodo); visualization, W.C., T.T. (Triwiyanto Triwiyanto) and V.P.; supervision, A.G. and T.T.(Tegoeh Tjahjowidodo); project administration, V.P.; funding acquisition, W.C., A.G. and T.T. (Tegoeh Tjahjowidodo). All authors have read and agreed to the published version of the manuscript.

Funding

The first author would like to thank to Universiti Brunei Darussalam for providing a research grant for this study through Research Grant No. UBD/RSCH/1.3/FICBF(b)/2019/007 and The APC was funded by Professor Dr. Adam Glowacz.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zhan, J.; Yu, S. Study on error compensation of machining force in aspheric surfaces polishing by profile-adaptive hybrid movement–force control. Int. J. Adv. Manuf. Technol. 2011, 54, 879–885. [Google Scholar] [CrossRef]
  2. Hauth, S.; Linsen, L. Cycloids for polishing along doublespiral tool paths in configuration space. Int. J. Adv. Manuf. Technol. 2012, 60, 343–356. [Google Scholar] [CrossRef]
  3. Tian, Y.B.; Zhong, Z.W.; Lai, S.T.; Ang, Y.J. Development of fixed abrasive chemical mechanical polishing process for glass disk substrates. J. Adv. Manuf. Technol. 2013, 68, 993–1000. [Google Scholar] [CrossRef]
  4. Cheung, C.F.; Kong, L.B.; Ho, L.T.; To, S. Modelling and simulation of structure surface generation using computer con-trolled ultra-precision polishing. Precis. Eng. 2011, 35, 574–590. [Google Scholar] [CrossRef]
  5. Sun, H.; Liu, Y.; Pan, J.; Zhang, J.; Ji, W. Enhancing cutting tool sustainability based on remaining useful life prediction. J. Clean. Prod. 2020, 244, 118794. [Google Scholar] [CrossRef]
  6. Liu, C.; Li, Y.; Hua, J.; Lu, N.; Mou, W. Real-time cutting tool state recognition approach based on machining features in NC machining process of complex structural parts. Int. J. Adv. Manuf. Technol. 2018, 97, 229–241. [Google Scholar] [CrossRef]
  7. Kothuru, A.; Nooka, S.P.; Liu, R. Application of deep visualization in CNN-based tool condition monitoring for end milling. Procedia Manuf. 2019, 34, 995–1004. [Google Scholar] [CrossRef]
  8. Zhou, Y.; Xue, W. Review of tool condition monitoring methods in milling processes. Int. J. Adv. Manuf. Technol. 2018, 96, 2509–2523. [Google Scholar] [CrossRef]
  9. Mohanraj, T.; Shankar, S.; Rajasekar, R.; Sakthivel, N.; Pramanik, A. Tool condition monitoring techniques in milling process—A review. J. Mater. Res. Technol. 2020, 9, 1032–1042. [Google Scholar] [CrossRef]
  10. Cheng, C.; Li, J.; Liu, Y.; Nie, M.; Wang, W. Deep convolutional neural network-based in-process tool condition monitoring in abrasive belt grinding. Comput. Ind. 2019, 106, 1–13. [Google Scholar] [CrossRef]
  11. Pandiyan, V.; Shevchik, S.; Wasmer, K.; Castagne, S.; Tjahjowidodo, T. Modelling and monitoring of abrasive finishing processes using artificial intelligence techniques: A review. J. Manuf. Process. 2020, 57, 114–135. [Google Scholar] [CrossRef]
  12. Pandiyan, V. Modelling and In-Process Monitoring of Abrasive Belt Grinding Process. Ph.D. Thesis, Nanyang Technological University, Singapore, 2019. [Google Scholar]
  13. Pandiyan, V.; Caesarendra, W.; Tjahjowidodo, T.; Praveen, G. Predictive Modelling and Analysis of Process Parameters on Material Removal Characteristics in Abrasive Belt Grinding Process. Appl. Sci. 2017, 7, 363. [Google Scholar] [CrossRef]
  14. Pandiyan, V.; Tjahjowidodo, T. In-process endpoint detection of weld seam removal in robotic abrasive belt grinding process. Int. J. Adv. Manuf. Technol. 2017, 93, 1699–1714. [Google Scholar] [CrossRef]
  15. Zhang, X.; Kuhlenkötter, B.; Kneupner, K. An efficient method for solving the Signorini problem in the simulation of free-form surfaces produced by belt grinding. Int. J. Mach. Tools Manuf. 2005, 45, 641–648. [Google Scholar] [CrossRef]
  16. Pandiyan, V.; Tjahjowidodo, T.; Samy, M.P. In-Process Surface Roughness Estimation Model for Compliant Abrasive Belt Machining Process. Procedia CIRP 2016, 46, 254–257. [Google Scholar] [CrossRef] [Green Version]
  17. Pandiyan, V.; Tjahjowidodo, T. Use of Acoustic Emissions to detect change in contact mechanisms caused by tool wear in abrasive belt grinding process. Wear 2019, 436–437, 203047. [Google Scholar] [CrossRef]
  18. Pandiyan, V.; Caesarendra, W.; Tjahjowidodo, T.; Tan, H.H. In-process tool condition monitoring in compliant abrasive belt grinding process using support vector machine and genetic algorithm. J. Manuf. Process. 2018, 31, 199–213. [Google Scholar] [CrossRef]
  19. Zhang, X.; Chen, H.; Xu, J.; Song, X.; Wang, J.; Chen, X. A novel sound-based belt condition monitoring method for robotic grinding using optimally pruned extreme learning machine. J. Mater. Process. Technol. 2018, 260, 9–19. [Google Scholar] [CrossRef]
  20. Pandiyan, V.; Murugan, P.; Tjahjowidodo, T.; Caesarendra, W.; Manyar, O.M.; Then, D.J.H. In-process virtual verification of weld seam removal in robotic abrasive belt grinding process using deep learning. Robot. Comput. Manuf. 2019, 57, 477–487. [Google Scholar] [CrossRef]
  21. Wang, N.; Zhang, G.; Ren, L.; Pang, W.; Wang, Y. Vision and sound fusion-based material removal rate monitoring for abrasive belt grinding using improved LightGBM algorithm. J. Manuf. Process. 2021, 66, 281–292. [Google Scholar] [CrossRef]
  22. Wang, N.; Zhang, G.; Pang, W.; Ren, L.; Wang, Y. Novel monitoring method for material removal rate considering quantitative wear of abrasive belts based on LightGBM learning algorithm. Int. J. Adv. Manuf. Technol. 2021, 114, 3241–3253. [Google Scholar] [CrossRef]
  23. Triwiyanto, T.; Pawana, I.P.A.; Purnomo, M.H. An Improved Performance of Deep Learning Based on Convolution Neural Network to Classify the Hand Motion by Evaluating Hyper Parameter. IEEE Trans. Neural Syst. Rehabilitation Eng. 2020, 28, 1678–1688. [Google Scholar] [CrossRef]
  24. Lindsay, G.W. Convolutional Neural Networks as a Model of the Visual System: Past, Present, and Future. J. Cogn. Neurosci. 2020, 1–15. [Google Scholar] [CrossRef] [Green Version]
  25. Demir, F.; Turkoglu, M.; Aslan, M.; Sengur, A. A new pyramidal concatenated CNN approach for environmental sound classification. Appl. Acoust. 2020, 170, 107520. [Google Scholar] [CrossRef]
  26. Cao, X.-C.; Chen, B.-Q.; Yao, B.; He, W.-P. Combining translation-invariant wavelet frames and convolutional neural network for intelligent tool wear state identification. Comput. Ind. 2019, 106, 71–84. [Google Scholar] [CrossRef]
  27. Zhang, Q.; Zhang, M.; Chen, T.; Sun, Z.; Ma, Y.; Yu, B. Recent advances in convolutional neural network acceleration. Neurocomputing 2019, 323, 37–51. [Google Scholar] [CrossRef] [Green Version]
  28. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  29. Tavakkoli, V.; Mohsenzadegan, K.; Kyamakya, K. A Visual Sensing Concept for Robustly Classifying House Types through a Convolutional Neural Network Architecture Involving a Multi-Channel Features Extraction. Sensors 2020, 20, 5672. [Google Scholar] [CrossRef] [PubMed]
  30. Zhao, Z.-Q.; Zheng, P.; Xu, S.-T.; Wu, X. Object Detection with Deep Learning: A Review. IEEE Trans. Neural Netw. Learn. Syst. 2019, 30, 3212–3232. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  31. Wang, Y.; Li, Y.; Song, Y.; Rong, X. The influence of the activation function in a convolution neural network model of facial expression recognition. Appl. Sci. 2020, 10, 1897. [Google Scholar] [CrossRef] [Green Version]
  32. Shrestha, A.; Mahmood, A. Review of Deep Learning Algorithms and Architectures. IEEE Access 2019, 7, 53040–53065. [Google Scholar] [CrossRef]
  33. Nguyen, P.T.; Ruscio, D.D.; Pierantonio, A.; Rocco, J.D.; Lovino, L. Convolutional neural networks for enhanced classification mechanisms of metamodels. J. Syst. Softw. 2021, 172, 110860. [Google Scholar] [CrossRef]
Figure 1. Binary conversion of feature map 1.
Figure 1. Binary conversion of feature map 1.
Electronics 10 01429 g001
Figure 2. Binary conversion of feature map 2.
Figure 2. Binary conversion of feature map 2.
Electronics 10 01429 g002
Figure 3. Binary conversion of feature map 3.
Figure 3. Binary conversion of feature map 3.
Electronics 10 01429 g003
Figure 4. Binary conversion of feature map 4.
Figure 4. Binary conversion of feature map 4.
Electronics 10 01429 g004
Figure 5. Binary conversion of feature map 25.
Figure 5. Binary conversion of feature map 25.
Electronics 10 01429 g005
Figure 6. The architecture of convolutional layer (binary conversion) from feature map 1 to 25.
Figure 6. The architecture of convolutional layer (binary conversion) from feature map 1 to 25.
Electronics 10 01429 g006
Figure 7. ReLU layer function applied in convolutional layer process.
Figure 7. ReLU layer function applied in convolutional layer process.
Electronics 10 01429 g007
Figure 11. Step-by-step process of CNN from convolutional layer, pooling layer to flattening layer.
Figure 11. Step-by-step process of CNN from convolutional layer, pooling layer to flattening layer.
Electronics 10 01429 g011
Figure 12. CNN model and architecture for 3-axis acceleration and 3-axis force data training and prediction.
Figure 12. CNN model and architecture for 3-axis acceleration and 3-axis force data training and prediction.
Electronics 10 01429 g012
Figure 13. A customized abrasive belt-grinding experiment and sensors position (accelerometer and dynamometer).
Figure 13. A customized abrasive belt-grinding experiment and sensors position (accelerometer and dynamometer).
Electronics 10 01429 g013
Figure 14. Abrasive tool path on the mild-steel workpiece.
Figure 14. Abrasive tool path on the mild-steel workpiece.
Electronics 10 01429 g014
Figure 15. Raw signals of x-axis sensor direction: (a) vibration and (b) force.
Figure 15. Raw signals of x-axis sensor direction: (a) vibration and (b) force.
Electronics 10 01429 g015
Figure 16. Vibration signals: (a) X-axis, (b) Y-axis and (c) Z-axis.
Figure 16. Vibration signals: (a) X-axis, (b) Y-axis and (c) Z-axis.
Electronics 10 01429 g016
Figure 17. Force signals: (a) X-axis, (b) Y-axis and (c) Z-axis.
Figure 17. Force signals: (a) X-axis, (b) Y-axis and (c) Z-axis.
Electronics 10 01429 g017
Figure 18. Data input preparation for CNN algorithm. The data consist of 3-axes vibration and 3-axes force signal.
Figure 18. Data input preparation for CNN algorithm. The data consist of 3-axes vibration and 3-axes force signal.
Electronics 10 01429 g018
Figure 19. (a) Model accuracy and loss curves; and (b) confusion matrix of triple-axis vibration data (please see Table 2 for detailed information on classification label).
Figure 19. (a) Model accuracy and loss curves; and (b) confusion matrix of triple-axis vibration data (please see Table 2 for detailed information on classification label).
Electronics 10 01429 g019
Figure 20. (a) Model accuracy and loss; and (b) Confusion matrix of triple-axis force data. (Please see Table 2 for detailed information on the classification label).
Figure 20. (a) Model accuracy and loss; and (b) Confusion matrix of triple-axis force data. (Please see Table 2 for detailed information on the classification label).
Electronics 10 01429 g020
Figure 21. Box plot of ‘training time’ for all sensor axis of the accelerometer.
Figure 21. Box plot of ‘training time’ for all sensor axis of the accelerometer.
Electronics 10 01429 g021
Figure 22. Box plot of ‘testing time’ for all sensor axis of the accelerometer.
Figure 22. Box plot of ‘testing time’ for all sensor axis of the accelerometer.
Electronics 10 01429 g022
Figure 23. Box plot of ‘accuracy on test data’ for all sensor axis of the accelerometer.
Figure 23. Box plot of ‘accuracy on test data’ for all sensor axis of the accelerometer.
Electronics 10 01429 g023
Figure 24. Box plot of ‘loss on test data’ for all sensor axis of the accelerometer.
Figure 24. Box plot of ‘loss on test data’ for all sensor axis of the accelerometer.
Electronics 10 01429 g024
Figure 25. Box plot of ‘training time’ for all sensor axis of the dynamometer.
Figure 25. Box plot of ‘training time’ for all sensor axis of the dynamometer.
Electronics 10 01429 g025
Figure 26. Box plot of ‘testing time’ for all sensor axis of the dynamometer.
Figure 26. Box plot of ‘testing time’ for all sensor axis of the dynamometer.
Electronics 10 01429 g026
Figure 27. Box plot of ‘accuracy on test data’ for all sensor axis of the dynamometer.
Figure 27. Box plot of ‘accuracy on test data’ for all sensor axis of the dynamometer.
Electronics 10 01429 g027
Figure 28. Box plot of ‘loss on test data’ for all sensor axis of the dynamometer.
Figure 28. Box plot of ‘loss on test data’ for all sensor axis of the dynamometer.
Electronics 10 01429 g028
Table 1. Experimental condition.
Table 1. Experimental condition.
ParametersDescription
Belt typeAluminium oxide with grit size 60
Force controlConstant force of 20 N using ATI force sensor
Grinding speedConstant at feed rate of 50 mm/s
WorkpieceMild steel with uniform surface roughness
Contact wheelShore A Hardness 80
Table 2. Four states of abrasive belt under monitoring.
Table 2. Four states of abrasive belt under monitoring.
Belt No.Abrasive Belt Grinding LabelUsage Time (min)
1Brand-new belt0
25-min cycle time belt5
315-min cycle time belt15
4Worn-out belt30
Table 3. Training curves and confusion matrix of single-axis vibration data.
Table 3. Training curves and confusion matrix of single-axis vibration data.
Model Accuracy and LossConfusion Matrix
(a)
Vibration data of
X-axis
Electronics 10 01429 i001 Electronics 10 01429 i002
(b)
Vibration data of
Y-axis
Electronics 10 01429 i003 Electronics 10 01429 i004
(c)
Vibration data of
Z-axis
Electronics 10 01429 i005 Electronics 10 01429 i006
Note: please see Table 2 for detailed information on the classification label.
Table 4. Training curves and confusion matrix of double-axis vibration data.
Table 4. Training curves and confusion matrix of double-axis vibration data.
Model Accuracy and LossConfusion Matrix
(a)
Vibration data of
XY-axis
Electronics 10 01429 i007 Electronics 10 01429 i008
(b)
Vibration data of
XZ-axis
Electronics 10 01429 i009 Electronics 10 01429 i010
(c)
Vibration data of
YZ-axis
Electronics 10 01429 i011 Electronics 10 01429 i012
Note: please see Table 2 for detailed information on the classification label.
Table 5. Model accuracy and loss and confusion matrix of single-axis force data.
Table 5. Model accuracy and loss and confusion matrix of single-axis force data.
Model Accuracy and LossConfusion Matrix
(a)
Force data of X-axis
Electronics 10 01429 i013 Electronics 10 01429 i014
(b)
Force data of Y-axis
Electronics 10 01429 i015 Electronics 10 01429 i016
(c)
Force data of Z-axis
Electronics 10 01429 i017 Electronics 10 01429 i018
Note: please see Table 2 for detailed information on the classification label.
Table 6. Training curves and confusion matrix of double-axis force data.
Table 6. Training curves and confusion matrix of double-axis force data.
Model Accuracy and LossConfusion Matrix
(a)
Force data of XY-axis
Electronics 10 01429 i019 Electronics 10 01429 i020
(b)
Force data of XZ-axis
Electronics 10 01429 i021 Electronics 10 01429 i022
(c)
Force data of YZ-axis
Electronics 10 01429 i023 Electronics 10 01429 i024
Note: please see Table 2 for detailed information on the classification label.
Table 7. CNN summary of ‘training time’ (sec).
Table 7. CNN summary of ‘training time’ (sec).
Data No.#XYZXYXZYZXYZ
120.2118.2917.9831.1127.9116.0618.20
218.0518.1017.9129.1127.9914.6018.00
318.1818.0618.1629.0927.9714.5718.57
418.2717.9718.0229.6427.8215.0318.37
519.2318.1218.0929.0028.0014.6418.43
618.3118.0818.1029.0728.0414.3018.14
718.4518.1918.4029.0627.6914.4618.23
818.1617.9418.0329.2627.6914.4318.65
917.9718.3517.9129.0727.8114.4718.35
1018.2117.8818.0328.9727.7614.5618.53
Table 8. CNN summary of ‘testing time’ (s).
Table 8. CNN summary of ‘testing time’ (s).
Data No.XYZXYXZYZXYZ
10.280.280.290.310.350.210.22
20.290.280.300.310.320.220.23
30.290.300.290.310.350.230.23
40.290.290.290.320.320.240.23
50.290.290.280.350.360.210.24
60.290.300.300.320.310.240.22
70.290.310.120.310.310.210.23
80.300.290.290.320.300.210.24
90.280.290.290.350.310.230.25
100.290.280.290.340.320.240.23
Table 9. CNN summary of ‘accuracy on test data’.
Table 9. CNN summary of ‘accuracy on test data’.
Data No.XYZXYXZYZXYZ
11.000.930.991.000.991.000.99
20.990.960.961.000.980.991.00
31.000.980.951.001.001.001.00
41.000.940.991.001.000.981.00
51.000.940.971.001.001.001.00
60.980.950.991.001.001.000.99
70.990.980.981.001.001.000.99
80.990.890.971.001.000.991.00
91.000.940.941.001.001.001.00
101.000.970.991.001.000.991.00
Table 10. CNN summary of ‘loss on test data’.
Table 10. CNN summary of ‘loss on test data’.
Data No.XYZXYXZYZXYZ
10.030.230.070.020.040.040.05
20.050.210.180.020.070.110.02
30.030.130.150.030.010.020.07
40.040.240.110.010.020.150.04
50.040.210.170.020.020.040.07
60.090.210.060.030.020.020.08
70.060.120.120.020.010.050.05
80.040.290.150.030.010.070.04
90.040.230.150.020.020.050.05
100.040.160.110.020.020.050.04
Table 11. CNN summary of ‘training time’ (s).
Table 11. CNN summary of ‘training time’ (s).
No.XYZXYXZYZXYZ
130.1729.6932.4915.6818.5830.9730.63
230.7230.6830.9313.9715.1228.7028.18
331.7030.7431.4213.9315.4128.4428.34
430.3930.1229.5413.9515.4928.6628.50
529.5029.6429.7713.9915.6328.7228.26
630.2831.5030.7913.9915.5028.4128.44
731.6629.7329.7213.9615.5729.1928.45
829.8229.6031.5614.1015.5828.7128.19
930.0629.3429.2714.3615.4128.3228.26
1031.2829.4229.8214.3615.3928.4028.33
Table 12. CNN summary of ‘testing time’ (s).
Table 12. CNN summary of ‘testing time’ (s).
No.XYZXYXZYZXYZ
10.330.360.340.330.220.310.31
20.310.370.340.350.210.310.31
30.340.340.380.340.210.310.33
40.330.390.350.350.210.320.31
50.320.330.340.350.210.320.34
60.340.360.320.330.220.320.31
70.330.340.330.330.220.310.32
80.330.340.350.340.220.320.32
90.360.320.330.340.220.310.35
100.370.330.330.330.210.320.31
Table 13. CNN summary of ‘accuracy on test data’.
Table 13. CNN summary of ‘accuracy on test data’.
Data No.XYZXYXZYZXYZ
10.980.980.980.991.001.001.00
20.980.940.991.001.000.990.97
30.990.930.991.001.000.991.00
40.990.960.991.001.001.000.99
50.990.940.970.991.001.001.00
60.970.940.970.991.001.001.00
70.990.960.990.991.001.001.00
80.940.960.961.001.000.991.00
90.990.940.991.001.000.951.00
100.980.950.991.001.001.001.00
Table 14. CNN summary of ‘loss on test data’.
Table 14. CNN summary of ‘loss on test data’.
Data No.XYZXYXZYZXYZ
10.100.090.100.040.030.020.02
20.080.180.060.020.020.040.07
30.060.180.060.020.020.050.02
40.050.130.060.030.030.020.03
50.060.180.120.030.020.030.02
60.090.190.110.040.020.020.02
70.050.110.070.030.020.020.02
80.150.140.140.030.020.070.02
90.050.180.080.030.030.100.03
100.100.150.060.030.010.030.03
Table 15. Accuracy summary of CNN results for three-axis accelerometer and its combination.
Table 15. Accuracy summary of CNN results for three-axis accelerometer and its combination.
Belt No.XYZXYXZYZXYZ
CICICICICICICI
17960795179067960796079607951
27964798279197973798279827946
3798271783748528000800080008000
479316701247197578867940794076529
Total3183729802103048142318193188231882315436
Note: C is the correct prediction (classification), and I is the miss-classification. For ‘Belt No.’ information, please refer to Table 2. This table is extracted from all confusion matrix of CNN results in Section 4.1.
Table 16. Accuracy summary of CNN results for three-axis dynamometer and its combination.
Table 16. Accuracy summary of CNN results for three-axis dynamometer and its combination.
Belt No.XYZXYXZYZXYZ
CICICICICICICI
17933794279427960794279427951
27973796479827973797379827946
3798277129788128000793779468000
474747764307859788679407931776529
Total31355531256531652531819317812317927315436
Note: C is the correct prediction (classification), and I is the miss-classification. For ‘Belt No.’ information, please refer to Table 2. This table is extracted from all confusion matrix of CNN results in Section 4.2.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Caesarendra, W.; Triwiyanto, T.; Pandiyan, V.; Glowacz, A.; Permana, S.D.H.; Tjahjowidodo, T. A CNN Prediction Method for Belt Grinding Tool Wear in a Polishing Process Utilizing 3-Axes Force and Vibration Data. Electronics 2021, 10, 1429. https://0-doi-org.brum.beds.ac.uk/10.3390/electronics10121429

AMA Style

Caesarendra W, Triwiyanto T, Pandiyan V, Glowacz A, Permana SDH, Tjahjowidodo T. A CNN Prediction Method for Belt Grinding Tool Wear in a Polishing Process Utilizing 3-Axes Force and Vibration Data. Electronics. 2021; 10(12):1429. https://0-doi-org.brum.beds.ac.uk/10.3390/electronics10121429

Chicago/Turabian Style

Caesarendra, Wahyu, Triwiyanto Triwiyanto, Vigneashwara Pandiyan, Adam Glowacz, Silvester Dian Handy Permana, and Tegoeh Tjahjowidodo. 2021. "A CNN Prediction Method for Belt Grinding Tool Wear in a Polishing Process Utilizing 3-Axes Force and Vibration Data" Electronics 10, no. 12: 1429. https://0-doi-org.brum.beds.ac.uk/10.3390/electronics10121429

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop