Next Article in Journal
Numerical Simulation of the Influence of Non-Uniform ζ Potential on Interfacial Flow
Previous Article in Journal
Analytical Investigation of Replica-Molding-Enabled Nanopatterned Tribocharging Process on Soft-Material Surfaces
Previous Article in Special Issue
Self-Powered Flow Rate Sensing via a Single-Electrode Flowing Liquid Based Triboelectric Nanogenerator
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

IC Packaging Material Identification via a Hybrid Deep Learning Framework with CNN–Transformer Bidirectional Interaction

1
School of Information Engineering, Guangdong University of Technology, Guangzhou 510006, China
2
China Electronic Product Reliability and Environmental Testing Research Institute, Guangzhou 510006, China
3
School of Mechanical and Electrical Engineering, Guangdong University of Technology, Guangzhou 510006,China
*
Author to whom correspondence should be addressed.
Submission received: 29 February 2024 / Accepted: 19 March 2024 / Published: 21 March 2024
(This article belongs to the Special Issue Functional Materials and Microdevices)

Abstract

:
With the advancement of micro- and nanomanufacturing technologies, electronic components and chips are increasingly being miniaturized. To automatically identify their packaging materials for ensuring the reliability of ICs, a hybrid deep learning framework termed as CNN–transformer interaction (CTI) model is designed on IC packaging images in this paper, in which several cascaded CTI blocks are designed to bidirectionally capture local and global features from the IC packaging image. Each CTI block involves a CNN branch with two designed convolutional neural networks (CNNs) for CNN local features and a transformer branch with two transformers for transformer global features and transformer local-window features. A bidirectional interaction mechanism is designed to interactively transfer the features in channel and spatial dimensions between the CNNs and transformers. Experimental results indicate that the hybrid framework can recognize three types of IC packaging materials with a good performance of 96.16% F1-score and 97.92% accuracy, which is superior to some existing deep learning methods.

1. Introduction

Nowadays, integrated circuit (IC) components and chips are widely used in micro- and nanomanufacturing, and their reliability highly influences the regular functions of the manufacturing equipment. IC packaging isolates the core circuits in the ICs from external environments to ensure IC reliable functions, which is commonly made of metal, ceramic, and plastic materials [1]. Decapsulation with preservation of the internal structures for ICs is essential for the failure analysis of IC chips [2], which is precisely performed using techniques such as nano-scale microscopy and precise mechanical cutting. It is noted that IC packaging materials necessitate unique decapsulation approaches. For instance, plastic packages are typically decapsulated by chemical agents, such as concentrated sulfuric acid, while metal packages may be decapsulated by a laser-based technology, and ceramic packages sealed with an epoxy often require mechanical decapsulation by utilizing sophisticated cutting techniques. Improper decapsulation methods can lead to inaccurate outcomes of the failure analysis and inflated costs. So, identification of IC packaging materials is performed before its decapsulation. However, it is commonly implemented by human eyes, which brings a large number of human labors. Consequently, it is preferrable to develop an automatic method for the identification of IC packaging materials.
Some analysis methods have been widely used to analyze the materials of IC packages, involving X-ray fluorescence spectrometry, electrochemical impedance spectroscopy, differential scanning calorimetry, and nuclear magnetic resonance. These methods will meet several issues during the material identification of IC packages, although almost no errors emerge for their identifications. The first is that professional knowledge is required for analytic devices and material science. Second, a large amount of time is required for material analysis. Third, sampling inspection is usually utilized due to its low efficiency. Fourth, these analytic devices are significantly expensive. Nowadays, with the development of computer vision, automatic optical inspection has been widely employed in many industrial inspections due to its advantages of no contact, low economic cost, high efficiency, and a little professional knowledge. In particular, with the rapid development of deep learning, it has been successfully employed in the inspection of IC packaging. Hu et al. [3] designed a hierarchical convolutional neural network (CNN) to classify IC components, in which multiple deep features were efficiently merged in Convolutional Auto-Encoder (CAE) layers to reduce computational complexity. Cai et al. [4] designed a novel cascaded CNN for Surface Mount Technology (SMT) solder joint inspection, in which each layer adaptively learned the Region of Interest (ROI) of the IC solder joint image. Guo et al. [5] modified the YOLOv4 network with several strategies to identify electronic components in real time. One was to eliminate the downsampling stages within the feature pyramid framework. The other was to reformulate the loss function by the weighted sum of classification loss, confidence loss and localization loss. Ram et al. [6] developed an improved deep CNN to identify the defective wafers. They employed resampling to solve the problem of data imbalance and various optimizers to train the model. An et al. [7] designed a transformer-based classification approach to detect PCB defects, termed Label Robust and Patch Correlation Enhanced ViT (LPViT). Kao et al. [8] designed a multi-scale GAN model with an embedded transformer for surface defect inspection of IC packages, which incorporated a novel feature extraction scheme and a cross-scale feature fusion module into a multi-scale CNN encoder. Feng et al. [9] developed a transformer-based deep learning model for the classification of PCBs, which utilized masked region prediction to discern relationships among different areas in the features. Chen et al. [10] integrated a feature pyramid structure with the transformer as the backbone into YOLOv5 to effectively classify PCB components.
Although these deep learning methods can easily implement the identification task by capturing the feature differences between different IC components or between the defect-free and defective IC packages, they may not be suitable for the identification of IC packaging materials. This is because IC components and their surface defects are quite different in their appearances, such as shapes and sizes. Comparatively, IC components made of different packaging materials usually exhibit highly similar overall appearances, necessitating distinction through a combination of features such as external structure, shape, and color. For example, plastic packages and metal packages typically have longer and fewer pins [11], with the latter also featuring identifiers that are laser-marked and embedded into the package [12], whereas ceramic packages have shorter and denser pins. Moreover, these comprehensive characteristics are possibly subtly different from each other. Thus, automatic identification of IC packaging materials is significant and challenging before IC decapsulation.
In this paper, a hybrid deep learning framework is designed for automatic identification of IC packaging materials. To capture subtle appearance differences and some similar characteristics between IC packages with different materials, a CNN branch with two designed CNNs and a transformer branch with two transformers are constructed to focus on local/global features of IC packages with different materials, respectively. Since several characteristics of IC packages with various materials are related to each other, a bidirectional interaction strategy is designed to interactively transfer local/global features in channel and spatial dimensions between the CNN branch and the transformer branch.

2. Methodology

2.1. Architecture of the Proposed Framework

Inspired by Efficientnet [13] and CoatNets [14], the proposed hybrid deep learning framework is designed as a feature pyramid architecture, which is composed of a convolutional stem, a feature-extracting backbone network, and a classification head. As shown in Figure 1, the convolutional stem is successively composed of a 3 × 3 convolutional layer, a 1 × 1 convolutional layer, and a 3 × 3 convolutional layer, which can expand the feature channels and make global attention feasible. The backbone is composed of several convolution–transformer interaction (CTI) blocks and four 2 × 2 convolutional layers, which are alternatively deployed in the backbone. The backbone is designed to comprehensively capture features from images of IC packages made of various materials. The classification head produces a probability vector that corresponds to different types of IC packaging materials, which is successively composed of a projection layer, a Sigmoid layer, and a fully connected (FC) layer.
Specifically, the feature channels for the input electronic component image are increased from 3 to C 1 through three successive convolution layers in the convolutional stem. Then, the feature maps successively pass through five downsampling modules in the backbone, each of which involves {2, 2, 6, 14, 2} CTI blocks successively and a 2 × 2 convolutional layer with the stride of 2. That is, the feature maps are downsampled with the rates of 1 ,   2 ,   4 ,   8 ,   16 through five downsampling modules. And their channels are progressively increased to C 5 , that is, {64, 96, 192, 384, 768}. Next, in the classification head, the feature channels are expanded to 1280 after the feature maps pass through a projection layer and a Sigmoid layer. Finally, the identification is achieved when the activated feature maps pass through an FC layer in the classification head, which indicates metallic, plastic, or ceramic package.

2.2. CTI Block

As illustrated in Figure 2, each CTI block is composed of three parallel 1 × 1 convolutional layers, a CNN branch, and a transformer branch. Specifically, the feature maps pass through three parallel 1 × 1 convolutional layers to generate the respective Q, K, and V feature maps [15], which can not only fulfill the requirements of the transformer branch but also add an extra feature extraction step for the CNN branch. Then, local/global features can be extracted by the CNN/transformer branches, respectively. It is noted that there is a bidirectional interaction between the two branches, which bidirectionally transfers local/global features to each other in channel and spatial dimensions. Finally, the feature maps output from two branches are adaptively fused through weighted summarization.
As depicted in Figure 3, the transformer branch with a traditional transformer and a local-window self-attention integrates global information into local information through a broadcast mechanism. Concurrently, the CNN branch combines local and channel information by means of the residual design. The outputs from the two branches are then coalesced via a weighted summarization to produce the feature maps output by the CTI block. This structure harnesses the global receptive field of the transformer to capture wide-ranging dependencies and causes the CNN to focus on local spatial and channel features.

2.2.1. Transformer Branch

Inspired by the “Adaptive Attention Span” [16,17], the attention is augmented by a masking function to enhance the transformer’s understanding of long-distance dependencies in the transformer branch, which can encourage each patch in the self-attention module to focus more on distant patches.
In the standard attention mechanism, for a given patch m , similarities b m n are calculated between this patch and other patches at positions n     S m , where S denotes the contextual window for patch m , formulated as
b m n = Q m K n T
where Q and K represent the Query and Key matrices, respectively. These are subjected to an activation function to obtain the attention weights a m n as
a m n = s o f t m a x b m n = exp b m n i = m s m 1 exp b m i
Then, they are multiplied by the Value matrix to derive the attention vector y m as
y m = i = m s m 1 a m i v i
where v i is the Value matrix for the i -th patch. Subsequently, a masking function M e x is introduced to the original attention to implement the “Adaptive Attention Span”, formulated as
M e x = c l a m p 0 , 1 e x H x + H e , 1
where H is a hyperparameter that governs the slope of the curve, which is empirically set to 0.5 in this work. e is differentiable and learnable. This masking function maps the distances between distinct patches into the interval of [0,1], which is graphically depicted in Figure 4.
Thus, a learnable token Z can be acquired by incorporating the masking function into the original attention, formulated as
Z = M A t t e n t i o n Q , K , V = M e m n exp b m n i = m s m 1 M e m i exp b m i
The token Z can possess an expanded receptive field to learn more robust global dependencies. It is calculated from the Q K V matrices only at the beginning of each downsampling module and will replace the Q matrix for propagation in the subsequent module, formulated as
Z ^ = M A t t e n t i o n Z , K , V
Furthermore, to address the limitation of self-attention modeling capacity along the channel dimension, the convolution-to-transformer interaction (C-to-T interaction) is designed to transfer the v matrix in the local feature maps output by the CNN-1 in the CNN branch to Z ^ , formulated as
V ^ = Z ^ v v
As indicated in (7), v can be considered as the inter-channel attention weights. Thus, the interacted Value matrix V ^ is combined with the transferred Q and K matrices to construct the mask attention, i.e., global attention, formulated as
X g l o b a l = M A t t e n t i o n Z ^ q , Z ^ k , V ^
To simultaneously capture local attention of distant dependencies, a local window self-attention [18] is combined with global attention to acquire the outputs X a t t of the transformer branch, formulated as
X a t t = X l o c a l + X g l o b a l
where X l o c a l is the outputs of the local window self-attention, in which a 7 × 7 window is employed to balance performance with efficiency.

2.2.2. CNN Branch

Q, K, and V feature maps severally pass through three parallel pooling layers, each of which is followed by a 3 × 3 convolution layer. After activation by a R e L U layer, the activated feature maps pass through a 3 × 3 convolution layer, a pixel-wise summation, and a S i g m o i d layer to produce a probability feature map v . This probability feature map with the dimensions B , C , 1 , 1 reveals the correlational information of features in channel dimension, where B stands for batch size and C for the number of channels of the features maps. The acquisition of the probability feature map v can be formulated as
v = σ W 2 δ W 1 z 1 + σ W 2 δ W 1 z 2 + σ W 2 δ W 1 ( z 3 ) z 1 = A v g P o o l F z 2 = M a x P o o l F z 3 = M i x P o o l F
where F denotes the input feature maps. W 1 C 0 × C and W 2 C × C 0 represent convolutional kernels with C 0 < C . σ () and δ () represent the Sigmoid/ReLU activation functions, respectively. To extract key features from the feature maps in the spatial dimension, a convolution-based spatial attention mechanism has been incorporated into the CNN-2 module, formulated as
X c n n = σ f 7 × 7 A v g P o o l v ^ ; M a x P o o l v ^ ; x ^ a t t v ^ = F s c a l e F , v
where x ^ a t t represents a single-channel feature map obtained by the dimensionality reduction of X a t t that is transferred from the transformer branch via the transformer-to-CNN interaction (T-to-C interaction). Thus, rich global information can be transferred into the CNN branch. f 7 × 7 signifies a 7 × 7 convolution to boast a larger receptive field compared with the 3 × 3 convolution, which can place a greater focus on the spatial dimension. X c n n denotes the output of the CNN branch.

2.2.3. Bidirectional Interaction

Typically, a self-attention mechanism employs a weight-sharing strategy across the channels of features, with a focus on the weights in the spatial dimensions, which tends to overlook inter-channel dependencies for weak dimensional modeling. This problem can be solved by sharing channel-related weights [19]. That is, in our designed framework, the CNN branch can offer “clues” to the transformer branch in the form of channel-wise attention weights. Conversely, the outputs of the transformer branch can supplement the CNN branch with spatial-attention clues. Thus, this bidirectional interactive design can enhance the modeling capacities of the framework in the channel-spatial dimensions.
As mentioned in Section 2.2.1 and Section 2.2.2, a bidirectional interaction interactively transfers global/local feature maps between the two branches, involving a C-to-T interaction for local information transmission and a T-to-C interaction for global information transmission. As indicated in (7), the C-to-T interaction transfers local feature information v into the transformer branch to enrich probability distribution of local features along the channel dimension. Similarly, the T-to-C interaction transfers global feature information X a t t acquired by the transformer branch into the CNN branch to enrich the probability distribution of global features along the spatial dimension, formulated as
x ^ a t t = σ M i x P o o l δ M i x P o o l X a t t
where MixPool() [20] denotes an adaptive selection of maxpooling or average pooling.

3. Experiments and Discussion

3.1. Dataset and Experimental Environments

All the images for IC packages with different materials were acquired by the industrial microscope (LV150N/LV100ND) in China Electronic Product Reliability and Environmental Testing Research Institute (Guangzhou, China). During the image acquisition, the microscope utilized the LV-S64 6 × 4 stage (Stroke: 150 × 100 mm with glass plate, ESD compatible) to place the electronic components, the LV-TI3 trinocular eyepiece tube ESD (Erected image, FOV: 22/25) as the eyepiece tube, the LV-NU5 U5 ESD as the objective revolver, and the LV-UEPI-N as the built-in reflected light illuminator. All the parameters of the microscope were elaborately configured to achieve sufficiently clear visibility of the markings on the IC packages. Then, 945 IC packaging images of 420 × 400 to 1280 × 960 pixels were acquired for three types of packaging materials such as metallic, ceramic, and plastic packages. Specifically, IC packages with each type of material involve 315 samples, some examples of which are illustrated in Figure 5. Different from pixel-level annotations, only category annotations are required for this study. Since all the acquired images were stored in three folders according to packaging materials during the image acquisition, we could directly utilize the folder names as the image labels, which indicated that no additional labeling task was required.
All the acquired images were rotated by ±30° and ±15° for data augmentation. Thus, 4725 images were utilized to construct the experimental dataset for the identification of IC packaging materials, which were all resized to 224 × 224 pixels to standardize the input for the proposed model. The dataset was divided into training, validation, and test sets in a completely random method at a ratio of 6:3:1, which was illustrated in Table 1.
In this study, all the weights were initialized as a Gaussian distribution with a mean of 0 and a variance of 0.01. The learning rate was set to 10 4 , and the Adam was chosen as the optimizer. Momentum and weight decay were set to 0.997 and 5 × 10−5, respectively. The loss function used cross-entropy loss. All the experiments were conducted on a computer equipped with an Intel(R) Core(TM) i7-9750H CPU @ 2.60 GHz and an Nvidia GeForce RTX 2080 12 GB GPU. The CPU was produced by Intel Corporation, located in Santa Clara, CA, USA, and the GPU was sourced from NVIDIA Corporation situated in Santa Clara, CA, USA. Both of the hardware were purchased in Guangzhou, China. Several commonly used metrics were employed for evaluation, involving accuracy, precision, recall, and F1-score [21].

3.2. Comparisons with Other Deep Learning Models

To the best of our knowledge, no studies have focused on the identification of IC packaging materials. Thus, to validate the proposed framework, it is compared with some state-of-the-art (SOTA) deep learning models for image classification, involving five hybrid models (MixFormer [22], Acmix [15], CoAtNet_4 [14], SMT [23], and FastVit_V3 [24]), four CNN models (ConvNeXt_S [25], FasterNet [26], ResNet_152 [27], and Wafer classification [6]), and three transformer models (Swin-Transformer_l [28], ITPN [29], and LPViT [7]). It is noted that all the models were equipped with the same classification head and were retrained on our training set. The experimental results are summarized in Table 2.
As indicated in Table 2, the three transformer models perform quite differently in identifying IC packaging materials. Specifically, the ITPN [29] achieves the worst identification performance with the second heaviest computational complexity among all the models, with the performance of 37.27% F1-score, 40.90% accuracy, and 59 G FLOPs. This means that the ITPN is of low efficiency in model inference and prediction. This is because its multi-layer feature pyramid structure excessively focuses on ultra-small objects, which possibly neglects distant dependencies existing in IC packages with different materials. Since the LPViT [7] directly employs the traditional transformer to extract global features from the images, it will neglect the local information of the IC package image, which involves some important hints to distinguish subtle differences between the IC packages with different materials. Thus, it also achieves a bad classification performance of 66.54% F1-score and 73.57% accuracy. The Swin Transformer [28] almost achieves the best identification performance among all the single models, with the performance of 82.13% F1-score and 82.21% accuracy. This can be attributed to its shift-window mechanism, which allows each local window to learn feature information from surrounding windows to effectively combine local details with global information. Also, this shift-window mechanism decreases the computational complexity while increasing the model size compared to the traditional transformer, with the parameter size of 39 M and 34 G FLOPs.
Although the CNN can capture local information from the image, the ability of local information extraction increases with the increase in the network depth. That is, if the network depth is not large enough, the CNN cannot efficiently extract local information at a high probability, which will significantly degrade the classification ability of the network. So, the Wafer classification [6] performs the second worst identification and has the lowest parameter size among all the methods due to its very shallow network architecture, with the performance of 54.24% F1-score, 59.22% accuracy, and 10 M parameters. Another three CNN models, namely ConvNeXt_S [25], FasterNet [26], and ResNet_152 [27], perform fairly good identification tasks for IC packages with different materials, even better than some hybrid models. This is because they can capture some subtle differences that exist in the IC packages with different materials, which is beneficial for identification. In particular, the FasterNet combines partial convolutions and point-wise convolutions [30] to speed up the model inference with a good model prediction, with an identification performance of 80.09% F-score and 83.57% accuracy, and a computational complexity of 4.4 G FLOPs. Since the ConvNeXt elaboratively integrates some previously published effective modules to fully mine the potential of the CNN framework to classification, it costs the most computational resources (61 G FLOPs) to establish a fair classification model for IC packaging material identification with a performance of 81.48% F1-score and 81.56% accuracy.
Since the CNN is good at capturing local features in the image, while the transformer is adept at understanding wide-range dependencies, the combination of the CNN and transformer can ideally result in better identification tasks being performed than most of the single models [31] (CNNs or transformers). As indicated in Table 1, three hybrid models (Acmix [15], MixFormer [22], and Ours) are the three best models for IC packaging material identification among all the models. However, the other three hybrid models (SMT [23], FastVit_V3 [24], and CoAtNet_4 [14]) do not exhibit superiority to single models. The two entirely different identification results can be contributed to their different hybrid schemes for the combination of CNNs and transformers. The former three hybrid models integrate CNNs and transformers into the total framework in a parallel mode, while the latter three hybrid ones sequentially cascade CNNs and transformer blocks. In particular, the FastVit_V3 achieves the second worst identification performance of 62.71% F1-score and 61.76% accuracy. This is because, except for the sequential cascade, it substitutes RepMixer and channel-wise and point-wise channel convolutions for skip connection and dense convolutions, respectively, to decrease its parameter size and computational complexity, which also to a great extent degrades its prediction performance.
Our designed framework achieves the best identification performance for IC packages with different materials at a reasonable inference speed and an acceptable model size. Its identification performance of 96.16% F1-score and 97.92% accuracy demonstrates its potential in real applications, and for which there are several reasons. One is its parallel combination of CNNs and transformers, which can simultaneously capture local and global features from the IC packaging image. Second, the elaboratively designed bidirectional interaction can integrate local information into the transformer branch and global information into the CNN branch, which can simultaneously capture local/global features in channel and spatial dimensions. Third, the combination of local-window self-attention and mask attention can together capture global features in global and local attentions, which can further improve the global feature representations for IC packaging images.

3.3. Ablation Experiment

We conducted an ablation experiment to validate the two branches, the bidirectional interaction, and weighted summarization, as illustrated in Table 3.
As indicated in Table 3, the simple combination of CNN and transformer branches performs a better identification of IC packaging materials than only a single branch, with the classification performance of 79.35% F1-score and 81.56% accuracy. This demonstrates that CNN and transformer branches both contribute to the model since they provide local/global feature information, respectively. When a one-way interaction is deployed between the two branches, the identification performance is more or less improved. In particular, if only the C-to-T interaction is incorporated into the proposed framework to transmit channel information provided by the CNN branch to the transformer branch, the classification performance significantly improves to more than 7% F1-score and 8% accuracy compared to the two-branch model without any interactions. This indicates that the two information interactions are beneficial for the fusion of global and local features in the IC packaging image, which can potentially provide deep subtle differences between different IC packaging materials. Thus, when the information interactions are bidirectional for the two branches, the model with a bidirectional interaction performs much better than those with a one-way interaction, which indicates that this bidirectional interaction can interactively transmit local/global features between the two branches to enhance their abilities of feature extraction. Since the two branches contribute differently to the designed framework, the model with weighted summarization of the branches (96.16% F1-score, 97.92% accuracy) improves the F1-score and the accuracy by almost 3% and 4%, respectively, compared to that with fixed fusion of the branches (93.19% F1-score, 94.08% accuracy).

3.4. Application for the Other Classification Task

To assess the application ability of our proposed model for the other specific task, we constructed an electronic component dataset from the Findchips website [32] for another classification task. The Findchips website is a powerful search engine that aggregates and normalizes data from the leading electronic part distributors in the electronic component market, which involves the attribute data, models, and schematics of the electronic components. There are 35 categories of electronic components on this website, such as resistor, capacitor, connector, transistor, and diode, each of which involves more than 1000 images.
To construct the Findchips dataset for the assessment of our proposed model, 200 images were randomly selected from each category on the Findchips website. Thus, a total of 7000 electronic component images were involved in the Findchips dataset, which were divided into training, validation, and test sets in a completely random method at a ratio of 6:3:1. Our proposed framework with the parameters configured in Section 3.1 was retrained, validated, and tested on the Findchips dataset. As illustrated in Table 4, the proposed model can also implement the other classification-specific task, although it is originally designed to identify IC packaging materials. However, since it is not specifically designed for this 35-category classification, its classification performance decreases to an 83.25% F1-score, compared to a 94.93% F1-score for 3-category IC packaging materials. This reveals the common shortcoming of deep learning, which is that task-specific deep learning models can be directly utilized but do not function well in the other tasks.

3.5. Repeated Experiments

To verify the reproductivity of the proposed framework, 10 repeated experiments were conducted on the augmented data with the sizes of 224 × 224 pixels, that is, 4725 images of IC packages with different materials. In each experiment, these images were randomly divided into training, validation, and test sets at a ratio of 6:3:1. This indicates that the samples in the three sets are different for each experiment.
As indicated in Table 5, the proposed framework achieves different identification performances for different experiments due to their different sample distributions. However, this difference is relatively small, with the largest and the smallest F1-scores being 96.16% and 91.98%, and the largest and the smallest accuracies being 97.92% and 94.83%, correspondingly. The classification performances of most of the repeated experiments approach the average metrics of 94.84% for precision, 95.23% for recall, 94.93% for F1-score, and 96.52% for accuracy, which demonstrates that the proposed framework has a fairly a good reproductivity ability.

4. Conclusions

In micro- and nanomanufacturing, accurate identification of IC packaging materials is crucial for quality control, since it is significant for packaging decapsulation. Misidentification may destroy the internal structures of IC components and chips during its decapsulation.
To identify IC packaging materials, a hybrid deep learning framework is designed in this paper, which adaptively integrates CNNs and transformers into an entire framework. The framework features two branches (i.e., CNN and transformer branches) with a bidirectional interaction to effectively capture local/global features in the IC packaging image to deeply mine the differences of IC packaging materials, which are validated by the ablation experiment. Comparative experiments indicate that the designed hybrid framework can better identify three types of IC packaging materials for the similar appearances of IC packages than the existing deep learning methods, with the performance of 94.84% precision, 95.23% recall, 94.93% F1-score, and 96.52% accuracy.
Although several CTI blocks improve the identification ability of the proposed model, they bring additional computational resources and computational burden due to their complex network design. In particular, the traditional vision transformer architecture in the CTI block further aggravates this issue. This limits the application of the proposed method in the assembly lines of real industries, since time cost and economic cost are sensitive in these scenarios. In the future, we will study the lightweight implementation of the proposed model to decrease the network size and speed up the network inference, such as the substitution of the mobile vision transformer for the traditional one. Since all the electronic components are provided by real industries, the experimental data acquired in the lab are prone to industrial ones, which indicates that the proposed model can be deployed in real industries for online identification in the future. However, the operation environments in real industries are quite different from lab ones, so the model will be retrained to adapt the data acquired in real industrial environments. If these issues are solved, our proposed model can be deployed in the assembly lines of real industries to automatically screen IC packages with different materials in the future, which can improve the efficiency of purpose-specific surface mounting, decapping, or recycling.

Author Contributions

Conceptualization, N.C. and C.Z.; data curation, C.Z., S.Z. and X.Z.; formal analysis, N.C. and H.W.; methodology, N.C. and C.Z.; resources, N.C. and H.W.; writing—original draft preparation, C.Z.; writing—review and editing, N.C. and C.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of Guangdong Province, China (No. 2021A1515011908).

Data Availability Statement

The original contributions presented in the study are included in the article, further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Wong-Ng, W. Ceramic materials. Int. Tables Crystallogr. 2019, H, 804–827. [Google Scholar]
  2. Morikawa, Y. Challenges of Semiconductor Microfabrication Technology for 3D Chiplet Integration. In Proceedings of the 2023 IEEE CPMT Symposium Japan (ICSJ), Kyoto, Japan, 15–17 November 2023; pp. 149–151. [Google Scholar]
  3. Hu, X.; Xu, J.; Wu, J. A novel electronic component classification algorithm based on hierarchical convolution neural network. IOP Conf. Ser. Earth Environ. Sci. 2020, 474, 052081. [Google Scholar] [CrossRef]
  4. Cai, N.; Cen, G.; Wu, J.; Li, F.; Wang, H.; Chen, X. SMT solder joint inspection via a novel cascaded convolutional neural network. IEEE Trans. Compon. Packag. Manuf. Technol. 2018, 8, 670–677. [Google Scholar] [CrossRef]
  5. Guo, C.; Lv, X.L.; Zhang, Y.; Zhang, M.L. Improved YOLOv4-tiny network for real-time electronic component detection. Sci. Rep. 2021, 11, 22744. [Google Scholar] [CrossRef] [PubMed]
  6. Ram, G.C.; Subbarao, M.V.; Varma, D.R.; Krishna, A.S. Enhanced Deep Convolutional Neural Network for Identifying and Classification of Silicon Wafer Faults in IC Fabrication Industries. In Proceedings of the 2023 International Conference on Wireless Communications Signal Processing and Networking (WiSPNET), Chennai, India, 29–31 May 2023; pp. 1–6. [Google Scholar]
  7. An, K.; Zhang, Y. LPViT: A transformer based model for PCB image classification and defect detection. IEEE Access 2022, 10, 42542–42553. [Google Scholar] [CrossRef]
  8. Kao, S.-X.; Chien, C.-F. Deep Learning-Based Positioning Error Fault Diagnosis of Wire Bonding Equipment and an Empirical Study for IC Packaging. IEEE Trans. Semicond. Manuf. 2023, 36, 619–628. [Google Scholar] [CrossRef]
  9. Feng, B.; Cai, J. PCB Defect Detection via Local Detail and Global Dependency Information. Sensors 2023, 23, 7755. [Google Scholar] [CrossRef]
  10. Chen, W.; Huang, Z.; Mu, Q.; Sun, Y. PCB Defect Detection Method Based on Transformer-YOLO. IEEE Access 2022, 10, 129480–129489. [Google Scholar] [CrossRef]
  11. Khan, R.A.; Muhaisin, M.M.; Roberts, G.W. An in-situ technique for measuring the individual contact resistance between the pins of an IC package and the board of a flexible hybrid electronic system. In Proceedings of the 2020 IEEE International Symposium on Circuits and Systems (ISCAS), Seville, Spain, 12–14 October 2020; pp. 1–5. [Google Scholar]
  12. Guin, U.; Huang, K.; DiMase, D.; Carulli, J.M.; Tehranipoor, M.; Makris, Y. Counterfeit integrated circuits: A rising threat in the global semiconductor supply chain. Proc. IEEE 2014, 102, 1207–1228. [Google Scholar] [CrossRef]
  13. Tan, M.; Le, Q. Efficientnet: Rethinking model scaling for convolutional neural networks. In Proceedings of the International Conference on Machine Learning, Long Beach, CA, USA, 9–15 June 2019; pp. 6105–6114. [Google Scholar]
  14. Dai, Z.; Liu, H.; Le, Q.V.; Tan, M. Coatnet: Marrying convolution and attention for all data sizes. Adv. Neural Inf. Process. Syst. 2021, 34, 3965–3977. [Google Scholar]
  15. Pan, X.; Ge, C.; Lu, R.; Song, S.; Chen, G.; Huang, Z.; Huang, G. On the integration of self-attention and convolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 815–825. [Google Scholar]
  16. Sukhbaatar, S.; Grave, E.; Bojanowski, P.; Joulin, A. Adaptive attention span in transformers. arXiv 2019, arXiv:1905.07799. [Google Scholar]
  17. Dai, Z.; Yang, Z.; Yang, Y.; Carbonell, J.; Le, Q.V.; Salakhutdinov, R. Transformer-xl: Attentive language models beyond a fixed-length context. arXiv 2019, arXiv:1901.02860. [Google Scholar]
  18. Vaswani, A.; Ramachandran, P.; Srinivas, A.; Parmar, N.; Hechtman, B.; Shlens, J. Scaling local self-attention for parameter efficient visual backbones. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 12894–12904. [Google Scholar]
  19. Li, D.; Hu, J.; Wang, C.; Li, X.; She, Q.; Zhu, L.; Zhang, T.; Chen, Q. Involution: Inverting the inherence of convolution for visual recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 12321–12330. [Google Scholar]
  20. Yu, D.; Wang, H.; Chen, P.; Wei, Z. Mixed pooling for convolutional neural networks. In Proceedings of the Rough Sets and Knowledge Technology: 9th International Conference (RSKT 2014), Shanghai, China, 24–26 October 2014; Proceedings 9. Springer International Publishing: Berlin/Heidelberg, Germany, 2014; pp. 364–375. [Google Scholar]
  21. Davis, J.; Goadrich, M. The relationship between Precision-Recall and ROC curves. In Proceedings of the 23rd International Conference on Machine learning, Pittsburgh, PA, USA, 25–29 June 2006; pp. 233–240. [Google Scholar]
  22. Chen, Q.; Wu, Q.; Wang, J.; Hu, Q.; Hu, T.; Ding, E.; Cheng, J.; Wang, J. Mixformer: Mixing features across windows and dimensions. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 5249–5259. [Google Scholar]
  23. Lin, W.; Wu, Z.; Chen, J.; Huang, J.; Jin, L. Scale-aware modulation meet transformer. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 6015–6026. [Google Scholar]
  24. Vasu, P.K.A.; Gabriel, J.; Zhu, J.; Tuzel, O.; Ranjan, A. FastViT: A Fast Hybrid Vision Transformer using Structural Reparameterization. arXiv 2023, arXiv:2303.14189. [Google Scholar]
  25. Woo, S.; Debnath, S.; Hu, R.; Chen, X.; Liu, Z.; Kweon, I.S.; Xie, S. Convnext v2: Co-designing and scaling convnets with masked autoencoders. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 16133–16142. [Google Scholar]
  26. Chen, J.; Kao, S.H.; He, H.; Zhuo, W.; Wen, S.; Lee, C.H.; Chan, S.H.G. Run, Don’t Walk: Chasing Higher FLOPS for Faster Neural Networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 12021–12031. [Google Scholar]
  27. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  28. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 10012–10022. [Google Scholar]
  29. Tian, Y.; Xie, L.; Wang, Z.; Wei, L.; Zhang, X.; Jiao, J.; Wang, Y.; Tian, Q.; Ye, Q. Integrally Pre-Trained Transformer Pyramid Networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 18610–18620. [Google Scholar]
  30. Chollet, F. Xception: Deep learning with depthwise separable convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1251–1258. [Google Scholar]
  31. Fan, Q.; Huang, H.; Guan, J.; He, R. Rethinking Local Perception in Lightweight Vision Transformer. arXiv 2023, arXiv:2303.17803. [Google Scholar]
  32. Website Findchips. Available online: https://www.findchips.com (accessed on 25 February 2024).
Figure 1. Architecture of the proposed hybrid deep learning framework. The input image dimensions are resized to 224 × 224. The convolutional stem expands the feature channels and makes global attention feasible. The backbone comprises five downsampling modules involving different CTI blocks ({2, 2, 6, 14, 2} successively) and a 2 × 2 convolutional layer with a stride of 2, serving as the downsampling module to halve the dimensions of the feature maps. The classification head yields a probability vector corresponding to the types of IC packaging materials.
Figure 1. Architecture of the proposed hybrid deep learning framework. The input image dimensions are resized to 224 × 224. The convolutional stem expands the feature channels and makes global attention feasible. The backbone comprises five downsampling modules involving different CTI blocks ({2, 2, 6, 14, 2} successively) and a 2 × 2 convolutional layer with a stride of 2, serving as the downsampling module to halve the dimensions of the feature maps. The classification head yields a probability vector corresponding to the types of IC packaging materials.
Micromachines 15 00418 g001
Figure 2. Sketch of the CTI block. The “C-to-T interaction” introduces the channel-wise features learned by the CNN into the transformer branch in a probabilistic fashion. Similarly, the global information learned by the transformer is integrated back into the CNN branch as probabilistic feature maps through the “T-to-C interaction”.
Figure 2. Sketch of the CTI block. The “C-to-T interaction” introduces the channel-wise features learned by the CNN into the transformer branch in a probabilistic fashion. Similarly, the global information learned by the transformer is integrated back into the CNN branch as probabilistic feature maps through the “T-to-C interaction”.
Micromachines 15 00418 g002
Figure 3. Detailed design of the two branches in the CTI block. The inter-channel feature maps extracted by CNN-1 are multiplied with (that is, the symbol * in Figure 3) the value matrix in global attention, compensating for the attention mechanism’s oversight of inter-channel information. The feature maps from the fusion of global and local attentions serve as the output for the transformer branch and as the input for T-to-C interaction, engaging in global information exchange with CNN-2.
Figure 3. Detailed design of the two branches in the CTI block. The inter-channel feature maps extracted by CNN-1 are multiplied with (that is, the symbol * in Figure 3) the value matrix in global attention, compensating for the attention mechanism’s oversight of inter-channel information. The feature maps from the fusion of global and local attentions serve as the output for the transformer branch and as the input for T-to-C interaction, engaging in global information exchange with CNN-2.
Micromachines 15 00418 g003
Figure 4. The masking function, which is designed to enhance the attention to focus on distant features.
Figure 4. The masking function, which is designed to enhance the attention to focus on distant features.
Micromachines 15 00418 g004
Figure 5. Some examples of IC packages with different materials. The first/second/third rows refer to IC packages with plastic/ceramic/metallic materials, correspondingly.
Figure 5. Some examples of IC packages with different materials. The first/second/third rows refer to IC packages with plastic/ceramic/metallic materials, correspondingly.
Micromachines 15 00418 g005
Table 1. Statistics of the dataset.
Table 1. Statistics of the dataset.
DatasetPlasticCeramicMetallicTotal
Training9119519632835
Validation4554754811401
Test209149131489
Total1575157515754725
Table 2. Comparisons of different deep learning models for identification of IC packaging materials.
Table 2. Comparisons of different deep learning models for identification of IC packaging materials.
MethodPrecision (%)Recall (%)F1-Score (%)Accuracy (%)FPSParams (M)FLOPs (G)
ITPN [29]76.4547.0437.2740.901.022959
Swin-Transformer_l [28]82.8284.4282.1382.211.363934
LPViT [7]74.3466.8166.5473.571.122841
ResNet_152 [27]83.8480.7078.4879.352.543212
FasterNet [26]74.3486.8180.0983.571.41254.4
ConvNeXt_S [25]84.5383.4981.4881.561.533561
Wafer classification [6]64.1659.3454.2459.223.311018
SMT [23]84.1883.8381.2781.181.763041
FastVit_V3 [24]79.7068.0262.7161.762.282227
CoAtNet_4 [14]75.5174.4570.9871.781.593855
ACmix [15]89.7583.1286.3988.391.62294.5
MixFormer [22]91.1986.0988.3490.671.47569.6
Ours96.1296.2696.1697.921.994423
Table 3. Ablation experiment.
Table 3. Ablation experiment.
Transformer
Branch
CNN
Branch
C-to-T
Interaction
T-to-C
Interaction
Weighted
Summarization
F1-Score
(%)
Accuracy
(%)
73.9979.40
78.3981.39
79.3581.56
86.6990.39
79.8387.56
93.1994.08
96.1697.92
Table 4. Classification results for the Findchips dataset via the proposed model.
Table 4. Classification results for the Findchips dataset via the proposed model.
MethodPrecision (%)Recall (%)F1-Score (%)Accuracy (%)
Ours-190.8576.8383.2587.61
Table 5. Ten repeated experiments for the proposed framework.
Table 5. Ten repeated experiments for the proposed framework.
IndexPrecision (%)Recall (%)F1-Score (%)Accuracy (%)
195.9495.9895.9996.55
294.1595.8894.9996.41
394.9595.7895.3896.58
495.3594.9795.1896.66
593.7994.8394.4395.97
695.4195.7995.5895.31
796.1296.2696.1697.92
894.9995.5295.3996.96
991.9691.4491.9894.96
1095.7495.8595.8197.89
Average94.8495.2394.9396.52
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, C.; Zhou, X.; Cai, N.; Zhou, S.; Wang, H. IC Packaging Material Identification via a Hybrid Deep Learning Framework with CNN–Transformer Bidirectional Interaction. Micromachines 2024, 15, 418. https://0-doi-org.brum.beds.ac.uk/10.3390/mi15030418

AMA Style

Zhang C, Zhou X, Cai N, Zhou S, Wang H. IC Packaging Material Identification via a Hybrid Deep Learning Framework with CNN–Transformer Bidirectional Interaction. Micromachines. 2024; 15(3):418. https://0-doi-org.brum.beds.ac.uk/10.3390/mi15030418

Chicago/Turabian Style

Zhang, Chengbin, Xuankai Zhou, Nian Cai, Shuai Zhou, and Han Wang. 2024. "IC Packaging Material Identification via a Hybrid Deep Learning Framework with CNN–Transformer Bidirectional Interaction" Micromachines 15, no. 3: 418. https://0-doi-org.brum.beds.ac.uk/10.3390/mi15030418

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop