Next Article in Journal
Diagnosis of Epilepsy with Functional Connectivity in EEG after a Suspected First Seizure
Next Article in Special Issue
Toward Evaluating Critical Factors of Extubation Outcome with XCSR-Generated Rules
Previous Article in Journal
A Survey on Physiological Signal-Based Emotion Recognition
Previous Article in Special Issue
Integrated Clinical Environment Security Analysis Using Reinforcement Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Estimation of Caenorhabditis Elegans Lifespan Stages Using a Dual-Path Network Combining Biomarkers and Physiological Changes

1
School of Computer Science and Technology, Wuhan University of Science and Technology, Wuhan 430081, China
2
Hubei Province Key Laboratory of Intelligent Information Processing and Real-Time Industrial System, Wuhan 430065, China
3
Department of Health Administration and Policy, College of Public Health, George Mason University, Fairfax, VA 22030, USA
*
Authors to whom correspondence should be addressed.
Submission received: 30 September 2022 / Revised: 8 November 2022 / Accepted: 9 November 2022 / Published: 14 November 2022

Abstract

:
Assessing individual aging has always been an important topic in aging research. Caenorhabditis elegans (C. elegans) has a short lifespan and is a popular model organism widely utilized in aging research. Studying the differences in C. elegans life stages is of great significance for human health and aging. In order to study the differences in C. elegans lifespan stages, the classification of lifespan stages is the first task to be performed. In the past, biomarkers and physiological changes captured with imaging were commonly used to assess aging in isogenic C. elegans individuals. However, all of the current research has focused only on physiological changes or biomarkers for the assessment of aging, which affects the accuracy of assessment. In this paper, we combine two types of features for the assessment of lifespan stages to improve assessment accuracy. To fuse the two types of features, an improved high-efficiency network (Att-EfficientNet) is proposed. In the new EfficientNet, attention mechanisms are introduced so that accuracy can be further improved. In addition, in contrast to previous research, which divided the lifespan into three stages, we divide the lifespan into six stages. We compared the classification method with other CNN-based methods as well as other classic machine learning methods. The results indicate that the classification method has a higher accuracy rate (72%) than other CNN-based methods and some machine learning methods.

Graphical Abstract

1. Introduction

1.1. Biological Background

Caenorhabditis elegans is one of the most important invertebrate model organisms in biological research. It has the characteristics of a short life cycle, a simple physiological structure, and a transparent worm body which makes for easy observation. Since the early 1960s, it has been widely used as a popular model organism [1]. Its research spans multiple disciplines, including large-scale gene function and characterization research [2], the complete lineage tracing of whole-body cells, the structural construction of the animal nervous system connection group [3], etc. Adult C. elegans are about 1 mm long and live in the soil. Under normal conditions, most are hermaphrodites and a few are males. The proportion of males can be greatly increased under special circumstances. C. elegans can grow and reproduce at 12 to 25 °C. In an environment of 25 °C, most C. elegans have an average lifespan of 12.1 days, with a standard deviation of 2.3 days of incubation [4]. C. elegans also provides an ideal model for studying the variability inducement that leads to differences in individual health and lifespan: the relative variability reflected in the lifespan cycle of about two weeks is almost as great as that of human beings from birth to 80 years old. In recent years, with the application of cutting-edge technologies, such as machine learning and artificial intelligence, in biological research, many researchers have used methods such as deep learning in biological research. Hua et al. developed an end-to-end ECG classification algorithm to help classify ECG signals and reduce the workload of physicians [5]. He et al. developed a novel evolvable adversarial framework for COVID-19 infection segmentation [6]. Mu et al. proposed a progressive global perception and local polishing (PCPLP) network to automatically segment pneumonia infection caused by COVID-19 in computed tomography (CT) images [7]. Zhao et al. developed a deep learning model combining a feature pyramid with a U-Net++ model for the automatic segmentation of coronary arteries in ICA [8]. Liu et al. proposed a new method for nuclei segmentation [9]. The proposed method performs end-to-end segmentation of pathological tissue sections using a deep fully convolutional neural network. Cao and Liu proposed a method for segmenting the terminal bulb of C. elegans based on the U-Net network [10]. This method solves the problem encountered with traditional single-stage networks that are not suitable for small samples. Lin et al. presented a quantitative method for measuring physiological age in C. elegans using a convolutional neural network (CNN) [11]. Fudickar et al. proposed an image acquisition system [12]. This system was used to create large datasets containing entire dishes of C. elegans. At the same time, the authors used the object detection framework Mask R-CNN to localize, classify, and predict the outlines of nematodes.
Regarding the lifespan assessment of C. elegans, there are currently two main research directions: one is to use physiological changes for evaluation; the other is to use biomarkers for evaluation. The physiological changes here refer to the physiological changes in C. elegans that can be directly observed, such as the swallowing rate of the pharynx, measurement of image entropy, measurement of appearance, measurement of exercise capacity, and measurement of auto-fluorescence. In the study of Zhang et al., based on the view that the aging of C. elegans is a plasticity process discovered by predecessors, a large number of comparative experiments were carried out on the aging of C. elegans [13]. Experiments have found that there is a big difference in lifespan among individuals of C. elegans. Short-lived C. elegans have a lifespan of about 12 days, while long-lived C. elegans can survive for more than 20 days. Through more detailed experiments, the authors determined the difference between the long-lived and short-lived C. elegans in the physiological process of aging. The difference in the lifespan of C. elegans is mainly concentrated in the time period between reproductive maturity and death, which is less related to the time of larval growth. Although it takes an average of 2.1 days for larvae to develop (accounting for 17.3% of the average lifespan), the variability in the development time is less than 0.1% of their total lifespan. In the subsequent lifespan process, the physiological health status and the rates of physiological health changes are related to the total lifespan of C. elegans. At the same time, it is also to be pointed out that the differences in physiological changes between different C. elegans at the same stage are much smaller than the differences in physiological changes between different C. elegans at the same absolute time. Stroustrup et al. designed a set of lifespan machines to judge and predict the lifespan of C. elegans by detecting the movement status and movement ability of C. elegans and achieved better results [14]. Martineau et al. extracted hundreds of morphological, postural, and behavioral features from C. elegans activity videos and used support vector machines (SVM) to analyze their direct relationship with C. elegans lifespan [15]. Lin et al. presented quantitative methods to measure the physiological age of C. elegans with convolution neural networks (CNNs), which measured ages with a granularity of days and achieved a mean absolute error (MAE) of less than 1 day [11]. Furthermore, they proposed two models: one was based on linear regression analysis and the other was based on logistic regression. The linear-regression-based model achieved a test MAE of 0.94 days, while the logistic-regression-based model achieved an accuracy of 84.78 percent with an error tolerance of 1 day. The advantage of using physiological changes for assessment is that this method has a higher accuracy rate and is applicable to various C. elegans mutants. However, because the research is limited to the C. elegans body and lacks the possibility of technological migration, the significance for human research is relatively limited.
Compared with physiological changes, biomarkers mainly consist of life-related genes or microRNA promoters carrying fluorescent proteins. The relevant signal pathway mechanism behind the genes is clear, and there is the possibility of technology migration, which has potential guiding significance for the assessment of human aging [16]. For example, Wan et al. designed a C. elegans life prediction algorithm based on Naive Bayes [17]. They investigated the relationship between C. elegans gene sequencing information, protein expression information, and lifespan. They proposed a feature selection method based on Naive Bayes to predict the effects of C. elegans genes on biological life. However, obtaining gene sequencing information and protein expression information for C. elegans leads to the death of the worms, which is not conducive to the verification of the test set and the further development of research. At the same time, the cost of extracting gene sequencing information from C. elegans is expensive, and it is not suitable for repeated experiments. Saberi-bosari et al. also selected biomarkers as the research objects, in order to use the Mask R-CNN [18] algorithm to identify the neurodegenerative sub-cellular processes that appear after the senescence of the C. elegans PVD neurons, and they used this information to determine current lifespan stages of C. elegans [19]. The biological state was divided into three states: young and old adults, cold-shocked and non-shocked nematodes, and cold-shocked and aged worms. Finally, a classification accuracy of 85% was obtained. However, in actual research, it has been found that the currently used biomarkers have the following two problems. On the one hand, the overall performance of biomarkers is relatively poor, which may be due to the limited influence of a single gene on lifespan [20]. On the other hand, some endogenous genes have a certain evaluative power in the wild type, but they often have poor evaluative potential in specific mutant strains (such as daf-16). This is because the genes used in the evaluation are often limited to specific signaling pathways, and the phenomenon of aging is jointly regulated by multiple signaling pathways [21].
We selected proteostasis as a life-related indicator because most biological activities are dependent on protein function and many life-related signal pathways in C. elegans show the regulation of proteostasis. With the aging of C. elegans, protein accumulation will gradually increase [22]. At the same time, in the process of human aging, proteostasis is also related to many senile diseases [23], such as Alzheimer’s disease [24], Parkinson’s disease [25], and so on. Intrinsic protein aggregation is a biomarker of aging. Knowing how to regulate it will help understand the underlying mechanisms of aging and protein aggregation diseases [26]. For the detection of proteostasis imbalance, there are also mature visual biomarkers in C. elegans metastable protein [27]. In summary, the physiological index of proteostasis change is closely related to aging, involves multiple signaling pathways, is conservative among species, and is easy to detect, so it is a good candidate for lifespan estimation. To minimize the impact on C. elegans, the response is closest to the aging process in the natural state. We selected firefly luciferase protein, which has not been reported as being related to pathological processes associated with a variety of metastable proteins. C. elegans carrying multiple copies of the firefly luciferase gene will not exhibit premature aging and paralysis phenotypes. This is the first study to use protein aggregation as a biomarker to estimate the lifespan of C. elegans.

1.2. Convolutional Network Background

In recent years, neural networks, representing an emerging deep learning method, have continuously enabled breakthroughs in various fields. Among them, convolutional neural networks (CNNs) [28], as deep feed-forward neural networks involving convolutional calculations, can better obtain spatial position and shape information from an image. They are widely used in many fields, including object tracking, pose estimation, text detection and recognition, visual saliency detection, action recognition, and scene labeling.
Generally, a convolutional neural network consists of a convolutional layer, a pooling layer, and a fully connected layer. The input image is respectively processed by the convolutional layer, the pooling layer, and the fully connected layer, and the feature map is finally outputted. Limited by computer performance and the acquisition of datasets, the convolutional neural network model was unutilized for decades, until the proposal of AlexNet [29] in 2012, which stimulated the frenzy of convolutional neural network research. In this study, we chose VGG16 [30], Inceptionv3 [31], MobileNetV3 [32], ResNet50 [13], and DenseNet [33] as candidate benchmark CNN models.
Given the shortcomings of the above two methods, we hope to design an evaluation and prediction method to maximize accuracy and have better mobility for future research on the human lifespan. To this end, in this paper, we propose an estimation method for C. elegans lifespan stages combining a C. elegans biomarker (protein aggregation) and a high-efficiency attention improved network (Att-efficientNet). Compared to using shape and motion trajectory features, this is a new attempt. The dual-path feature fusion model based on deep neural networks can extract local features of C. elegans through neural networks and compensate for the loss of global features by calculating fluorescent protein aggregation information and finally output the multi-stage classification results for C. elegans lifespan stages. In this paper, we divide the lifespan of C. elegans into six stages. This method has basically met the needs of biological research to predict the lifespan stages of C. elegans. This work makes several main contributions:
  • The study is the first to estimate the lifespan stages of C. elegans using the six-stage standard from fluorescence microscope images;
  • A dual-path network combining biomarker and physiological changes is proposed to jointly estimate the lifespan stages of C. elegans from fluorescence microscope images;
  • An Att-EfficientNet was developed to extract physiological changes and a unique fluorescent protein feature extraction technology was developed to calculate protein aggregation degrees;
  • We evaluated the proposed method on a dataset with 4593 fluorescence microscope images of C. elegans and achieved promising results in lifespan stage estimation compared with several other machine learning methods.

2. Methods

2.1. Materials

The images used in this article were images of live C. elegans carrying exogenous transfection of firefly luciferase fusion protein taken by the Sino-French Joint Laboratory of the School of Life Sciences, Huazhong University of Science and Technology, under a fluorescence microscope. In the previous literature, there is no direct evidence regarding the effect of luciferase overexpression on the lifespan of nematodes. Previous studies regarded the luciferase molecule as a protein homeostasis detection biosensor [34,35,36] and did not discuss whether it causes additional protein homeostasis pressure in the worm. In contrast to pathological infectious proteins that can produce significant characterization changes in nematodes and greatly reduce lifespan and to detection proteins that are restricted to specific tissues, luciferase is independent of the nematode gene regulatory network [37] and has a stable effect on the protein network. The fluctuations in its state have high sensitivity and can be expressed in almost all nematode tissues, so it has corresponding advantages. To further reduce the possible proteostasis pressure on the worm, we improved the multi-copy strain used in previous studies using single-copy cloning technology for lifespan prediction. We must admit that due to the metastability of the luciferase protein molecule, nematode individuals may experience prolonged lifespans due to microprotein stress due to its stimulation [34] or shortened lifespans due to the deterioration of protein homeostasis [37]. Therefore, compared with wild-type lifespan experiments, even if there is no significant difference in lifespan, it may be due to the superposition of the two, and potential differences in the regulation of protein homeostasis cannot be ruled out. Therefore, we accepted these limitations and focused on the correlation between protein homeostasis and lifespan itself.
The images used were from 26 batches of about 40 C. elegans each. Images were taken of the head, tail, and torso of C. elegans and were evenly distributed across each life stage of C. elegans. Each C. elegans was cultured independently in a single Petri dish at the L4 stage and photographed daily from the first day after adulthood. A total of 4593 image samples were obtained, and 80% of the images were used as the training set and 20% of the images were used as the test set. In order to obtain accurate lifespan data for C. elegans and to try to avoid inaccurate lifespan data for the C. elegans caused by damage to the worms during the shooting process, the C. elegans were alive at the time the images were taken and no immobilization measures were taken. At the same time, in order to obtain a clear fluorescent protein bright spot during shooting, exposure times of 1/20 s, 1/40 s, and 1/80 s were used. Most previous research on lifespan perception tended to focus on short-term worm phenotype observation. Most of the worms were cultured at 25 °C. Under this condition, C. elegans lifespan was mostly concentrated within 15 days. Using the remaining days as an indicator can allow better distinguishment of the degree of aging. However, when cultured at 20 °C, there is a huge difference in the lifespans of C. elegans of the same genotype, ranging from 10 to 30 days, with a wider lifespan distribution and lower concentration. If this indicator is used, there will be too great a difference in the actual degree of aging within the same group. For example, a C. elegans with a lifespan of 25 days has a remaining lifespan of 5 days at 20 days. C. elegans with a lifespan of only 10 days are still in the egg-laying period on the 5th day and have only 5 days of remaining life. Obviously, there is a huge difference in physiology. The other method is to classify directly by the number of days of age, but this method can only assess the age of C. elegans and cannot assess the actual degree of aging. To solve such problems, we determined the lifespan of each individual worm in the image set and divided each of the lifespans into 25 stages, 0–4% being the first group, 4%–8% being the second group, and so on. For the C. elegans in the dataset, we achieved knowledge of their ultimate lifespans. The assumption behind this method of dividing lifespans according to the proportion of life processes undergone is that the steady-state modes of aggregation and formation of various C. elegans proteins are similar, only the rates are different. This assumption can facilitate data division and processing. When the data were insufficient, we often combined different groups. At this time, dividing by proportion can be shown to be convenient, as is shown in Figure 1. In this paper, we divided the dataset into 6 groups, which were merged from the 25 stages. These groups included early, middle, and late stages, and each period included two groups. The 6 groups were formed by combining groups 1–5, 6–9, 10–13, 14–17, 18–21, and 22–25. The number of data in each category in the divided dataset is shown in Table 1. The dataset is shown in Figure 1.

2.2. Model Architecture

2.2.1. CNN Model

The effective feature areas in the C. elegans images are small, and the images of C. elegans at different life stages have close similarities. Compared with natural image classification tasks, C. elegans image classification tasks involve more attention being paid to feature information at a fine-grained level. In a traditional convolutional neural network, when the characteristic information is transmitted between the convolutional layer and the fully connected layer, there is some loss of information. Increasing the network depth [38] is a frequently used method for training many neural networks because it can allow the capture of richer and more complex features and adaptation to new tasks for learning. However, increasing the depth of the network will bring about the problem of gradient disappearance. The width of the network is the number of channels in the feature map. Increasing the width of the network [39] means that the number of channels in the feature map increases, and more convolution kernels can obtain more rich features, which enhances the characterization ability of the network. Small-size models require less network width, and wider networks can often learn richer features and are easier to train. However, it is difficult for networks with too wide network structures and shallow depths to learn higher-level features in the process of feature extraction. Convolutional neural networks can also capture fine-grained features for high-resolution input images, which can enrich the receptive field of the network to improve the network.
EfficientNets are a series of models (EfficientNet-B0 to B7) that are obtained by scaling up the basic network (usually called EfficientNet-B0). For all dimensions of the network—width, depth, and resolution—a composite scaling method is used. EfficientNets have attracted much attention due to their advantages in performance. This series of models have surpassed all previous convolutional neural network models in terms of efficiency and accuracy. The width refers to the number of channels in any layer, the depth refers to the number of layers in the CNN, and the resolution is related to the size of the image. To systematically expand the size of a network, composite scaling uses a composite coefficient, which controls how many resources are available for model scaling, and the dimensions are scaled in the following way through the composite coefficient:
Depth :   d = α ϕ Width :   w = β ϕ Resolution :   r = γ ϕ s . t . α β 2 γ 2 2 α 1 , β 1 , γ 1
EfficientNet successfully scales the classification model in three dimensions through the scaling factor and adaptively optimizes the network structure. In this way, during the training process, the training parameters are greatly reduced, and the computational complexity is also reduced.
In this paper, EfficientNet was used for feature extraction from the images of C. elegans, and the network is expressed as follows:
N = i = 1 , 2 , , s F L i ( X [ H i , W i , C i ] )
where N represents classification network,   represents the convolution operation, X represents the input tensor, F represents the basic network layer, i represents the number of convolution layers, and L i represents the depth of the network. The network adjusts 3 dimensions (height (H), width (W), and number of channels (C)) for optimization. It is necessary to find the optimal scaling parameters in 3 dimensions. When the model parameters and the amount of calculation are maximized, the accuracy of the model is improved. The maximum accuracy of the model is denoted as Acc max ( N ( d , w , r ) ) ; the specific formula is as follows:
N ( d , w , r ) = F ^ d × L ^ i ( X [ r × H ^ i , r × W ^ i , w × C ^ i ] ) i = 1 , 2 , , s
where depth d = α φ , width w = β φ , and resolution r = γ φ . The relationship between the variables α , β , and γ is:
α 2 × β 2 × γ 2 2 , α 1 , β 1 , γ 1
To obtain the three-dimensional parameters which can satisfy Formula (3), the composite parameter φ was used to optimize the depth, width, and resolution of the network. First, we set φ = 1 , then we found the optimal α , β , and γ parameters satisfying Formula (4) through a grid search. After the experimental adjustments, we obtained α = 2.3 ,   β = 1.5 , and   γ = 1.18 . With Formulas (2)–(4), EfficientNet was used to extract image features, and the features of C. elegans images were fused in multiple dimensions. The M1 sub-model designed to achieve the above processing is shown in Figure 2.

2.2.2. Attention Mechanism

As the images of C. elegans contained a lot of noise, some images presented problems, such as ghosting, which may interfere with decision making. For example, in live-cell imaging, to prevent damage to cells caused by light, low-light and long-exposure shooting methods are usually used, which make the signal-to-noise ratios of fluorescent bright spot microscopic images low, and thus fluorescent bright spots are difficult to detect, even for experienced biologists. To solve this issue, an attention mechanism was introduced into the network. With the attention mechanism added, the network is able to automatically select the area that needs attention when performing feature extraction. The attention mechanism network mainly takes the output feature map F of EfficientNet as the input and then puts it into three 1 × 1 convolutional layers and adds the activation functions ReLU and Sigmoid to convert the input into nonlinear features. The details are shown in Figure 3.
The purpose of joining this network is to generate an attention map A . Multiplying the feature map F and the attention map A will generate a mask M of the image. To reduce the parameters of the network and avoid overfitting, global average pooling (GAP) is used on the image mask M and the attention map A . Finally, the division operation is used to obtain the weight of the image and to filter out irrelevant information. The output of the attention mechanism is:
O = G A P ( A l ) / G A P ( A l × F l )
where A l and F l represent the l-th layer attention map and the l-th layer feature map, respectively. Before entering the M1 module, a C. elegans image needs to be averagely pooled to compress the resolution of the image from 6000 × 4000 to 600 × 400 pixels.

2.3. Fluorescent Protein Feature Extraction

The feature vector obtained from the C. elegans image through the M1 module contains rich semantic information, while the macro-level information, such as ROI contour information, is relatively rough. A lot of smaller fluorescent protein spots will be lost because the image needs to be averagely pooled before entering the M1 module (the resolution of the image will be compressed from the original size of 6000 × 4000 pixels to 600 × 400 pixels). For feature extraction from a C. elegans image, although the abstract high-level semantic information is important, features such as the protein distribution density of the C. elegans cannot be ignored either. Therefore, traditional image feature extraction algorithms are employed to extract features based on the density of the spots in the image, which will be fused with the features from the M1 module to construct the final feature vector for lifespan prediction. In the microscopic images of C. elegans, because the bright spots of fluorescent protein have strong correlations with the lifespan stages of the C. elegans, we first detect the spots and then extract the features from the detected spots.
We developed an image processing technique to detect the spots. Since long exposure during image acquisition could cause noise and blur the image, the acquired images are preprocessed before the spots are detected. The preprocess includes two steps: histogram equalization and noise reduction. Histogram equalization is employed to enhance the image and a low-pass filter is employed to reduce the noise. After preprocessing, spot detection is performed. From experiments, we found that the bright spots of the fluorescent protein belong to the high-frequency parts of images, while C. elegans tissues and the surrounding backgrounds belong to low-frequency regions. Thus, a high-pass filter is employed to remove the tissue regions and surrounding backgrounds from the images. To high-pass-filter an image, the image is transformed to the frequency domain and then inverse transformed back to the spatial domain after being passed through a high-pass filter. The filtered image is used to detect the fluorescent protein spots. A watershed spot detection algorithm based on local extrema is used.
OpenCV [40] provides a function called SimpleBlobDetector, which is used to achieve the task. SimpleBlobDetector can screen and mark irregular spots of a specified size and with a limited grayscale range. It has a high detection accuracy through the setting and adjustment of its parameters. As the shape size and grayscale of the fluorescent protein spots in a C. elegans image are relatively uniform and the background noise in the image is almost eliminated after low-pass filtering, the algorithm can achieve good results. For the experimental data, the settings for the parameters of the SimpleBlobDetector function are shown in Table 2. The settings were obtained from our experimental tests. The parameter ThresholdStep represents the step value which is the span of the threshold when the algorithm starts to perform threshold segmentation and was set to 8. minThreshold and maxThreshold are the parameters that control binarization, and they were set to 0 and 255, respectively. The parameters minArea and maxArea were set to 10 and 2500; they are used to describe the areas of the spots. minCircularity represents the minimum roundness of a spot, and its value is obtained using the formula 4 π ( area / permeter 2 ) . When the roundness is 1, this means that the shape of the spot is a perfect circle; when the roundness is 0, this means that the shape of the spot is a gradually elongated polygon. The minConvexity parameter is the value describing the minimum convexity of a spot, and the formula is a r e a / a r e a   o f   C o n v e x H u l l . minInertiaRatio describes the minimum inertia rate of the spot, which is the ratio of the minimum diameter to the maximum diameter of the ellipse. If maxInertiaRatio is not specified, the default value is 1. These parameters describe the basic characteristics of the C. elegans fluorescent protein bright spots and can be used to determine whether a blob is a spot or not. SimpleBlobDetector will return the coordinates and radius of each spot, and we denote the i-th protein spot by A i ( x i , y i , r i ) , where i , = 1 , N and N is the number of spots detected. ( x i , y i ) are the coordinates of the i-th protein spot and r i is the radius of the spot. After we have obtained the coordinates and radius of each spot, we can use them to compute the distance between any two spots. For N spots, we will obtain ( N ( N 1 ) ) / 2 distance values C j . The distance values C j will be used to construct an n-dimensional feature vector F 2 = ( P 1 , P 2 , P 3 , , P n ) . The algorithm for the computation of P i and the construction of F 2 is described in Algorithm 1. In the experiments, we only computed the value of P i when i was less than 12. If i was greater than 12, P i took the maximum value of 5000. The algorithm used to extract the features from the aggregation of the fluorescent protein bright spots is shown in Figure 4.
Algorithm 1. Calculate Aggregation information pi
Input: y = Ai(xi, yi, ri)∨n//Ai(xi, yi, ri):Fluorescent protein bright spot coordinate set, n: the number of clustering features.
Output: feature vector containing aggregation degree information F2
Ensure: F2//Feature vector containing aggregation information
1:F2 <= Initialize(F2)//Randomly initialize F2
2:fori = 0, 1, …, n do//n iterations
3:  p <= Traverse(Ai(xi, yi, ri))//Traverse the fluorescent protein bright spot coordinate set (Ai(xi, yi, ri))
4:ifp > 2 then
5:  =>9//Go to step 9
6:else
7:  =>11
8:end if
9:  Cj <= Dist(Random(Ai(xi, yi, ri))//Calculate the distance between any two points in set Ai(xi, yi, ri), getting point spacing set Cj
10:end for
11:Cj = Sort(Cj))//Sort the point spacing set Cj from smallest to largest
12:M Number(Cj)//The amount of data in the point spacing set Cj
13:ifM > n then
14:  =>18
15:else
16:  =>19
17:end if
18:F2 (F2, …, Cj)//Add all items of point spacing set Cj to the end of feature vector F2
19:F2 (F2, …, 5000)//Add the set maximum value of 5000 to the end of the feature vector F2
In this paper, a convolutional-neural-network-based two-way feature fusion model is proposed to solve the problem of the estimation of C. elegans lifespan stages. A new additional attribute—aggregation information—is introduced as a global feature to improve classification accuracy. The overall framework of the model is shown in Figure 5, which is divided into two main modules: the CNN feature extraction module M 1 and the aggregation feature extraction module M 2 . The sub-module M 1 is an EfficientNet network with an attention mechanism. The feature vector F 1 is obtained after the last layer through global average pooling. After the M 2 module, the feature vector F 2 is obtained. The feature vector F 1 is concatenated with the feature vector F 2 , and finally the evaluation results for the C. elegans lifespan stages are outputted through the Softmax classifier [41].

2.4. Model Setups

To prevent network overfitting and improve the prediction performance of the test set, dropout was added to the dense layers [42]. This method can randomly eliminate the units in a network during the training process, reduce the errors in the network results introduced by local features, and make the model more robust. Early stopping and fine-tuning were applied to the model to shorten the training and debugging time and improve the performance of the model. In addition, when training deep learning models, the choice of the optimizer is also very important. A good optimizer can significantly speed up the training process while avoiding partial optimization of the results. In this paper, the Adam optimizer with a good adaptive learning rate was used to optimize the classification model.
Regarding Table 1, where y ( i ) { 1 , 2 , , k } , k is determined according to the number of categories in the actual training dataset, which was taken as 6 in this paper. For a given test input x , the output is a k -dimensional vector representing the probabilities that a sample belongs to each category. The formula is:
h ω ( x ) = [ p ( y ( i ) = 1 x ( i ) ; ω ) p ( y ( i ) = 2 x ( i ) ; ω ) p ( y ( i ) = k x ( i ) ; ω ) ] = 1 j = 1 k e ω j T ( i ) [ e ω T x ( i ) e ω T x ( i ) e ω k T x ( i ) ]
where ω 1 , ω 2 , , ω k n + 1 are the weighting parameters. The coefficient product term on the right side of the formula normalizes the probability distribution so that the sum of the probabilities for all categories is 1.
When using a convolutional neural network to train a model, choosing an appropriate loss function can improve the accuracy and robustness of the model. The two-way feature fusion model proposed in this paper is a multi-class network model, so Categorical Cross Entropy [43] was used as the loss function. The true label of the i-th sample is y j i ; the predicted value label is y ^ j i .
The hardware configuration used in the experimental environment of this paper was an i7-8700K processor, the GPU model was NVIDIA GeForce GTX 1080, and the video memory was 8 GB. The operating system was Ubuntu Server 16.04 64-bit, and the programming language used was python 3.6.0.

3. Experiments and Results

3.1. Evaluation Metrics

In the experiments, we used Accuracy (AC), Recall, F1 score, and a confusion matrix to evaluate the results; the formulae are as follows:
A C = T P + T N T P + T N + F P + F N
R e c a l l = T P T P + F N
P = T P T P + F P
F 1 = 2 P R P + R
where TP means true positive, which means that the prediction is positive and the actual is also positive; FP means false positive, which means that the prediction is positive but the actual is negative; FN means false negative, which means the predictive is negative but the actual is positive; and TN refers to samples that are predicted to be negative and actually negative. The confusion matrix is a kind of evaluation classification model used to subdivide categories.

3.2. Experiments

3.2.1. The Influence of M1 and M2

To analyze the influence of different network branches in the model, we compared the performance of the networks with or without the M 1 sub-module or the M 2 sub-module. Three networks were created: the network with both the M 1 sub-module and the M 2 sub-module (denoted by M 1 + M 2 ), the network with only the M 1 sub-module, and the network with only the M 2 sub-module. The experimental results are shown in Table 3. It can be seen that each sub-module has a certain independent classification ability. Figure 6 shows the confusion matrix using the network with only one sub-module and the complete model ( M 1 + M 2 ) of this paper. Compared with the network with only the M 2 sub-module, the network with only the M 1 module showed a better performance in classification. However, the M 1 + M 2 dual-path method achieved an accuracy score of 0.726, a recall of 0.724, and an F1 score of 0.740, which scores are higher than those of both of the networks with only one module. These results show the effectiveness of the proposed dual-path method.

3.2.2. Benefits of the M2 Module

We further investigated the performance of lifespan stage estimation with different popular networks by adding the M 2 module. The investigated networks were VGG16, InceptionV3, MobileNetV3, ResNet50, and DenseNet. During the experiments, we first replaced the M 1 module in Figure 5 with a network listed above and obtained a new hybrid model. After that, the obtained hybrid model was trained, and the trained model was used for the estimation of lifespan stages. For each network, we did the same. The results are shown in Table 4. In Table 4, it can be seen that the hybrid model can generate more accurate results than networks without the M 2 module. These results suggested the effectiveness of the proposed protein aggregation degree strategy used in the M 2 sub-module. Moreover, it can be seen that all the hybrid models can obtain higher accuracy than the original networks, which indicates that the M 2 module proposed in this paper has a certain degree of versatility.

3.2.3. Influence of the Attention Mechanism

To further verify the effectiveness of the attention mechanism with respect to classification networks, the attention mechanism was also added to the VGG16, InceptionV3, MobileNetV3, ResNet50, and DenseNet networks, and the resulting classification result data are shown in Table 5. It can be seen from Table 5 that after the attention mechanism was added more attention could be paid to the main features in the image during classification, which played a positive auxiliary role in the feature extraction of the network. When the attention mechanism was added to EfficientNet, the accuracy rate was increased by 3.5%, which was higher than the result of 0.691 obtained without the attention mechanism, and the classification performance was improved. For other networks, after the attention mechanism was added the classification effect was also improved. However, the improvement of the proposed EfficientNet with the attention mechanism was slightly higher than the improvements achieved with the other classic neural network models.

3.2.4. Comparison with Random Forests and Support Vector Machines

Our method combines traditional machine learning methods and deep learning methods. In traditional machine learning models, random forests and support vector machines are common machine learning methods, and we compared them with the proposed network. It should be pointed out that, in the comparison, the information inputted to the two models only contained the numerical information after the blob processing, not the picture information. The performances are shown in Table 6 and Table 7. From the tables, it can be seen that the performance of the random forest model was generally slightly lower than that of the proposed model, and it performed poorly especially in the 3–5 lifespan stages. The results for the support vector machine model were generally not ideal, and the classification accuracy in stages 5–6 was 0. Overall, the proposed network showed significant advantages over the traditional machine learning methods.

4. Discussion and Conclusions

In this study, we explored deep learning methods for the assessment of lifespan stages of C. elegans. Compared with traditional machine learning algorithms [44,45], deep learning algorithms offer many advantages. A two-way feature fusion model based on convolutional neural networks and biomarker changes was proposed to estimate the lifespan stages of C. elegans. The model can be divided into an EfficientNet-based sub-module with an attention mechanism and an aggregation feature extraction module that uses traditional image processing algorithms to calculate the aggregation information for fluorescent protein bright spots. The experimental results showed that, compared with other classification networks, the model introduced in this paper can effectively improve classification accuracy and meet the needs of biological research for the estimation of C. elegans lifespan stages.
The focus of this paper is the study of lifespan stage estimation of C. elegans based on convolutional neural networks and biomarker changes. The aging assessment system designed has a higher possibility of technology migration than previous studies. In the absence of interventions or restricted conditions for the cultivation of C. elegans, a high accuracy rate can be maintained, reflecting the aging process of C. elegans under natural conditions. The results for this system show that in C. elegans the degree of protein aggregation is a good indicator of aging, while machine learning and deep learning can maximize the use of biological data. The research results can be used as references for the future development of human lifespan evaluation and prediction methods. It should be noted that there are some deficiencies and unresolved problems with the research which need to be further explored and studied. First, for the detection of fluorescent protein bright spots of C. elegans, new methods could be investigated to segment the C. elegans images. For example, we could explore more target detection models in the computer vision field for the detection of fluorescent protein bright spots. Second, for the C. elegans lifespan stage estimation model based on convolutional neural networks, different levels of features could be optimized, because there is a certain amount of redundant information connecting different levels of features. Third, this paper only analyzed images containing a single C. elegans, and the actual application scenarios are limited. As this method relies on high-resolution images to capture fluorescent bright spots, it is currently unable to work on a large number of C. elegans images. In the future, we plan to investigate lifespan estimation and analysis of multiple C. elegans in a single image. The model framework can also be transported to mobile platforms, such as TensorFlow Lite, so that it can be applied in actual work scenarios.

Author Contributions

Data curation, Y.S. and J.L.; Investigation, J.L.; Methodology, Y.S., Y.Y. and J.T.; Project administration, J.L.; Software, Y.S. and Y.Y.; Supervision, J.L. and J.T.; Writing—original draft, Y.S. and J.L.; Writing—review & editing, Y.Y. and J.T. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

No new data were created or analyzed in this study. Data sharing is not applicable to this article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Antoshechkin, I.; Sternberg, P.W. The versatile worm: Genetic and genomic resources for Caenorhabditis elegans research. Nat. Rev. Genet. 2007, 8, 518–532. [Google Scholar] [CrossRef] [PubMed]
  2. Goldstein, B. Sydney Brenner on the Genetics of Caenorhabditis elegans. Genetics 2016, 204, 1–2. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Cook, S.J.; Jarrell, T.A.; Brittin, C.A.; Wang, Y.; Bloniarz, A.E.; Yakovlev, M.A.; Nguyen, K.C.Q.; Tang, L.T.-H.; Bayer, E.A.; Duerr, J.S.; et al. Whole-animal connectomes of both Caenorhabditis elegans sexes. Nature 2019, 571, 63–71. [Google Scholar] [CrossRef] [PubMed]
  4. Melov, S. Uncovering the Dark Energy of Aging. Cell Syst. 2016, 3, 328–330. [Google Scholar] [CrossRef] [Green Version]
  5. Hua, X.; Han, J.; Zhao, C.; Tang, H.; He, Z.; Chen, Q.; Tang, S.; Tang, J.; Zhou, W. A novel method for ECG signal classification via one-dimensional convolutional neural network. Multimed. Syst. 2020, 28, 1387–1399. [Google Scholar] [CrossRef]
  6. He, J.; Zhu, Q.; Zhang, K.; Yu, P.; Tang, J. An evolvable adversarial network with gradient penalty for COVID-19 infection segmentation. Appl. Soft Comput. 2021, 113, 107947. [Google Scholar] [CrossRef] [PubMed]
  7. Mu, N.; Wang, H.; Zhang, Y.; Jiang, J.; Tang, J. Progressive global perception and local polishing network for lung infection segmentation of COVID-19 CT images. Pattern Recognit. 2021, 120, 108168. [Google Scholar] [CrossRef] [PubMed]
  8. Zhao, C.; Vij, A.; Malhotra, S.; Tang, J.; Tang, H.; Pienta, D.; Xu, Z.; Zhou, W. Automatic extraction and stenosis evaluation of coronary arteries in invasive coronary angiograms. Comput. Biol. Med. 2021, 136, 104667. [Google Scholar] [CrossRef]
  9. Liu, X.; Guo, Z.; Cao, J.; Tang, J. MDC-net: A new convolutional neural network for nucleus segmentation in histopathology images with distance maps and contour information. Comput. Biol. Med. 2021, 135, 104543. [Google Scholar] [CrossRef]
  10. Cao, S.; Liu, J. Terminal Bulb Segmentation of Caenorhabditis Elegans under Small Samples Based on Two-stage U-Net Network. In Proceedings of the 2019 3rd International Conference on Computer Science and Artificial Intelligence, Beijing, China, 6–8 December 2019. [Google Scholar] [CrossRef]
  11. Lin, J.-L.; Kuo, W.-L.; Huang, Y.-H.; Jong, T.-L.; Hsu, A.-L.; Hsu, W.-H. Using Convolutional Neural Networks to Measure the Physiological Age of Caenorhabditis elegans. IEEE/ACM Trans. Comput. Biol. Bioinform. 2020, 18, 2724–2732. [Google Scholar] [CrossRef]
  12. Fudickar, S.; Nustede, E.; Dreyer, E.; Bornhorst, J. Mask R-CNN Based C. elegans Detection with a DIY Microscope. Biosensors 2021, 11, 257. [Google Scholar] [CrossRef]
  13. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  14. Stroustrup, N.; Anthony, W.E.; Nash, Z.M.; Gowda, V.; Gomez, A.; López-Moyado, I.F.; Apfeld, J.; Fontana, W. The temporal scaling of Caenorhabditis elegans ageing. Nature 2016, 530, 103–107. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  15. Martineau, C.N.; Brown, A.E.; Laurent, P. Multidimensional phenotyping predicts lifespan and quantifies health in C. elegans. bioRxiv 2019. bioRxiv:681197. [Google Scholar]
  16. Karp, X.; Hammell, M.; Ow, M.C.; Ambros, V. Effect of life history on microRNA expression during C. elegans development. RNA 2011, 17, 639–651. [Google Scholar] [CrossRef] [Green Version]
  17. Wan, C.; Freitas, A.A.; De Magalh, J.P. Predicting the pro-longevity or anti-longevity effect of model organism genes with new hierarchical feature selection methods. IEEE/ACM Trans. Comput. Biol. Bioinform. 2014, 12, 262–275. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  18. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask R-CNN. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2961–2969. [Google Scholar]
  19. Saberi-Bosari, S.; Flores, K.B.; San-Miguel, A. Deep learning-enabled analysis reveals distinct neuronal phenotypes induced by aging and cold-shock. BMC Biol. 2020, 18, 130. [Google Scholar] [CrossRef]
  20. Zhang, N.; Li, Y.; Gao, T.; Ma, L.; Li, G. Research Progress of C. elegansas a Human Disease Model. Chin. J. Food Hyg. 2014, 26, 398–403. [Google Scholar]
  21. Libina, N.; Berman, J.R.; Kenyon, C. Tissue-specific activities of C. elegans DAF-16 in the regulation of lifespan. Cell 2003, 115, 489–502. [Google Scholar] [CrossRef] [Green Version]
  22. Cornaglia, M.; Krishnamani, G.; Mouchiroud, L.; Sorrentino, V.; Lehnert, T.; Auwerx, J.; Gijs, M.A.M. Automated longitudinal monitoring of in vivo protein aggregation in neurodegenerative disease C. elegans models. Mol. Neurodegener. 2016, 11, 17. [Google Scholar] [CrossRef] [Green Version]
  23. Kaufman, D.M.; Wu, X.; Scott, B.A.; Itani, O.A.; Van Gilst, M.R.; Bruce, J.E.; Crowder, C.M. Ageing and hypoxia cause protein aggregation in mitochondria. Cell Death Differ. 2017, 24, 1730–1738. [Google Scholar] [CrossRef] [Green Version]
  24. Son, J.H.; Shim, J.H.; Kim, K.-H.; Han, J.Y. Neuronal autophagy and neurodegenerative diseases. Exp. Mol. Med. 2012, 44, 89–98. [Google Scholar] [CrossRef] [PubMed]
  25. Twelves, D.; Perkins, K.S.; Counsell, C. Systematic review of incidence studies of Parkinson’s disease. Mov. Disord. Off. J. Mov. Disord. Soc. 2003, 18, 19–31. [Google Scholar] [CrossRef] [PubMed]
  26. David, D.C.; Ollikainen, N.; Trinidad, J.C.; Cary, M.P.; Burlingame, A.L.; Kenyon, C. Widespread Protein Aggregation as an Inherent Part of Aging in C. elegans. PLoS Biol. 2010, 8, e1000450. [Google Scholar] [CrossRef] [PubMed]
  27. Vázquez, T.S.; Askjaer, P. NanoBiT based toolkit to study protein-protein interactions in C. elegans. Biosaia 2018, 7, 55. [Google Scholar]
  28. LeCun, Y.; Boser, B.; Denker, J.S.; Henderson, D.; Howard, R.E.; Hubbard, W.; Jackel, L.D. Backpropagation Applied to Handwritten Zip Code Recognition. Neural Comput. 1989, 1, 541–551. [Google Scholar] [CrossRef]
  29. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. NIPS 2012, 60, 84–90. [Google Scholar] [CrossRef] [Green Version]
  30. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  31. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  32. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar]
  33. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
  34. Kern, A.; Ackermann, B.; Clement, A.M.; Duerk, H.; Behl, C. HSF1-Controlled and Age-Associated Chaperone Capacity in Neurons and Muscle Cells of C. elegans. PLoS ONE 2010, 5, e8568. [Google Scholar] [CrossRef] [Green Version]
  35. Nussbaum-Krammer, C.I.; Neto, M.F.; Brielmann, R.M.; Pedersen, J.S.; Morimoto, R.I. Investigating the Spreading and Toxicity of Prion-like Proteins Using the Metazoan Model Organism C. elegans. J. Vis. Exp. 2015, 95, e52321. [Google Scholar] [CrossRef]
  36. Walther, D.M.; Kasturi, P.; Zheng, M.; Pinkert, S.; Vecchi, G.; Ciryam, P.; Morimoto, R.I.; Dobson, C.M.; Vendruscolo, M.; Mann, M.; et al. Widespread Proteome Remodeling and Aggregation in Aging C. elegans. Cell 2015, 161, 919–932. [Google Scholar] [CrossRef] [Green Version]
  37. Gupta, R.; Kasturi, P.; Bracher, A.; Loew, C.; Zheng, M.; Villella, A.; Garza, D.; Hartl, F.U.; Raychaudhuri, S. Firefly luciferase mutants as sensors of proteome stress. Nat. Methods 2011, 8, 879–884. [Google Scholar] [CrossRef]
  38. Huang, Y.; Cheng, Y.; Bapna, A.; Firat, O.; Chen, D.; Chen, M.; Lee, H.; Ngiam, J.; Le, Q.V.; Wu, Y. Gpipe: Efficient training of giant neural networks using pipeline parallelism. Adv. Neural Inf. Process. Syst. 2019, 32, 103–112. [Google Scholar]
  39. Zagoruyko, S.; Komodakis, N. Wide residual networks. arXiv 2016, arXiv:1605.07146. [Google Scholar]
  40. Bradski, G.; Kaehler, A. OpenCV. Dr. Dobb’s J. Softw. Tools 2000, 3, 120. [Google Scholar]
  41. Jang, E.; Gu, S.; Poole, B. Categorical reparameterization with gumbel-softmax. arXiv 2016, arXiv:1611.01144. [Google Scholar]
  42. Srivastava, N.; Hinton, G.; Krizhevsky, A.; Sutskever, I.; Salakhutdinov, R. Dropout: A simple way to prevent neural networks from overfitting. J. Mach. Learn. Res. 2014, 15, 1929–1958. [Google Scholar]
  43. Zhang, Z.; Sabuncu, M. Generalized cross entropy loss for training deep neural networks with noisy labels. In Proceedings of the 32nd Conference on Neural Information Processing Systems, Montreal, QC, Canada, 3–8 December 2018. [Google Scholar]
  44. Tang, J.; Liu, X.; Cheng, H.; Robinette, K.M. Gender Recognition Using 3-D Human Body Shapes. IEEE Trans. Syst. Man Cybern. Part C (Applications Rev. 2011, 41, 898–908. [Google Scholar] [CrossRef]
  45. Xu, J.; Cao, Y.-Y.; Sun, Y.; Tang, J. Absolute Exponential Stability of Recurrent Neural Networks With Generalized Activation Function. IEEE Trans. Neural Netw. 2008, 19, 1075–1089. [Google Scholar] [CrossRef]
Figure 1. Representative pictures from each class. The serial number in the upper-left area of each picture is the class each picture belongs to. Each class represents a 4% period of time of the whole lifespan of C. elegans (e.g., the picture taken on day 5 of an individual with a 15-day lifespan would be assigned to class 9). The original photo was 4000 × 6000 pixels. We manually intercepted 300 × 300 pixels of the region with protein aggregation. The pictures in the figure are representative for each class (not from the same individual). Scale bar: 5 μm (the actual length of the white line is 5 μm).
Figure 1. Representative pictures from each class. The serial number in the upper-left area of each picture is the class each picture belongs to. Each class represents a 4% period of time of the whole lifespan of C. elegans (e.g., the picture taken on day 5 of an individual with a 15-day lifespan would be assigned to class 9). The original photo was 4000 × 6000 pixels. We manually intercepted 300 × 300 pixels of the region with protein aggregation. The pictures in the figure are representative for each class (not from the same individual). Scale bar: 5 μm (the actual length of the white line is 5 μm).
Bioengineering 09 00689 g001
Figure 2. The EfficientNetB0 network is outlined by red dashed lines. Similar to the original EfficientNet model, Stem is a common structure. Module1 contains DepthwiseConv2D, BatchNormalization, and an activation function. Module2 consists of two module1 units which are connected by the Zero Padding layer in the middle. Module3 consists of a global pooling layer, rescaling, and two Conv2Ds. The attention module is outlined by blue dashed lines.
Figure 2. The EfficientNetB0 network is outlined by red dashed lines. Similar to the original EfficientNet model, Stem is a common structure. Module1 contains DepthwiseConv2D, BatchNormalization, and an activation function. Module2 consists of two module1 units which are connected by the Zero Padding layer in the middle. Module3 consists of a global pooling layer, rescaling, and two Conv2Ds. The attention module is outlined by blue dashed lines.
Bioengineering 09 00689 g002
Figure 3. Detailed structure of the attention module. The attention mechanism network mainly takes the output feature map F of the efficient net as input, which is then put into three 1 × 1 convolutional layers. The activation functions ReLU and Sigmoid are added after the batchnorm operation to convert the input into nonlinear features.
Figure 3. Detailed structure of the attention module. The attention mechanism network mainly takes the output feature map F of the efficient net as input, which is then put into three 1 × 1 convolutional layers. The activation functions ReLU and Sigmoid are added after the batchnorm operation to convert the input into nonlinear features.
Bioengineering 09 00689 g003
Figure 4. Informational route for fluorescent protein bright spot aggregation degree. (a) Image preprocessing. (b) Bright spot information extraction and aggregation degree calculation.
Figure 4. Informational route for fluorescent protein bright spot aggregation degree. (a) Image preprocessing. (b) Bright spot information extraction and aggregation degree calculation.
Bioengineering 09 00689 g004
Figure 5. Overall framework for the CNN dual-path feature fusion model.
Figure 5. Overall framework for the CNN dual-path feature fusion model.
Bioengineering 09 00689 g005
Figure 6. Confusion matrix. (a) M1 module classification effect. (b) M2 module classification effect. (c) The classification effect of the method in this paper.
Figure 6. Confusion matrix. (a) M1 module classification effect. (b) M2 module classification effect. (c) The classification effect of the method in this paper.
Bioengineering 09 00689 g006
Table 1. Number of datasets.
Table 1. Number of datasets.
GroupNumber of ImagesTrainingTest
1742593149
2976780196
3587469118
4677541136
5822657165
6778622156
Table 2. SimpleBlobDetector parameter settings.
Table 2. SimpleBlobDetector parameter settings.
Parameter TypeParameter Value
thresholdStep8
minArea10
maxArea2500
minCircularity0.2
minConvexity0.75
minInertiaRatio0.1
Table 3. Ablation experiment results for each module.
Table 3. Ablation experiment results for each module.
ModelACRecallF1
M1 0.6400.6470.643
M2 0.5260.5420.534
M1 + M2 0.7260.7420.734
Table 4. Comparison of the method introduced in this paper with other classification models with respect to the M2 module. We performed two experiments for each model. The first row of experimental results for each model presents the results without the M2 module. The second row presents the experimental results using the M2 module.
Table 4. Comparison of the method introduced in this paper with other classification models with respect to the M2 module. We performed two experiments for each model. The first row of experimental results for each model presents the results without the M2 module. The second row presents the experimental results using the M2 module.
ModelM2ACRecallF1
VGG16 0.4780.4860.477
0.5560.5740.583
InceptionV3 0.5540.5740.569
0.6020.6140.615
MobileNetV3 0.5110.5290.513
0.5710.5930.580
ResNet50 0.5580.5240.522
0.6530.6500.657
DenseNet 0.5750.5570.565
0.7130.7130.723
Ours 0.6400.6470.643
0.7260.7420.734
Table 5. Comparison of the method introduced in this paper with other classification models with respect to the M2 module. The first row of experimental results presents the results without the attention module. The second row presents the experimental results using the attention module.
Table 5. Comparison of the method introduced in this paper with other classification models with respect to the M2 module. The first row of experimental results presents the results without the attention module. The second row presents the experimental results using the attention module.
ModelAttentionACRecallF1
VGG16 0.5080.5570.496
0.5560.5740.583
InceptionV3 0.5430.5960.548
0.6020.6140.615
MobileNetV3 0.5830.5950.566
0.5710.5930.580
ResNet50 0.4730.4990.490
0.6530.6500.657
DenseNet 0.6660.7020.698
0.7130.7130.723
Ours 0.6090.7030.714
0.7260.7420.734
Table 6. Statistical results for the random forest method.
Table 6. Statistical results for the random forest method.
ACRecallF1
10.820.820.82
20.650.710.68
30.340.310.33
40.370.420.39
50.420.400.41
60.630.570.60
Overall accuracy0.50--
Table 7. Statistical results for the SVM method.
Table 7. Statistical results for the SVM method.
ACRecallF1
10.540.260.35
20.340.380.36
30.170.040.07
40.250.960.39
50.0000
60.0000
Overall accuracy0.27--
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Song, Y.; Liu, J.; Yin, Y.; Tang, J. Estimation of Caenorhabditis Elegans Lifespan Stages Using a Dual-Path Network Combining Biomarkers and Physiological Changes. Bioengineering 2022, 9, 689. https://0-doi-org.brum.beds.ac.uk/10.3390/bioengineering9110689

AMA Style

Song Y, Liu J, Yin Y, Tang J. Estimation of Caenorhabditis Elegans Lifespan Stages Using a Dual-Path Network Combining Biomarkers and Physiological Changes. Bioengineering. 2022; 9(11):689. https://0-doi-org.brum.beds.ac.uk/10.3390/bioengineering9110689

Chicago/Turabian Style

Song, Yao, Jun Liu, Yanhao Yin, and Jinshan Tang. 2022. "Estimation of Caenorhabditis Elegans Lifespan Stages Using a Dual-Path Network Combining Biomarkers and Physiological Changes" Bioengineering 9, no. 11: 689. https://0-doi-org.brum.beds.ac.uk/10.3390/bioengineering9110689

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop