Next Article in Journal
Extraction and Application of Hydraulic Support Safety Valve Characteristic Parameters Based on Roof Pressure Data
Next Article in Special Issue
An Autoscaling System Based on Predicting the Demand for Resources and Responding to Failure in Forecasting
Previous Article in Journal
Multi-Scene Mask Detection Based on Multi-Scale Residual and Complementary Attention Mechanism
Previous Article in Special Issue
Ensemble Approach to Combining Episode Prediction Models Using Sequential Circadian Rhythm Sensor Data from Mental Health Patients
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Cat Monitoring System Based on Concept Drift Adaptive Machine Learning Architecture

1
Research and Development Department, Codevision Inc., Seoul 03722, Republic of Korea
2
Development Department, Valiantx Co., Ltd., Bucheon 14553, Republic of Korea
3
Department of Culture Techno, Changwon National University, Changwon 51140, Republic of Korea
*
Author to whom correspondence should be addressed.
Submission received: 31 August 2023 / Revised: 29 October 2023 / Accepted: 30 October 2023 / Published: 31 October 2023

Abstract

:
In multi-cat households, monitoring individual cats’ various behaviors is essential for diagnosing their health and ensuring their well-being. This study focuses on the defecation and urination activities of cats, and introduces an adaptive cat identification architecture based on deep learning (DL) and machine learning (ML) methods. The architecture comprises an object detector and a classification module, with the primary focus on the design of the classification component. The DL object detection algorithm, YOLOv4, is used for the cat object detector, with the convolutional neural network, EfficientNetV2, serving as the backbone for our feature extractor in identity classification with several ML classifiers. Additionally, to address changes in cat composition and individual cat appearances in multi-cat households, we propose an adaptive concept drift approach involving retraining the classification module. To support our research, we compile a comprehensive cat body dataset comprising 8934 images of 36 cats. After a rigorous evaluation of different combinations of DL models and classifiers, we find that the support vector machine (SVM) classifier yields the best performance, achieving an impressive identification accuracy of 94.53%. This outstanding result underscores the effectiveness of the system in accurately identifying cats.

1. Introduction

The increasing trend of households choosing to adopt pet cats (Felis catus) is evident in modern society. The surge in cat ownership has brought about greater awareness and concern for the well-being and overall healthcare of these feline companions. This growing interest has, in turn, led to a spike in scientific and veterinary research aimed at proper diagnosis, health management, and treatment of diseases pertaining to cats [1,2,3,4]. Nevertheless, as the emphasis on cat healthcare increases, specific challenges have arisen, particularly in settings where multiple cats coexist. For households with more than one cat or in locations such as cat shelters, closely monitoring the health and well-being of each individual cat has become a considerably more complex undertaking. Each cat possesses unique personalities, behaviors, and health profiles, necessitating additional attention and diligence in their monitoring. In addition to ensuring that the cats are fed and sheltered, it is also crucial to spot any subtle signs of health issues, which can often go unnoticed in a multi-cat environment.
Consequently, systems for remote monitoring in multi-cat households are continuously being researched. For instance, Majid et al. [5] studied an IoT-based cat feeding and monitoring system, identifying which cat ate the food through RFID tags attached to the cats’ collars. Eagan et al. [6] researched a computer vision-based marker tracking system for cats in shelters, tracking the behavior of individual cats. They identified the cats by detecting 2D ArUco markers printed on paper collars attached to the cats’ necks. While these contact-based systems can continuously monitor the location and behavior of the target pet cats, there are concerns regarding distress and safety for things like collars attached to the cats [7]. In this study, leveraging the advancements in computer vision and artificial intelligence technologies, we propose a cat identification system for monitoring the health of pet cats.
We introduce a novel two-step system that relies on machine learning (ML) methods, specifically integrating an object detector and a classification module. We focus on identifying the optimal combination within our classification module, aiming to achieve the highest accuracy in individual cat identification. Specifically, this study focuses on monitoring defecation and urination activities, which serve as clues for diagnosing the health of cats, and conceptualizes and experiments with a cat identification system in litter boxes. One of the common diseases that can occur in domestic cats is feline lower urinary tract disease (FLUTD), which refers to diseases affecting the cat’s bladder or urethra, presenting symptoms such as pollakiuria, periuria, stranguria, and hematuria [4,8]. If a cat shows pollakiuria, we may suspect diseases such as cystitis or urinary stones. Meanwhile, diseases such as feline chronic colitis can affect a cat’s defecation frequency [9]. In multi-cat households, recognizing cat activities in the litter box and determining the frequency of individual cats’ defecation and urination activities are effective in diagnosing these cat digestive and excretory organ conditions.
In the case of multi-cat households, the composition of the cat population may change over time—whether due to the adoption of new cats or other factors. Additionally, cats undergo significant morphological transformations as they transition from kittens to adults—a process that occurs within a relatively short span of time. Given the complexity of this pattern and its potential impact on cat identification, we design an algorithm specifically tailored to make our model highly adaptive to these appearance shifts. This adaptability is crucial not only for recognizing evolving cat features but also for addressing the dynamic environments in which cats reside. For instance, the location and lighting conditions of a litter box can vary. Our retraining mechanism ensures that the monitoring system remains updated, factoring in these variable environments and the changing appearances of cats. This continuous recalibration empowers our system to consistently deliver precise identification.
Furthermore, we compile and utilize a cat body dataset obtained through monocular cameras placed in litter boxes. This dataset includes bounding box labels for both the cat’s face and body. For the body bounding box, there might be instances where the face is not present, resulting in a higher degree of freedom for specific entities. Despite using these body labels for training, our model exhibits good identification accuracy. This success not only underscores the robustness of our ML algorithms, but also highlights the potential of leveraging diverse data sources to enhance diagnostic precision.
One of the main contributions of this study is the determination of the optimal combination for the classification module of our proposed monitoring system. The usability of the model is enhanced by training it with a body image dataset that captures the entire cat body, offering a higher degree of freedom, rather than focusing solely on the cat’s face. In addition, we propose a model-retraining algorithm to ensure consistent performance, especially for cats whose appearances may change rapidly.
The remainder of this paper is organized as follows. Section 2 reviews previous research relates to animal identification and model retraining. Section 3 outlines our method for investigating the best approach to cat identification across the entire architecture. Section 4 presents the experimental results and identifies the optimal combination for the classification module. Section 5 and Section 6 provide a discussion and summary of this study, along with our contributions and suggestions for future research.

2. Related Works

2.1. ML-Based Identification Systems

Recent studies have proposed deep learning (DL)- and ML-based architectures for the individual identification of various animal species. Hou et al. [10] conducted research on recognizing the faces of 25 giant pandas (Ailuropoda melanoleuca) using an architecture composed of a convolutional neural network (CNN)-based model, the Visual Geometry Group Network (VGGNet) [11], with a softmax layer as the classifier. Hitelman et al. [12] designed a biometric identification system for sheep (Ovis aries) by employing a two-step approach involving detection and classification. They utilized a Faster R-CNN [13] for sheep-face detection, and compared seven CNN classification models trained with the ArcFace loss function [14]. Schofield et al. [15] focused on the facial recognition of wild chimpanzees (Pan troglodytes verus) using a single-shot detector (SSD) model [16] as the detector and a VGG-M [17] based network for identification. Clapham et al. [18] concentrated on the facial recognition of 132 brown bears (Ursus arctos) by adopting Schroff et al.’s [19] approach and implementing the overall structure using the dlib toolkit [20]. They utilized a sliding-window-based CNN and an ensemble of regression trees for face detection and alignment in object detection. Following face reorientation and cropping, they generated bear face embeddings via ResNet-34 [21]. Identity classification was carried out using a linear support vector machine (SVM), and the encoding model was trained utilizing pairwise hinge loss.
Similarly, in this study, we explore a CNN-based detector and feature extractor to recognize cats. However, unlike previous studies, we focus on a system that identifies cat body images rather than only cat faces. For the detection model, we utilize YOLOv4 [22], which is known for its fast and accurate performance among the CNN-based detectors. For the classification module, we design an architecture based on [19]. We use a CNN model, EfficientNetV2 [23], as the feature extractor of our architecture. EfficientNetV2 is a powerful CNN model that specifically focuses on efficiency in terms of parameters, floating-point operations or FLOPs, and training speed. Subsequently, ML-based classifiers are used to identify individual cats. Table 1 summarizes existing DL- and ML-based animal identification methods.

2.2. Model Adaptation for Concept Drift

In ML and data science, the evolving appearance of pet cats as they mature and grow can be equated to a form of concept drift. Concept drift refers to the phenomenon where the statistics of a target variable in data-based learning models change after the initial training [24]. In various real-world domains, shifts in the data distribution can trigger concept drift, which degrades the model performance over time.
Typically, concept drift adaptive models are initially trained on the target variable, detect drift from the classification accuracy or the statistical characteristics of the data distribution, and are subsequently retrained to accommodate the detected drift. For example, the drift detection method [25] analyzes the error rate of input data to detect abrupt drifts. In the case of adaptive windowing [26], it assumes that there is no change in the distribution of the input data and the combined mean of the two sub-windows is compared when new data are used.
In animal monitoring, the impact of concept drift is significant if there is no control over the external environment or the subjects being monitored. Moallem et al. [27] proposed a system for detecting wild birds by using a two-stage deep neural network pipeline. With a particular focus on the changes in the background of the data, they suggested retraining the model if the average of the images collected throughout the day deviated from the mean images from any single day in the training set.
We propose a periodic retraining method without drift detection. With this method the model continuously adapts to the concept drift stemming from cat class changes, appearance changes in a cat’s life cycle, and environmental factors like lighting or backgrounds. Considering the memory efficiency of the recognition server, the number of embedding vectors is fixed when retraining the ML-based classifier. This iterative refinement ensures that the model remains robust and accurate, capturing the intricacies of a cat’s evolving appearance and the dynamic conditions under which they are observed.

3. Proposed Method

3.1. Cat Identification Architecture

Our cat identification architecture comprises two key components: the object detector in an embedded computer within the cat litter box, and the classification module hosted on the server. The overall architecture for our system is shown in Figure 1.
When the cat enters the litter box, a RGB camera equipped with an embedded computer detects its defecation activity and records it as a video. The object detector in the litter box focuses on capturing the entire body of the cat during these activities. When a series of body images for an activity are detected, they are transmitted to a remote server for identification of the given cat body images.
We use YOLOv4 for the detector model, which enables accurate real-time detection within an embedded computer. The structure of the YOLOv4 network is illustrated in Figure 2. YOLOv4 incorporates various optimizations into the YOLOv3 [28] model. Notably, YOLOv4 adopted the cross-stage hierarchy approach of the cross stage partial network (CSPNet) [29] to change the Darknet53 from YOLOv3 to CSPDarknet53. This approach significantly reduced the computational cost of each layer in the backbone, resulting in improved performance during training and inference. Additionally, YOLOv4 leveraged the structures of the spatial pyramid pooling network (SPPNet) [30] and the path aggregation network (PANet) [31] in its neck module to increase the receptive field and enhance detection performance by augmenting different backbone paths. Based on the exceptional performance and learning efficiency of YOLOv4, we adopt it as the detector in our cat identification model.
In the identification server, the final inference is achieved through with two model components: a feature extractor and a classifier. When the feature extractor receives a body image as input from the litter box, it extracts an embedding vector from the given image. To accomplish this, the feature extractor employs EfficientNetV2. The overall structure of the feature extractor is illustrated in Figure 3. EfficientNetV2 is an improved version of the EfficientNet [32] model, which was derived from a mobile neural architecture search network (MnasNet) [33] and utilizes a neural architecture search (NAS) [33,34] to determine the optimal model compound scaling dimensions (depth, width, and resolution). By incorporating mobile inverted bottleneck convolution (MBConv) blocks with squeeze and excitation blocks, it reduced computational complexity while enhancing the performance. EfficientNet demonstrated significantly higher performance with a much smaller number of parameters than traditional CNN models. EfficientNetV2 focused on enhancing training efficiency. This was achieved by using Fused-MBConv blocks early in the model, which replaced depth-wise 3 × 3 convolutions with regular 3 × 3 convolutions, reducing the overhead associated with depth-wise convolution GPU operations. Additionally, rather than employing the traditional compound scaling method that uniformly scales the model size, EfficientNetV2 utilized non-uniform scaling in the later stages of the model, placing more emphasis on scaling to find a more efficient model structure for training. These advancements in EfficientNetV2 have contributed to its superior efficiency and performance compared to its predecessors. Considering the limited server resources and the need for frequent retraining, we select EfficientNetV2 for its lightweight characteristics and rapid training capabilities.
This network extracts a 128-dimensional embedding, following the architecture design outlined in [19], which is then fed into the ML-based embedding classifier. The classifier uses these feature vectors as inputs and subsequently determines the class of cats present in each input image. Once a class is determined by the classifier, the server stores and aggregates data regarding the cat’s activity. A user can check the daily or weekly litter box usage status for each pet cat, along with the stored videos. By understanding the frequency and condition of defecation and urination, the user can gauge the state of the cat’s digestive and excretory organs.

3.2. Classification Module Adaptation

Within the litter box, the overall environment observed by the camera can continuously change over time. The background may shift due to factors such as relocating the litter box or altering the lighting conditions at the litter box location. Additionally, foreground distribution can undergo significant changes owing to changes in the composition of cats or individual transformations in the appearance of the cats themselves. In such dynamic situations, retraining vision DL models and embedding-based ML models is crucial for consistently maintaining high identification performance.
In this context, we design a classification module to effectively adapt to changes in the composition or appearance of a user’s pet cats through an interactive scenario between the user and the server. This allows the server to continuously update and refine its identification capabilities based on user input and feedback, ensuring the accurate and personalized identification of pet cats over time.
Initially, a user registers the information of the litter box and cats on the server. Since the model classifier cannot make predictions from litter box images without any cues, a set of images taken by the user for each cat is also transmitted to the server. After registration, a two-step retraining process is designed, comprising embedding vector selection and fine-tuning. A flowchart illustrating these steps is shown in Figure 4.
As the server continues to collect images over a specific period (T), the user verifies whether the cat identification is properly conducted and relabels the classes of these images. At this time, the user labels one video in which the cat entered and exited, and all images corresponding to the labeled video are labeled at once. Then, the model classifier is retrained with these relabeled litter box images to obtain more accurate identification. As a certain number (M) of new images for each cat is gathered, a mean vector is calculated using both the existing embedding vectors and the collected vectors. This mean vector serves as a representation of the cat’s characteristics. To optimize the training data, the embedding vectors closest to this mean vector are selected (N-prioritized vectors), while the remaining vectors are discarded. Through this approach, we can remove outlier data for a specific class and stabilize the amount of training data for that class, thereby maintaining memory efficiency on the server. In addition, when a new cat is registered, the embedding vector selection for that cat starts by learning the user’s image for the cat and importing the images during period T. If a user deregisters a specific cat, the embedding vector for that cat is deleted.
Once a sufficient number (C) of images for all cats are collected, the user again relabels the images. The feature extractor is fine-tuned using the relabeled dataset. A fine-tuned feature extractor is used to extract new embedding vectors from the training sample images. By retraining the extracted embedding vectors back into the classifier, the model can adapt to changes in the appearance of pet cats within a particular household.
This iterative and interactive approach ensures the continuous enhancement and refinement of the system performance, as it continuously adapts to changes in the configuration and individual appearance of a user’s pet cats, and to changes in the external environment. By leveraging the new data and user feedback, our retraining process aims to achieve the accurate and personalized identification of pet cats over time, ultimately improving the overall efficiency and effectiveness of the system.

4. Results

4.1. Environment

The overall experiment is performed on a desktop computer equipped with an Intel® Core™ i7-12700KF and 16.0 GB RAM. The architecture is trained using an NVIDIA RTX A6000. All source codes, including training and testing, are implemented employing Python 3.8 and PyTorch 11.3 libraries with the CUDA toolkit on Ubuntu 21.04.

4.2. Dataset

To demonstrate the proposed cat monitoring system, we first create our own cat body dataset. Using customized litter boxes (515 × 695 × 475 mm, with a hole in front) with a monocular camera, we gather diverse cat image data by recording videos of their activities. For experiments on the identification module, bounding boxes for the face and body are labeled and used for the collected images. The litter boxes are positioned in various environments such as cat cafés, streets, and shelters.
To make the architecture robust to changes in the color temperature, we diversify the dataset by applying four color filters: warm white (2700 K), natural white (4100 K), cool white (6500 K), and LED light (10,000 K) by 1:1:1:1, as shown in Figure 5. It should be noted that when we acquire the data, we assume that the color temperature environment was cool white. So there is no difference before and after passing through the filter for cool white.
The dataset comprises 8934 RGB Full HD images of 36 cats. We divided the dataset into 8:2 ratios using stratified sampling for model training and evaluation.

4.3. Classification Module Training

A metric learning technique is used to enable the feature extractor to learn the similarity of data in the embedding space. We train the feature extractor using triplet loss [10]. The loss function employed in this study demonstrated excellent performance in facial recognition tasks. It utilizes a technique that learns the structure of the feature representation by separating the positive pairs and negative pairs. This approach aims to enhance the discriminative power of feature embeddings, allowing the model to effectively distinguish between similar and dissimilar instances, thereby improving the accuracy and effectiveness of the recognition system.
The overall training process comprises three stages; each stage utilizes semi-hard, hard, and hardest triplets. The margin for calculating the loss is set at 0.2. We also add a global orthogonal regularization term [35] to the loss function, to spread the features throughout the embedding space. Adam is used as the model optimizer, with the learning rate and momentum parameters β1 and β2 being set to 0.001, 0.9, and 0.999, respectively.
To train the feature extractor, 7162 images of 36 cats are used, which are split into 8:2 ratios for training and validation. If the validation loss does not decrease for 5 epochs during a stage, the training for that stage is stopped early; otherwise, it continues for up to 100 epochs. The loss graph for each training stage is shown in Figure 6.
Once the feature extractor is trained, we obtain the embedding vectors from all training and validation data and use them to fit the classifier. By comparing the identification accuracies of different ML classifiers, we select the one with the highest accuracy as the final model classifier.

4.4. Evaluation Metrics

In this experiment, we assess the classification module’s performance using several evaluation metrics, including accuracy, confusion matrix, receiver operating characteristic (ROC) curves, and precision-recall (PR) curves.
The accuracy is the percentage of samples that the classifier classifies from a given sample into the correct class:
Accuracy = Correct predictions/All predictions.
The confusion matrix visualizes the performance of the classification algorithm and comprises true positives (TP), false negatives (FN), false positives (FP), and true negatives (TN). TP is the number of samples which accurately classify a category of interest as a category of interest. FN is the number of samples that misclassify a category of interest as not a category of interest. FP is the number of samples that misclassify non-interest categories as interest categories, and TN is the number of samples that accurately classify non-interest categories.
The ROC curve expresses the relationship between the true positive rate (TPR) and false positive rate (FPR) as practice values for trainees, which are calculated as follows:
TPR = TP/(TP + FN),
FPR = FP/(FP + TN).
The PR curve expresses the relationship between recall and precision as a threshold for class-specific probability; recall and precision are calculated as follows:
Recall (R) = TP/(TP + FN),
Precision (P) = TP/(TP + FP).

4.5. Experimental Results

4.5.1. Performance Comparison with Different Classifiers

In this study, the performance of various classification methods is evaluated using a test dataset comprising 1772 images of 36 different cats. Each test image is classified based on its embedding vector, using a complete set of embedding vectors from the training dataset of 36 cats. We examine classifiers such as K-nearest neighbors (KNN), random forest, and SVMs. The identification accuracies of the classifiers are listed in Table 2. Notably, the SVM classifier equipped with a linear kernel achieves the highest identification accuracy of 94.53%. Other performance metrics, such as the confusion matrix, ROC curves, and PR curves, are illustrated in Figure 7 and Figure 8. The additional inference results are presented in Figure 9.

4.5.2. Performance Comparison with Different Data

To confirm the superiority of our system’s performance when using cat body labels, we conduct an additional experiment to compare the performance when identifying cats using only the cat face, similar to other existing animal recognition methods. Within the complete image dataset, some images comprise only body without a face. For a fair performance comparison, we use 5115 images from 35 types after excluding those without faces; for class 13, no images include faces. Consistent with the previous experiment, the dataset is divided into an 8:2 ratio using stratified sampling for model training and evaluation. The training of the classification module is executed in the same manner as described in Section 4.3., and an SVM classifier with a linear kernel is used for the final accuracy measurement. Both models show a high performance of over 90%, and even with a higher degree of freedom such as the body, the performance difference is minimal at about 0.69%. The overall identification accuracy and confusion matrix for each dataset are presented in Table 3 and Figure 10 and Figure 11.

5. Discussion

In this study, we train and evaluate the proposed algorithm using a body dataset of 36 different cat species. During evaluation, we compute the identification accuracy for the test dataset with all 36 class embeddings trained in the classifier. During the training and evaluation phases, we observe a high identification accuracy, which is a promising indicator of the system’s potential effectiveness. In particular, the combination of EfficientNetV2-S with the SVM classifier yields outstanding results with an impressive accuracy of 94.53%.
Notably, even in scenarios where the data could be confusing, such as images that include not only the cat’s face but also its body, our architecture demonstrates only a slight performance decrease of approximately 0.69% compared with the face recognition model. This success underscores the ability of the model to handle diverse and complex visual inputs.

6. Conclusions

We focus on the integration of a neural-network-based feature extractor with existing ML-based classifiers to develop a cat monitoring system. Additionally, we propose a model retraining algorithm with the aim of enabling the DL-based model to adapt seamlessly to the dynamic appearance changes that occur during a cat’s life cycle and variations in the camera environment.
To validate the monitoring performance of the proposed system, we collect a pet cat body dataset through a monocular camera inside the litter box. Instead of the face commonly used for animal identification, our system uses the entire body for identification (where the face may not be included), making the scenario correspond to a more challenging task. We perform testing on the classification module and compare the performance of various ML-based classifiers. The combination of EfficientNetV2-S and the SVM classifier demonstrates a high identification accuracy of 94.53% on the entire cat body dataset, and the identification performance receiving a body dataset with face is only about 0.69% lower than when receiving a face dataset. This indicates that our cat body identification system has high applicability in monitoring the defecation and urination activities of pet cats in multi-cat households.
In future research, there are several avenues for further improvement. Exploring different stage combinations, optimizers, and advanced loss functions during training can lead to performance enhancement. In addition, the evaluation and development of various models for the detector part of the system will contribute to a more comprehensive and refined architecture. Meanwhile, a mathematical validation of the model and retraining method presented in this study is required. Especially, since a mathematical analysis of the recognition system has not been addressed, there is a need to ascertain the solution for the mathematical model of this research system and its stability [36,37]. Lastly, in addition to the vision-based identification system, it is an important consideration in the future to analyze the cat’s defecation and urination status more precisely and comprehensively by measuring the cat’s litter box activity time or integrating more diverse sensors, such as weight, humidity and pH of urine and feces.
The cat litter box monitoring system developed in this study contributes to managing the urinary health of individual cats in multi-cat households. Moreover, the identification architecture of this system is not limited to cat litter boxes and can be expanded to other situations, aiding in personalized monitoring and diagnosis for multi-cat households.

Author Contributions

Conceptualization, Y.C. and E.S.; methodology, Y.C.; software, Y.C.; validation, Y.C., E.S. and Y.J.; formal analysis, Y.C.; investigation, Y.C.; resources, S.Y. (Saetbyeol Yang) and S.P.; data curation, Y.J., T.K., S.P. and D.B.; writing–original draft preparation, Y.C.; writing–review and editing, E.S., Y.J. and S.Y. (Sunjin Yu); visualization, Y.C.; supervision, E.S. and S.Y. (Saetbyeol Yang); project administration, Y.C.; funding acquisition, E.S. and S.Y. (Saetbyeol Yang). All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Ministry of Science and ICT(MSIT, Korea), grant number RS_2023_00227552.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data sharing is not applicable to this article.

Acknowledgments

This work was partly supported by Institute of Information & Communications Technology Planning & Evaluation(IITP) grant funded by the Korea government(MSIT) (No.2022-0-00910, Development of automatic litter box system with artificial intelligence cat detection/recognition algorithm) and Institute of Information & Communications Technology Planning & Evalua-tion(IITP) grant funded by the Korea government(MSIT) (No.RS_2023_00227552, Development of artificial intelligence video background removal SaaS service using domestic semiconductor 64 TOPS).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Dubey, J.P.; Cerqueira-Cézar, C.K.; Murata, F.H.A.; Kwok, O.C.H.; Yang, Y.R.; Su, C. All about Toxoplasmosis in Cats: The Last Decade. Vet. Parasitol. 2020, 283, 109145. [Google Scholar] [CrossRef] [PubMed]
  2. Vojtkovská, V.; Voslářová, E.; Večerek, V. Methods of Assessment of the Welfare of Shelter Cats: A Review. Animals 2020, 10, 1527. [Google Scholar] [CrossRef]
  3. Tan, S.M.L.; Stellato, A.C.; Niel, L. Uncontrolled Outdoor Access for Cats: An Assessment of Risks and Benefits. Animals 2020, 10, 258. [Google Scholar] [CrossRef] [PubMed]
  4. Piyarungsri, K.; Tangtrongsup, S.; Thitaram, N.; Lekklar, P.; Kittinuntasilp, A. Prevalence and Risk Factors of Feline Lower Urinary Tract Disease in Chiang Mai, Thailand. Sci. Rep. 2020, 10, 196. [Google Scholar] [CrossRef] [PubMed]
  5. Majid, A.Y.; Nurmansyah, R.F.; Pratama, M.L.A.; Susanti, H.; Prihatiningrum, N. IoT-Based Cat Feeding and Monitoring System. In Proceedings of the 2023 8th International Conference on Instrumentation, Control, and Automation (ICA), Jakarta, Indonesia, 9 August 2023; IEEE: Piscataway, NJ, USA, 2023; pp. 160–165. [Google Scholar] [CrossRef]
  6. Eagan, B.H.; Eagan, B.; Protopopova, A. Behaviour Real-Time Spatial Tracking Identification (BeRSTID) Used for Cat Behaviour Monitoring in an Animal Shelter. Sci. Rep. 2022, 12, 17585. [Google Scholar] [CrossRef] [PubMed]
  7. Arhant, C.; Heizmann, V.; Schauberger, G.; Windschnurer, I. Risks and Benefits of Collar Use in Cats (Felis Catus); a Literature Review. J. Vet. Behav. 2022, 55–56, 35–47. [Google Scholar] [CrossRef]
  8. Lund, H.S.; Eggertsdóttir, A.V. Recurrent Episodes of Feline Lower Urinary Tract Disease with Different Causes: Possible Clinical Implications. J. Feline Med. Surg. 2019, 21, 590–594. [Google Scholar] [CrossRef] [PubMed]
  9. Bonagura, J.D.; Twedt, D.C.; Kirk, R.W. Kirk’s Current Veterinary Therapy XIV, 14th ed.; Elsevier Saunders: St. Louis, MI, USA, 2009. [Google Scholar]
  10. Hou, J.; He, Y.; Yang, H.; Connor, T.; Gao, J.; Wang, Y.; Zeng, Y.; Zhang, J.; Huang, J.; Zheng, B.; et al. Identification of Animal Individuals Using Deep Learning: A Case Study of Giant Panda. Biol. Conserv. 2020, 242, 108414. [Google Scholar] [CrossRef]
  11. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. In Proceedings of the 3rd International Conference on Learning Representations (ICLR 2015), San Diego, CA, USA, 7–9 May 2015; Computational and Biological Learning Society. pp. 1–14. [Google Scholar]
  12. Hitelman, A.; Edan, Y.; Godo, A.; Berenstein, R.; Lepar, J.; Halachmi, I. Biometric Identification of Sheep via a Machine-Vision System. Comput. Electron. Agric. 2022, 194, 106713. [Google Scholar] [CrossRef]
  13. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef] [PubMed]
  14. Deng, J.; Guo, J.; Xue, N.; Zafeiriou, S. ArcFace: Additive Angular Margin Loss for Deep Face Recognition. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 4685–4694. [Google Scholar] [CrossRef]
  15. Schofield, D.; Nagrani, A.; Zisserman, A.; Hayashi, M.; Matsuzawa, T.; Biro, D.; Carvalho, S. Chimpanzee Face Recognition from Videos in the Wild Using Deep Learning. Sci. Adv. 2019, 5, eaaw0736. [Google Scholar] [CrossRef] [PubMed]
  16. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.-Y.; Berg, A.C. SSD: Single Shot MultiBox Detector. In Computer Vision—ECCV 2016; Leibe, B., Matas, J., Sebe, N., Welling, M., Eds.; Springer International Publishing: Cham, Switzerland, 2016; pp. 21–37. [Google Scholar] [CrossRef]
  17. Chatfield, K.; Simonyan, K.; Vedaldi, A.; Zisserman, A. Return of the Devil in the Details: Delving Deep into Convolutional Nets. In Proceedings of the British Machine Vision Conference 2014, Nottingham, UK, 1–5 September 2014; British Machine Vision Association. pp. 1–12. [Google Scholar]
  18. Clapham, M.; Miller, E.; Nguyen, M.; Darimont, C.T. Automated Facial Recognition for Wildlife That Lack Unique Markings: A Deep Learning Approach for Brown Bears. Ecol. Evol. 2020, 10, 12883–12892. [Google Scholar] [CrossRef] [PubMed]
  19. Schroff, F.; Kalenichenko, D.; Philbin, J. FaceNet: A Unified Embedding for Face Recognition and Clustering. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 815–823. [Google Scholar] [CrossRef]
  20. King, D.E. Dlib-Ml: A Machine Learning Toolkit. J. Mach. Learn. Res. 2009, 10, 1755–1758. [Google Scholar]
  21. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 26 June–1 July 2016; pp. 770–778. [Google Scholar] [CrossRef]
  22. Bochkovskiy, A.; Wang, C.-Y.; Liao, H.-Y.M. YOLOv4: Optimal Speed and Accuracy of Object Detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
  23. Tan, M.; Le, Q. EfficientNetV2: Smaller Models and Faster Training. In Proceedings of the 38th International Conference on Machine Learning, Online, 18–24 July 2021; Volume 139, pp. 10096–10106. [Google Scholar]
  24. Iwashita, A.S.; Papa, J.P. An Overview on Concept Drift Learning. IEEE Access 2019, 7, 1532–1547. [Google Scholar] [CrossRef]
  25. Gama, J.; Medas, P.; Castillo, G.; Rodrigues, P. Learning with drift detection. In Proceedings of the Advances in Artificial Intelligence—SBIA 2004: 17th Brazilian Symposium on Artificial Intelligence, Sao Luis, Maranhao, Brazil, 29 September–1 October 2004; Proceedings 17. Springer: Berlin, Germany, 2004; pp. 286–295. [Google Scholar]
  26. Bifet, A.; Gavaldà, R. Learning from Time-Changing Data with Adaptive Windowing. In Proceedings of the 2007 SIAM International Conference on Data Mining, Philadelphia, PA, USA, 26 April 2007; Society for Industrial and Applied Mathematics. pp. 443–448. [Google Scholar] [CrossRef]
  27. Moallem, G.; Pathirage, D.D.; Reznick, J.; Gallagher, J.; Sari-Sarraf, H. An Explainable Deep Vision System for Animal Classification and Detection in Trail-Camera Images with Automatic Post-Deployment Retraining. Knowl. Based Syst. 2021, 216, 106815. [Google Scholar] [CrossRef]
  28. Redmon, J.; Farhadi, A. YOLOv3: An Incremental Improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  29. Wang, C.Y.; Mark Liao, H.Y.; Wu, Y.H.; Chen, P.Y.; Hsieh, J.W.; Yeh, I.H. CSPNet: A New Backbone that can Enhance Learning Capability of CNN. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Seattle, WA, USA, 14–19 June 2020; pp. 1571–1580. [Google Scholar] [CrossRef]
  30. He, K.; Zhang, X.; Ren, S.; Sun, J. Spatial Pyramid Pooling in Deep Convolutional Networks for Visual Recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 37, 1904–1916. [Google Scholar] [CrossRef]
  31. Liu, S.; Qi, L.; Qin, H.; Shi, J.; Jia, J. Path Aggregation Network for Instance Segmentation. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–22 June 2018; pp. 8759–8768. [Google Scholar] [CrossRef]
  32. Tan, M.; Le, Q. EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks. In Proceedings of the 36th International Conference on Machine Learning, Long Beach, CA, USA, 10–15 June 2019; Volume 97, pp. 6105–6114. [Google Scholar]
  33. Tan, M.; Chen, B.; Pang, R.; Vasudevan, V.; Sandler, M.; Howard, A.; Le, Q.V. MnasNet: Platform-Aware Neural Architecture Search for Mobile. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; IEEE: Piscataway, NJ, USA, 2019; Volume 2019, pp. 2815–2823. [Google Scholar] [CrossRef]
  34. Zoph, B.; Le, Q.V. Neural Architecture Search with Reinforcement Learning. In Proceedings of the 5th International Conference on Learning Representations, Toulon, France, 24–26 April 2017. ICLR 2017—Conference Track Proceedings. [Google Scholar]
  35. Zhang, X.; Yu, F.X.; Kumar, S.; Chang, S.-F. Learning Spread-Out Local Feature Descriptors. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 4605–4613. [Google Scholar] [CrossRef]
  36. Zhao, K. Local Exponential Stability of Several Almost Periodic Positive Solutions for a Classical Controlled GA-Predation Ecosystem Possessed Distributed Delays. Appl. Math. Comput. 2023, 437, 127540. [Google Scholar] [CrossRef]
  37. Zhao, K. Existence and Stability of a Nonlinear Distributed Delayed Periodic AG-Ecosystem with Competition on Time Scales. Axioms 2023, 12, 315. [Google Scholar] [CrossRef]
Figure 1. Overall architecture of the proposed cat identification model.
Figure 1. Overall architecture of the proposed cat identification model.
Sensors 23 08852 g001
Figure 2. YOLOv4 network structure for cat body detection.
Figure 2. YOLOv4 network structure for cat body detection.
Sensors 23 08852 g002
Figure 3. Modified EfficientNetV2-S network structure for the feature extractor in the classification module.
Figure 3. Modified EfficientNetV2-S network structure for the feature extractor in the classification module.
Sensors 23 08852 g003
Figure 4. Model retraining algorithm for the feature extractor of the classification module.
Figure 4. Model retraining algorithm for the feature extractor of the classification module.
Sensors 23 08852 g004
Figure 5. Color temperature diversification examples. (a) Image on warm white (2700 K); (b) image on natural white (4100 K); (c) image on cool white (6500 K); and (d) image on LED light (10,000 K).
Figure 5. Color temperature diversification examples. (a) Image on warm white (2700 K); (b) image on natural white (4100 K); (c) image on cool white (6500 K); and (d) image on LED light (10,000 K).
Sensors 23 08852 g005
Figure 6. Loss graphs of each training stage for the feature extractor. Each stage utilizes semi-hard, hard, and hardest triplet minings.
Figure 6. Loss graphs of each training stage for the feature extractor. Each stage utilizes semi-hard, hard, and hardest triplet minings.
Sensors 23 08852 g006
Figure 7. The confusion matrices for the SVM classifier with a linear kernel. Each left and right matrix shows the confusion matrix without/with normalization by the number of test set images in each class. The color of each cell in the matrix represents the number of images or normalized ratio, with darker colors corresponding to larger values.
Figure 7. The confusion matrices for the SVM classifier with a linear kernel. Each left and right matrix shows the confusion matrix without/with normalization by the number of test set images in each class. The color of each cell in the matrix represents the number of images or normalized ratio, with darker colors corresponding to larger values.
Sensors 23 08852 g007
Figure 8. The ROC and PR curves for the SVM classifier with a linear kernel. For each class, AUC (area under ROC curve) and AP (average precision) values are indicated.
Figure 8. The ROC and PR curves for the SVM classifier with a linear kernel. For each class, AUC (area under ROC curve) and AP (average precision) values are indicated.
Sensors 23 08852 g008
Figure 9. Examples of identification results for specific cat images of our dataset. Each row corresponds to TP, FN, FP, and TN from the top to the bottom. For each image, if the identification result is correct then the color of bounding box and information are displayed in blue; otherwise, they are displayed in red.
Figure 9. Examples of identification results for specific cat images of our dataset. Each row corresponds to TP, FN, FP, and TN from the top to the bottom. For each image, if the identification result is correct then the color of bounding box and information are displayed in blue; otherwise, they are displayed in red.
Sensors 23 08852 g009
Figure 10. The confusion matrix evaluated on the cat body dataset. Each left and right matrix shows the confusion matrix without/with normalization by the number of test set images in each class. The color of each cell in the matrix represents the number of images or normalized ratio, with darker colors corresponding to larger values.
Figure 10. The confusion matrix evaluated on the cat body dataset. Each left and right matrix shows the confusion matrix without/with normalization by the number of test set images in each class. The color of each cell in the matrix represents the number of images or normalized ratio, with darker colors corresponding to larger values.
Sensors 23 08852 g010
Figure 11. The confusion matrix evaluated on the cat face dataset. Each left and right matrix shows the confusion matrix without/with normalization by the number of test set images in each class. The color of each cell in the matrix represents the number of images or normalized ratio, with darker colors corresponding to larger values.
Figure 11. The confusion matrix evaluated on the cat face dataset. Each left and right matrix shows the confusion matrix without/with normalization by the number of test set images in each class. The color of each cell in the matrix represents the number of images or normalized ratio, with darker colors corresponding to larger values.
Sensors 23 08852 g011
Table 1. A summary of existing animal identification methods. (OD = object detector, CL = classifier).
Table 1. A summary of existing animal identification methods. (OD = object detector, CL = classifier).
MethodRegion of InterestArchitecture
Hou et al. [10]FaceCL: VGGNet [11]
Hitelman et al. [12]FaceOD: Faster R-CNN [13]
CL: CNN-based models
Schofield et al. [15]FaceOD: SSD [16]
CL: VGG-M [17]
Clapham et al. [18]FaceOD: CNN-based model
CL: ResNet-34 [21] + SVM
OursBodyOD: YOLOv4 [22]
CL: EfficientNetV2 [23] + ML-based models
Table 2. Identification accuracy evaluated on the test dataset with different classifiers.
Table 2. Identification accuracy evaluated on the test dataset with different classifiers.
ClassifierAccuracy
KNN (K = 3)94.24%
KNN (K = 5)94.41%
KNN (K = 7)94.19%
Random forest94.13%
SVM (RBF kernel)94.41%
SVM (Linear kernel)94.53%
Table 3. Identification accuracy evaluated on the cat body dataset and cat face dataset.
Table 3. Identification accuracy evaluated on the cat body dataset and cat face dataset.
DatasetAccuracy
Cat body with face93.08%
Cat face only93.77%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Cho, Y.; Song, E.; Ji, Y.; Yang, S.; Kim, T.; Park, S.; Baek, D.; Yu, S. Multi-Cat Monitoring System Based on Concept Drift Adaptive Machine Learning Architecture. Sensors 2023, 23, 8852. https://0-doi-org.brum.beds.ac.uk/10.3390/s23218852

AMA Style

Cho Y, Song E, Ji Y, Yang S, Kim T, Park S, Baek D, Yu S. Multi-Cat Monitoring System Based on Concept Drift Adaptive Machine Learning Architecture. Sensors. 2023; 23(21):8852. https://0-doi-org.brum.beds.ac.uk/10.3390/s23218852

Chicago/Turabian Style

Cho, Yonggi, Eungyeol Song, Yeongju Ji, Saetbyeol Yang, Taehyun Kim, Susang Park, Doosan Baek, and Sunjin Yu. 2023. "Multi-Cat Monitoring System Based on Concept Drift Adaptive Machine Learning Architecture" Sensors 23, no. 21: 8852. https://0-doi-org.brum.beds.ac.uk/10.3390/s23218852

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop