Next Article in Journal
Development of Digital Competence for Research
Previous Article in Journal
AI Makes Crypto Evolve
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Spherical Tree-Structured SOM and Its Application to Hierarchical Clustering

1
Graduate School of Science and Engineering, Saga University, Saga 840-8502, Japan
2
Faculty of Science and Engineering, Saga University, Saga 840-8502, Japan
*
Author to whom correspondence should be addressed.
Appl. Syst. Innov. 2022, 5(4), 76; https://0-doi-org.brum.beds.ac.uk/10.3390/asi5040076
Submission received: 28 June 2022 / Revised: 25 July 2022 / Accepted: 27 July 2022 / Published: 30 July 2022

Abstract

:
When analyzing high-dimensional data with many elements, a data visualization that maps the data onto a low-dimensional space is often performed. By visualizing the data, humans can intuitively understand the structure of the data in the high-dimensional space. The self-organizing map (SOM) is one such data visualization method. We propose a spherical tree-structured SOM (S-TS-SOM), which speeds up the search for winner nodes and eliminates the unevenness of learning due to the position of the winner nodes by placing the nodes on a sphere and applying the tree search method. In this paper, we confirm that the S-TS-SOM can achieve the same results as a normal spherical SOM while reducing the learning time. In addition, we confirm the granularity of clustering on the tree structure of the S-TS-SOM.

1. Introduction

Because of the recent developments in data science, various types of data are now analyzed in many fields. When analyzing data, if the structure of the data is unknown in advance, it is necessary to first understand the structure of the data and then select the analysis methods according to the task. If the data to be analyzed consist of few elements, it is easy to understand the structure of the data. However, in the case of high-dimensional data with many elements, it is difficult for humans to grasp the structure of the data. Therefore, data visualization that maps high-dimensional data onto low-dimensional space with two or three dimensions is often performed. By visualizing these data, humans can intuitively understand the structure of the data in the high-dimensional space. Various methods have been proposed for data visualization, and a self-organizing map (SOM) is one such data visualization method [1,2].
The SOM maps high-dimensional data onto a low-dimensional map and expresses the similarity between data points using distance on the map. Similar data are placed close to each other on the map, and dissimilar data are placed far from each other on the map. A brief description of the SOM algorithm is as follows. First, the SOM arranges nodes that have reference vectors initialized with random values in a two-dimensional grid. In the learning phase, a SOM determines the winner node, which is the reference vector closest to the input vector, and updates the winner node’s reference vector and its neighbors’ reference vectors so that they are closer to the input vector. Learning in the SOM is performed by repeatedly determining the winner node and updating the reference vectors of the winner node and its neighbors.
In recent years, large amounts of data have accumulated because of the development of various web services, and hence the number of data to be analyzed has increased. Generally, when visualizing a large number of data using a SOM, the number of nodes in the map is set to a large number. However, as the number of nodes increases, the SOM needs more time to search for the winner nodes. In addition, because the SOM usually arranges nodes in a two-dimensional grid, there are edges to the map. Therefore, according to the position of winner nodes, e.g., the center or edge of the map, the update area sizes are different. This means that learning is not uniform but depends on the positions of the winner nodes. To solve both these problems, we propose a spherical tree-structured SOM (S-TS-SOM). The S-TS-SOM arranges nodes on a spherical surface and applies a tree search method to speed up the search for the winner nodes while eliminating the edges of the map. We compare the effectiveness of the S-TS-SOM with a conventional spherical SOM (S-SOM) using the benchmark data MNIST [3].
The S-TS-SOM builds a tree structure and determines the winner nodes using a tree search method. In addition to speeding up the search, the advantage of the tree structure is that we can obtain different granularities of clustering in the tree structure. In this paper, we examine whether the granularity of clustering can be found in the tree structure of the S-TS-SOM. The main contributions of this paper can be summarized as follows.
(1)
We propose the S-TS-SOM, which applies a tree search method to the S-SOM, to speed up the search for winner nodes and eliminate the edges of the map.
(2)
We examine the effectiveness of the S-TS-SOM by comparing it with the S-SOM using a benchmark dataset.
(3)
We examine whether the granularity of clustering can be determined using the tree structure of the S-TS-SOM.
The remainder of this paper is organized as follows. Section 2.1 introduces work related to this study. Section 2.2 presents the proposed method, the S-TS-SOM. In Section 3.1 and Section 3.2, we evaluate the effectiveness of the S-TS-SOM by comparing it with the S-SOM using the benchmark dataset MNIST. Section 3.3 examines whether the granularity of clustering can be found using the tree structure of the S-TS-SOM. Section 4 presents the discussion and future work.

2. Materials and Methods

2.1. Related Work

First, we describe a SOM and other methods used for data visualization. A SOM is a type of artificial neural network proposed by Kohonen. It maps high-dimensional data onto a low-dimensional map and expresses the similarity of the data using distance. On the map, similar data are placed close to each other, and dissimilar data are placed far from each other. This characteristic enables the SOM to be used for data visualization. By visualizing the data, humans can intuitively grasp the relationships among the data in a high-dimensional space. WEBSOM is an example of an application that is used as a tool for data visualization. WEBSOM makes it easy to search for documents that are similar to a query document by arranging many documents onto a two-dimensional map according to similarity [4].
We describe the SOM algorithm as follows: Figure 1 shows the structure of the SOM. It has nodes m i i = 1 , I with reference vector mi. First, each reference vector is initialized with a random value. In the SOM, the node with the reference vector that is closest to the input vector x j j = 1 , J that was selected from the input set is determined to be the winner node m c . After the winner node is determined, the reference vectors of the winner node and its neighborhood nodes are updated so that they are closer to the value of the input vector using the following formula:
m i t + 1 = m i t + h c i x j t m i t
h c i t = α t exp d c i 2 2 σ 2 t ,
where h c i is a neighborhood function determined by the positions of winner node m c and node m i , d c i is the distance between the nodes m c and m i on the map, α t is the learning rate, and σ t is a function that determines the size of the update area; α t and σ t are set to decrease as the number of updates increases. By repeatedly determining the winner node and updating the reference vectors, the SOM is trained. The algorithm described above is a common technique called online learning. By contrast, batch learning that does not depend on the order of input has also been proposed [5]. In batch learning, after the winner node of all inputs is determined, the map is updated using the following formula:
m i t + 1 = j = 1 J h c j , i x j j = 1 J h c j , i
h c j , i t = exp d c j , i 2 2 σ 2 t .
In addition to the SOM, principal component analysis (PCA) and t-distributed stochastic neighbor embedding (t-SNE) are common dimensional reduction or compression methods [6,7]. PCA searches for the axes with the large variances in the data and removes axes with small variances to reduce the dimensions. Note that PCA is based on the assumption that the magnitude of the variance represents the characteristics of the data. t-SNE regards the distance between data points as a probability and compresses the dimensions so that the distances between data points in the high- and low-dimensional spaces are the same. Here, we introduced PCA and t-SNE as typical methods. However, many other methods continue to be proposed [8,9]. Similar to the SOM, PCA and t-SNE are mainly used for data visualization. The major difference between the SOM and other dimensional reduction or compression methods such as PCA and t-SNE is that the SOM uses nodes for learning. Therefore, using a SOM, we obtain not only the coordinates of the input data in the low-dimensional space but also the codebook vectors from the nodes.
As mentioned in the Introduction, if the number of nodes in a SOM increases, the number of search nodes needed to determine the winner nodes increase, and the calculation time increases. Therefore, a tree-structured SOM (TS-SOM), which uses the tree search method in the SOM, was proposed [10]. The structure of a TS-SOM is shown on the left-hand side of Figure 2. A TS-SOM is a SOM composed of multiple layers and has a structure in which the number of nodes increases from the upper layers to the lower layers. The nodes of the upper layer have links to the nodes of lower layer, and the winner nodes determined in the upper layer assist the search for the lower layer’s winner nodes to speed up the search for the winner nodes. The right-hand side of Figure 2 shows how to determine the winner node in a TS-SOM. The lower winner node is determined from the child node group that has a link from the winner node determined in the upper layer. As in a SOM, the nodes of each layer in a TS-SOM are arranged in a two-dimensional grid pattern. Therefore, a TS-SOM also has map edges.
Next, we describe the shape of the SOM. Normally, the SOM arranges nodes on a two-dimensional grid pattern for learning. Therefore, a SOM is a map with edges. The presence of the edges of the map causes a difference in the size of the update area of a winner node at the center of the map and a winner node at the edge of the map. Figure 3 shows the update area according to the position of the winner node. In addition, because of the characteristics of the SOM, if the map has an edge, the SOM tends to collect data at the edges of the map. Therefore, to eliminate differences in the size of the update area, the torus SOM, which has a map without edges, and the spherical SOM (S-SOM), which arranges nodes on a spherical surface, were proposed [11,12]. Figure 4 shows the map shapes of the torus SOM and the S-SOM. In the torus SOM, by connecting the edges of the map as shown in the Figure 4, the edges of the map are eliminated. The torus SOM can learn uniformly regardless of the position of the winner nodes. However, when the map is visualized, it is necessary to consider the connections between the edges of the map, which make it difficult for humans to intuitively understand the result. In this study, we chose to arrange nodes on a sphere for intuitive visibility. Intuitive visibility is achieved by drawing the spherical map in three dimensions and rotating the spherical map for viewing.
When clustering data, how finely the data should be clustered strongly depends on the target task. For example, if the data are uniformly distributed, it is difficult to automatically determine the optimal number of clusters. Furthermore, for data with small categories within large categories, the hyperparameters need to be adjusted depending on the size of the target category. Generally, for such data, a dendrogram is created by performing hierarchical clustering, and the number of clusters is arbitrarily determined according to the task. In addition to hierarchical clustering, a SOM with multiple layers can be used to arbitrarily determine the number of clusters. The TS-SOM, which is the basis of the method proposed in this paper, is composed of multiple layers, and has a structure in which the number of nodes increases from the upper layers to the lower layers. Therefore, the data are classified more finely toward the lower layers, and clustering results with different granularities can be obtained in each layer. In addition to the TS-SOM, the hierarchal feature map (HFM) has also been proposed [13,14]. Moreover, the GH-SOM, which enables the HFM to learn adaptively, and a method that makes the TS-SOM learn adaptively, have been proposed [15,16,17].

2.2. S-TS-SOM

In this section, we describe the S-TS-SOM. The basic structure is the same as that of a TS-SOM. However, in an S-TS-SOM, nodes are evenly arranged on a spherical surface to eliminate the edges of the map. In this paper, we use polyhedrons inscribed in a sphere to arrange the nodes on the spherical surface. The nodes are arranged as follows. First, we divide the icosahedron and stack it, as shown in Figure 5. Next, nodes are placed on each surface of the polyhedrons, and links are added from the nodes of the inner layer to the nodes of the outer layer. Using this process, the tree structure is constructed for each face of the polyhedron, as shown in Figure 6. In the method proposed in this paper, except for the outermost layer’s nodes, the tree structure is created such that each node has four links to the outer layer’s nodes. The direction of layering is not important because it does not change the underlying concept. However, for the sake of explanation, we create the structure such that the outer layer has more nodes.
An S-TS-SOM has multiple layers, each layer has a node m i l i l = 1 , I l , and node m i l has a reference vector m i l . Here, l indicates the level of the layer, and l = 0 , 1 , 2 , L in order from the inner layer to the outer layer. In addition, i l indicates the node number of the l th layer. We used an icosahedron for the innermost polyhedron, and because we divide the icosahedron to increase the number of layers, the number of nodes in each layer is expressed by the following formula.
I l = 20 × 4 l
The learning algorithm of the S-TS-SOM is as follows. First, we initialize the reference vectors of the 0th layer nodes with random values. The 0th layer is trained using the SOM algorithm. In a SOM, the node with the reference vector that is closest to the input vector that was selected from the input set is determined to be the winner node. After determining the winner node, the reference vectors of the winner node and its neighborhood nodes are updated so that they are closer to the value of the input vector. By repeatedly determining the winner node and updating the reference vectors, the SOM is trained. In this paper, Euclidean distance is used to measure the similarity between the input and reference vectors. For layers other than the 0th layer, the winner node is determined using the winner node of the inner layer. If the winner node in the l th layer of x j j = 1 , J selected from the input vector set is m c l ,   j , the winner node in the l + 1 th layer is determined from the child node set c l ,   j , which has links from node m c l ,   j . The update method of each layer is the same as that used in a SOM. Update methods include online learning and batch learning methods. We adopt batch learning because it is not affected by the order of the input data. After determining the winner nodes of all the input vectors in the l th layer, the reference vectors of the l th layer are updated by the following formula.
m i l t + 1 = j = 1 J h c l j , i l x j j = 1 J h c l j , i l
h c l j , i l t = exp θ c l j , i l 2 2 σ 2 t ,
Here, h c l j , i l is a neighborhood function obtained from the positions of winner node m c l ,   j and node m i l . In addition, θ c l j , i l is the angle between winner node m c l ,   j and node m i l , as seen from the center of the sphere; σ t is a function that determines the size of the update area; σ t is set to decrease as the number of updates increases. When the learning of the l th layer is completed, the nodes of the l + 1 th layer are initialized by the following formula, and the l + 1 th layer is trained:
m i l + 1 = i l = 1 I l h i l , i l + 1   m i l i l = 1 I l h i l , i l + 1
h i l ,   i l + 1 = exp θ i l , i l + 1 2 2 σ l + 1 2
Here, σ l + 1 is the parameter for initializing the reference vectors of the l + 1 th layer.
When the number of updates of each layer is defined as T l , the number of nodes searched during learning for one input vector is expressed by the following formula.
20 T 0 + 1 + l = 1 L 4 ( T l + 1 )
We use this formula in Section 3 to compare the performance of the S-TS-SOM and S-SOM.
The learning process of the S-TS-SOM is described below.
(1)
Train the 0th layer using the SOM algorithm.
(2)
Add a competitive layer.
(3)
Search for the winner nodes in the added layer.
(4)
Update the reference vectors of the added layer using the neighborhood function.
(5)
Repeat steps 3 to 4 a certain number of times.
(6)
If the number of layers is the same as a pre-determined number, the learning is finished. Otherwise, the process returns to step 2.

3. Results

3.1. Visualization Experiment

First, we evaluate the S-TS-SOM as a visualization tool. Here, we use the MNIST benchmark dataset. MNIST is a dataset of images consisting of 10 categories, with 60,000 training data, and 10,000 test data. Handwritten numbers from “0” to “9” are drawn on each image. Each image in MNIST is 28 × 28 pixels. Similar to SOM, S-TS-SOM needs to convert the data to a vector format. Therefore, 28 × 28 data are inputted by transforming them into a vector of 784 elements.
For reasons of visualization, we used a total of 1000 images (100 extracted from each category) as the training data. We set the number of layers of the S-TS-SOM to L = 4 . Parameter σ l + 1 was set to decrease as the inner layers become outer layers. ( σ 1 , σ 2 , σ 3 , σ 4 = 90 ° , 45 ° , 22.5 ° , 11.25 ° ). In addition, the σ t of each layer was set to decay linearly with the number of updates, with σ l + 1 as the initial value. The σ t of the 0th layer was set with 180° as the initial value. We set the number of updates for each layer to 100. Figure 7 shows the result. For comparison, we show the result of the S-SOM in Figure 8. We placed the nodes of the S-SOM on the faces of the polyhedron, set the number of the S-SOM updates to 100, and updated the map by batch learning. The σ t of the S-SOM was set with 180° as the initial value. For the sake of clarity, we added colors to the MNIST images according to each category. The results show that similar images are gathered close to each other on the maps generated by both the S-TS-SOM and S-SOM. These results indicate that the S-TS-SOM is working properly as a visualization tool.

3.2. Quantitative Evaluation of Clustering

Here, we quantitatively compare the clustering performance of the S-TS-SOM and S-SOM. We used the purity and normalized mutual information (NMI) clustering indicators to evaluate the performance [18]. These two indicators take a value from 0 to 1, and values close to 1 indicate a high performance. Purity is calculated by counting the most dominant labels in each cluster and dividing by the number of data. Purity is simple to calculate. However, purity can easily be increased to 1 by increasing the number of clusters. By contrast, the calculation of NMI is complicated. However, it is calculated considering the number of clusters. NMI is close to 1 when the labels for each cluster are not mixed, and the number of clusters is close to the number of original categories. To calculate the purity and NMI clustering indicators, the true category labels are required. We used a total of 70,000 MNIST images as the experimental data. Therefore, in this experiment, the purity and NMI will be high if the data are correctly classified.
We set the number of outermost nodes to 80, 320, 1280, and 5120 for both the S-TS-SOM and S-SOM. We calculated the purity and NMI by treating the data placed on the same winner node as belonging to one cluster. We set the number of S-TS-SOM updates to 100 for each layer and the number of S-SOM updates to 100. We set σ l + 1 and σ t in the same way as in Section 3.1. The 0th layer of the S-TS-SOM and S-SOM were initialized with random values. Therefore, we present the average of the results of the five trials in Table 1. This table reveals that the S-TS-SOM and S-SOM have almost the same performance with respect to purity and NMI.
In addition to the calculation of purity and NMI, we compared the learning time of the S-TS-SOM and S-SOM to examine whether the S-TS-SOM could be trained more quickly than the S-SOM. We used Python for model creation and calculation and used a PC with an Intel(R) Core (TM) i7-10870 2.20 GHz/5.0 GHz and 64 GB DDR4 memory. The “Time” column in Table 1 lists the learning time of each model. Because the number of updates for each layer of the S-TS-SOM was set to 100, the number of nodes searched during learning for one input vector is (20 + 4L) × 101 from Formula (10). By contrast, in the S-SOM, the number of nodes searched for one input vector is 20 × 4L × 101.
In addition to the difference in the number of searched nodes, the S-TS-SOM initializes and updates the reference vectors in multiple layers. Moreover, because of the differences in the algorithms, the number of searched nodes does not directly reflect the difference in the learning times of the S-TS-SOM and S-SOM. However, the learning time of the S-TS-SOM was shorter than the learning time of the S-SOM. Moreover, as the number of nodes increased, the difference between the S-TS-SOM and S-SOM learning times increased.
From the experiments, we confirmed that an S-TS-SOM has almost the same clustering performance as a S-SOM but a shorter learning time.

3.3. Data Clustering Utilizing the Tree Structure of the S-TS-SOM

The S-TS-SOM proposed in this paper has a structure in which the number of nodes increases from the inner to outer layers. Therefore, we believe that by checking the results for each layer, we can obtain results with different data classification granularity. We consider that the layers closer to the inner layer have a smaller number of nodes, so the data are more roughly classified. Moreover, the layers closer to the outer layer have a larger number of nodes, so the data are more finely classified. In this section, we check the results of each layer of the S-TS-SOM to determine if the adjustment of the number of clusters is successful.
We used the Zoo dataset of UCI as the experimental data [19]. The Zoo dataset is composed of 101 species of animals, where each animal has 17 attributes. We used 16 attributes as the input and excluded the “type” attribute. Therefore, each animal was treated as a vector with 16 elements. Because the “legs” attribute is not a Boolean value, we used the value divided by the maximum value in the “legs” attribute as the input. The “type” attribute classifies each animal into an animal type using a number from one to seven. We set the number of layers of the S-TS-SOM to L = 2 and the number of updates for each layer to 100. Figure 9 shows the results. In Section 3.1, we presented only the outermost layer of the S-TS-SOM. However, the other layers are also shown in Figure 9, and the labels are color-coded according to the “type” attribute. In Figure 9, the top row shows the front of the map, and the bottom row shows the back of the map. We plotted the data placed on the same winner node in the same direction as seen from the center of the sphere. If only one data point was placed at a node in a layer other than the outermost layer, the branch growth was stopped for the sake of visibility.
In the results of Figure 9, in the 0th layer, the clusters are roughly formed according to each “type” attribute. As the number of layers increases, we can see that the clusters branch into smaller clusters. Note that some data have stopped branching because the input values are the same. Table A1 shows the groups of data with the same input value. From this experiment, we confirmed that the S-TS-SOM can extract the granularity of clustering on the tree structure by checking each layer after the data were learned.

4. Conclusions

In this paper, we proposed the S-TS-SOM, which uses a tree search method on a spherical surface to eliminate the edges of the map in the SOM and speed up the search for winner nodes. In addition, we confirmed the effectiveness of the S-TS-SOM by comparing the visualization performance and clustering performance with the S-SOM using the MNIST dataset. As a result, the S-TS-SOM placed similar images close to each other on the map equally as well as the S-SOM. Additionally, when we checked purity and NMI while changing the number of nodes to measure clustering performance, the difference between the values of the S-TS-SOM and S-SOM was within 0.1. When we measured the learning time, the ratio of the learning time of the S-TS-SOM to that of the S-SOM became shorter at 0.915, 0.435, 0.215, and 0.121 as the number of nodes in the map increased. The results show that the S-TS-SOM obtained the same performance as the S-SOM while speeding up the search for winner nodes.
We furthermore confirmed that the granularity of clustering could be found on each layer of the S-TS-SOM that learned the Zoo dataset. Hence, the S-TS-SOM is able to organize the map into a tree structure that reflects the granularity of the clustering.
As a limitation, the S-TS-SOM has more hyperparameters than the S-SOM because the former initializes and updates the map of each layer. In this paper, we did not investigate the influence of the hyperparameters of the S-TS-SOM. In the future, we will investigate the influence of the hyperparameters and compare the proposed method with other data visualization and clustering methods. In addition, we will develop a data search system that makes use of the tree structure of the S-TS-SOM.

Author Contributions

Conceptualization, K.Y. and H.D.; methodology, K.Y. and H.D; software, K.Y.; validation, K.Y.; formal analysis, K.Y.; investigation, K.Y.; writing—original draft preparation, K.Y.; writing—review and editing, K.Y. and H.D.; visualization, K.Y.; supervision, H.D.; project administration, H.D. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

Not applicable.

Acknowledgments

We thank Edanz (https://jp.edanz.com/ac, accessed on 26 July 2022) for editing a draft of this manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Table A1. Animals with the same input values.
Table A1. Animals with the same input values.
aardvark, bear
antelope, buffalo, deer, elephant, giraffe, oryx
bass, catfish, chub, herring, piranha
boar, cheetah, leopard, lion, lynx, mongoose, polecat, puma, raccoon, wolf
calf, goat, pony, reindeer
chicken, dove, parakeet
crayfish, lobster
crow, hawk
dogfish, pike, tuna
dolphin, porpoise
flea, termite
fruitbat, vampire
gull, skimmer, skua
haddock, seahorse, sole
hare, vole
housefly, moth
lark, pheasant, sparrow, wren
mole, opossum
slug, worm

References

  1. Kohonen, T. The self-organizing map. Proc. IEEE 1990, 9, 78. [Google Scholar] [CrossRef]
  2. Cottrell, M.; Olteanu, M.; Rossi, M.; Villa-Vialaneix, N. Self-Organizing Maps, theory and applications. Rev. Investig. Oper. 2018, 39, 1–22. [Google Scholar]
  3. MNIST Database. Available online: http://yann.lecun.com/exdb/mnist/ (accessed on 26 June 2022).
  4. Honkela, T.; Lagus, K.; Kohonen, T. WEBSOM D Self-organizing maps of document collections. Neurocomputing 1998, 21, 101–107. [Google Scholar]
  5. Matsushita, H.; Nishio, Y. Batch-learning self-organizing map with weighted connections avoiding false-neighbor effects. In Proceedings of the 2010 International Joint Conference on Neural Networks, Barcelona, Spain, 18–23 July 2010. [Google Scholar]
  6. Jolliffe, I.T.; Cadima, J. Principal component analysis: A review and recent developments. Philos. Trans. R. Soc. A Math. Phys. Eng. Sci. 2016, 374, 202. [Google Scholar] [CrossRef] [PubMed]
  7. Van der Maaten, L.; Hinton, G. Visualizing data using t-SNE. J. Mach. Learn. Res. 2008, 9, 2579–2605. [Google Scholar]
  8. Espadoto, M.; Martins, R.M.; Kerren, A.; Hirata, N.S.; Telea, A.C. Toward a quantitative survey of dimension reduction techniques. IEEE Trans. Vis. Comput. Graph. 2019, 27, 2153–2173. [Google Scholar] [CrossRef] [PubMed]
  9. Anowar, F.; Samira, S.; Selim, B. Conceptual and empirical comparison of dimensionality reduction algorithms (PCA, KPCA, LDA, MDS, SVD, LLE, ISOMAP, LE, ICA, t-SNE). Comput. Sci. Rev. 2021, 40, 100378. [Google Scholar] [CrossRef]
  10. Oja, E.; Kaski, S. Kohonen Maps; Elsevier: Amsterdam, The Netherlands, 1999; pp. 121–130. [Google Scholar]
  11. Masahiro, I.; Tsutomu, M.; Hiroshi, M. The characteristics of the torus Self Organizing Map. In Proceedings of the of 6th International Conference on Soft Computing, San Jose, CA, USA, 11–14 October 2000. [Google Scholar]
  12. Wu, Y.; Takatsuka, M. Spherical self-organizing map using efficient indexed geodesic data structure. Neural Netw. 2006, 19, 900–906. [Google Scholar] [CrossRef] [PubMed]
  13. Astudillo, C.A.; Oommen, B.J. Topology-oriented self-organizing maps: A survey. Pattern Anal. Appl. 2014, 17, 9. [Google Scholar] [CrossRef]
  14. Merkl, D. Exploration of text collections with hierarchical feature maps. In Proceedings of the 20th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, Philadelphia, PA, USA, 37–31 July 1997. [Google Scholar]
  15. Rauber, A.; Merkl, D.; Dittenbach, M. The growing hierarchical self-organizing map: Exploratory analysis of high-dimensional data. IEEE Trans. Neural Netw. 2002, 13, 1331–1341. [Google Scholar] [CrossRef] [PubMed]
  16. Qu, X.; Yang, L.; Guo, K.; Ma, L.; Sun, M.; Ke, M.; Li, M. A survey on the development of self-organizing maps for unsupervised intrusion detection. Mob. Netw. Appl. 2021, 26, 808–829. [Google Scholar] [CrossRef]
  17. Yamaguchi, T.; Ichimura, T.; Mackin, K.J. Adaptive learning algorithm in tree-structured self-organizing feature map. In Proceedings of the Joint 5th International Conference on Soft Computing and Intelligent Systems and 11th International Symposium on Advanced Intelligent, Okayama, Japan, 8–12 December 2010. [Google Scholar]
  18. Evaluation of Clustering. Available online: https://nlp.stanford.edu/IR-book/html/htmledition/evaluation-of-clustering-1.html (accessed on 26 June 2022).
  19. Machine Learning Repository. Available online: https://archive.ics.uci.edu/ml/datasets/zoo (accessed on 26 June 2022).
Figure 1. The structure of the SOM.
Figure 1. The structure of the SOM.
Asi 05 00076 g001
Figure 2. Structure of the TS-SOM and how to determine the winner node.
Figure 2. Structure of the TS-SOM and how to determine the winner node.
Asi 05 00076 g002
Figure 3. Difference in the update area depending on the positions of the winner nodes.
Figure 3. Difference in the update area depending on the positions of the winner nodes.
Asi 05 00076 g003
Figure 4. Map shapes the extended SOMs.
Figure 4. Map shapes the extended SOMs.
Asi 05 00076 g004
Figure 5. Layers of the S-TS-SOM.
Figure 5. Layers of the S-TS-SOM.
Asi 05 00076 g005
Figure 6. Structure of the S-TS-SOM.
Figure 6. Structure of the S-TS-SOM.
Asi 05 00076 g006
Figure 7. Result of the S-ST-SOM.
Figure 7. Result of the S-ST-SOM.
Asi 05 00076 g007
Figure 8. Result of the S-SOM.
Figure 8. Result of the S-SOM.
Asi 05 00076 g008
Figure 9. Result of data clustering utilizing the S-TS-SOM.
Figure 9. Result of data clustering utilizing the S-TS-SOM.
Asi 05 00076 g009
Table 1. Performance of the S-TS-SOM and S-SOM.
Table 1. Performance of the S-TS-SOM and S-SOM.
(1) S-T S-SOM(2) S-SOM(1)/(2)
Number of Outermost NodesNumber of Search NodesPurityNMITime (s)Number of Search NodesPurityNMITime (s)PurityNMITime (s)
8024 × 1010.7610.481215.1380 × 1010.7870.508235.070.9670.9480.915
32028 × 1010.8220.443409.06320 × 1010.8230.449941.170.9980.9860.435
128032 × 1010.8640.409789.741280 × 1010.8440.4013678.231.0241.0200.215
512036 × 1010.8950.3812011.44 5120 × 1010.8580.36516,560.321.0431.0460.121
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Yoshioka, K.; Dozono, H. Spherical Tree-Structured SOM and Its Application to Hierarchical Clustering. Appl. Syst. Innov. 2022, 5, 76. https://0-doi-org.brum.beds.ac.uk/10.3390/asi5040076

AMA Style

Yoshioka K, Dozono H. Spherical Tree-Structured SOM and Its Application to Hierarchical Clustering. Applied System Innovation. 2022; 5(4):76. https://0-doi-org.brum.beds.ac.uk/10.3390/asi5040076

Chicago/Turabian Style

Yoshioka, Koki, and Hiroshi Dozono. 2022. "Spherical Tree-Structured SOM and Its Application to Hierarchical Clustering" Applied System Innovation 5, no. 4: 76. https://0-doi-org.brum.beds.ac.uk/10.3390/asi5040076

Article Metrics

Back to TopTop