Skip to main content

ORIGINAL RESEARCH article

Front. Genet., 01 July 2022
Sec. RNA
This article is part of the Research Topic Machine Learning-Based Methods for RNA Data Analysis - Volume III View all 11 articles

D3K: The Dissimilarity-Density-Dynamic Radius K-means Clustering Algorithm for scRNA-Seq Data

Guoyun LiuGuoyun Liu1Manzhi Li,
Manzhi Li1,2*Hongtao WangHongtao Wang1Shijun LinShijun Lin1Junlin XuJunlin Xu3Ruixi LiRuixi Li4Min TangMin Tang5Chun LiChun Li1
  • 1School of Mathematics and Statistics, Hainan Normal University, Haikou, China
  • 2Key Laboratory of Data Science and Smart Education, Ministry of Education, Hainan Normal University, Haikou, China
  • 3College of Information Science and Engineering, Hunan University, Changsha, China
  • 4Geneis Beijing Co., Ltd., Beijing, China
  • 5School of Life Sciences, Jiangsu University, Zhenjiang, China

A single-cell sequencing data set has always been a challenge for clustering because of its high dimension and multi-noise points. The traditional K-means algorithm is not suitable for this type of data. Therefore, this study proposes a Dissimilarity-Density-Dynamic Radius-K-means clustering algorithm. The algorithm adds the dynamic radius parameter to the calculation. It flexibly adjusts the active radius according to the data characteristics, which can eliminate the influence of noise points and optimize the clustering results. At the same time, the algorithm calculates the weight through the dissimilarity density of the data set, the average contrast of candidate clusters, and the dissimilarity of candidate clusters. It obtains a set of high-quality initial center points, which solves the randomness of the K-means algorithm in selecting the center points. Finally, compared with similar algorithms, this algorithm shows a better clustering effect on single-cell data. Each clustering index is higher than other single-cell clustering algorithms, which overcomes the shortcomings of the traditional K-means algorithm.

1 Introduction

Since the start of genome Project, genome sequencing has been carried out rapidly, and a large amount of genome data has been mined. In order to obtain the information needed by people, bioinformatics emerges as The Times require (Li and Wong, 2019; Liu et al., 2021). It is an interdisciplinary subject composed of life science and computer science, which can dig out the biological significance contained in the chaotic biological data (Sun et al., 2022). Transcriptome is an important research field in bioinformatics, which can study gene function and gene structure from an overall level, and reveal specific biological processes and molecular mechanisms in the process of disease occurrence (Qi et al., 2021; Tang et al., 2020). In order to study the transcriptome, it must be sequenced first, but traditional sequencing techniques ignore the critical differences of individual cells, which will mask the heterogeneous expression between cells and make it difficult to detect subtle potential changes (Huang et al., 2017; Liu et al., 2020). To solve this problem, the single cell RNA sequencing (scrNA-SEQ) technology was developed (Qiao et al., 2017).

scRNA-seq is a powerful method for analyzing gene expression patterns and quickly determining the correct gene expression patterns of thousands of single cells (Potter, 2018). By analyzing scRNA-seq data, we can identify rare cell populations, find subgroup types with different functions, and reveal the regulatory relationship between genes. scRNA-seq can not only show the complexity of single-cell horizontal structure but also improve biomedical research and solve various problems in biology (Yang et al., 2019).

Although the research prospect of scRNA-seq is comprehensive, it also brings new problems and challenges (Kiselev et al., 2019). The scRNA-seq data are high-dimensional and noisy (Xu et al., 2020). Therefore, many clustering methods have been proposed to deal with high-dimensional data structures and noise distribution (Jiang et al., 2018; Zhang et al., 2021; Zhuang et al., 2021). Most of the existing scRNA-seq clustering methods can be divided into unsupervised or semi-supervised clustering (Chen et al., 2016). Zhang et al., (2018) et al. proposed an improved K-means algorithm based on density canopy to find the appropriate center point by calculating the density of the sample data set; Li et al. proposed a new improved algorithm based on T-SNE and density canopy algorithm, called density-canopy-K-means (Li et al., 2019). Compared with similar methods, this clustering algorithm shows stable and efficient clustering performance on single-cell data, thus overcoming the shortcomings of traditional methods; Dong and Zhu, (2020) et al. calculated the dissimilarity parameter between each model by calculating the dissimilarity function between samples and selected the maximum dissimilarity parameter value as the initial clustering center point; Zhu (Zhuang et al., 2021) et al. proposed a new sparse subspace clustering method, which can describe the relationship between cells in a subspace; Ruiqing (Zheng et al., 2019) et al. proposed a method for detecting scRNA-seq cell types based on similarity learning. Wang et al., (2022) propose the scHFC, which is a hybrid fuzzy clustering method optimized by natural computation based on Fuzzy C Mean (FCM) and Gath-Geva (GG) algorithms. The FCM algorithm is optimized by simulated annealing algorithm, and the genetic algorithm is applied to cluster the data to output a membership matrix. Gan et al., (2022). propose a new deep structural clustering method for sc RNA-seq data, named scDSC, which integrates the structural information into deep clustering of single cells. The study by Gan et al., (2022) not only explained the cell typing method behaviors under different experimental settings but also provided a general guideline for the choice of the method according to the scientific goal and dataset properties. Li et al., (2019) Surrogate-Assisted Evolutionary Deep Imputation Model (SEDIM) is proposed to automatically design the architectures of deep neural networks for imputing gene expression levels in scRNA-seq data without any manual tuning. Yu et al., (2022)propose a single-cell model-based deep graph embedding clustering (scTAG) method, which simultaneously learns cell–cell topology representations and identifies cell clusters based on a deep graph convolutional network. Li et al., (2021) propose a multiobjective evolutionary clustering based on adaptive non-negative matrix factorization (MCANMF) for multiobjective single-cell RNA-seq data clustering. Peng et al., (2020) compared 12 single-cell clustering methods and found that most of them improved based on the K-means algorithm.

The K-means algorithm (Macqueen, 1966; Lloyd, 1982) was first proposed by Steinhaus in 1955, Lloyd in 1957, Ball and Hall in 1965, and McQueen in 1967 in different scientific fields. Once the algorithm is put forward, it is widely used in various areas because of its simple principle and easy implementation. At the same time, it is also commonly used in scRNA-seq clustering. However, the K-means algorithm still has some problems. Including that the value of K is difficult to determine, the clustering result depends on the selection of the initial center point, and it is easy to fall into the optimal local solution. In addition, the K-means algorithm is sensitive to noise points and outliers, and it is not practical for nonconvex data sets or data with too significant differences in category size. These problems will have a particular impact on the clustering results. To solve this problem, many workers have carried out a lot of research.

Due to the high-dimensional characteristics of single cells, we reduce the dimension of data sets and then cluster them, which can not only improve the clustering effect but also visually analyze the clustering results. This technology has been widely used in scRNA-seq clustering. Common dimensionality reduction algorithms include Principal Components Analysis (PCA), Locality Preserving Projections (LPP), t-distributed Stochastic Neighbor Embedding (t-SNE), Multidimensional Scaling (MDS), Isometric feature mapping (Isomap), and Locally Linear Embedding (LLE).

Based on dimension reduction, we propose a scRNA-seq clustering method: The dissimilarity-Density-Dynamic Radius-K-means algorithm. The algorithm obtains a set of initial center points by calculating the product of dissimilarity density ρ, average dissimilarity of candidate clusters α, and disparity of candidate clusters s. At the same time, the algorithm can optimize the clustering results by adjusting the dynamic radius parameters. We apply this algorithm to single-cell data sets, and the obtained indicators (NMI, FMeasure_node, Accuracy, and RandIndex) are superior to those of other algorithms. They can be used as an effective tool for scRNA-seq clustering.

The main significance of this study lies in the establishment of a clustering model based on single-cell sequencing data, which can be used to cluster cells with similar gene expression patterns into the same cell type so as to infer cell functions and understand the correlation between diseases and genomic characteristics. A more precise and unbiased classification of cells would have a huge impact in oncology, genetics, immunology, and other research fields.

2 Materials and Methods

2.1 Theoretical Presentation

The K-means algorithm will randomly select K points as initial center points when clustering, which will make the algorithm fall into optimal local solution, and the obtained clustering distribution is not optimal. It is possible to divide a smaller group into one cluster and a large cluster into several small groups. Therefore, the initial center point of the optimal group should meet the following requirements: the difference between the initial center point and other sample points in the group should be as slight as possible; The difference to sample points between the groups is as large as possible.

In thisarticle , the concept of dissimilarity is used when selecting the center point. The so-called dissimilarity is the dissimilarity between two objects, and its expression form is a n×n matrix

[d(a1,a1)d(a1,a2)d(a2,a1)d(a2,a2)d(a1,a2)d(a2,an)d(an,a1)d(an,a2)d(an,an)],

where d(ai,aj) represents the degree of dissimilarity between objects ai and aj, which is usually a non-negative value. The more similar the two things are, the closer the case is to 0; Otherwise, the closer the matter is to 1. We find that if the dissimilarity density ρ of a point is greater, the fact is more likely to become the initial center point.

The dissimilarity density ρ of sample points xi is the number of samples whose dissimilarity with sample objects xi less than the dynamic radius r. Because there are often some noise points in single-cell data sets, if the average dissimilarity is taken as the radius, this fixed-radius algorithm will make the dissimilarity density ρ inaccurate and affect the selection of the initial center point. At the same time, the fixed radius will also cause the number of clusters to be unsatisfactory. Therefore, the traditional fixed-radius method is no longer suitable for single-cell clustering. If it is set to the dynamic radius r, it can effectively solve this problem and is more conducive to single-cell data clustering. The dynamic radius here is the ratio of the average dissimilarity between samples to the dynamic parameters T. The degree of dissimilarity is a model which fully considers the comprehensive distance and dynamic radius, constructed the dissimilarity matrix, and converts the single-cell data into a phase dissimilarity matrix. It can be used to better judge the differences between cells, not just by the distance between them.

2.2 Basic Definitions

X={x1,x2,,xn}is set as the sample data set to be clustered, where x={xi1,xi2,,xip},i{1,2,,n}, and ρ is the number of attributes.

DEFINITION 1. Dissimilarity dij between sample points xi and xj:

dij=s=1pdij(s),(1)

among them

dij(s)=|xis-xjs|max{xrs}-min{xrs}(2)

represents the dissimilarity of the s th attribute between the sample point and, xrs is all the values of the s th attribute.

DEFINITION 2. Constructing dissimilarity matrix d:

d=[0d12d210d13d1nd23d2nd31d32dn1dn20d3ndn300],(3)

where dij represents the dissimilarity between the sample points xi and xj.

DEFINITION 3. Dynamic radius r of data set X:

r=Mean_r(d)T,(4)

among them

Mean_r(d)=1n2i=1nj=1ndij,(5)

T is the dynamic radius parameter, and the value is as follows:

T=0.423+0.328K1.211mead(d)+0.662max(d)+1.631min(d),(6)

where K represents the number of data categories; mean means the average of dissimilarity; max represents the maximum phase dissimilarity; and min represents the minimum phase dissimilarity.

DEFINITION 4. Sample dissimilarity density ρ:

ρ=j=1nδ(dij-r),(7)

where δ(z)={1,         z00,      others, ρi represents the dissimilarity density of the sample object xi, which is the number of points satisfied d1i<r.The sample point dissimilarity density is the number of points that satisfy d1i<r. As shown in Figure 1, the conditions are d11,d12,d13,d14. so the dissimilarity density of sample point 1 is 4. Similarly, the dissimilarity density of red dots is 6; the dissimilarity density of yellow dots is 8; the dissimilarity density of blue dots is 7. It is to be noted that the points in the intersection of two great circles can be calculated repeatedly.

FIGURE 1
www.frontiersin.org

FIGURE 1. Dissimilarity density ρ.

DEFINITION 5. According to Definition 4, ρ is the number of samples whose dissimilarity with the sample object xi is less than the dynamic radius r. Samples meeting the conditions form a candidate cluster, where the average dissimilarity between the samples of the candidate cluster is

α(i)=1n2i=1nj=1ndij,(8)

DEFINITION 6. The dissimilarity si of candidate clusters represents the dissimilarity between sample xi objects xj, which satisfies the following formula

si={min(dij),p(j)>p(i)max(dij),p(j)≤p(i),(9)

As shown on the left of Figure 2, the dissimilarity density of sample point 1 is 5, and there is a dissimilarity density larger than it, so the smallest dissimilarity is selected as the candidate cluster dissimilarity of sample point 1; as shown in the right of Figure 2, the dissimilarity density of sample point 1 is 11, and there is no dissimilarity density larger than it. Therefore, the biggest dissimilarity is selected as the candidate cluster dissimilarity of sample point 1.By analyzing Definitions 5, 6, when the candidate cluster is formed with xi as the center point, if the average dissimilarity value α(i) between samples of the candidate cluster is smaller, the dissimilarity of the cluster is very small, and the similarity is very high; similarly, the greater the value of si, the greater the dissimilarity between samples. Therefore, the dissimilarity density ρ, the average dissimilarity α(i), and the dissimilarity value si of candidate clusters can be taken as the standard to measure the initial center point, which is specifically defined as follows:

DEFINITION 7. The dissimilarity weight formula for selecting the cluster center point is as follows:

ωi=ρi1αisi,(10)

among them, the point with the most significant weight of dissimilarity is the initial center point.

FIGURE 2
www.frontiersin.org

FIGURE 2. Dissimilarity of candidate clusters si.

2.3 Algorithm Flow and Block Diagram

2.3.1 Algorithm Flow

The Dissimilarity-Density-Dynamic Radius-K-means algorithm calculates the dissimilarity density ρ of sample points, the average dissimilarity α of candidate clusters, and the dissimilarity value s of candidate clusters to obtain the dissimilarity weight ω of sample points and determine a group of initial center points. Then, the obtained center point is used as the initial center point of K-means for clustering. The flow of the Dissimilarity-Density-Dynamic Radius-K-means algorithm is as follows:

1) Giving a data set X={x1,x2,,xn};

2) Calculating the dissimilarity density of all points in x is in accordance with the definition and form a set ρ;

3) Finding that point corresponding to the maximum value from the dissimilarity set ρ; if the number of the value is 1, the point is taken as the first initial clustering center point; if the number of the maximum value is not 1, the calculated sum(i)=j=1ndij, wherein dijr,j=1,2,,n, and form the set S, that satisfying sum(i)=min(S) point is taken as the first initial center point;

4) Obtaining a first initial clustering center point at this time, recording C1, and putting it in the set C at that time C={C1}. Then, points satisfying d1i<r are then removed from the data set X;

5) Calculating the weight value ωi of the dissimilarity of the remaining point according to the definition, wherein the second initial center point is the point with the maxumun weight value of the distinction and is recorded C2 and put in set C at that time C={C1,C2}. Then, deleting the points that meet the criteria;

6) Repeating step 5 until that data set is empty, C={C1,C2,,Ck};

7) At this time, a group of initial center points C and the number K of clustering have been obtained, and the parameters are brought into the k-means algorithm for clustering;

8) Calculating the distance between each point in the sample and the initial center point, classifying the space into the cluster where the center point with the smallest distance between each other is located, and calculating the new center points of each group;

9) Repeating the step 8 until the division condition of all sample points remain unchanged or the central point does not change;

10) Output that clustering result.

2.3.2 Algorithm Block Diagram

The algorithm block diagram is shown in Figure 3.

FIGURE 3
www.frontiersin.org

FIGURE 3. Algorithm block diagram.

2.4 The Necessity of Setting the Dynamic Radius Parameter T

When introducing the D3K algorithm, we put forward the definition of dynamic radius R; the so-called dynamic radius is the ratio of average dissimilarity and dynamic radius parameter T. The distribution of the data set is not uniform. If the distribution of the data set is too scattered or too close, the average dissimilarity will be too large or too small. If the average dissimilarity is taken as the radius, the clustering result will be inaccurate, which will affect the selection of the initial center point and result in an inaccurate clustering result. If the dynamic radius parameter is added, the radius can be adjusted flexibly according to the data characteristics so as to optimize the clustering result. As shown in Figure 4 below:

FIGURE 4
www.frontiersin.org

FIGURE 4. Clustering at a fixed radius. (A): The radius is too small; (B): The radius is too large; (C): The radius is appropriate.

As shown in Figure 4C, for clustering results under ideal conditions, appropriate radii are set and clusters are divided reasonably. However, if the average dissimilarity is taken as the radius, the average dissimilarity will be too small for some overly tight data sets, which will make the radius smaller, and the original cluster will be divided into two or more clusters, as shown in Figure 4A. For some data sets that are too scattered or have noise points, the average dissimilarity will be very large. In this case, taking the average dissimilarity as the radius will make the radius very large so that originally different clusters can be divided into one cluster, as shown in Figure 4B. Therefore, adding the dynamic radius parameter T into the model can reasonably adjust the radius size according to the data characteristics and optimize the result of cluster division.

The dynamic radius parameter T is considered from multiple perspectives, including the maximum, minimum, average, and the number of clusters K. Considering many aspects, we get the optimal solution through the greedy algorithm and then fit the equation of the dynamic radius parameter T through a large amount of data. Among them, the dissimilarity between each point and itself is 0, so the dissimilarity between each point and itself should be removed when selecting the minimum value of phase dissimilarity, that is, the value with the smallest foreign phase dissimilarity except 0. By observing the equation of dynamic radius parameter T, it is found that the coefficient of K value of the number of clusters is only 0.328, indicating that although the dynamic radius parameter T is related to the number of clusters, it does not account for the main factor, and the optimal solution of T is in an interval, so the equation can be satisfied without a particularly accurate K value.

3 Result

To verify the algorithm, we selected nine groups of single-cell data sets for experiments, namely, Kolod, Pollen, Ting, Ioh, Goolam, Usoskin, Xin, Zeisel, and Macosko data sets. Table 1 shows the details of the data set.

TABLE 1
www.frontiersin.org

TABLE 1. Summary of six scRNA-seq data sets used in this study.

Clustering the data in Table 1 after dimension reduction can improve the clustering effect and visually analyze the clustering results. We compare the effects of six dimensionality reduction methods on single-cell data and visually examine the clustering results and find out an algorithm suitable for dimensionality reduction of single-cell data. At the same time, to verify the quality of the algorithm, we compare it with other single-cell clustering algorithms and finally confirm the selection of parameter T in this study.

3.1 Dimension Reduction

To find a dimension reduction algorithm suitable for single-cell data sets, we preprocess single-cell data with different dimension reduction algorithms and then cluster the reduced data to obtain clustering results. Here, we compare six dimensionality reduction algorithms: T-SNE, PCA, MDS, LPP, and LLE Isomap. By reducing dimensions in clustering, we obtain the data in Table 2:

TABLE 2
www.frontiersin.org

TABLE 2. Clustering indexes after dimensionality reduction.

By analyzing the data in Table 2, it can be found that after dimensionality reduction is used, the values of each index of clustering have been significantly improved, indicating that dimensionality reduction is very important for clustering, which can not only greatly increase the accuracy of clustering but also reduce the calculation time. At the same time, it can be found that in most of the data, the t-SNE algorithm has the best improvement effect. Therefore, the T-SNE algorithm can be used as an effective tool for single-cell clustering.

In the previous experiments, we have concluded that the t-SNE algorithm is more suitable for single-cell data dimension reduction, but how many dimensions to reduce the dimension is more suitable for clustering is still a problem to be discussed. To this end, we set up the following experiments: The t-SNE algorithm with the best dimensional reduction effect for single-cell data was selected, and six groups of single-cell data were reduced to 3, 10, 20, 50, and 100 dimensions for K-means clustering, and the clustering index results in different dimensions were analyzed. In order to compare the differences between different dimensions more clearly, the results are presented in a broken line graph. As shown in Figure 5:

FIGURE 5
www.frontiersin.org

FIGURE 5. Clustering index values of different dimension clustering.

Through the analysis of Figure 5, it is found that each data set has an inflection point in three dimensions, that is to say, the data will be reduced to three-dimension clusterings, and the clustering result will be significantly improved. Although some data still improve after three-dimension clustering, the increase is very small and can be almost ignored. Therefore, we can conclude that the t-SNE algorithm has the best clustering effect when the data are reduced to three dimensional ones. Therefore, in the following experiments, we uniformly used the t-SNE algorithm to reduce single-cell data to three dimensional ones for clustering.

3.2 Comparison With Other Clustering Algorithms

To verify the effectiveness of the D3K algorithm, we selected seven single-cell clustering algorithms to compare with it, namely, DCK (Zhang et al., 2018), S3C2 (Zhuang et al., 2021), sinNLRR (Zheng et al., 2019), Corr (Dong et al., 2018), Max-Min (Sen et al., 2018), K-means, and DBSCAN algorithm.

The nine groups of single-cell data in Table 1 were clustered by the single-cell clustering algorithm described above, and each index (NMI, FMeasure_node, Accuracy, RandIndex) of the clustering result was obtained to obtain Figure 6 as follows:

FIGURE 6
www.frontiersin.org

FIGURE 6. Index of the D3K algorithm in single-cell data aggregation class.

Compared with other clustering algorithms, the D3K algorithm is obviously higher than different algorithms in various indexes, and the results of multiple indexes are basically above 0.8, among which the effects of multiple indexes of the Pollen data set can reach above 0.95, especially Ting data set, and the results all are 1. It can be seen that the D3K algorithm can achieve ideal clustering results for both small and large data sets and can be used as a clustering model for single-cell data.

Visual analysis of clustering results can not only clearly display complex data in the form of images but also intuitively observe the differences between clusters and the size of differences within clusters. For single-cell data, this study first constructs its dissimilarity degree matrix and then obtains the cluster label of single-cell data through clustering. According to the cluster label, visual analysis of the dissimilarity matrix can not only show the clustering results of single cells after clustering but also make the distance within the same cluster smaller and the distance between different clusters larger. The following Figure 7 is a visual analysis of the clustering results of six groups of single-cell data, and the clustering results of the D3K algorithm are displayed in the form of images.

FIGURE 7
www.frontiersin.org

FIGURE 7. D3K algorithm visualization analysis.

As shown in Figure 7, the visualization results of the D3K algorithm after clustering 10 groups of single-cell data are shown. It can be seen that the D3K algorithm can perfectly divide these data into different cell types according to the labels after clustering and make the differences within clusters after clustering very small, but the differences between clusters are very large.

3.3 Validation of Parameter T

When introducing the D3K algorithm, we propose the definition of dynamic radius r, and the so-called dynamic radius is the ratio of the average degree of difference to the dynamic radius parameter. The distribution of the dataset is not uniform, and if the dataset distribution is too scattered or too tight, it will cause the average difference to be too large or too small. If the radius is based on the average degree of difference, it will affect the selection of the initial center point, resulting in inaccurate clustering results. By adding the dynamic radius parameter, you can find the right radius for each set of data to optimize clustering results.

In order to explain the necessity of the dynamic radius parameter T more rigorously, we set up the following experiment and ninesets of single-cell data were taken and clustered using D3K. The dynamic radius parameter T is not added to the first cluster, and the dynamic radius parameter T is added to the second cluster to compare the difference between the results. The result is shown in Figure 8:

FIGURE 8
www.frontiersin.org

FIGURE 8. Clustering results when T and not T are set.

As shown in Figure 8, the comparison of clustering results of the D3K algorithm when T is set and not set is shown. The abscissa of each of these plots represents the dataset, and the ordinate coordinate represents the values of each metric. The black polyline represents the clustering result when T is set, and the green polyline represents the clustering result when T is not set. The analysis found that the clustering results when setting T were better than the clustering results when T was not set. It is to be noted that setting the T value can optimize the clustering results and make the clustering results more accurate.

3.4 Genetic Markers

The task of single-cell scrNA-SEQ sequencing is not only to cluster single-cell sequencing data but also to cluster cells with similar gene expression patterns into the same cell type. Extraction of gene markers from the single-cell level of single-cell RNA-SEQ and cell identification is also an important part because it can assist in subsequent analysis of gene interactions. As shown in Figure 9, after annotation of the Deng data cluster class, its marker genes can be determined. The Deng marker genes include Early-2cell, mid-2cell, late-2cell, 4cell, 8cell, 16cell, and Zygoto. By clustering single-cell data, gene markers can be realized more effectively, which is convenient for further research on a single cell.

FIGURE 9
www.frontiersin.org

FIGURE 9. Deng data set gene marker results.

4 Discussion

scRNA-seq can quickly determine the precise gene expression patterns of thousands of single cells and reveal the complexity of the horizontal structure of single cells, thus improving biomedical research and solving various problems in biology. However, due to the high dimension and multi-noise characteristics of single-cell sequencing data sets, it brings significant challenges to the traditional clustering algorithm. In this study, we propose a Dissimilarity-Density-Dynamic Radius-K-means clustering algorithm. By selecting the dynamic radius, the algorithm effectively calculates the dissimilarity density ρ of the data set, the average dissimilarity α of candidate clusters, and the dissimilarity s of candidate clusters, finds a group of high-quality initial center points, and achieves the purpose of improving the K-means algorithm.

We use the Dissimilarity-Density-Dynamic Radius-K-means clustering algorithm to cluster some single-cell data sets and evaluate the clustering results. Experiments show that the Dissimilarity-Density-Dynamic Radius-K-means clustering algorithm has good performance for single-cell data clusters. At the same time, we also compared with other single-cell clustering algorithms. Experiments show that the Dissimilarity-Density-Dynamic Radius-K-means clustering algorithm is superior to other single-cell clustering algorithms.

Data Availability Statement

Publicly available datasets were analyzed in this study. These datasets can be found here: This study used data from the GEO database and can be found below: https://www.ncbi.nlm.nih.gov/geo/.

Author Contributions

GL and ML proposed the algorithm, wrote the code, and wrote the manuscript. HW revised the original manuscript. SL ran other single-cell clustering algorithms. JX, RL, and MT discussed the proposed algorithm and carried out further research.

Funding

This work was supported by the National Natural Science Foundation of China (No. 61903106), the Hainan Province Natural Science Foundation (No. 621MS0773, No. 118QN231), Key Laboratory of Data Science and Smart Education, Ministry of Education, the Project of Hainan Key Laboratory for Computational Science and Application, and Hainan Normal University for the funding of the Ph.D.

Conflict of Interest

RL was employeed by Geneis Beijing Co., Ltd.

The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s Note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors, and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Acknowledgments

The authors appreciate Binbin Ji, Lingyu Cui, and others for useful discussion.

References

Belkin, M., and Niyogi, P. (2001). Laplacian Eigenmaps and Spectral Techniques for Embedding and Clustering. Adv. Neural Inf. Process. Syst. 14, 585–591. doi:10.7551/mitpress/2F1120.003.0080

CrossRef Full Text | Google Scholar

Chen, L., Xu, Z., Wang, H., and Liu, S. (2016). An Ordered Clustering Algorithm Based on K-Means and the Promethee Method. Int. J. Mach. Learn. Cyber. 9, 917–926. doi:10.1007/s13042-016-0617-9

CrossRef Full Text | Google Scholar

Dong, J., Hu, Y., Fan, X., Wu, X., Mao, Y., Hu, B., et al. (2018). Single-cell RNA-Seq Analysis Unveils a Prevalent Epithelial/mesenchymal Hybrid State during Mouse Organogenesis. Genome Biol. 19, 31. doi:10.1186/s13059-018-1416-2

PubMed Abstract | CrossRef Full Text | Google Scholar

Dong, Q., and Zhu, Z. (2020). A New K-Means Algorithm for Selecting Initial Clustering Center. Statistics Decis. 36, 32–35. doi:10.13546/j.cnki.tjyjc.16.007

CrossRef Full Text | Google Scholar

Gan, Y., Huang, X., Zou, G., Zhou, S., and Guan, J. (2022). Deep Structural Clustering for Single-Cell RNA-Seq Data Jointly through Autoencoder and Graph Neural Network. Briefings Bioinforma. 23 (2), 1467–5463. doi:10.1093/bib/bbac018

CrossRef Full Text | Google Scholar

Huang, L., Li, X., Guo, P., Yao, Y., Liao, B., Zhang, W., et al. (2017). Matrix Completion with Side Information and its Applications in Predicting the Antigenicity of Influenza Viruses. Bioinformatics 33, 3195–3201. doi:10.1093/bioinformatics/btx390

PubMed Abstract | CrossRef Full Text | Google Scholar

Jiang, H., Sohn, L. L., Huang, H., and Chen, L. (2018). Single Cell Clustering Based on Cell-Pair Differentiability Correlation and Variance Analysis. Bioinformatics 34, 3684–3694. doi:10.1093/bioinformatics/bty390

PubMed Abstract | CrossRef Full Text | Google Scholar

Kiselev, V. Y., Andrews, T. S., and Hemberg, M. (2019). Challenges in Unsupervised Clustering of Single-Cell RNA-Seq Data. Nat. Rev. Genet. 20, 273–282. doi:10.1038/s41576-018-0088-9

PubMed Abstract | CrossRef Full Text | Google Scholar

Li, M., Wang, H., Long, H., Xiang, J., and Yang, J. (2019). Community Detection and Visualization in Complex Network by the Density-Canopy-Kmeans Algorithm and MDS Embedding, IEEE Access, 7, 120616–120625. doi:10.1109/ACCESS.2936248

CrossRef Full Text | Google Scholar

Li, X., Li, S., Huang, L., Zhang, S., and Wong, K.-c. (2021). High-throughput Single-Cell RNA-Seq Data Imputation and Characterization with Surrogate-Assisted Automated Deep Learning. Briefings Bioinforma. 23 (1), 1. doi:10.1093/bib/bbab368

CrossRef Full Text | Google Scholar

Li, X., and Wong, K.-C. (2019). Single-Cell RNA Sequencing Data Interpretation by Evolutionary Multiobjective Clustering. IEEE/ACM Trans. Comput. Biol. Bioinf. 17, 1773–1784. doi:10.1109/TCBB.2019.2906601

PubMed Abstract | CrossRef Full Text | Google Scholar

Liu, C., Wei, D., Xiang, J., Ren, F., Huang, L., Lang, J., et al. (2020). An Improved Anticancer Drug-Response Prediction Based on an Ensemble Method Integrating Matrix Completion and Ridge Regression. Mol. Ther. - Nucleic Acids 21, 676–686. doi:10.1016/j.omtn.2020.07.003

PubMed Abstract | CrossRef Full Text | Google Scholar

Liu, H., Qiu, C., Wang, B., Bing, P., Tian, G., Zhang, X., et al. (2021). Evaluating DNA Methylation, Gene Expression, Somatic Mutation, and Their Combinations in Inferring Tumor Tissue-Of-Origin. Front. Cell Dev. Biol. 9, 619330. doi:10.3389/fcell.2021.619330

PubMed Abstract | CrossRef Full Text | Google Scholar

Lloyd, S. (1982). Least Squares Quantization in PCM. IEEE Trans. Inf. Theory 28, 129–137. doi:10.1109/TIT.1982.1056489

CrossRef Full Text | Google Scholar

Macqueen, J. B. (1966). Some Methods for Classification and Analysis of Multivariate Observations 1 (14), 281–297.

Google Scholar

Peng, L., Tian, X., Tian, G., Xu, J., Huang, X., Weng, Y., et al. (2020). Single-cell RNA-Seq Clustering: Datasets, Models, and Algorithms. RNA Biol. 17, 765–783. doi:10.1080/15476286.2020.1728961

PubMed Abstract | CrossRef Full Text | Google Scholar

Potter, S. S. (2018). Single-cell RNA Sequencing for the Study of Development, Physiology and Disease. Nat. Rev. Nephrol. 14, 479–492. doi:10.1038/s41581-018-0021-7

PubMed Abstract | CrossRef Full Text | Google Scholar

Qi, R., Wu, J., Guo, F., Xu, L., and Zou, Q. (2021). A Spectral Clustering with Self-Weighted Multiple Kernel Learning Method for Single-Cell RNA-Seq Data. Brief. Bioinform 22 (4), bbaa216. doi:10.1093/bib/bbaa216

PubMed Abstract | CrossRef Full Text | Google Scholar

Qiao, L., Efatmaneshnik, M., Ryan, M., and Shoval, S. (2017). Product Modular Analysis with Design Structure Matrix Using a Hybrid Approach Based on MDS and Clustering. J. Eng. Des. 28, 433–456. doi:10.1080/09544828.2017.1325858

CrossRef Full Text | Google Scholar

Sen, X. U., Hua, X., Jing, X. U., Xiufang, X. U., Gao, J., and Jing, A. N. (2018). Cluster Ensemble Approach Based on T-Distributed Stochastic Neighbor Embedding. J. Electron. Inf. Technol. 40 (6), 1316–1322. doi:10.11999/JEIT170937

CrossRef Full Text | Google Scholar

Sun, X., Lin, X., Li, Z., and Wu, H. (2022). A Comprehensive Comparison of Supervised and Unsupervised Methods for Cell Type Identification in Single-Cell RNA-Seq. Briefings Bioinforma. 23. doi:10.1093/bib/bbab567

CrossRef Full Text | Google Scholar

Tang, X., Cai, L., Meng, Y., Xu, J., Lu, C., and Yang, J. (2020). Indicator Regularized Non-negative Matrix Factorization Method-Based Drug Repurposing for COVID-19. Front. Immunol. 11, 603615. doi:10.3389/fimmu.2020.603615

PubMed Abstract | CrossRef Full Text | Google Scholar

Wang, J., Xia, J., Tan, D., Lin, R., Su, Y., and Zheng, C.-H. (2022). scHFC: a Hybrid Fuzzy Clustering Method for Single-Cell RNA-Seq Data Optimized by Natural Computation. Briefings Bioinforma. 23 (2), bbaa588. doi:10.1093/bib/bbab588

CrossRef Full Text | Google Scholar

Xu, J., Cai, L., Liao, B., Zhu, W., and Yang, J. (2020). CMF-impute: an Accurate Imputation Tool for Single-Cell RNA-Seq Data. Bioinformatics 36, 3139–3147. doi:10.1093/bioinformatics/btaa109

PubMed Abstract | CrossRef Full Text | Google Scholar

Yang, J., Liao, B., Zhang, T., and Xu, Y. (2019). Editorial: Bioinformatics Analysis of Single Cell Sequencing Data and Applications in Precision Medicine. Front. Genet. 10, 1358. doi:10.3389/fgene.2019.01358

PubMed Abstract | CrossRef Full Text | Google Scholar

Yu, Z., Lu, Y., and Wang, Y. (2022). ZINB-based Graph Embedding Autoencoder for Single-Cell RNA-Seq Interpretations.

Google Scholar

Zhang, G., Zhang, C., and Zhang, H. (2018). Improved K-Means Algorithm Based on Density Canopy. Knowledge-Based Syst. 145, 289–297. doi:10.1016/j.knosys.2018.01.031

CrossRef Full Text | Google Scholar

Zhang, Z., Cui, F., Wang, C., Zhao, L., and Zou, Q. (2021). Goals and Approaches for Each Processing Step for Single-Cell RNA Sequencing Data. Briefings Bioinforma. 22, bbaa314. doi:10.1093/bib/bbaa314

CrossRef Full Text | Google Scholar

Zheng, R., Li, M., Liang, Z., Wu, F.-X., Pan, Y., and Wang, J. (2019). SinNLRR: a Robust Subspace Clustering Method for Cell Type Detection by Non-negative and Low-Rank Representation. Bioinformatics 35, 3642–3650. doi:10.1093/bioinformatics/btz139

PubMed Abstract | CrossRef Full Text | Google Scholar

Zhuang, J., Cui, L., Qu, T., Ren, C., Xu, J., Li, T., et al. (2021). A Streamlined scRNA-Seq Data Analysis Framework Based on Improved Sparse Subspace Clustering. IEEE Access 9, 9719–9727. doi:10.1109/ACCESS.2021.3049807

CrossRef Full Text | Google Scholar

Keywords: Dissimilarity matrix, density, dynamic radius, ScRNA-seq, K-means

Citation: Liu G, Li M, Wang H, Lin S, Xu J, Li R, Tang M and Li C (2022) D3K: The Dissimilarity-Density-Dynamic Radius K-means Clustering Algorithm for scRNA-Seq Data. Front. Genet. 13:912711. doi: 10.3389/fgene.2022.912711

Received: 04 April 2022; Accepted: 25 April 2022;
Published: 01 July 2022.

Edited by:

Lihong Peng, Hunan University of Technology, China

Reviewed by:

Xiangtao Li, Jilin University, China
Jujuan Zhuang, Dalian Maritime University, China
Qi Ren, Tianjin University, China

Copyright © 2022 Liu, Li, Wang, Lin, Xu, Li, Tang and Li. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Manzhi Li, lmz20031979@163.com

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.