Skip to main content

ORIGINAL RESEARCH article

Front. Bioeng. Biotechnol., 05 July 2021
Sec. Bionics and Biomimetics
Volume 9 - 2021 | https://doi.org/10.3389/fbioe.2021.698390

An Optimization Algorithm for Computer-Aided Diagnosis of Breast Cancer Based on Support Vector Machine

  • 1Network Information Center, Tianjin Baodi Hospital, Tianjin, China
  • 2Baodi Clinical College, Tianjin Medical University, Tianjin, China

As one of the most vulnerable cancers of women, the incidence rate of breast cancer in China is increasing at an annual rate of 3%, and the incidence is younger. Therefore, it is necessary to conduct research on the risk of breast cancer, including the cause of disease and the prediction of breast cancer risk based on historical data. Data based statistical learning is an important branch of modern computational intelligence technology. Using machine learning method to predict and judge unknown data provides a new idea for breast cancer diagnosis. In this paper, an improved optimization algorithm (GSP_SVM) is proposed by combining genetic algorithm, particle swarm optimization and simulated annealing with support vector machine algorithm. The results show that the classification accuracy, MCC, AUC and other indicators have reached a very high level. By comparing with other optimization algorithms, it can be seen that this method can provide effective support for decision-making of breast cancer auxiliary diagnosis, thus significantly improving the diagnosis efficiency of medical institutions. Finally, this paper also preliminarily explores the effect of applying this algorithm in detecting and classifying breast cancer in different periods, and discusses the application of this algorithm to multiple classifications by comparing it with other algorithms.

Introduction

Health is the foundation of all-round development of human beings. The incidence rate of breast cancer worldwide has been increasing since the end of 1970s. Breast Cancer is a malignant tumor of abnormal breast cell division and proliferation. The incidence of breast cancer is more prominent in female patients. A United States survey shows that in 2016, 16,85,210 cases of new cancer and 595 cases of cancer were found. Among 690 cancer deaths, breast cancer is the main cause of cancer death in women aged 20–59 (Siegel et al., 2016). Each year, the number of new breast cancer cases and deaths in China account for 12.2 and 9.6% of the world’s total, respectively. In view of this serious social reality, there is an urgent need to carry out research on the risk of breast cancer, including the cause analysis and prediction of breast cancer risk diagnosis based on historical data (Li Y. et al., 2020). In the examination, the characteristics of cell size, shape, and mass thickness are considered as the criteria to distinguish benign from malignant tumors, while the characteristics of age, tumor size, menopause, number of lymph nodes involved and radiotherapy are considered as the factors influencing the recurrence of breast cancer. It is difficult for doctors to manually determine whether breast cancer is benign or not and the recurrence of breast cancer according to the complex characteristic data, but computer technology can analyze and predict the existing data.

Artificial intelligence (AI) is the product of the rapid development of computer technology. It has a profound impact on the development of human society and the progress of science and technology. At this stage, artificial intelligence has been widely used in clinic. With the development of technology and the availability of big data, the application and development of artificial intelligence in medical disease diagnosis has become a research hotspot in today’s era. As one of the important means in artificial intelligence, in 1959, Arthur Samuel proposed the concept of machine learning, that is, using algorithms to make machines learn from a large number of data, to obtain the method of new data analysis and research (Skoff, 2017).

At present, researchers have used deep learning or machine learning methods to study different breast cancer data. Khan et al. (2019) used the method of combining transfer learning and deep learning to detect and classify breast cancer cells, and achieved high accuracy. Abbass (2002) proposed a neural network method based on differential evolution algorithm and local search to predict breast cancer, and the standard deviation of its test accuracy is 0.459 lower than that of Fogel et al. (1995). Abdikenov et al. (2019) used the evolutionary algorithm NSGA III (non-dominated sorting genetic algorithm – III) to initialize the deep neural network and optimize its super parameters for the prognosis of breast cancer. Liu et al. (2019) proposed an end-to-end deep learning system combined with full convolution network to extract breast region data, and the results are highly correlated with the diagnosis made by pathologists. Lu et al. (2019) proposed a novel genetic algorithm based online gradient boosting (GAOGB) model to predict the diagnosis and prognosis of breast cancer in real time through online learning (Oza, 2005) technology. The above research shows that the application of artificial intelligence in the medical field is practical and effective. The application of existing machine learning methods in the medical field helps medical workers improve work efficiency and reduce work burden. People are trying to improve the traditional algorithm while applying computer technology to the medical field.

In this paper, Support Vector Machine (SVM) is taken as a breakthrough point. The choice of penalty parameter c and g in SVM kernel function is directly related to the effectiveness and accuracy of SVM algorithm in solving dichotomy. According to previous research methods, there are mainly 5 optimization methods for the above two important parameters, namely, empirical selection method, grid selection method, genetic optimization algorithm, particle swarm optimization algorithm, and ant colony optimization algorithm so on (Ali and Abdullah, 2020; Kouziokas, 2020; Li X. et al., 2020; Arya Azar et al., 2021; Ramkumar et al., 2021). Although these optimization algorithms have been applied to some extent and achieved some effects, they all have problems of different degrees. For example, the empirical selection method is highly experienced by users and highly dependent on samples, which lacks sufficient theoretical support. The disadvantage of grid selection method lies in the step size selection. If the step size selection is too large, it is easy to fall into the local optimum; if the step size selection is too small, the calculation amount will be too large. The genetic optimization algorithm needs to go through three steps of selection, crossover and mutation. The parameter setting is relatively complex, the convergence speed is slow, and it is easy to fall into the local optimal solution. Particle swarm optimization (PSO) SVM has the advantage of faster convergence speed and fewer parameters, but it is also easy to fall into local optimal. The combination of genetic or particle swarm optimization and simulated annealing (SA) to optimize SVM parameters improves the convergence speed and improves the poor local optimization ability to some extent. However, poor stability may occur in some practical applications. Therefore, how to use the advantages of three heuristic algorithms to optimize the selection of parameters in support vector machines, so that the algorithm to achieve the best classification performance is the focus of this paper.

Conceptual Principle

Support Vector Machine

Support vector machine was proposed by Vapnik (1995). The basic idea of the algorithm is to map the input data into a high-dimensional space through non-linear transformation and establish the optimal linear classification surface to classify the two sample categories correctly. Based on the principle of structural risk minimization, the SVM model is classified by calculating the optimal separating hyperplane (OSH) (Zhou et al., 2018). The larger the interval between the optimal hyperplanes, the stronger the generalization ability of the established SVM model. Suppose that the training sample set {(xi, yi), i = 1,2,…,l} with the size of 1, its data samples can only be divided into two categories. If it belongs to the first type of samples, it is recorded as positive (yi = 1), otherwise it belongs to the second category and is recorded as a negative value (yi = −1). At this time, we need to construct a discriminant function to make the function classify the test data samples as correctly as possible. If there is a classification hyperplane

wx+b=0(2-1)

bring

{wxi+b1,yi=1wxi+b-1,yi=-1,i=1,2,,l(2-2)

We call the training sample set is linearly separable. wx is called the inner product of vector wRN and vector xRN, and wRN and bR in formula (2-1) and formula (2-2) are normalized. For formula (2-2), it can be rewritten as follows:

yi(wxi+b)1,i=1,2,,l(2-3)

According to the definition of the optimal hyperplane, the following discriminant functions can be obtained

y(x)=sign(wx+b)(2-4)

Its generalization ability is the best, and sign (⋅) is the symbol function. The solution of the optimal hyperplane needs to maximize 2/||w||, that is to say it can be transformed into the following quadratic programming problem composed of objective function and constraint conditions

minw,b||w||22s.t.yi(wxi+b)1,i=1,2,,l(2-5)

When the training sample set is linear and indivisible, it is necessary to introduce a non-negative parameter, i.e., relaxation variable ξi, i = 1,2,…, l. at this time, the optimization problem of classification hyperplane is transformed into the form shown in formula (2-6).

minw,b,ξ||w||22+ci=1lξis.t.{yi(wxi+b)1-ξiξi0,i=1,2,,l(2-6)

Where c is the constraint parameter, also known as the penalty parameter. The higher the value of c, the greater the penalty for error classification. Using Lagrange multiplier method to solve the problem

maxminα,β,w,b,ξ{Lp=||w||22+ci=1lξi-i=1lαi[yi(wxi+b)-1+ξi]-i=1lβiξi}s.t.{αi0βi0(2-7)

Where αi and βi are Lagrange multipliers

Lpw=0w=i=1lαiyixi(2-8)
Lpb=0i=1lαiyi=0(2-9)
Lpξi=0c-αi-βi=0(2-10)

By substituting formula (2-8) to (2-10) into formula (2-7), the dual optimization problem form is obtained

maxα{LD=i=1lαi-12i=1lj=1lαiαjyiyjxixj}s.t.{0αici=1lαiyi=0(2-11)

The αi obtained by optimization may be (a) αi = 0; (b) 0 < αi < c; (c) αi = c. According to formula (2-8), only when the support vector has a positive effect on the optimal hyperplane and discriminant function, the corresponding learning method is called support vector machine algorithm. In support vector, xi corresponding to c is called boundary support vector (BSV), which is actually the training sample points that are misclassified; (b) The corresponding xi is called normal support vector (NSV). According to Karush–Kuhn–Tucher condition (Chauhan and Ghosh, 2021), the product between Lagrange multiplier and corresponding constraint is equal to 0 when the sample point is optimal

{αi[yi(wxi+b)-1+ξi]=0βiξi=0(2-12)

For the standard support vector (0 < αi < c), βi > 0 is obtained from formula (2-10). Therefore, βi = 0 can be obtained from formula (2-12). Therefore, it can be seen that all the criteria satisfy the following requirements for any standard support vector xi,

yi(wxi+b)=1(2-13)

the parameter b is calculated

b=yi-wxi=yi-xiNSV,xjSVαjyjxjxi(2-14)

The value of b is calculated for all standard support vectors, and then the average value of the results is obtained

b=1NNSVxiNSV(yi-xiSVαjyjxjxi)(2-15)

Where NNSV is the number of the standard support vectors. According to formula (2-13), the support vector machine model is the sample data that meets the requirements of formula (2-3).

Kernel Function Selection for Support Vector Machine Algorithm

The use of support vector machines to solve pattern classification problems usually requires the selection of an appropriate kernel function. Since the low-dimensional space vector sample set is usually difficult to divide, we usually use to map the low-dimensional space vector sample set into the high-dimensional feature space, but the consequent problem is to increase the computational complexity, and the emergence of the kernel function is a good solution to the problem. Theoretically, any function that can satisfy the Merce condition can be used as the kernel function of a support vector machine algorithm, but the different choices of kernel functions can lead to different algorithms and directly lead to different performance of their classifiers. Therefore, the selection of the appropriate kernel function is crucial to effectively improve the distribution of feature vectors in the high-dimensional feature space, thus making the structure of the classifier simpler; at the same time, even if a certain kernel function is selected, the selection of the corresponding parameters in the kernel function, such as the order in the polynomial kernel function and the width parameter in the Gaussian kernel function, also needs to be deliberated.

The most studied kernel functions are mainly of the following types, one is linear kernel function, as shown in formula (2-16), which mainly solves linear classification problems.

K(xi,xj)=xixj(2-16)

Second, the polynomial kernel function, as shown in formula (2-17), is obtained as a polynomial classifier of order q.

K(xi,x)=(xix+1)d(2-17)

Third, the radial basis function (referred to as the RBF kernel function), as shown in formula (2-18).

K(x,xi)=exp(-||x-xi||2σ2)(2-18)

The resulting classifier differs from the traditional RBF method in that it has a support vector corresponding to the center of each basis function, where the weights of the output are determined automatically by the algorithm. A Sigmoid function can also be used as the inner product, i.e.,

K(x,xi)=tanh(v(xxi)+c)(2-19)

The support vector machine algorithm implemented in this case is equivalent to a multilayer perceptron network with hidden layers, in which the number of hidden layer nodes is also determined automatically by the algorithm, and it is also able to better solve the problem of local minima in neural networks. Based on this, and also considering that the SVM algorithm is not sensitive to the selection of the kernel, this paper uses the radial basis kernel function, which is also called Gaussian Kernel. The classification accuracy factor σ in the RBF kernel is the parameter that needs to be adjusted, and the different values of σ will also have a great impact on the nature of the classifier and the correct recognition rate, etc.

Optimization Algorithm

For the improvement of local optimization and global optimization, this paper uses the genetic algorithm and particle swarm optimization algorithm in the algorithm to determine the respective population optimal solution, so as to seek the global optimal solution as the parameter input of SVM, so as to achieve a good balance between global and local search optimization. Through the assignment between the optimal particle and the worst chromosome or between the worst particle and the optimal chromosome, the two search algorithms complement each other and accelerate the convergence speed of the algorithm.

For the improvement that particle swarm optimization algorithm is easy to fall into local optimum, this paper takes into account the role of inertia factor ω in particle velocity and position update in formula (2-20), Because ω reflects the ability of particles to inherit the previous velocity, when the value is large, the particle swarm optimization has strong search ability in the early stage, but it is not conducive to ensure the optimal solution when the search enters the late stage; When the value of ω is small, the effect is just the opposite. When the value is small, the search ability of particle swarm optimization is enhanced, but the ability of global search for optimal solution is decreased. Therefore, in order to improve this deficiency, the harmonic inertia factor is adopted, as shown in formula (2-22).

vi,j(t+1)=ωvi,j(t)+c1r1[pi,j-xi,j(t)]+c2r2[pg,j-xi,j(t)](2-20)
xi,j(t+1)=xi,j(t)+vi,j(t+1)(2-21)

The meanings of parameters in the above two formulas are as follows:

ω represents the inertia weight of particles, and c1 and c2 represent the self-learning factor and global learning factor of particles, respectively. r1 and r2 represent random numbers between [0–1]. In order to make particles search in effective space, it is generally necessary to limit the search space of particles, that is to limit the position to [xmin, xmax]. At the same speed, a range [vmin, vmax] should be set instead of blindly optimizing. This setting can control the movement of particles.

ωm=ω1-(ω1-ω2)(m-1)2t2(2-22)

Where m is the number of iterations and t is the maximum evolution algebra.

In order to improve the local search ability of particle swarm optimization (PSO), a simulated annealing algorithm is introduced in this paper. Metropolis criterion (Wang et al., 2019) is used to determine whether to accept the new location of particles, suppose that the change of fitness of the particle in the new position is Δf, if Δf ≥ 0, then accept the new position of the particle at time t;If Δf < 0, the acceptance probability is calculated according to formula (2-23). By comparing with the threshold value, it is a standard normal distribution random quantity with a mean value of 0 and a standard deviation of 1. When Pa > P the bad position is accepted.

Pa=exp(Δft),t=KT(2-23)

Where t is the control parameter, K is the Boltzmann constant in physics, and T is the temperature of the material.

Both genetic algorithm and particle swarm optimization belong to the branch of evolutionary algorithm. Both of them are suitable for solving discrete problems, especially 0–1 non-linear optimization, integer programming and mixed integer programming. Therefore, this paper selects GA and PSO as basic algorithms. At the same time, in order to make full use of the local search solution space of GA and the fast convergence ability of PSO algorithm, and to improve the poor local search ability of PSO algorithm in the later stage, the simulated annealing algorithm is introduced f or optimization. In this paper, a new algorithm combining three classical algorithms to optimize support vector machine (GSP_SVM) is proposed. By comparing the population optimal solutions obtained from GA and PSO algorithm, the overall optimal solution is found. In this paper, the accuracy of training classification is taken as fitness value. If the fitness of PSO optimal solution is higher than that of GA, it is regarded as the global optimal solution and assigned to the worst chromosome in GA. however, if the fitness of PSO optimal solution is lower than that of GA, the chromosome with the highest fitness is regarded as the global optimal solution and assigned to the particle with the worst fitness, and then iterative calculation is carried out until the algorithm is implemented Termination. The overall framework of the algorithm is shown in Figure 1.

FIGURE 1
www.frontiersin.org

Figure 1. Flow chart of GSP_SVM algorithm.

Experiment

Data Set

In order to verify the effectiveness and feasibility of the gsposvm algorithm proposed in this paper, we use the breast cancer data set1 provided by Dr. William H. wolberg of the Wisconsin Medical School in the United States. Each data sample in the medical data set has 10 attribute variables, which are case code number, tumor thickness value, cell size uniformity, cell shape uniformity, edge viscosity, single epithelial cell size, naked nucleus, boring chromosome, normal nucleus and mitotic number. Except the case code number, the values of the other 9 attributes were all [1,10]. The binary variable was to judge the characteristics of breast cancer. 1 was malignant and 2 was benign. In order to get a better prediction effect, this section makes a study on the original data set of “breast cancer”- wisconsin.data to preserve the authenticity of the data, the redundant attributes are removed 16 data samples were eliminated, and the final experimental data samples were 683. Finally, in order to reduce the value range of some attributes which are too large while others are too small, so that the large number will submerge the decimal. At the same time, in order to avoid the difficulties in numerical calculation due to the calculation of kernel function, the data of training set and test set are normalized, and the data is scaled to [0,1].

Evaluating Indicators

In order to better explain the evaluation index used in this paper, we first give a confusion matrix about binary classification problem, as shown in Table 1.

TABLE 1
www.frontiersin.org

Table 1. Contingency table for binary classification problems.

Based on the Precision and recall rate, the receiver characteristic curve, namely AUC, F-measure, total accuracy G are used to evaluate the application effect of the proposed optimization algorithm in unbalanced data sets.

Precision: refers to the ratio of the number of records that the classifier can correctly determine as the category and the total number of records that should be determined as the category. As shown in formula (3-1), the precision rate represents the classification accuracy of the classifier itself. If the TPi is larger and the FPi is smaller, the precision value will be larger, which means that the probability of the classifier’s misclassification on this category will be smaller.

Precision=TPiTPi+FPi×100%(3-1)

Recall: refers to the ratio of the number of records that can be correctly determined by the classifier to the total number of records in the classification records that should be the category. As shown in formula (3-2), recall reflects the completeness of the classification results of the classifier. If the greater the TPi is, the smaller the FNi is, the greater the recall value is, which means that the fewer records should have been missed by the classification system.

Recall=TPiTPi+FNi×100%(3-2)

Sensitivity: the proportion of correct number of multi class discrimination in all multi class samples, and the calculation method is consistent with the calculation formula of recall rate.

Specificity: the proportion of the correct number of minority discrimination in all minority samples. The calculation method is shown in formula (3-3).

Specificity=TNiTNi+FPi×100%(3-3)

Total accuracy G: considering the classification performance of minority and majority records, the calculation method is the geometric average of specificity and sensitivity. It can be seen from formula (3-4) for details. Therefore, G is also called geometric average, and the accuracy increases monotonically with the values of specificity and sensitivity in [0,1].

G=Specificity*Sensitivity(3-4)

Fβ: considering the difference between precision rate and recall rate, the formula is as follows:

Fβ=(β2+1)×Precision×Recallβ2×Precision+Recall(3-5)
F-measure=2*Precision*RecallPrecision+Recall=2*Precision*SensitivityPrecision+Sensitivity(3-6)

The Fβ measure value represents the trade-off between accuracy and recall when evaluating the performance of classifiers. β is used to adjust the proportion of precision and recall in the formula. Usually, when it is used in practice, it is taken as β = 1 to get the performance evaluation index F-measure of our common classifier. The calculation formula is as follows (3-6). F-measure is the harmonic mean of precision and recall. When the accuracy and recall are both high, the F-measure value will also increase. This index takes into account the recall and precision of minority records. Therefore, any change of any value can affect the size of F-measure. Therefore, it can show the classification effect of the classifier on the majority class and minority class, but it focuses on the classification effect of minority records is also discussed.

MCC: Matthew’s correlation coeffcient (3-7):

MCC=TP×TN-FP×FN(TP+FN)(TP+FP)(TN+FP)(TN+FN)(3-7)

AUC (area under the ROC curve): the area under the ROC curve, between 0.1 and 1. It can quantify the ROC curve and present the algorithm performance more intuitively. The larger the value is better. The larger the value is, the more likely the positive samples will be placed before the negative samples, so as to better classify.

Results

Discussion of the Binary Classification Problems

In this paper, we use radial basis function, which is also known as Gaussian kernel function. The classification accuracy factor in RBF kernel is a parameter σ that needs to be adjusted. Different σ values will have a great impact on the properties of classifier and recognition accuracy. In this paper, a heuristic search method is used to find the optimal parameters in the model selection, so as to achieve the optimal performance for the classification and prediction.

In order to verify the effect of different optimization algorithms on the optimization of support vector machine parameters, this paper uses several algorithms for experimental comparison: (1) Based on the most original support vector machine algorithm; (2) Based on principal component analysis support vector machine algorithm (PCA_SVM) (Tao and Cuicui, 2020); (3) Support vector machine algorithm based on grid search optimization (GS_SVM) (Fayed and Atiya, 2019); (4) Support vector machine algorithm based on genetic algorithm optimization (GA_SVM) (Guan et al., 2021); (5) Particle swarm optimization based support vector machine algorithm (PSO_SVM) (Zhang and Su, 2020), in order to compare the genetic algorithm, particle swarm optimization algorithm and simulated annealing algorithm based on the fusion algorithm to optimize the parameters of support vector machine (GSP_SVM).

For better performance comparison and algorithm verification, we randomly take 50, 60, 70, and 80% of the data as labeled data and training data, and the remaining data as unlabeled sample data and test sample set. In order to balance the random effect, the average value of 10 repeated independent running results is used for the reported experimental results. The specific results are shown in Table 2.

TABLE 2
www.frontiersin.org

Table 2. Experimental results.

Since the parameters of SVM and PCA_SVM algorithms are set to fixed values, c is 100 and g is 4, all the other algorithms are optimized for SVM parameters except these two algorithms. On the whole, with the increase of training sample data, most of the values of the evaluation matrix have a positive growth trend. It can be seen from the above table that among all the optimization algorithms, the support vector machine algorithm (GSP_SVM) based on the fusion of three classical optimization algorithms has improved the value of each evaluation index to varying degrees compared with other algorithms. The accuracy rate, recall rate, sensitivity, F-measure measurement value and other four evaluation indicators are presented in 60, 70, and 80% training data, respectively, the best result, in AUC, the best is in 70 and 80% training data.

This paper also investigates the accuracy, the most basic evaluation index of classification algorithm. The calculation method is shown in formula (3-8). No matter which category, as long as the prediction is correct, the number is placed on the numerator, and the denominator is the number of all the data. It shows that the accuracy is the judgment of all the data, and it is the evaluation index that can directly reflect the advantages and disadvantages of the algorithm. The accuracy of each algorithm on the training data set with different proportions can be seen from Table 3.

Accuracy=TP+TNTP+TN+FP+FN(3-8)
TABLE 3
www.frontiersin.org

Table 3. The experimental results of classification accuracy of algorithms.

According to the above table, the results of the experiment on 50, 60, 80, and 90% training proportion data sets are the best, especially in the 90%, the training proportion data set reaches 0.9853. Therefore, combined with the experimental results of evaluation indexes in Table 2, it can be concluded that the algorithm proposed in this paper can obtain the optimal parameter values and classify more accurately.

Discussion of the Multiclass Problems

In the above study, we considered the effectiveness of the algorithm for evaluation on the dichotomous classification problem, and next, in this paper, we will initially explore the classification of the algorithm on the multiclassification problem. We use the dataset proposed by M. Zwitter and M. Soklic from the Institute of Oncology, University of Ljubljana, Yugoslavia (Bennett et al., 2002), which has 286 instances, each containing 10 attributes such as tumor size, number of invaded lymph nodes, presence or absence of nodal adventitious, mass location, etc., all of which are of enumerated type, according to which we ask experts to manually annotate The defective instances accounted for only 0.3% of the total data set, so they were directly discarded. A total of 277 instances consisting of 10 independent variables and 1 multicategorical variable were finally used for the experiment.

For the binary classification problem, we have many evaluation metrics because there are only two types of positive and negative classes, but they are not applicable for the multi-classification problem. In this paper, we choose the following evaluation metrics for the multi-classification problem: (1) Accuracy_score, which is the ratio of the total number of correctly classified data in the classification result. (2) Precision_score, i.e., the proportion of positive cases in the prediction results. (3) Recall_score, i.e., the proportion of true positive cases that are finally predicted to be positive. (4) F1_score, as a combination of accuracy and recall, is often used as a metric for multi-classification model selection. (5) Hamming_loss, which is a measure of the distance between the predicted label and the true label, takes a value between 0 and 1, the smaller the value the better, and a distance of 0 indicates that the predicted result is exactly the same as the true result. (6) Cohen_kappa_score, the value range is [0,1], the higher the value of this coefficient, the higher the accuracy of the classification achieved by the model. It is calculated as k = (PoPe)/(1 − Pe), Where Po denotes the overall classification accuracy and Pe denotes SUM (the number of true samples in class i number of samples predicted in class i)/total number of samples squared. (7) Jaccard_score, which is used to compare the similarity and difference between the true and predicted values. The larger the coefficient value, the higher the sample similarity, indicating the more accurate prediction.

The experiments in this section are in the form of ten-fold cross-validation, and the average of five experiments is calculated as the indicator results, where for the average price indicators (2)–(4) a micro-averaging approach is used, i.e., a global confusion matrix is established for each instance in the dataset without categorizing the statistics, and then the corresponding indicators are calculated. The experimental results are shown in Table 4.

TABLE 4
www.frontiersin.org

Table 4. Evaluation results of multicategorical indicators.

Discussion

Discussion of the Binary Classification Problems

In this experiments, we all assume that the range of penalty factor c is [0.1, 100], which is mainly used to control the tradeoff between model complexity and approximation error of classification model. If the penalty factor c is larger, the better the fitting degree of the algorithm is, but at the same time, the generalization ability of the algorithm will be reduced, which is not conducive to the popularization and application of the algorithm. At the same time, we also assume that the value range of parameter g in the selected Gaussian kernel function is [0.01,1000], which determines the classification accuracy of the algorithm. Through parameter optimization, we get the optimal parameter values of each algorithm, as shown in Table 5. Figure 2 also shows the visualization of iterative optimization of parameters c and g based on the algorithm of optimizing support vector machine parameters based on GA_SVM and PSO_SVM.

TABLE 5
www.frontiersin.org

Table 5. The optimal parameters of algorithms.

FIGURE 2
www.frontiersin.org

Figure 2. The visualization of iterative optimization.

From the above analysis, this paper uses the advantages of GA, PSO and SAA to improve the parameter optimization algorithm of support vector machine, which can balance the difference between global search optimization and local search optimization. Through the mutual assignment between the optimal particle and the worst chromosome or between the worst particle and the optimal dye, the genetic algorithm and particle swarm optimization algorithm complement each other. In the later stage, the local search ability of the sub group algorithm is insufficient, and the simulated annealing method is used to enhance it, the Metropolis criterion is used to select whether to accept new particles. According to the experimental results, we can also see that the improved SVM optimization algorithm can show good performance in the case of small samples and non-linear, and its robustness is high, the generalization ability is strong, and there is no problem of under fitting and over fitting.

Extension to Multiclass Problems

From Table 4, it can be seen that, overall, the performance of this paper’s algorithm is optimal compared with other algorithms on data with 50, 60, and 90% training share, and the algorithm of this paper applied on 70% of the training dataset is consistent with Accuracy_score, Precision_score, Recall_score, and F1_score metrics with PCA_SVM and GA_SVM exhibit consistent results with SVM and PCA_SVM on 80% of the training set datasets. As the most commonly used metrics in evaluating multi-classification problems, the smaller the value of Hamming_loss, the closer the predicted label is to the true label, and the higher the value of Cohen_kappa_score, the better the classification accuracy of the algorithm. The algorithm in this paper shows optimal results in both metrics, especially in the 90% training data share, the Hamming_loss decreases to 0.0148 and Cohen_kappa_score reaches 97.64%, and Figure 3 shows the classification results of the algorithm in this paper on different training sets. Therefore, the algorithm proposed in this paper can also show better classification results when extended to multi-classification problems.

FIGURE 3
www.frontiersin.org

Figure 3. (A–D) Classification results under 60, 70, 80, and 90% training percentages, respectively.

Conclusion

In this paper, through the simultaneous interpretation of traditional optimization algorithms and machine learning methods, an algorithm combining three classical algorithms for optimization of support vector machines is proposed and trained and tested based on different breast cancer datasets, and the experimentally obtained classification accuracy, MCC, AUC, and other indexes reach high levels on the binary dataset. On the multiclassification dataset, the experimentally obtained metrics such as Hamming_loss minimum, Cohen_kappa_score and classification accuracy are optimal, which fully demonstrate that the method can provide decision support for breast cancer assisted diagnosis and thus significantly improve the diagnostic efficiency of medical institutions. Our research work will be based on this and will be developed into a breast cancer diagnosis recognition system, using artificial intelligence methods and combined with computer visualization to provide an auxiliary diagnostic basis for clinicians’ decision making through a graphical interface.

From the perspective of medical risk, in order to maximize the accuracy of the classification of malignant tumors, further research can be done on the combination of more complex kernel functions for different classifications. At the same time, medical institutions need to collect typical sample data purposefully to prevent the serious asymmetry of the two types of sample data. Of course, if we want to comprehensively improve the level of computer-aided diagnosis of diseases in medical institutions, we need to do further research on other high-risk diseases.

Data Availability Statement

The original contributions generated for this study are included in the article/Supplementary Material, further inquiries can be directed to the corresponding author.

Author Contributions

YD and WM conceived and designed the experimental protocol. YD involved in the analysis and designed the model of improved SVM. WM performed operations. YD wrote the first draft of the manuscript. WM reviewed and revised the manuscript. Both authors read and approved the final manuscript.

Funding

This work was supported by the Scientific Research Project of Tianjin Health Information Association under Grant No. TJHIA-2020-001.

Conflict of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Acknowledgments

We thank Tianjin Baodi Hospital for providing the experimental platform.

Supplementary Material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fbioe.2021.698390/full#supplementary-material

Footnotes

  1. ^ http://www.csie.ntu.edu.tw/~cjlin/libsvmtools/datasets/binary.html#breast-cancer

References

Abbass, H. A. (2002). An evolutionary artificial neural networks approach for breast cancer diagnosis. Artif. Intell. Med. 25, 265–281. doi: 10.1016/s0933-3657(02)00028-3

CrossRef Full Text | Google Scholar

Abdikenov, B., Iklassov, Z., Sharipov, A., Hussain, S., and Jamwal, P. K. (2019). Analytics of heterogeneous breast cancer data using neuroevolution. IEEE Access 7, 18050–18060. doi: 10.1109/access.2019.2897078

CrossRef Full Text | Google Scholar

Ali, A. H., and Abdullah, M. Z. (2020). A parallel grid optimization of SVM hyperparameter for big data classification using spark radoop. Karbala Int. J. Modernence 6:3.

Google Scholar

Arya Azar, N., Ghordoyee Milan, S., and Kayhomayoon, Z. (2021). The prediction of longitudinal dispersion coefficient in natural streams using LS-SVM and ANFIS optimized by Harris hawk optimization algorithm. J. Contam. Hydrol. 240:103781. doi: 10.1016/j.jconhyd.2021.103781

PubMed Abstract | CrossRef Full Text | Google Scholar

Bennett, K. P., Demiriz, A., and Maclin, R. (2002). “Exploiting unlabeled data in ensemble methods,” in Proceedings of the Eighth ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD ‘02), (New York, NY: Association for Computing Machinery), 289–296.

Google Scholar

Chauhan, R. S., and Ghosh, D. (2021). An erratum to “Extended Karush-Kuhn-Tucker condition for constrained interval optimization problems and its application in support vector machines”. Inform. Sci. 559, 309–313. doi: 10.1016/j.ins.2020.12.034

CrossRef Full Text | Google Scholar

Fayed, H. A., and Atiya, A. F. (2019). Speed up grid-search for parameter selection of support vector machines. Appl. Soft Comput. 80, 202–210. doi: 10.1016/j.asoc.2019.03.037

CrossRef Full Text | Google Scholar

Fogel, D. B., Wasson, E. C. III, and Boughton, E. M. (1995). Evolving neural networks for detecting breast cancer. Cancer Lett. 96, 49–53. doi: 10.1016/0304-3835(95)03916-k

CrossRef Full Text | Google Scholar

Guan, S., Wang, X., Hua, L., and Li, L. (2021). Quantitative ultrasonic testing for near-surface defects of large ring forgings using feature extraction and GA-SVM. Appl. Acoust. 173:107714. doi: 10.1016/j.apacoust.2020.107714

CrossRef Full Text | Google Scholar

Khan, S. U., Islam, N., Jan, Z., Din, I. U., and Rodrigues, J. J. (2019). A novel deep learning based framework for the detection and classification of breast cancer using transfer learning. Pattern Recognit. Lett. 125, 1–6. doi: 10.1016/j.patrec.2019.03.022

CrossRef Full Text | Google Scholar

Kouziokas, G. N. (2020). SVM kernel based on particle swarm optimized vector and Bayesian optimized SVM in atmospheric particulate matter forecasting. Appl. Soft Comput. 93:106410. doi: 10.1016/j.asoc.2020.106410

CrossRef Full Text | Google Scholar

Li, X., Guo, Y., and Li, Y. (2020). Particle swarm optimization-based SVM for classification of cable surface defects of the cable-stayed bridges. IEEE Access 8, 44485–44492. doi: 10.1109/access.2019.2961755

CrossRef Full Text | Google Scholar

Li, Y., Chen, S. X., Jia, H., and Wang, X. (2020). Prediction of breast cancer based on C-AdaBoost model. Comput. Eng. Sci. 42, 1414–1422.

Google Scholar

Liu, J., Xu, B., Zhang, C., Gong, Y., Garibaldi, J., Soria, D., et al. (2019). An end-to-end deep learning histochemical scoring system for breast cancer tissue microarray. IEEE Trans. Med. Imaging 38, 617–628. doi: 10.1109/tmi.2018.2868333

PubMed Abstract | CrossRef Full Text | Google Scholar

Lu, H., Wang, H., and Yoon, S. W. (2019). A dynamic gradient boosting machine using genetic optimizer for practical breast cancer prognosis. Expert Syst. Appl. 116, 340–350. doi: 10.1016/j.eswa.2018.08.040

CrossRef Full Text | Google Scholar

Oza, N. C. (2005). “Online bagging and boosting,” in Proceedings of the 2015 IEEE International Conference on Systems, Man and Cybernetics, (Waikoloa, HI), 2340–2345.

Google Scholar

Ramkumar, M., Babu, C. G., Priyanka, G. S., and Sarath Kumar, R. (2021). Ecg arrhythmia signals classification using particle swarm optimization-support vector machines optimized with independent component analysis. IOP Conf. Ser. Mater. Sci. Eng. 1084:012009. doi: 10.1088/1757-899x/1084/1/012009

CrossRef Full Text | Google Scholar

Siegel, R. L., Miller, K. D., and Jemal, A. (2016). Cancer statistics. CA Cancer J. Clin. 66, 7–30.

Google Scholar

Skoff, D. N. (2017). Exploring potential flaws and dangers involving machine learning technology. S & T’s Peer to Peer 1:4.

Google Scholar

Tao, Y., and Cuicui, L. (2020). Recognition system for leaf diseases of Ophiopogon japonicus based on PCA-SVM. Plant Dis. Pests 11, 11–15.

Google Scholar

Vapnik, V. N. (1995). “Controlling the generalization ability of learning processes,” in The Nature of Statistical Learning Theory, (New York, NY: Springer), 89–118. doi: 10.1007/978-1-4757-2440-0_5

CrossRef Full Text | Google Scholar

Wang, C., Wang, Y., Wang, K., Yang, Y., and Tian, Y. (2019). An improved biogeography/complex algorithm based on decomposition for many-objective optimization. Int. J. Mach. Learn. Cybern. 10, 1961–1977. doi: 10.1007/s13042-017-0728-y

CrossRef Full Text | Google Scholar

Zhang, H., and Su, M. (2020). Hand gesture recognition of double-channel EMG signals based on sample entropy and PSO-SVM. J. Phys. Conf. Ser. 1631:012001. doi: 10.1088/1742-6596/1631/1/012001

CrossRef Full Text | Google Scholar

Zhou, S., Qian, S., Chang, W., Xiao, Y., and Cheng, Y. (2018). A novel bearing multi-fault diagnosis approach based on weighted permutation entropy and an improved SVM ensemble classifier. Sensors 18:1934. doi: 10.3390/s18061934

PubMed Abstract | CrossRef Full Text | Google Scholar

Keywords: breast cancer, computer-aided diagnosis, support vector machine, optimization, machine learning, classification

Citation: Dou Y and Meng W (2021) An Optimization Algorithm for Computer-Aided Diagnosis of Breast Cancer Based on Support Vector Machine. Front. Bioeng. Biotechnol. 9:698390. doi: 10.3389/fbioe.2021.698390

Received: 21 April 2021; Accepted: 11 June 2021;
Published: 05 July 2021.

Edited by:

Zhiwei Luo, Kobe University, Japan

Reviewed by:

Krishna Chandra Persaud, The University of Manchester, United Kingdom
Laurent Simon, New Jersey Institute of Technology, United States

Copyright © 2021 Dou and Meng. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Wentao Meng, network0828@163.com

Download