Lontar - Template LONTAR KOMPUTER VOL. 13, NO. 2 AUGUST 2022 p-ISSN 2088-1541 DOI : 10.24843/LKJITI.2022.v13.i02.p01 e-ISSN 2541-5832 Accredited Sinta 2 by RISTEKDIKTI Decree No. 158/E/KPT/2021 72 Implementation of Sample Sample Bootstrapping for Resampling Pap Smear Single Cell Dataset Anita Desiania1, Sugandi Yahdina2, Azhar Kholiq Affandib3, Shania Putri Andhinia4, Yuli Andirania5, Muhammad Arhamic6 aMathematics Department, Mathematics and Natural Science faculty, Universitas Sriwijaya Jl. Palembang-Prabumulih Km 32, Inderalaya, Sumatera Selatan, Indonesia 1anita_desiani@unsri.ac.id (Corresponding author) 2sugandi@unsri.ac.id 4shania.andhini@gmail.com 5yuliandrianii@unsri.ac.id bPhysics Department, Mathematics and Natural Science faculty, Universitas Sriwijaya Jl. Palembang-Prabumulih Km 32, Inderalaya, Sumatera Selatan, Indonesia 3azharka@unsri.ac.id cInformatics Engineering Department, Politeknik Negeri Lhokseumawe Jl. Medan - Banda Aceh, Kota Lhokseumawe, Aceh, Indonesia 6muhammad.arhami@pnl.ac.id Abstract The purpose of this study was to determine how the effect of using Bootstrapping Samples for resampling the Harlev dataset in improving the performance of single-cell pap smear classification by dealing with the data imbalance problem. The Harlev dataset used in this study consists of 917 data with 20 attributes. The number of classes on the label had data imbalance in the dataset that affected single-cell pap smear classification performance. The data imbalance in the classification causes machine learning algorithms to produce poor performance in the minority class because they were overwhelmed by the majority class. To overcome it, The resampling data could be used with Sample Bootstrapping. The results of the Sample Bootstrapping were evaluated using the Artificial Neural Network and K-Nearest Neighbors classification methods. The classification used was seven classes and two classes. The classification results using these two methods showed an increase in accuracy, precision, and recall values. The performance improvement reached 10.82% for the two classes classification and 35% for the seven classes classification. It was concluded that Sample Boostrapping was good and robust in improving the classification method. Keywords: Pap Smear, Imbalance Data, Sample Boostrapping, Artificial Neural Networks, K- Nearest Neighbor. 1. Introduction The imbalance in the classification data causes machine learning algorithms to produce poor performance in the minority class because they are overwhelmed by the majority class [1]. Several studies have addressed data imbalance in several ways. The first is to change the class distribution through various resampling, and the second is to set different priorities by modifying the algorithm structure [2]. The problem of unbalanced data like that often occurs in some machine learning application research [1], [3]–[7]. The problem of data imbalance is related to the accuracy of predictions because predictions are biased towards the majority class, while sometimes, prediction accuracy in the minority class is also required. One solution to overcome the data imbalance is to use the resampling technique [7]. Resampling techniques have become a concern, especially in big data [8]–[11]. There are many ways to increase the accuracy of the LONTAR KOMPUTER VOL. 13, NO. 2 AUGUST 2022 p-ISSN 2088-1541 DOI : 10.24843/LKJITI.2022.v13.i02.p01 e-ISSN 2541-5832 Accredited Sinta 2 by RISTEKDIKTI Decree No. 158/E/KPT/2021 73 minority class based on the resampling technique because resampling can balance the number of minority classes with the majority class [3], [12], [13]. The Harlev dataset contains single-cell pap smear data, which has seven diagnostic classes. However, the dataset has a problem with an unbalanced number of classes, where the majority class is very far from the minority class. The Superficial Epithelial class has 74 data, the Intermediate Epithelial class has 70 data, the Columnar Epithelial class has 98 data, the Mild Light Dysplasia class has 182 data, the Severe Dysplasia class has 146 data, Moderate Dysplasia class has 197 data, and the last Carcinoma In Situ class has 150 data. The seven classes of labels can be grouped into two normal and abnormal groups. The use of 7 classes shows that the majority class is Moderate Dysplasia and the minority class is Epithelial which has a far data range of 127 data. Unbalanced data affects classification performance, accuracy, precision, and recall because it is difficult to find information on minority classes [14]. Meanwhile, the use of 2 classes also has a disproportionate amount of 242 for the normal category and 675 for the abnormal category. Several studies have shown that the use of resampling techniques can improve classification performance. One of the methods commonly used in resamples technique is Sample Bootstrapping. The Sample Bootstrapping method has several advantages. The method does not require any assumptions about the distribution of the data. It can resample the sample data up to thousands of times even though the number of samples was limited, and the method has simple calculation [15], [16]. Thanathamathee and Lursinap [12] used the Sample Bootstrapping method to resample data for classification on the Monk2 dataset. This study showed that the resample Sample Bootstrapping technique increased the accuracy value from 82.13% to 85.96%. Research from Al-Luhaybi et al. [17] also used resampling with the Sample Bootstrapping method to classify student datasets at Brunel University. The accuracy increased after the resampling technique was carried out from 75.59% to 93.1%. Several other studies also used sample bootstrapping as a resampling method for classification [18]–[21]. Research on the Herlev dataset was conducted by Kurniawati [22] that research applied SVM to cervical cancer classification with seven classes without using the resampling technique. This study resulted in a low accuracy value of 78.67%. The research from Kusy et al. [23] also classified cervical cancer data using an artificial neural network without resampling and resulted in an accuracy value that was still not good at 71.87%. Several studies on pap smears for detecting cervical cancer disorders used two classes, including Bora et al., study [24], which applied the KNN method. The results indicated that the accuracy, precision, and recall values were excellent above 80%. Likewise, research from Oka et al., [25] uses two classes with the Artificial Neural Network method gave excellent results of 88.8%. From this research, it could be seen that the performance of the classification method was influenced by an imbalance data problem. The Herlev data set had an unbalanced number of classes. This study focused on applying the resample technique using the Sample Bootstrapping method to classify cervical cancer. The Sample Bootstrapping method was applied to the Herlev pap smear single-cell data to classify the types of cervical cancer disorders. The results of the Sample Bootstrapping application were evaluated by the Artificial Neural Network and K-Nearest Neighbors classification methods to determine the extent to which Sample Bootstrapping was able to improve the performance of the classification method. 2. Research Methods For the training and testing process, the data was split by 10-fold cross-validation. Algorithm performance was measured based on the accuracy, precision, and recall of ANN and KNN methods. 2.1. Dataset The dataset used was the Harlev dataset developed by the pathology department of Harlev University Hospital with the Danish Technical University Automation department. This dataset consisted of 917 single cell images, which have been classified into seven classes by cyto- LONTAR KOMPUTER VOL. 13, NO. 2 AUGUST 2022 p-ISSN 2088-1541 DOI : 10.24843/LKJITI.2022.v13.i02.p01 e-ISSN 2541-5832 Accredited Sinta 2 by RISTEKDIKTI Decree No. 158/E/KPT/2021 74 technicians and specialists [26]. The dataset consists of 20 attributes which were described in Table 1. Table 1. Attribute Description On Harlev Dataset Attribute Attribute Description Data Type Kerne_A The area of the nucleus Real Cyto_A The area of the cytoplasm Real K/C Comparison of the area of the nucleus and cytoplasm Real Kerne_Ycol Nuclear light intensity degree Real Cyto_Ycol Cytoplasmic light intensity Real KerneShort The shortest diameter of the nucleus Real KerneLong The longest diameter of the nucleus Real KerneElong Nuclear stretching Real Kerne Rund Nuclear roundabout Real CytoShort The shortest diameter of the cytoplasm Real CytoLong The longest diameter of the cytoplasm Real CytoElong Cytoplasm stretching Real CytoRund Cytoplasmic roundness Real KernePeri Nucleus Limit Real KernePos Nucleus position Real KerneMax Maximum number of nuclear pixels Integer KerneMin Minimum number of pixel nucleus Integer Cyto_Max Maximum number of cytoplasmic pixels Integer CytoMin Maximum number of cytoplasmic pixels Integer Class Diagnosis (Cell type) Polynomial As previously mentioned, there were seven classes in this dataset, namely Superficial Epithelial, Intermediate Epithelial, Columnar Epithelial, Mild Light Dysplasia, Severe Dysplasia, Moderate Dysplasia, and Carcinoma In Situ. Cervical cells were grouped into two categories, namely normal and abnormal categories, the types of cells that fall into the normal and abnormal categories could be seen in Table 2 [27]. Table 2. Cell Type in The Harlev Dataset Categories Class Cell Type Normal 1 Superficial Epithelial 2 Intermediate Epithelial 3 Columnar Epithelial Abnormal 4 Mild Light Dysplasia 5 Severe Dysplasia 6 Moderate Dysplasia 7 Carcinoma In Situ From Table 1, the normal category has three types of cell types, and the abnormal category has four types of cell types. Although some cells have the same category, each cell type has a different cell shape. Cell shapes for each category can be seen in Figures 1 and 2. a b C Figure 1. Images of Cervical Cells in Normal Category (a)Superficial Epithelial (b) Intermediate Epithelial (c) Columnar Epithelial LONTAR KOMPUTER VOL. 13, NO. 2 AUGUST 2022 p-ISSN 2088-1541 DOI : 10.24843/LKJITI.2022.v13.i02.p01 e-ISSN 2541-5832 Accredited Sinta 2 by RISTEKDIKTI Decree No. 158/E/KPT/2021 75 Figure 1 showed the shape of cells that fall into the normal category, which consisted of 3 classes. Then Figure 2 showed the shape of cells that fall into the abnormal category, which consisted of 4 classes. a b c d Figure 2. Images of Cervical Cells in Abnormal Category (a) Mild Light Dyplasia (b) Severe Dysplasia (c) Moderate Dysplasia (d) Carcinoma In Situ. 2.2. Implementation of Sample Sample Bootstrapping The implementation of Sample Sample Boostrapping (SB) was in the pre-processing stage, which occurred before the data entered the classification process. Sample Bootstrapping was a method used to estimate the deviation from the standard error [28]. Sample Boostrapping used statistical procedures by changing the data from the existing sample and replicating the sample data (resampling) randomly to get new simulation data. The Sample Boostrapping took samples with the replacement method, which replaced the original data randomly with a specific label. The data in the process had an equal chance of being selected. The data could be re-selected in the following process [29]. Based on several studies, the advantages of Sample Boostrapping were the ability to study any statistic of interest and handle sampling error by creating a specific model [30]. The working steps in Sample Boostrapping could not reduce data errors but only estimated standard errors in the data [20]. The steps on the Sample Boostrapping method were [31]: a. Construct a distribution of n Sample Bootstrapping sample (�̂�) by assigning a probability of 1/n to each data (Xi ) for i=1,2,3,…,n. b. Take a Sample Bootstrapping sample of size n at random with the return of the distribution of stage 1. c. Choose the replication of each sample ( �̂�) statistic from the Sample Bootstrapping sample was referred to as �̂�1 ∗ . d. Repeat steps 2 and 3 until B times, so you got �̂�1 ∗, �̂�2 ∗, …, �̂�𝐵 ∗ . e. Estimate the standard error (seB) using the standard deviation B times with Equation 1. 𝑠𝑒𝐵 = {∑ [ �̂� ∗ (𝑏) − �̂� ∗ (. )] 2𝐵 𝑏=1 } 1/2 (1) Where �̂�𝑃∗𝑝(. ) = ∑ �̂� ∗ (𝑏)/𝐵𝐵𝑏=1 (2) 2.3. Evaluation of Sample Bootstrapping Using Classification Methods To evaluate the performance of Sample Boostrapping samples, it would be implemented to classification methods and analyzed the performance result of the classification. Classification is the process of forming a model to predict an unknown class pattern [32]. In this study, the methods used in the classification for evaluation were the ANN and KNN method. 2.3.1. Artificial Neural Networks Artificial Neural Networks (ANN) was a method that could process large data [33]. The ANN consisted of several layers, namely the input layer, hidden layer, and output layer. Hidden layers used were one hidden layer with learning rates of 0.01 and training cycles of 200. The form of the ANN architecture used could be seen in Figure 3. LONTAR KOMPUTER VOL. 13, NO. 2 AUGUST 2022 p-ISSN 2088-1541 DOI : 10.24843/LKJITI.2022.v13.i02.p01 e-ISSN 2541-5832 Accredited Sinta 2 by RISTEKDIKTI Decree No. 158/E/KPT/2021 76 Figure 3. ANN Architecture on Pap smear Single Cell Classification This study used ANN Backpropagation. This algorithm was one of the most frequently used algorithms [34]. The stages of the ANN Backpropagation method were [35]: a. Initialize all weights with a small random number. b. Enter to feedforward stage. c. Forward Input Data ( i = 1,2,…,n ) from input layer to the hidden layer. d. Calculate the forwarded data (zj) on the hidden layer with Equation 3. 𝑧_𝑖𝑛𝑗 = 𝑣0𝑗 + ∑ 𝑣𝑖𝑗 𝑥𝑖 𝑛 𝑖=1 (3) Where 𝑧_𝑖𝑛𝑗 was the j-th data input (j=1,2,3,…,n) in the hidden layer, 𝑣0𝑗 was the weight value for bias for unit zj and 𝑣𝑖𝑗 is the weight value for unit xi. Then the data that came out of the hidden layer to the output layer was calculated by Equation 4. 𝑧𝑗 = 𝑓(𝑧_𝑖𝑛𝑗 ) (4) where s was the activation function used in the hidden layer. After all the data was calculated, then proceed to the next layer. e. Calculate the forwarded data on the output layer with Equation 5. 𝑦_𝑖𝑛𝑘 = 𝑤0𝑘 + ∑ 𝑤𝑗𝑘 𝑧𝑗 𝑚 𝑗=1 (5) where 𝑦_𝑖𝑛𝑘 Was the k-th data input at the output layer d, and w0k was the weight for bias to the output unit. f. Calculate the data that comes out as output with Equation 6. 𝑦𝑘 = 𝑓(𝑦_𝑖𝑛𝑘 ) (6) g. Prepare for the backpropagation stage. h. Calculate all output data (𝑦𝑘 , k=1,2,3,…,m) for all target patterns. Calculate the factor error (𝛿) with Equation 6. 𝛿𝑘 = (𝑡𝑘 − 𝑦𝑘 )𝑓′(𝑦_𝑖𝑛𝑘 ) (7) Where 𝛿𝑘 was the error used when the layer weight changes, 𝑡𝑘 was the output target. Next, update the weight value of 𝑤𝑗𝑘 by calculating the change in weight using acceleration 𝛼 using Equation 8. ∆𝑤𝑗𝑘 = 𝛼𝛿𝑘 𝑧𝑗 (8) Update the value of the bias b by calculating the value of the change in bias using Equation 9. ∆𝑤𝑘0 = 𝛼𝛿𝑘 (9) Then the calculated value was sent to the previous layer. i. Calculate each input data from the output layer with Equation (10). 𝛿_𝑖𝑛𝑗 = ∑ 𝛿𝑘 𝑚 𝑘=1 𝑤𝑗𝑘 (10) The input data that has been obtained would be multiplied by the inverse function of the activation function using equation 11. 𝛿𝑗 = 𝛿𝑖𝑛𝑗 𝑓 ′ (𝑧𝑖𝑛 𝑗 ) (11) j. Calculate the weight changes with equation 12 and the value of the change in bias to update the weight and bias values in the hidden layer. ∆𝑣𝑗𝑖 = 𝛼𝛿𝑗 𝑥𝑖 (12) ∆𝑣𝑗0 = 𝛼𝛿𝑗 (13) LONTAR KOMPUTER VOL. 13, NO. 2 AUGUST 2022 p-ISSN 2088-1541 DOI : 10.24843/LKJITI.2022.v13.i02.p01 e-ISSN 2541-5832 Accredited Sinta 2 by RISTEKDIKTI Decree No. 158/E/KPT/2021 77 k. Update the bias and weight values on the output layer with equation 14 and the hidden layer with equation 15. 𝑤𝑗𝑘 (𝑛𝑒𝑤) = 𝑤𝑗𝑘 (𝑝𝑟𝑒𝑣𝑖𝑜𝑢𝑠) + ∆𝑤𝑘𝑗 (14) 𝑣𝑗𝑖 (𝑛𝑒𝑤) = 𝑣𝑗𝑖 (𝑝𝑟𝑒𝑣𝑖𝑜𝑢𝑠) + ∆𝑣𝑗𝑖 (15) l. Repeat Steps a-i for each training data. m. Perform steps a-j for each iteration. 2.3.2. K-Nearest Neighbors The second method used for classification was K-Nearest Neighbors (KNN). KNN was an algorithm that worked on the shortest distance from the query instance to the training sample. The goal was to classify an object based on attributes and training samples. For this research, the value of k used was k=5. The following are the steps in the KNN algorithm [36]: a. Determine the parameter k to be used. b. Calculate the distance between the new data and all training data using Euclidean Distance with Equation 16 [32]. 𝑑𝑖 = √∑ (𝑥2𝑖 − 𝑥1𝑖 ) 2𝑛 𝑖=1 (16) Where di was defined as the distance between 𝑥1 and 𝑥2, 𝑥1 was the sample data and 𝑥2 was the test data, i was the data variable. c. Sort the distance calculation results from the smallest to the largest and determine the nearest neighbor based on the kth minimum distance. d. Claim the class wher it was taken based on the highest number of class. 2.4. Algorithm Performance Assessment Algorithm performance assessment was based on the confusion matrix that appeared after the training and testing process. There were two classes in the confusion matrix, namely positive class and negative class. The true positive (TP) was a positive class that was guessed correctly. The False positive (FP) was a positive class that was guessed wrong. As well as true negative (TN) was a negative class that was guessed right, and The False Negative (FN)was a negative class that was guessed wrong. If the case had more than two classes, one class became a positive class, and the rest became a negative one. The confusion matrix calculated the accuracy value used to measure the accuracy of the classification results. In addition, from the confusion matrix, the precision value used to calculate the accuracy of the prediction results against the requested information could also be calculated and the recall value used to calculate the ratio of the selected relevant items to the actual value[37]. Accuracy = 𝑇𝑃+𝑇𝑁 𝑇𝑃+𝐹𝑃+𝑇𝑁+𝐹𝑁 x 100% (17) Precision = 𝑇𝑃 (𝐹𝑃+𝑇𝑃) x 100% (18) Recall = 𝑇𝑃 (𝑇𝑃+𝐹𝑁) x 100% (19) 3. Result and Discussion In this study, sample Sample Bootstrapping was used for resampling the single-cell pap smear dataset before classification using ANN and KNN. The steps taken were the research method described above, preprocessing using the SB method for sampling. Then, the classification method used was ANN and KNN. in the sampling process, the parameters used include relative or ration (0-1). Furthermore, the resampled data was validated using n-fold cross-validation. The value of n used was 10-fold. The data set was divided into ten partitions, nine partitions as training data, and one partition for test data. This process was repeated ten times for each section so that every part of the ten sections had become testing data. These stages could be seen in the form of a flowchart which could be seen in Figure 4. LONTAR KOMPUTER VOL. 13, NO. 2 AUGUST 2022 p-ISSN 2088-1541 DOI : 10.24843/LKJITI.2022.v13.i02.p01 e-ISSN 2541-5832 Accredited Sinta 2 by RISTEKDIKTI Decree No. 158/E/KPT/2021 78 A comparative test of results was carried out in this study to find out the result of applying the Sample Bootstrapping method. The results were compared by comparing the test results on the original data (which does not use Sample Bootstrapping) with the test results obtained in the Sample Bootstrapping implementation. This test was carried out on a dataset group with two classes and a group of 7 classes. In the classification of 7 classes, the data were grouped into seven classes based on all types of pap smear cell images in table 2. in the classification of 2 classes, it grouped from 7 classes into two categories normal and abnormal in table 2. Figure 4. The Flowchart on The Proposed Method 3.1. Classification without Sample Sample Bootstrapping The classification method was used to analyze the result of sample Sample Bootstrapping(SB) was implemented. To be fair, all parameters in both methods used the same parameters when using sample Sample Boostrapping and without it. The number of data before using SB could be seen in Figure 5. Figure 5. The Number of Data by Cell Type on Original Single Cell Pap Smear Dataset The number of data by categories on pap smear data set on the normal class had 242 data, and the abnormal class had 675 data. By 10-fold cross-validation, the classification results using the original dataset or without using the SB could be seen in Table 4. Table 4. Classification Result Without Sample Sample Boostrapping (%) Classification ANN KNN Accuracy Precision Recall Accuracy Precision Recall 2 classes 93,90 92,83 92,16 91,28 90,76 86,72 7 classes 62,70 67,74 67,43 50,59 53,90 52,87 LONTAR KOMPUTER VOL. 13, NO. 2 AUGUST 2022 p-ISSN 2088-1541 DOI : 10.24843/LKJITI.2022.v13.i02.p01 e-ISSN 2541-5832 Accredited Sinta 2 by RISTEKDIKTI Decree No. 158/E/KPT/2021 79 From Table 5, it could be seen that the 7 class classifications had smaller accuracy, precision, and recall values. It could have happened because the data spread in 7 classes was smaller than two classes. 3.2. Classification with Sample Bootstrapping The resampled dataset had the same size as the original dataset. In this study, the number of samples was determined, namely relative to the ratio 1. Resampling with SB was done five times. The number of each class after passing the SB process could be seen in Figure 6. Figure 6. Amount of Data by Cell Type After Five Times Resampling with Sample Sample Bootstrapping The results of the Sample Boostrapping sample performance on the ANN and KNN methods with two types of classification could be seen in Table 5. Based on Figure 6, it could be seen that the minority class experienced an increase in the amount of data so that the number between classes was not too far apart even though the majority class still had a wide range of values. The number of data by categories became 269 for normal class and 648 for abnormal class. Table 5. Classification Result Using Sample Sample Boostrapping (%) Classification ANN KNN Accuracy Precision Recall Accuracy Precision Recall 2 classes 96,61 96,52 95,52 97,49 96,63 97,54 7 classes 81,13 84,87 82,42 86,49 87,77 87,54 Based on Table 5, classification in 7 classes still produced lower performance scores than two classes, but the results were much better than before. A comparison of results on the use of SB was discussed further in the next section. 3.3. Comparison of Results To see more clearly, the comparison on the use of SB was be divided into 2 Tables. There was a Table for comparison of 2 classes and the other for a comparison of 7 classes. For comparison in the two classes could be seen in Table 6 below. Table 6. Comparison of the results of the Single Cell Pap Smear classification in 2 classes (%) Classification Methods ANN KNN Accuracy Precision Recall Accuracy Precision Recall Without SB 93,90 92,83 92,16 91,28 90,76 86,72 With SB 96,61 96,52 95,52 97,49 96,63 97,54 Difference 2,71 3,69 3,36 6,21 5,87 10,82 The SB method of classification could increase the accuracy value, especially the recall value, because the higher the recall value, the better the machine in finding information about a class. The recall value increased due to the machine recognizing the minority class, which was previously biased towards the majority class. Based on the classification method used, the SB LONTAR KOMPUTER VOL. 13, NO. 2 AUGUST 2022 p-ISSN 2088-1541 DOI : 10.24843/LKJITI.2022.v13.i02.p01 e-ISSN 2541-5832 Accredited Sinta 2 by RISTEKDIKTI Decree No. 158/E/KPT/2021 80 method worked well on KNN compared to ANN. It could be seen from the value of the increase that occurred in KNN reaching more than 5% in all performance values. In Table 6, it could be seen in the classification of 2 classes. Furthermore, to see the comparison of the seven classes could be seen in Table 7. Table 7. The Comparison of the results of the Single Cell Pap Smear classification in 7 classes (%) Classification Methods ANN KNN Accuracy Precision Recall Accuracy Precision Recall Without SB 62,70 67,74 67,43 50,59 53,90 52,87 With SB 81,13 84,87 82,42 86,49 87,77 87,54 Difference 18,43 17,44 14,99 35,9 33,87 34,67 In Table 7, the SB method was much better in the 7 class classification because it could be seen from the difference, which was quite far up to 35%. Significant accuracy, precision, and recall values indicate that resampling using SB greatly improved classification results on unbalanced data. Although the amount of data generated (Figure 6) was still not very balanced, it has made an excellent classification. Based on the method used, KNN also had a better performance value than ANN. It could be seen that the highest difference was in the KNN method. Although the performance in 7 classes has increased, the numbers produced were not as good as the 2 class classification. However, the SB method worked very well on the KNN method on both classifications because it has increased the accuracy value considerably. This showed that SB was very good at improving the performance of the KNN method for seven classes. To analyze the results of this study further, a comparison of the results with previous studies was carried out. The Comparison of research results for single-cell pap smear classification could be seen in Table 8. Table 8. The Comparison of Results on Classification of Single Cell Pap Smear Author / Dataset / Methods The Increasing Accuracy The Increasing Presicion The Increasing Recall Zughrat et. Al (2014)/Rail Dataset/SB-SVM [1] 18,2% - - Thanathamathee and Lursinsap(2013)/ Monk2/SB-ANN [12] 8,98% - - Thanathamathee and Lursinsap(2013)/ Abalone/SB-ANN [12] 3,14% - - Saez et. Al (2015)/Abalone/Smote-C.45 [38] 3% - - Arifin and Rachman(2020)/Harlev Dataset 2 Classes/DecisionTree-PSO [39] 5,37% - - Proposed Method/Harlev Dataset 2 Classes/SB-ANN 2,71% 3,69% 3,36% Proposed Method/Harlev Dataset 2 Classes/SB-KNN 6,21% 5,87% 10,82% Proposed Method/Harlev Dataset 7 Classes/SB-ANN 18,43% 17,44% 14,99% Proposed Method/Harlev Dataset 7 Classes/SB-KNN 35,9% 33,87% 34,67% In Table 8, it could be seen that several studies used resampling techniques for unbalanced data. In other studies, it only showed differences in accuracy values as it was known that the accuracy value was not enough to determine an algorithm works well. If there is so much the number of majority classes, the machine could only predict the majority class. In contrast, a good accuracy value could occur because the number of minorities that could not be predicted is few. From Table 8, it could also be seen that only the proposed method could display other performance values LONTAR KOMPUTER VOL. 13, NO. 2 AUGUST 2022 p-ISSN 2088-1541 DOI : 10.24843/LKJITI.2022.v13.i02.p01 e-ISSN 2541-5832 Accredited Sinta 2 by RISTEKDIKTI Decree No. 158/E/KPT/2021 81 and was the advantage of this research. In addition, this study had the highest improvement value compared to the others. Although the classification of 2 classes using the ANN method was still lower than Arifin and Rachman's [39] research, this study had an increase in other performance values not shown in that study. From this comparison, it could be concluded that SB was very good and robust in improving the classification method. 4. Conclusion The sample Sample Bootstrapping method was very good and robust for resampling on an imbalanced data problem. That is indicated by the improved classification performance of the study. The highest increase occurred using the KNN method in the classification of 2 classes and seven classes. The highest difference value is the KNN method on seven classes classification, with an increasing value is 35.9% for accuracy, 33.87% for precision, and 34.67% for recall. With a significant increase, it can be concluded that sample Sample Bootstrapping can improve the classification of labels that have many classes. References [1] A. Zughrat, M. Mahfouf, Y. Y. Yang, and S. Thornton, “Support Vector Machines for Class Imbalance Rail Data Classification with Sample Boostrapping- based Over-Sampling and Under-Sampling, ” IFAC Proceedings Volumes, vol. 47, no. 3. IFAC, 2014. [2] A. Tharwat, T. Gabel, and T. Gabel, “Parameters optimization of support vector machines for imbalanced data using social ski driver algorithm,” Neural Computing and Applications, vol. 0123456789, 2019, doi: 10.1007/s00521-019-04159-z. [3] R. Ghorbani and R. Ghousi, “Comparing Different Resampling Methods in Predicting Students ’ Performance Using Machine Learning Techniques,” IEEE Access, vol. 8, pp. 67899–67911, 2020, doi: 10.1109/ACCESS.2020.2986809. [4] J. . Sanz, D. Bernardo, F. Herrera, H. Bustince, and H. Hagras, “A compact evolutionary interval-valued fuzzy rule-based classification system for the modeling and prediction of real- world financial applications with imbalanced data,” IEEE Transactions on Fuzzy Systems, vol. 23, no. 4, pp. 973–990, 2015. [5] W. Wei, J. Li, L. Cao, Y. Ou, and J. Chen, “Effective detection of sophisticated online banking fraud on extremely imbalanced data,” World Wide Web, vol. 16, no. 4, pp. 449–475, 2013. [6] H. Yu, J. Ni, and J. Zhao, “ACOSampling: An ant Colony Optimizaiton-Based Undersampling Method for Classifying Imbalanced DNA Microarray Data,” Neurocomputing, vol. 101, pp. 309–318, 2013. [7] T. Sasada, Z. Liu, T. Baba, K. Hatano, and Y. Kimura, “A resampling method for imbalanced datasets considering noise and overlap,” Procedia Computer Science, vol. 176, pp. 420– 429, 2020, doi: 10.1016/j.procs.2020.08.043. [8] I. Triguero, S. del Rio, V. L´opez, J. Bacardit, J. . Ben´ıtez, and F. Herrera, “ROSEFW-RF: The winner algorithm for the ECBDL14 big data competition. An extremely imbalanced big data bioinformatics problem,” Knowledge-Based Systems, vol. 87, pp. 69–79, 2015. [9] M. Koziarski and M. Wozniak, “CCR: A combined cleaning and resampling algorithm for imbalanced data classification,” International Journal of Applied Mathematics and Computer Science, vol. 27, no. 4, pp. 727–736, 2017, doi: 10.1515/amcs-2017-0050. [10] T. . Hoens, R. Polikar, and N. . Chawla, “Learning from streaming data with concept drift and imbalance: An overview,” Progress in Artificial Intelligence, vol. 1, no. 1, pp. 89–101, 2012. [11] F. Fernandez-Navarro, C. Hervas-Martinez, and P. . Gutierrez, “A dynamic over-sampling procedure based on sensitivity for multi-class problems,” Pattern Recognition, vol. 44, no. 8, pp. 1821–1833, 2011. [12] P. Thanathamathee and C. Lursinsap, “Handling Imbalanced Data Sets with Synthetic Boundary Data Generation Using Sample Bootstrapping Re-sampling and AdaBoost Techniques,” Pattern Recognition Letters, vol. 34, no. 12, pp. 1339–1347, 2013, doi: LONTAR KOMPUTER VOL. 13, NO. 2 AUGUST 2022 p-ISSN 2088-1541 DOI : 10.24843/LKJITI.2022.v13.i02.p01 e-ISSN 2541-5832 Accredited Sinta 2 by RISTEKDIKTI Decree No. 158/E/KPT/2021 82 10.1016/j.patrec.2013.04.019. [13] A. Elhassan, M. Aljourf, F. Al-Mohanna, and M. Shoukri, “Classification of Imbalance Data using Tomek Link ( T-Link ) Combined with Random Under-sampling ( RUS ) as a Data Reduction Method Technology & Optimization,” Global Journal of Technology & Optimization, vol. 1, no. 111, pp. 1–11, 2017, doi: 10.4172/2229-8711.S1. [14] A. Desiani, S. Yahdin, and A. Kartikasari, “Handling the imbalanced data with missing value elimination SMOTE in the classification of the relevance education background with graduates employment,” IAES International Journal of Artificial Intelligence, vol. 10, no. 2, pp. 346–354, 2021, doi: 10.11591/ijai.v10.i2.pp346-354. [15] I. Ivanov, “Tenfold Boostrap Procedure for Support Vector Machine,” Computer Science, vol. 21, no. 2, pp. 253–268, 2020. [16] I. Rodliyah, “Perbandingan Metode Sample Bootstrapping Dan Jackknife ( Comparison of Sample Bootstrapping and Jackknife Methods To,” Jurnal Matematika dan Pendidikan Matematika, vol. 1, no. 1, pp. 76–86, 2016. [17] M. Al-Luhaybi, L. Yousefi, S. Swift, S. Counsell, and A. Tucker, Predicting academic performance: A Sample Boostrapping approach for learning dynamic bayesian networks, vol. 11625 LNAI. Springer International Publishing, 2019. [18] T. Agus, S. M. Adib, and A. Karomi, “Penerapan Metode Sample Sample Boostrapping untuk Meningkatkan Performa kNearest Neighbor pada Dataset Berdimensi Tinggi,” IC-Tech, vol. XII, no. 1, pp. 9–14, 2017. [19] T. A. Setiawan, R. Satria, and A. Syukur, “Integrasi Metode Sample Sample Boostrapping dan Weighted Principal Component Analysis untuk Meningkatkan Performa k Nearest Neighbor pada Dataset Besar,” Journal of Intelligent Systems, vol. 1, no. 2, pp. 76–81, 2015. [20] E. Siswanto, Suprapedi, and Purwanto, “Metode Sample Boostraping Pada K-Nearest Neighbor Untuk Klasifikasi Status Desa,” Jurnal Teknologi Informasi, vol. 14, no. 1, pp. 13– 23, 2018. [21] E. Jumiati and M. R. Kamal, “Integrasi Sample Sample Boostrapping Pada K-Nearest Neighbor untuk Klasifikasi Herregistrasi Calon Mahasiswa Baru,” IC-Tech, vol. 12, no. 1, pp. 23–32, 2017. [22] Y. E. Kurniawati, A. E. Permanasari, and S. Fauziati, “Comparative study on data mining classification methods for cervical cancer prediction using pap smear results,” Proceeding 2016 1st International Conference on Biomedical Engineering (IBIOMED) 2016, 2017, doi: 10.1109/IBIOMED.2016.7869827. [23] M. Kusy, B. Obrzut, and J. Kluska, “Application of gene expression programming and neural networks to predict adverse events of radical hysterectomy in cervical cancer patients,” Medical & Biological Engineering & Computing, vol. 51, no. 12, pp. 1357–1365, 2013, doi: 10.1007/s11517-013-1108-8. [24] K. Bora, M. Chowdhury, and L. B. Mahanta, “Automated classification of Pap smear images to detect cervical dysplasia,” Comput Methods Programs Biomed, vol. 138, pp. 31–47, 2017, doi: 10.1016/j.cmpb.2016.10.001. [25] N. P. A. Wiastini Oka, I. K. G. Darma Putra, and K. S. Wibawa, “Klasifikasi Sel Nukleus Pap Smear Menggunakan Metode Backpropagation Neural Network,” Jurnal Ilmiah Merpati, vol. 7, no. 3, pp. 182–192, 2019. [26] D. Riana, D. H. Widyantoro, T. Latifah, and R. Mengko, “Ekstraksi dan Klasifikasi Tekstur Citra Sel Nukleus Pap Smear,” Jurnal TICOM, vol. 1, no. 3, pp. 62–70, 2013. [27] Y. Ramdhani and D. Riana, “Hierarchical Decision Approach Based on Neural Network and Genetic Algorithm Method for Single Image Classification of Pap Smear,” Second International Conference on Informatics and Computing (ICIC), pp. 1–6, 2017, [Online]. Available: doi: 10.1109/IAC.2017.8280587. [28] R. E. McRoberts, S. Magnussen, E. O. Tomppo, and G. Chirici, “Parametric, Sample Bootstrapping, and jackknife variance estimators for the k-Nearest Neighbors technique with LONTAR KOMPUTER VOL. 13, NO. 2 AUGUST 2022 p-ISSN 2088-1541 DOI : 10.24843/LKJITI.2022.v13.i02.p01 e-ISSN 2541-5832 Accredited Sinta 2 by RISTEKDIKTI Decree No. 158/E/KPT/2021 83 illustrations using forest inventory and satellite image data,” Remote Sensing of Environment, vol. 115, no. 12, pp. 3165–3174, 2011, doi: 10.1016/j.rse.2011.07.002. [29] T. Siswanto, “Optimalisasi Sosial Media Sebagai Media Pemasaran Usaha Kecil Menengah,” Liquidity, vol. 2, no. 1, pp. 80–86, 2018, doi: 10.32546/lq.v2i1.134. [30] L. R. Zientek and B. Thompson, “Applying the Sample Bootstrapping to the multivariate case : Sample Bootstrapping component / factor analysis,” Behavior Research Methods, vol. 39, no. 2, pp. 318–325, 2007. [31] H. lin Shang, “Resampling Techniques for Estimating the Distribution of Descriptive Statistics of Functional Data,” Communication in Statistics-Simulation and Computation, vol. 44, no. 3, pp. 614–635, 2015, [Online]. Available: doi: 10.1080/03610918.2013.788703. [32] N. L. W. S. R. Ginantra, “Deteksi Batik Parang Menggunakan Fitur Co-Occurence Matrix Dan Geometric Moment Invariant Dengan Klasifikasi KNN,” Lontar Komputer : Jurnal Ilmiah Teknologi Informasi, vol. 7, no. 1, p. 40, 2016, doi: 10.24843/lkjiti.2016.v07.i01.p05. [33] M. Hasanipanah, M. Noorian-Bidgoli, D. Jahed Armaghani, and H. Khamesi, “Feasibility of PSO-ANN model for predicting surface settlement caused by tunneling,” Engineering with Computers, vol. 32, no. 4, pp. 705–715, 2016, doi: 10.1007/s00366-016-0447-0. [34] D. Kristianto, C. Fatichah, B. Amaliah, and K. Sambodho, “Prediction of Wave-induced Liquefaction using Artificial Neural Network and Wide Genetic Algorithm,” Lontar Komputer : Jurnal Ilmiah Teknologi Informasi, vol. 8, no. 1, p. 1, 2017, doi: 10.24843/lkjiti.2017.v08.i01.p01. [35] D. Graupe, Principles of Artificial Neural Networks (2nd Edition), vol. 53, no. 9. University of Illinois, Chicago,USA, 2007. [36] R. Apurb, S. Milan, A. Avi, and R. Dundigalla, “Heart disease prediction using machine learning classifiers,” International Journal of Advanced Science and Technology, vol. 29, no. 6, pp. 1700–1707, 2020, doi: 10.37200/IJPR/V24I6/PR260661. [37] S. Yahdin, A. Desiani, N. Gofar, K. Agustin, and D. Rodiah, “Application of the Relief-f Algorithm for Feature Selection in the Prediction of the Relevance Education Background with the Graduate Employment of the Universitas Sriwijaya,” Computer Engineering and Applications (ComEngApp), vol. 10, no. 2, pp. 71–80, 2021. [38] J. A. Saez, J. Luengo, J. Stefanowski, and F. Herrera, “SMOTE–IPF: Addressing the noisy and borderline examples problem in imbalanced classification by a re-sampling method with filtering,” Information Sciences, vol. 291, pp. 184–203, 2015. [39] T. Arifin and R. Rachman, “Optimasi Decision Tree Menggunakan Particle Swarm Optimization Untuk Klasifikasi Sel Pap Smear,” (JATISI) Jurnal Teknik Informatika dan Sistem Informasi, vol. 7, no. 3, pp. 572–579, 2020.