Microsoft Word - ETASR_V10_N6_pp6533-6541 Engineering, Technology & Applied Science Research Vol. 10, No. 6, 2020, 6533-6541 6533 www.etasr.com Bahig et al.: Speeding up the Multiplication Algorithm for Large Integers Speeding up the Multiplication Algorithm for Large Integers Hazem M. Bahig Computer Science and Information Dept. College of Computer Science and Engineering, University of Ha’il Ha’il, Saudi Arabia h.bahig@uoh.edu.sa Amer Alghadhban Electrical Engineering Dept. College of Engineering University of Ha’il Ha’il, Saudi Arabia a.alghadban@uoh.edu.sa Mohammed A. Mahdi Computer Science and Information Dept. College of Computer Science and Engineering, University of Ha’il Ha’il, Saudi Arabia m.mahdi@uoh.edu.sa Khaled A. Alutaibi Computer Engineering Dept. College of Computer Science and Engineering University of Ha’il Ha’il, Saudi Arabia alutaibi@uoh.edu.sa Hatem M. Bahig Mathematics Department Faculty of Science Ain Shams University Cairo, Egypt h.m.bahig@gmail.com Abstract-Multiplication is one of the basic operations that influence the performance of many computer applications such as cryptography. The main challenge of the multiplication operation is the cost of the operation as compared to other basic operations such as addition and subtraction, especially when the size of the numbers is large. In this work, we investigate the use of the window strategy for multiplying a sequence of large integers to design an efficient sequential algorithm in order to reduce the number of bit-multiplication operations involved in multiplying a sequence of large integers. In our implementation, several parameters are considered and measured for their effect on the proposed algorithm and the best-known sequential algorithm in the literature. These parameters are the size of the sequence of integers, the size of the integers, the size of the window, and the distribution of the data. The experimental results prove the effectiveness of the proposed algorithm are compared to the ones of the best-known sequential algorithm, and the proposed algorithm is able to achieve a reduction in computing time greater than 50% in most cases. Keywords-multiplication; big data; cryptography; algorithm performance; computer arithmetic I. INTRODUCTION Computer arithmetic plays an essential role in every layer of computing and it is an important consideration when developing computer solutions for many problems such as cryptography, image processing, and numerical computations. In computer arithmetic, we use different operations such as addition, subtraction, multiplication, and division to achieve the goal of computation. Among these, the operation that has particular significance for many applications is the multiplication operation [1]. The multiplication operation is important, mainly for three reasons. First, the time cost of performing the multiplication operation is greater than that of other operations such as addition and subtraction. For example, given two integer numbers of �-bits each, the addition and multiplication of the two integer numbers require ���� and ����� bit operations, respectively, using the Naïve method [1]. This means that there is a significant difference between the costs of the two operations. Second, many primitive and essential arithmetic operations, such as division, squaring, inverse multiplication, and modulo operations, are based on the multiplication operation. Therefore, the running time of the multiplication operation affects these operations. Third, several complex applications in computer science, such as cryptography and digital signal processing, are based on a huge number of multiplication operations [2-6]. For example, in RSA and El-Gamal public-key cryptosystem, the multiplication operation is necessary. So, a more efficient multiplication method would lead to the speeding up of the computation process in complex applications. For the above reasons, different strategies have been suggested to reduce the total number of operations required for multiplication. Two main research directions have been followed to improve the efficiency of the multiplication operation on a data set that consists of � integer numbers each of size �. The first direction has been to reduce the cost of multiplying two numbers, � and �, of size � each and thereby decrease the total cost of the multiplication operations for a data set. The second direction has been to reduce the cost of multiplying the data set by proposing an efficient strategy to multiply the � numbers. Regarding the first research direction, many methods have been proposed to reduce the time complexity of multiplying two integers in both sequential [7- 23] and parallel computation [24-29]. In the case of sequential computation, several techniques have been proposed such as the Naïve multiplication algorithm [1], Karatsuba’s algorithm Corresponding authors: Hazem M. Bahig Engineering, Technology & Applied Science Research Vol. 10, No. 6, 2020, 6533-6541 6534 www.etasr.com Bahig et al.: Speeding up the Multiplication Algorithm for Large Integers [1,15], the Toom–Cook multiplication algorithm [20], and a fast Fourier transform-based algorithm [18]. The time complexity of the Naïve multiplication method is ����� , whereas Karatsuba’s algorithm uses a divide and conquer strategy to multiply the two integers in ��� �, where = log� 3 ≈ 1.585. The Toom–Cook multiplication method is a generalization of Karatsuba’s algorithm using r-way multiplication and has a cost of � ������� ���� �⁄ !". In contrast, Schonhage and Strassen utilize the fast Fourier transform to reduce the time complexity of multiplication. They propose two algorithms, where the best one runs on ��� log� loglog�� and uses an arithmetical modulo operation. Another two algorithms are proposed to reduce the running time of the Schonhage-Strassen algorithm to achieve ��� log� 2�����∗ ��!, where log∗ � = �%�%�&�'%: )*+�,�� ≤ 2. and )*+�/�� = �. The first algorithm is based on arithmetic over complex numbers [12], while the other is based on modular arithmetic [10]. On the other hand, many different attempts have been made to parallelize the multiplication problem using different parallel models [24-29]. Most of these attempts have been based on the shared memory model, where the processors in this model communicate through shared memory. Also, some research studies have focused on implementing some parallel algorithms on real machines, such as FPGAs, GPUs, and multicore [25- 27]. In the case of the second research direction, a few algorithms have been proposed to reduce the time complexity of the multiplication of � integers in both sequential and parallel computation. In the case of sequential computation, the best-known algorithm is the Naïve method, which scans the sequence of integer numbers and multiplies one number in each iteration. In the case of parallel computation, a few strategies [30] have been proposed that use a shared memory model and are implemented on specific real machines such as the multicore system [31]. In this paper, we are interested in contributing to the second research direction that focuses on using sequential computation because, despite the progress that has been made toward developing an effective strategy, there is still room for improvement. Here, we present an efficient improvement sequential algorithm to multiply a large number of integers, each of large size. A comparison of the proposed algorithm and the best-known sequential algorithm indicates that, from a practical perspective, the proposed algorithm performs better. Our improved algorithm speeds up the best-known sequential algorithm when � and � are large. II. THE OPTIMAL ALGORITHM In this section, first we describe the optimal sequential algorithm that is used to multiply � numbers, 0 = ��/�� … �23��. Then we give an example to illustrate the number of multiplication operations required to multiply � numbers. A. The Optimal Multiplication Algorithm The optimal multiplication algorithm, denoted as OM, multiplies � numbers by sequentially scanning the input array of � − 1 iterations. In each iteration %, the algorithm multiplies the number �, with 5, where % ≥ 1 and 5 is initially equal to �/ and finally 5 = ∏ �823�8 9 / . In the light of the above, the algorithm requires ��� :2;<� time. The term �is is derived from � − 1 iterations, where ���� = ��� − 1�, while the term :2;< is the total number of operations required to multiply two numbers. The value of :2;< depends on the size of the two numbers. If we have two numbers of size n each: �, = ��,��3��,�,��3��, …,�,�,�,/! and �8 = ��8��3��, �8��3��, …,�8�,�8/!, then multiplying �, and �8 requires ����� multiplication operations using the Naïve method, while the best-known time complexity for :2;< is ��� log � 2����� ∗ ��! [10, 12]. To differentiate between the multiplication operation for the two integers �, and �8, and the multiplication operation for the two digits/bits �,< and �, , we name the second type of operation as the digit-multiplication or bit-multiplication operation. So, the multiplication of �, and �8 requires ����� digit-multiplication or bit-multiplication operations. The main problem when multiplying � numbers, each of size � bits, is the size of the result of the multiplication, which increases with the increase in the value of �. For example, when � = 4, initially 5 = �/ and in the first iteration, the algorithm computes 5 = �/ × ��, of size 2� using �� bit-multiplication operations in the worst case. The second iteration of the algorithm involves multiplying 5 = �/ × �� of size 2� with �� of size � to produce 5 = �/ × �� × �� of size 3� using 2�� bit- multiplication operations. In the last iteration, the algorithm computes 5 = �/ × �� × �� × �@ of size 4� using 3�� bit- multiplication operations. Hence, the overall number of bit- multiplication operations to compute 5 = �/ × �� × �� × �@ is �� + 2�� + 3�� = 6�� in the worst case. Furthermore, in general, the multiplication of � integer numbers, of size � bits each, requires �� + 2�� + 3�� + ⋯ + �� − 1��� bit- multiplication operations, which is equal to � � �� − 1�/2 in the worst case. B. Illustrative Example An illustrative example is given to explain the total number of digit-multiplication operations required to multiply m numbers using the sequential OM algorithm. The notation :∗ represents the total number of digit-multiplication operations. For simplicity, let us assume that we have an array of � = 20 integer numbers and that the value of each element in the array is 5, i.e.: 0 = �5,5,5,5,5,5,5,5,5,5,5,5,5,5,5,5,5,5,5,5� The execution of the optimal sequential algorithm on 0 is described below. Initially, 5 = �/ = 5 and :∗ = 0 , and the algorithm executes 19 iterations to compute 5 = 5�/. These iterations are represented in Figures 1–4. Each figure consists of many subfigures, where each subfigure represents an iteration in the execution of the OM algorithm. Three pieces of information are associated with each iteration. The first is the number of digit-multiplications for multiplying the two numbers and is denoted by ∗ � �, where is the number of digit-multiplication Engineering, Technology & Applied Science Research Vol. 10, No. 6, 2020, 6533-6541 6535 www.etasr.com Bahig et al.: Speeding up the Multiplication Algorithm for Large Integers operations. For example, in Figure 1(c), the algorithm multiplies 5 with the digits 5, 2, and 1, so the number of digit- multiplication operations is 3, and is denoted by ∗ �3�. The second piece of information is the total number of digit- multiplication operations from the start to the current iteration and is denoted by :∗. For example, in Figure 1(c), the total number of digit-multiplication operations is equal to the number of digit-multiplication operations for the current iteration, % � 3, plus the total number of digit-multiplication operations for the previous iterations. Therefore, :∗ � :∗ A∗ �3� � 3 A 3 � 6 . The third and last piece of information in each subfigure consists of the iteration number, %, and the value of 5 that is equal to ∏ �8 , 8 9 / . Figures 1-4 display the complete execution of the OM algorithm on 0 for iterations 1–5, 6–10, 11–14, and 15–19, respectively. Fig. 1. Number of operations of the iterations 1-5 for the OM algorithm. Fig. 2. Number of operations of the iterations 6-10 for the OM algorithm. Fig. 3. Number of operations of the iterations 11-14 for the OM algorithm. Fig. 4. Number of operations of the iterations 15-18 for the OM algorithm. III. THE MODIFIED ALGORITHM In this section, at first the idea and the steps of the suggested strategy for speeding up the execution time of the multiplication of � numbers, each of size �, where � is a large value are introduced. Then two examples are given to illustrate the proposed approach by applying it to the array discussed in Section II. A. The Algorithm The proposed algorithm is based on dividing the array into blocks, each of size F, where F G � . Then the algorithm computes the multiplication of the elements in each block independently. The result of the multiplication for each block, say :H�I, is accumulated to produce the final result of the multiplication, 5. So, after computing the multiplication of each block the value of 5 is updated by calculating 5 = 5 × :H�I, see Algorithm BM0. Initially, the algorithm computes the total number of blocks, �J , by calculating K� F⁄ L . Then the algorithm computes the multiplication of F elements in an auxiliary Engineering, Technology & Applied Science Research Vol. 10, No. 6, 2020, 6533-6541 6536 www.etasr.com Bahig et al.: Speeding up the Multiplication Algorithm for Large Integers variable :H�I , where the initial value of :H�I is the first element in the block. After that, the algorithm updates the value of the final result by multiplying it with the value of :H�I. After it has finished multiplying all the elements in all the blocks, the algorithm updates the final result by multiplying it with the remaining elements, � − K� F⁄ L × F . The time complexity of the proposed algorithm is equal to �K� F⁄ L× F + �� − K� F⁄ L × F�! × :2;