Move more comments that confuse indent
[openssl.git] / crypto / bn / rsaz_exp.c
1 /*****************************************************************************
2 *                                                                            *
3 *  Copyright (c) 2012, Intel Corporation                                     *
4 *                                                                            *
5 *  All rights reserved.                                                      *
6 *                                                                            *
7 *  Redistribution and use in source and binary forms, with or without        *
8 *  modification, are permitted provided that the following conditions are    *
9 *  met:                                                                      *
10 *                                                                            *
11 *  *  Redistributions of source code must retain the above copyright         *
12 *     notice, this list of conditions and the following disclaimer.          *
13 *                                                                            *
14 *  *  Redistributions in binary form must reproduce the above copyright      *
15 *     notice, this list of conditions and the following disclaimer in the    *
16 *     documentation and/or other materials provided with the                 *
17 *     distribution.                                                          *
18 *                                                                            *
19 *  *  Neither the name of the Intel Corporation nor the names of its         *
20 *     contributors may be used to endorse or promote products derived from   *
21 *     this software without specific prior written permission.               *
22 *                                                                            *
23 *                                                                            *
24 *  THIS SOFTWARE IS PROVIDED BY INTEL CORPORATION ""AS IS"" AND ANY          *
25 *  EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE         *
26 *  IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR        *
27 *  PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL INTEL CORPORATION OR            *
28 *  CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL,     *
29 *  EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO,       *
30 *  PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR        *
31 *  PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF    *
32 *  LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING      *
33 *  NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS        *
34 *  SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.              *
35 *                                                                            *
36 ******************************************************************************
37 * Developers and authors:                                                    *
38 * Shay Gueron (1, 2), and Vlad Krasnov (1)                                   *
39 * (1) Intel Corporation, Israel Development Center, Haifa, Israel            *
40 * (2) University of Haifa, Israel                                            *
41 *****************************************************************************/
42
43 #include "rsaz_exp.h"
44
45 /*
46  * See crypto/bn/asm/rsaz-avx2.pl for further details.
47  */
48 void rsaz_1024_norm2red_avx2(void *red,const void *norm);
49 void rsaz_1024_mul_avx2(void *ret,const void *a,const void *b,const void *n,BN_ULONG k);
50 void rsaz_1024_sqr_avx2(void *ret,const void *a,const void *n,BN_ULONG k,int cnt);
51 void rsaz_1024_scatter5_avx2(void *tbl,const void *val,int i);
52 void rsaz_1024_gather5_avx2(void *val,const void *tbl,int i);
53 void rsaz_1024_red2norm_avx2(void *norm,const void *red);
54
55 #if defined(__GNUC__)
56 # define ALIGN64        __attribute__((aligned(64)))
57 #elif defined(_MSC_VER)
58 # define ALIGN64        __declspec(align(64))
59 #elif defined(__SUNPRO_C)
60 # define ALIGN64
61 # pragma align 64(one,two80)
62 #else
63 /* not fatal, might hurt performance a little */
64 # define ALIGN64
65 #endif
66
67 ALIGN64 static const BN_ULONG one[40] = {
68         1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,
69         0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0
70 };
71 ALIGN64 static const BN_ULONG two80[40] = {
72         0,0,1<<22,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,
73         0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0
74 };
75
76 void RSAZ_1024_mod_exp_avx2(BN_ULONG result_norm[16],
77         const BN_ULONG base_norm[16], const BN_ULONG exponent[16],
78         const BN_ULONG m_norm[16], const BN_ULONG RR[16], BN_ULONG k0)
79 {
80         unsigned char    storage[320*3+32*9*16+64];     /* 5.5KB */
81         unsigned char   *p_str = storage + (64-((size_t)storage%64));
82         unsigned char   *a_inv, *m, *result;
83         unsigned char   *table_s = p_str+320*3;
84         unsigned char   *R2      = table_s;     /* borrow */
85         int index;
86         int wvalue;
87
88         if ((((size_t)p_str&4095)+320)>>12) {
89                 result = p_str;
90                 a_inv = p_str + 320;
91                 m = p_str + 320*2;      /* should not cross page */
92         } else {
93                 m = p_str;              /* should not cross page */
94                 result = p_str + 320;
95                 a_inv = p_str + 320*2;
96         }
97
98         rsaz_1024_norm2red_avx2(m, m_norm);
99         rsaz_1024_norm2red_avx2(a_inv, base_norm);
100         rsaz_1024_norm2red_avx2(R2, RR);
101
102         rsaz_1024_mul_avx2(R2, R2, R2, m, k0);
103         rsaz_1024_mul_avx2(R2, R2, two80, m, k0);
104
105         /* table[0] = 1 */
106         rsaz_1024_mul_avx2(result, R2, one, m, k0);
107         /* table[1] = a_inv^1 */
108         rsaz_1024_mul_avx2(a_inv, a_inv, R2, m, k0);
109
110         rsaz_1024_scatter5_avx2(table_s,result,0);
111         rsaz_1024_scatter5_avx2(table_s,a_inv,1);
112
113         /* table[2] = a_inv^2 */
114         rsaz_1024_sqr_avx2(result, a_inv, m, k0, 1);
115         rsaz_1024_scatter5_avx2(table_s,result,2);
116 #if 0
117         /* this is almost 2x smaller and less than 1% slower */
118         for (index=3; index<32; index++) {
119                 rsaz_1024_mul_avx2(result, result, a_inv, m, k0);
120                 rsaz_1024_scatter5_avx2(table_s,result,index);
121         }
122 #else
123         /* table[4] = a_inv^4 */
124         rsaz_1024_sqr_avx2(result, result, m, k0, 1);
125         rsaz_1024_scatter5_avx2(table_s,result,4);
126         /* table[8] = a_inv^8 */
127         rsaz_1024_sqr_avx2(result, result, m, k0, 1);
128         rsaz_1024_scatter5_avx2(table_s,result,8);
129         /* table[16] = a_inv^16 */
130         rsaz_1024_sqr_avx2(result, result, m, k0, 1);
131         rsaz_1024_scatter5_avx2(table_s,result,16);
132         /* table[17] = a_inv^17 */
133         rsaz_1024_mul_avx2(result, result, a_inv, m, k0);
134         rsaz_1024_scatter5_avx2(table_s,result,17);
135
136         /* table[3] */
137         rsaz_1024_gather5_avx2(result,table_s,2);
138         rsaz_1024_mul_avx2(result,result,a_inv,m,k0);
139         rsaz_1024_scatter5_avx2(table_s,result,3);
140         /* table[6] */
141         rsaz_1024_sqr_avx2(result, result, m, k0, 1);
142         rsaz_1024_scatter5_avx2(table_s,result,6);
143         /* table[12] */
144         rsaz_1024_sqr_avx2(result, result, m, k0, 1);
145         rsaz_1024_scatter5_avx2(table_s,result,12);
146         /* table[24] */
147         rsaz_1024_sqr_avx2(result, result, m, k0, 1);
148         rsaz_1024_scatter5_avx2(table_s,result,24);
149         /* table[25] */
150         rsaz_1024_mul_avx2(result, result, a_inv, m, k0);
151         rsaz_1024_scatter5_avx2(table_s,result,25);
152
153         /* table[5] */
154         rsaz_1024_gather5_avx2(result,table_s,4);
155         rsaz_1024_mul_avx2(result,result,a_inv,m,k0);
156         rsaz_1024_scatter5_avx2(table_s,result,5);
157         /* table[10] */
158         rsaz_1024_sqr_avx2(result, result, m, k0, 1);
159         rsaz_1024_scatter5_avx2(table_s,result,10);
160         /* table[20] */
161         rsaz_1024_sqr_avx2(result, result, m, k0, 1);
162         rsaz_1024_scatter5_avx2(table_s,result,20);
163         /* table[21] */
164         rsaz_1024_mul_avx2(result, result, a_inv, m, k0);
165         rsaz_1024_scatter5_avx2(table_s,result,21);
166
167         /* table[7] */
168         rsaz_1024_gather5_avx2(result,table_s,6);
169         rsaz_1024_mul_avx2(result,result,a_inv,m,k0);
170         rsaz_1024_scatter5_avx2(table_s,result,7);
171         /* table[14] */
172         rsaz_1024_sqr_avx2(result, result, m, k0, 1);
173         rsaz_1024_scatter5_avx2(table_s,result,14);
174         /* table[28] */
175         rsaz_1024_sqr_avx2(result, result, m, k0, 1);
176         rsaz_1024_scatter5_avx2(table_s,result,28);
177         /* table[29] */
178         rsaz_1024_mul_avx2(result, result, a_inv, m, k0);
179         rsaz_1024_scatter5_avx2(table_s,result,29);
180
181         /* table[9] */
182         rsaz_1024_gather5_avx2(result,table_s,8);
183         rsaz_1024_mul_avx2(result,result,a_inv,m,k0);
184         rsaz_1024_scatter5_avx2(table_s,result,9);
185         /* table[18] */
186         rsaz_1024_sqr_avx2(result, result, m, k0, 1);
187         rsaz_1024_scatter5_avx2(table_s,result,18);
188         /* table[19] */
189         rsaz_1024_mul_avx2(result, result, a_inv, m, k0);
190         rsaz_1024_scatter5_avx2(table_s,result,19);
191
192         /* table[11] */
193         rsaz_1024_gather5_avx2(result,table_s,10);
194         rsaz_1024_mul_avx2(result,result,a_inv,m,k0);
195         rsaz_1024_scatter5_avx2(table_s,result,11);
196         /* table[22] */
197         rsaz_1024_sqr_avx2(result, result, m, k0, 1);
198         rsaz_1024_scatter5_avx2(table_s,result,22);
199         /* table[23] */
200         rsaz_1024_mul_avx2(result, result, a_inv, m, k0);
201         rsaz_1024_scatter5_avx2(table_s,result,23);
202
203         /* table[13] */
204         rsaz_1024_gather5_avx2(result,table_s,12);
205         rsaz_1024_mul_avx2(result,result,a_inv,m,k0);
206         rsaz_1024_scatter5_avx2(table_s,result,13);
207         /* table[26] */
208         rsaz_1024_sqr_avx2(result, result, m, k0, 1);
209         rsaz_1024_scatter5_avx2(table_s,result,26);
210         /* table[27] */
211         rsaz_1024_mul_avx2(result, result, a_inv, m, k0);
212         rsaz_1024_scatter5_avx2(table_s,result,27);
213
214         /* table[15] */
215         rsaz_1024_gather5_avx2(result,table_s,14);
216         rsaz_1024_mul_avx2(result,result,a_inv,m,k0);
217         rsaz_1024_scatter5_avx2(table_s,result,15);
218         /* table[30] */
219         rsaz_1024_sqr_avx2(result, result, m, k0, 1);
220         rsaz_1024_scatter5_avx2(table_s,result,30);
221         /* table[31] */
222         rsaz_1024_mul_avx2(result, result, a_inv, m, k0);
223         rsaz_1024_scatter5_avx2(table_s,result,31);
224 #endif
225
226         /* load first window */
227         p_str = (unsigned char*)exponent;
228         wvalue = p_str[127] >> 3;
229         rsaz_1024_gather5_avx2(result,table_s,wvalue);
230
231         index = 1014;
232
233         while(index > -1) {     /* loop for the remaining 127 windows */
234
235                 rsaz_1024_sqr_avx2(result, result, m, k0, 5);
236
237                 wvalue = *((unsigned short*)&p_str[index/8]);
238                 wvalue = (wvalue>> (index%8)) & 31;
239                 index-=5;
240
241                 rsaz_1024_gather5_avx2(a_inv,table_s,wvalue);   /* borrow a_inv */
242                 rsaz_1024_mul_avx2(result, result, a_inv, m, k0);
243         }
244
245         /* square four times */
246         rsaz_1024_sqr_avx2(result, result, m, k0, 4);
247
248         wvalue = p_str[0] & 15;
249
250         rsaz_1024_gather5_avx2(a_inv,table_s,wvalue);   /* borrow a_inv */
251         rsaz_1024_mul_avx2(result, result, a_inv, m, k0);
252
253         /* from Montgomery */
254         rsaz_1024_mul_avx2(result, result, one, m, k0);
255
256         rsaz_1024_red2norm_avx2(result_norm, result);
257
258         OPENSSL_cleanse(storage,sizeof(storage));
259 }
260
261 /*
262  * See crypto/bn/rsaz-x86_64.pl for further details.
263  */
264 void rsaz_512_mul(void *ret,const void *a,const void *b,const void *n,BN_ULONG k);
265 void rsaz_512_mul_scatter4(void *ret,const void *a,const void *n,BN_ULONG k,const void *tbl,unsigned int power);
266 void rsaz_512_mul_gather4(void *ret,const void *a,const void *tbl,const void *n,BN_ULONG k,unsigned int power);
267 void rsaz_512_mul_by_one(void *ret,const void *a,const void *n,BN_ULONG k);
268 void rsaz_512_sqr(void *ret,const void *a,const void *n,BN_ULONG k,int cnt);
269 void rsaz_512_scatter4(void *tbl, const BN_ULONG *val, int power);
270 void rsaz_512_gather4(BN_ULONG *val, const void *tbl, int power);
271
272 void RSAZ_512_mod_exp(BN_ULONG result[8],
273         const BN_ULONG base[8], const BN_ULONG exponent[8],
274         const BN_ULONG m[8], BN_ULONG k0, const BN_ULONG RR[8])
275 {
276         unsigned char    storage[16*8*8+64*2+64];       /* 1.2KB */
277         unsigned char   *table = storage + (64-((size_t)storage%64));
278         BN_ULONG        *a_inv = (BN_ULONG *)(table+16*8*8);
279         BN_ULONG        *temp  = (BN_ULONG *)(table+16*8*8+8*8);
280         unsigned char   *p_str = (unsigned char*)exponent;
281         int index;
282         unsigned int wvalue;
283
284         /* table[0] = 1_inv */
285         temp[0] = 0-m[0];       temp[1] = ~m[1];
286         temp[2] = ~m[2];        temp[3] = ~m[3];
287         temp[4] = ~m[4];        temp[5] = ~m[5];
288         temp[6] = ~m[6];        temp[7] = ~m[7];
289         rsaz_512_scatter4(table, temp, 0);
290
291         /* table [1] = a_inv^1 */
292         rsaz_512_mul(a_inv, base, RR, m, k0);
293         rsaz_512_scatter4(table, a_inv, 1);
294
295         /* table [2] = a_inv^2 */
296         rsaz_512_sqr(temp, a_inv, m, k0, 1);
297         rsaz_512_scatter4(table, temp, 2);
298
299         for (index=3; index<16; index++)
300                 rsaz_512_mul_scatter4(temp, a_inv, m, k0, table, index);
301
302         /* load first window */
303         wvalue = p_str[63];
304
305         rsaz_512_gather4(temp, table, wvalue>>4);
306         rsaz_512_sqr(temp, temp, m, k0, 4);
307         rsaz_512_mul_gather4(temp, temp, table, m, k0, wvalue&0xf);
308
309         for (index=62; index>=0; index--) {
310                 wvalue = p_str[index];
311
312                 rsaz_512_sqr(temp, temp, m, k0, 4);
313                 rsaz_512_mul_gather4(temp, temp, table, m, k0, wvalue>>4);
314
315                 rsaz_512_sqr(temp, temp, m, k0, 4);
316                 rsaz_512_mul_gather4(temp, temp, table, m, k0, wvalue&0x0f);
317         }
318
319         /* from Montgomery */
320         rsaz_512_mul_by_one(result, temp, m, k0);
321
322         OPENSSL_cleanse(storage,sizeof(storage));
323 }