1 /* crypto/ripemd/rmd_dgst.c */
2 /* Copyright (C) 1995-1998 Eric Young (eay@cryptsoft.com)
5 * This package is an SSL implementation written
6 * by Eric Young (eay@cryptsoft.com).
7 * The implementation was written so as to conform with Netscapes SSL.
9 * This library is free for commercial and non-commercial use as long as
10 * the following conditions are aheared to. The following conditions
11 * apply to all code found in this distribution, be it the RC4, RSA,
12 * lhash, DES, etc., code; not just the SSL code. The SSL documentation
13 * included with this distribution is covered by the same copyright terms
14 * except that the holder is Tim Hudson (tjh@cryptsoft.com).
16 * Copyright remains Eric Young's, and as such any Copyright notices in
17 * the code are not to be removed.
18 * If this package is used in a product, Eric Young should be given attribution
19 * as the author of the parts of the library used.
20 * This can be in the form of a textual message at program startup or
21 * in documentation (online or textual) provided with the package.
23 * Redistribution and use in source and binary forms, with or without
24 * modification, are permitted provided that the following conditions
26 * 1. Redistributions of source code must retain the copyright
27 * notice, this list of conditions and the following disclaimer.
28 * 2. Redistributions in binary form must reproduce the above copyright
29 * notice, this list of conditions and the following disclaimer in the
30 * documentation and/or other materials provided with the distribution.
31 * 3. All advertising materials mentioning features or use of this software
32 * must display the following acknowledgement:
33 * "This product includes cryptographic software written by
34 * Eric Young (eay@cryptsoft.com)"
35 * The word 'cryptographic' can be left out if the rouines from the library
36 * being used are not cryptographic related :-).
37 * 4. If you include any Windows specific code (or a derivative thereof) from
38 * the apps directory (application code) you must include an acknowledgement:
39 * "This product includes software written by Tim Hudson (tjh@cryptsoft.com)"
41 * THIS SOFTWARE IS PROVIDED BY ERIC YOUNG ``AS IS'' AND
42 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
43 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
44 * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
45 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
46 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
47 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
48 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
49 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
50 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
53 * The licence and distribution terms for any publically available version or
54 * derivative of this code cannot be changed. i.e. this code cannot simply be
55 * copied and put under another distribution licence
56 * [including the GNU Public Licence.]
61 #include <openssl/opensslv.h>
63 char *RMD160_version="RIPE-MD160" OPENSSL_VERSION_PTEXT;
66 void ripemd160_block_x86(RIPEMD160_CTX *c, unsigned long *p,int num);
67 # define ripemd160_block ripemd160_block_x86
69 void ripemd160_block(RIPEMD160_CTX *c, unsigned long *p,int num);
71 void RIPEMD160_Init(RIPEMD160_CTX *c)
83 void RIPEMD160_Update(RIPEMD160_CTX *c, register unsigned char *data,
92 l=(c->Nl+(len<<3))&0xffffffffL;
93 if (l < c->Nl) /* overflow */
104 if ((c->num+len) >= RIPEMD160_CBLOCK)
109 for (; sw<RIPEMD160_LBLOCK; sw++)
114 len-=(RIPEMD160_CBLOCK-c->num);
116 ripemd160_block(c,p,64);
118 /* drop through and do the rest */
125 if ((sc+len) < 4) /* ugly, add char's to a word */
128 p_c2l_p(data,l,sc,len);
138 for (; sw < ew; sw++)
139 { c2l(data,l); p[sw]=l; }
149 /* we now can process the input data in blocks of RIPEMD160_CBLOCK
150 * chars and save the leftovers to c->data. */
152 if ((((unsigned long)data)%sizeof(ULONG)) == 0)
154 sw=(int)len/RIPEMD160_CBLOCK;
157 sw*=RIPEMD160_CBLOCK;
158 ripemd160_block(c,(ULONG *)data,sw);
165 while (len >= RIPEMD160_CBLOCK)
167 #if defined(L_ENDIAN) || defined(B_ENDIAN)
168 if (p != (unsigned long *)data)
169 memcpy(p,data,RIPEMD160_CBLOCK);
170 data+=RIPEMD160_CBLOCK;
172 for (sw=(RIPEMD160_LBLOCK/4); sw; sw--)
174 Endian_Reverse32(p[0]);
175 Endian_Reverse32(p[1]);
176 Endian_Reverse32(p[2]);
177 Endian_Reverse32(p[3]);
182 for (sw=(RIPEMD160_LBLOCK/4); sw; sw--)
184 c2l(data,l); *(p++)=l;
185 c2l(data,l); *(p++)=l;
186 c2l(data,l); *(p++)=l;
187 c2l(data,l); *(p++)=l;
191 ripemd160_block(c,p,64);
192 len-=RIPEMD160_CBLOCK;
198 sw=sc>>2; /* words to copy */
205 { c2l(data,l); *(p++)=l; }
212 void RIPEMD160_Transform(RIPEMD160_CTX *c, unsigned char *b)
215 #if !defined(L_ENDIAN)
220 #if defined(B_ENDIAN) || defined(L_ENDIAN)
224 for (i=(RIPEMD160_LBLOCK/4); i; i--)
226 Endian_Reverse32(q[0]);
227 Endian_Reverse32(q[1]);
228 Endian_Reverse32(q[2]);
229 Endian_Reverse32(q[3]);
235 for (i=(RIPEMD160_LBLOCK/4); i; i--)
244 ripemd160_block(c,p,64);
249 void ripemd160_block(RIPEMD160_CTX *ctx, register ULONG *X, int num)
251 register ULONG A,B,C,D,E;
256 A=ctx->A; B=ctx->B; C=ctx->C; D=ctx->D; E=ctx->E;
258 RIP1(A,B,C,D,E,WL00,SL00);
259 RIP1(E,A,B,C,D,WL01,SL01);
260 RIP1(D,E,A,B,C,WL02,SL02);
261 RIP1(C,D,E,A,B,WL03,SL03);
262 RIP1(B,C,D,E,A,WL04,SL04);
263 RIP1(A,B,C,D,E,WL05,SL05);
264 RIP1(E,A,B,C,D,WL06,SL06);
265 RIP1(D,E,A,B,C,WL07,SL07);
266 RIP1(C,D,E,A,B,WL08,SL08);
267 RIP1(B,C,D,E,A,WL09,SL09);
268 RIP1(A,B,C,D,E,WL10,SL10);
269 RIP1(E,A,B,C,D,WL11,SL11);
270 RIP1(D,E,A,B,C,WL12,SL12);
271 RIP1(C,D,E,A,B,WL13,SL13);
272 RIP1(B,C,D,E,A,WL14,SL14);
273 RIP1(A,B,C,D,E,WL15,SL15);
275 RIP2(E,A,B,C,D,WL16,SL16,KL1);
276 RIP2(D,E,A,B,C,WL17,SL17,KL1);
277 RIP2(C,D,E,A,B,WL18,SL18,KL1);
278 RIP2(B,C,D,E,A,WL19,SL19,KL1);
279 RIP2(A,B,C,D,E,WL20,SL20,KL1);
280 RIP2(E,A,B,C,D,WL21,SL21,KL1);
281 RIP2(D,E,A,B,C,WL22,SL22,KL1);
282 RIP2(C,D,E,A,B,WL23,SL23,KL1);
283 RIP2(B,C,D,E,A,WL24,SL24,KL1);
284 RIP2(A,B,C,D,E,WL25,SL25,KL1);
285 RIP2(E,A,B,C,D,WL26,SL26,KL1);
286 RIP2(D,E,A,B,C,WL27,SL27,KL1);
287 RIP2(C,D,E,A,B,WL28,SL28,KL1);
288 RIP2(B,C,D,E,A,WL29,SL29,KL1);
289 RIP2(A,B,C,D,E,WL30,SL30,KL1);
290 RIP2(E,A,B,C,D,WL31,SL31,KL1);
292 RIP3(D,E,A,B,C,WL32,SL32,KL2);
293 RIP3(C,D,E,A,B,WL33,SL33,KL2);
294 RIP3(B,C,D,E,A,WL34,SL34,KL2);
295 RIP3(A,B,C,D,E,WL35,SL35,KL2);
296 RIP3(E,A,B,C,D,WL36,SL36,KL2);
297 RIP3(D,E,A,B,C,WL37,SL37,KL2);
298 RIP3(C,D,E,A,B,WL38,SL38,KL2);
299 RIP3(B,C,D,E,A,WL39,SL39,KL2);
300 RIP3(A,B,C,D,E,WL40,SL40,KL2);
301 RIP3(E,A,B,C,D,WL41,SL41,KL2);
302 RIP3(D,E,A,B,C,WL42,SL42,KL2);
303 RIP3(C,D,E,A,B,WL43,SL43,KL2);
304 RIP3(B,C,D,E,A,WL44,SL44,KL2);
305 RIP3(A,B,C,D,E,WL45,SL45,KL2);
306 RIP3(E,A,B,C,D,WL46,SL46,KL2);
307 RIP3(D,E,A,B,C,WL47,SL47,KL2);
309 RIP4(C,D,E,A,B,WL48,SL48,KL3);
310 RIP4(B,C,D,E,A,WL49,SL49,KL3);
311 RIP4(A,B,C,D,E,WL50,SL50,KL3);
312 RIP4(E,A,B,C,D,WL51,SL51,KL3);
313 RIP4(D,E,A,B,C,WL52,SL52,KL3);
314 RIP4(C,D,E,A,B,WL53,SL53,KL3);
315 RIP4(B,C,D,E,A,WL54,SL54,KL3);
316 RIP4(A,B,C,D,E,WL55,SL55,KL3);
317 RIP4(E,A,B,C,D,WL56,SL56,KL3);
318 RIP4(D,E,A,B,C,WL57,SL57,KL3);
319 RIP4(C,D,E,A,B,WL58,SL58,KL3);
320 RIP4(B,C,D,E,A,WL59,SL59,KL3);
321 RIP4(A,B,C,D,E,WL60,SL60,KL3);
322 RIP4(E,A,B,C,D,WL61,SL61,KL3);
323 RIP4(D,E,A,B,C,WL62,SL62,KL3);
324 RIP4(C,D,E,A,B,WL63,SL63,KL3);
326 RIP5(B,C,D,E,A,WL64,SL64,KL4);
327 RIP5(A,B,C,D,E,WL65,SL65,KL4);
328 RIP5(E,A,B,C,D,WL66,SL66,KL4);
329 RIP5(D,E,A,B,C,WL67,SL67,KL4);
330 RIP5(C,D,E,A,B,WL68,SL68,KL4);
331 RIP5(B,C,D,E,A,WL69,SL69,KL4);
332 RIP5(A,B,C,D,E,WL70,SL70,KL4);
333 RIP5(E,A,B,C,D,WL71,SL71,KL4);
334 RIP5(D,E,A,B,C,WL72,SL72,KL4);
335 RIP5(C,D,E,A,B,WL73,SL73,KL4);
336 RIP5(B,C,D,E,A,WL74,SL74,KL4);
337 RIP5(A,B,C,D,E,WL75,SL75,KL4);
338 RIP5(E,A,B,C,D,WL76,SL76,KL4);
339 RIP5(D,E,A,B,C,WL77,SL77,KL4);
340 RIP5(C,D,E,A,B,WL78,SL78,KL4);
341 RIP5(B,C,D,E,A,WL79,SL79,KL4);
343 a=A; b=B; c=C; d=D; e=E;
345 A=ctx->A; B=ctx->B; C=ctx->C; D=ctx->D; E=ctx->E;
347 RIP5(A,B,C,D,E,WR00,SR00,KR0);
348 RIP5(E,A,B,C,D,WR01,SR01,KR0);
349 RIP5(D,E,A,B,C,WR02,SR02,KR0);
350 RIP5(C,D,E,A,B,WR03,SR03,KR0);
351 RIP5(B,C,D,E,A,WR04,SR04,KR0);
352 RIP5(A,B,C,D,E,WR05,SR05,KR0);
353 RIP5(E,A,B,C,D,WR06,SR06,KR0);
354 RIP5(D,E,A,B,C,WR07,SR07,KR0);
355 RIP5(C,D,E,A,B,WR08,SR08,KR0);
356 RIP5(B,C,D,E,A,WR09,SR09,KR0);
357 RIP5(A,B,C,D,E,WR10,SR10,KR0);
358 RIP5(E,A,B,C,D,WR11,SR11,KR0);
359 RIP5(D,E,A,B,C,WR12,SR12,KR0);
360 RIP5(C,D,E,A,B,WR13,SR13,KR0);
361 RIP5(B,C,D,E,A,WR14,SR14,KR0);
362 RIP5(A,B,C,D,E,WR15,SR15,KR0);
364 RIP4(E,A,B,C,D,WR16,SR16,KR1);
365 RIP4(D,E,A,B,C,WR17,SR17,KR1);
366 RIP4(C,D,E,A,B,WR18,SR18,KR1);
367 RIP4(B,C,D,E,A,WR19,SR19,KR1);
368 RIP4(A,B,C,D,E,WR20,SR20,KR1);
369 RIP4(E,A,B,C,D,WR21,SR21,KR1);
370 RIP4(D,E,A,B,C,WR22,SR22,KR1);
371 RIP4(C,D,E,A,B,WR23,SR23,KR1);
372 RIP4(B,C,D,E,A,WR24,SR24,KR1);
373 RIP4(A,B,C,D,E,WR25,SR25,KR1);
374 RIP4(E,A,B,C,D,WR26,SR26,KR1);
375 RIP4(D,E,A,B,C,WR27,SR27,KR1);
376 RIP4(C,D,E,A,B,WR28,SR28,KR1);
377 RIP4(B,C,D,E,A,WR29,SR29,KR1);
378 RIP4(A,B,C,D,E,WR30,SR30,KR1);
379 RIP4(E,A,B,C,D,WR31,SR31,KR1);
381 RIP3(D,E,A,B,C,WR32,SR32,KR2);
382 RIP3(C,D,E,A,B,WR33,SR33,KR2);
383 RIP3(B,C,D,E,A,WR34,SR34,KR2);
384 RIP3(A,B,C,D,E,WR35,SR35,KR2);
385 RIP3(E,A,B,C,D,WR36,SR36,KR2);
386 RIP3(D,E,A,B,C,WR37,SR37,KR2);
387 RIP3(C,D,E,A,B,WR38,SR38,KR2);
388 RIP3(B,C,D,E,A,WR39,SR39,KR2);
389 RIP3(A,B,C,D,E,WR40,SR40,KR2);
390 RIP3(E,A,B,C,D,WR41,SR41,KR2);
391 RIP3(D,E,A,B,C,WR42,SR42,KR2);
392 RIP3(C,D,E,A,B,WR43,SR43,KR2);
393 RIP3(B,C,D,E,A,WR44,SR44,KR2);
394 RIP3(A,B,C,D,E,WR45,SR45,KR2);
395 RIP3(E,A,B,C,D,WR46,SR46,KR2);
396 RIP3(D,E,A,B,C,WR47,SR47,KR2);
398 RIP2(C,D,E,A,B,WR48,SR48,KR3);
399 RIP2(B,C,D,E,A,WR49,SR49,KR3);
400 RIP2(A,B,C,D,E,WR50,SR50,KR3);
401 RIP2(E,A,B,C,D,WR51,SR51,KR3);
402 RIP2(D,E,A,B,C,WR52,SR52,KR3);
403 RIP2(C,D,E,A,B,WR53,SR53,KR3);
404 RIP2(B,C,D,E,A,WR54,SR54,KR3);
405 RIP2(A,B,C,D,E,WR55,SR55,KR3);
406 RIP2(E,A,B,C,D,WR56,SR56,KR3);
407 RIP2(D,E,A,B,C,WR57,SR57,KR3);
408 RIP2(C,D,E,A,B,WR58,SR58,KR3);
409 RIP2(B,C,D,E,A,WR59,SR59,KR3);
410 RIP2(A,B,C,D,E,WR60,SR60,KR3);
411 RIP2(E,A,B,C,D,WR61,SR61,KR3);
412 RIP2(D,E,A,B,C,WR62,SR62,KR3);
413 RIP2(C,D,E,A,B,WR63,SR63,KR3);
415 RIP1(B,C,D,E,A,WR64,SR64);
416 RIP1(A,B,C,D,E,WR65,SR65);
417 RIP1(E,A,B,C,D,WR66,SR66);
418 RIP1(D,E,A,B,C,WR67,SR67);
419 RIP1(C,D,E,A,B,WR68,SR68);
420 RIP1(B,C,D,E,A,WR69,SR69);
421 RIP1(A,B,C,D,E,WR70,SR70);
422 RIP1(E,A,B,C,D,WR71,SR71);
423 RIP1(D,E,A,B,C,WR72,SR72);
424 RIP1(C,D,E,A,B,WR73,SR73);
425 RIP1(B,C,D,E,A,WR74,SR74);
426 RIP1(A,B,C,D,E,WR75,SR75);
427 RIP1(E,A,B,C,D,WR76,SR76);
428 RIP1(D,E,A,B,C,WR77,SR77);
429 RIP1(C,D,E,A,B,WR78,SR78);
430 RIP1(B,C,D,E,A,WR79,SR79);
446 void RIPEMD160_Final(unsigned char *md, RIPEMD160_CTX *c)
451 static unsigned char end[4]={0x80,0x00,0x00,0x00};
452 unsigned char *cp=end;
454 /* c->num should definitly have room for at least one more byte. */
459 /* purify often complains about the following line as an
460 * Uninitialized Memory Read. While this can be true, the
461 * following p_c2l macro will reset l when that case is true.
462 * This is because j&0x03 contains the number of 'valid' bytes
463 * already in p[i]. If and only if j&0x03 == 0, the UMR will
464 * occur but this is also the only time p_c2l will do
465 * l= *(cp++) instead of l|= *(cp++)
466 * Many thanks to Alex Tang <altitude@cic.net> for pickup this
469 if ((j&0x03) == 0) p[i]=0;
475 /* i is the next 'undefined word' */
476 if (c->num >= RIPEMD160_LAST_BLOCK)
478 for (; i<RIPEMD160_LBLOCK; i++)
480 ripemd160_block(c,p,64);
483 for (; i<(RIPEMD160_LBLOCK-2); i++)
485 p[RIPEMD160_LBLOCK-2]=c->Nl;
486 p[RIPEMD160_LBLOCK-1]=c->Nh;
487 ripemd160_block(c,p,64);
495 /* clear stuff, ripemd160_block may be leaving some stuff on the stack
496 * but I'm not worried :-) */
498 /* memset((char *)&c,0,sizeof(c));*/
502 int printit(unsigned long *l)
508 for (ii=0; ii<8; ii++)
510 fprintf(stderr,"%08lx ",l[i*8+ii]);
512 fprintf(stderr,"\n");