OLD | NEW |
1 #!/usr/bin/env perl | 1 #!/usr/bin/env perl |
2 # | 2 # |
3 # ==================================================================== | 3 # ==================================================================== |
4 # Written by Andy Polyakov <appro@fy.chalmers.se> for the OpenSSL | 4 # Written by Andy Polyakov <appro@fy.chalmers.se> for the OpenSSL |
5 # project. The module is, however, dual licensed under OpenSSL and | 5 # project. The module is, however, dual licensed under OpenSSL and |
6 # CRYPTOGAMS licenses depending on where you obtain it. For further | 6 # CRYPTOGAMS licenses depending on where you obtain it. For further |
7 # details see http://www.openssl.org/~appro/cryptogams/. | 7 # details see http://www.openssl.org/~appro/cryptogams/. |
8 # ==================================================================== | 8 # ==================================================================== |
9 # | 9 # |
10 # 2.22x RC4 tune-up:-) It should be noted though that my hand [as in | 10 # 2.22x RC4 tune-up:-) It should be noted though that my hand [as in |
(...skipping 40 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
51 # preferred, as it provides better *all-round* performance]. | 51 # preferred, as it provides better *all-round* performance]. |
52 | 52 |
53 # Intel Core2 was observed to perform poorly on both code paths:-( It | 53 # Intel Core2 was observed to perform poorly on both code paths:-( It |
54 # apparently suffers from some kind of partial register stall, which | 54 # apparently suffers from some kind of partial register stall, which |
55 # occurs in 64-bit mode only [as virtually identical 32-bit loop was | 55 # occurs in 64-bit mode only [as virtually identical 32-bit loop was |
56 # observed to outperform 64-bit one by almost 50%]. Adding two movzb to | 56 # observed to outperform 64-bit one by almost 50%]. Adding two movzb to |
57 # cloop1 boosts its performance by 80%! This loop appears to be optimal | 57 # cloop1 boosts its performance by 80%! This loop appears to be optimal |
58 # fit for Core2 and therefore the code was modified to skip cloop8 on | 58 # fit for Core2 and therefore the code was modified to skip cloop8 on |
59 # this CPU. | 59 # this CPU. |
60 | 60 |
61 $output=shift; | 61 $flavour = shift; |
| 62 $output = shift; |
| 63 if ($flavour =~ /\./) { $output = $flavour; undef $flavour; } |
| 64 |
| 65 $win64=0; $win64=1 if ($flavour =~ /[nm]asm|mingw64/ || $output =~ /\.asm$/); |
62 | 66 |
63 $0 =~ m/(.*[\/\\])[^\/\\]+$/; $dir=$1; | 67 $0 =~ m/(.*[\/\\])[^\/\\]+$/; $dir=$1; |
64 ( $xlate="${dir}x86_64-xlate.pl" and -f $xlate ) or | 68 ( $xlate="${dir}x86_64-xlate.pl" and -f $xlate ) or |
65 ( $xlate="${dir}../../perlasm/x86_64-xlate.pl" and -f $xlate) or | 69 ( $xlate="${dir}../../perlasm/x86_64-xlate.pl" and -f $xlate) or |
66 die "can't locate x86_64-xlate.pl"; | 70 die "can't locate x86_64-xlate.pl"; |
67 | 71 |
68 open STDOUT,"| $^X $xlate $output"; | 72 open STDOUT,"| $^X $xlate $flavour $output"; |
69 | 73 |
70 $dat="%rdi"; # arg1 | 74 $dat="%rdi"; # arg1 |
71 $len="%rsi"; # arg2 | 75 $len="%rsi"; # arg2 |
72 $inp="%rdx"; # arg3 | 76 $inp="%rdx"; # arg3 |
73 $out="%rcx"; # arg4 | 77 $out="%rcx"; # arg4 |
74 | 78 |
75 @XX=("%r8","%r10"); | 79 @XX=("%r8","%r10"); |
76 @TX=("%r9","%r11"); | 80 @TX=("%r9","%r11"); |
77 $YY="%r12"; | 81 $YY="%r12"; |
78 $TY="%r13"; | 82 $TY="%r13"; |
79 | 83 |
80 $code=<<___; | 84 $code=<<___; |
81 .text | 85 .text |
82 | 86 |
83 .globl RC4 | 87 .globl RC4 |
84 .type RC4,\@function,4 | 88 .type RC4,\@function,4 |
85 .align 16 | 89 .align 16 |
86 RC4: or $len,$len | 90 RC4: or $len,$len |
87 jne .Lentry | 91 jne .Lentry |
88 ret | 92 ret |
89 .Lentry: | 93 .Lentry: |
| 94 push %rbx |
90 push %r12 | 95 push %r12 |
91 push %r13 | 96 push %r13 |
| 97 .Lprologue: |
92 | 98 |
93 add \$8,$dat | 99 add \$8,$dat |
94 movl -8($dat),$XX[0]#d | 100 movl -8($dat),$XX[0]#d |
95 movl -4($dat),$YY#d | 101 movl -4($dat),$YY#d |
96 cmpl \$-1,256($dat) | 102 cmpl \$-1,256($dat) |
97 je .LRC4_CHAR | 103 je .LRC4_CHAR |
98 inc $XX[0]#b | 104 inc $XX[0]#b |
99 movl ($dat,$XX[0],4),$TX[0]#d | 105 movl ($dat,$XX[0],4),$TX[0]#d |
100 test \$-8,$len | 106 test \$-8,$len |
101 jz .Lloop1 | 107 jz .Lloop1 |
(...skipping 24 matching lines...) Expand all Loading... |
126 | 132 |
127 xor ($inp),%rax | 133 xor ($inp),%rax |
128 add \$8,$inp | 134 add \$8,$inp |
129 mov %rax,($out) | 135 mov %rax,($out) |
130 add \$8,$out | 136 add \$8,$out |
131 | 137 |
132 test \$-8,$len | 138 test \$-8,$len |
133 jnz .Lloop8 | 139 jnz .Lloop8 |
134 cmp \$0,$len | 140 cmp \$0,$len |
135 jne .Lloop1 | 141 jne .Lloop1 |
136 ___ | 142 » jmp» .Lexit |
137 $code.=<<___; | |
138 .Lexit: | |
139 » sub» \$1,$XX[0]#b | |
140 » movl» $XX[0]#d,-8($dat) | |
141 » movl» $YY#d,-4($dat) | |
142 | 143 |
143 pop %r13 | |
144 pop %r12 | |
145 ret | |
146 .align 16 | 144 .align 16 |
147 .Lloop1: | 145 .Lloop1: |
148 add $TX[0]#b,$YY#b | 146 add $TX[0]#b,$YY#b |
149 movl ($dat,$YY,4),$TY#d | 147 movl ($dat,$YY,4),$TY#d |
150 movl $TX[0]#d,($dat,$YY,4) | 148 movl $TX[0]#d,($dat,$YY,4) |
151 movl $TY#d,($dat,$XX[0],4) | 149 movl $TY#d,($dat,$XX[0],4) |
152 add $TY#b,$TX[0]#b | 150 add $TY#b,$TX[0]#b |
153 inc $XX[0]#b | 151 inc $XX[0]#b |
154 movl ($dat,$TX[0],4),$TY#d | 152 movl ($dat,$TX[0],4),$TY#d |
155 movl ($dat,$XX[0],4),$TX[0]#d | 153 movl ($dat,$XX[0],4),$TX[0]#d |
156 xorb ($inp),$TY#b | 154 xorb ($inp),$TY#b |
157 inc $inp | 155 inc $inp |
158 movb $TY#b,($out) | 156 movb $TY#b,($out) |
159 inc $out | 157 inc $out |
160 dec $len | 158 dec $len |
161 jnz .Lloop1 | 159 jnz .Lloop1 |
162 jmp .Lexit | 160 jmp .Lexit |
163 | 161 |
164 .align 16 | 162 .align 16 |
165 .LRC4_CHAR: | 163 .LRC4_CHAR: |
166 add \$1,$XX[0]#b | 164 add \$1,$XX[0]#b |
167 movzb ($dat,$XX[0]),$TX[0]#d | 165 movzb ($dat,$XX[0]),$TX[0]#d |
168 test \$-8,$len | 166 test \$-8,$len |
169 jz .Lcloop1 | 167 jz .Lcloop1 |
170 » cmp» \$0,260($dat) | 168 » cmpl» \$0,260($dat) |
171 jnz .Lcloop1 | 169 jnz .Lcloop1 |
172 push %rbx | |
173 jmp .Lcloop8 | 170 jmp .Lcloop8 |
174 .align 16 | 171 .align 16 |
175 .Lcloop8: | 172 .Lcloop8: |
176 mov ($inp),%eax | 173 mov ($inp),%eax |
177 mov 4($inp),%ebx | 174 mov 4($inp),%ebx |
178 ___ | 175 ___ |
179 # unroll 2x4-wise, because 64-bit rotates kill Intel P4... | 176 # unroll 2x4-wise, because 64-bit rotates kill Intel P4... |
180 for ($i=0;$i<4;$i++) { | 177 for ($i=0;$i<4;$i++) { |
181 $code.=<<___; | 178 $code.=<<___; |
182 add $TX[0]#b,$YY#b | 179 add $TX[0]#b,$YY#b |
(...skipping 34 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
217 } | 214 } |
218 $code.=<<___; | 215 $code.=<<___; |
219 lea -8($len),$len | 216 lea -8($len),$len |
220 mov %eax,($out) | 217 mov %eax,($out) |
221 lea 8($inp),$inp | 218 lea 8($inp),$inp |
222 mov %ebx,4($out) | 219 mov %ebx,4($out) |
223 lea 8($out),$out | 220 lea 8($out),$out |
224 | 221 |
225 test \$-8,$len | 222 test \$-8,$len |
226 jnz .Lcloop8 | 223 jnz .Lcloop8 |
227 pop %rbx | |
228 cmp \$0,$len | 224 cmp \$0,$len |
229 jne .Lcloop1 | 225 jne .Lcloop1 |
230 jmp .Lexit | 226 jmp .Lexit |
231 ___ | 227 ___ |
232 $code.=<<___; | 228 $code.=<<___; |
233 .align 16 | 229 .align 16 |
234 .Lcloop1: | 230 .Lcloop1: |
235 add $TX[0]#b,$YY#b | 231 add $TX[0]#b,$YY#b |
236 movzb ($dat,$YY),$TY#d | 232 movzb ($dat,$YY),$TY#d |
237 movb $TX[0]#b,($dat,$YY) | 233 movb $TX[0]#b,($dat,$YY) |
238 movb $TY#b,($dat,$XX[0]) | 234 movb $TY#b,($dat,$XX[0]) |
239 add $TX[0]#b,$TY#b | 235 add $TX[0]#b,$TY#b |
240 add \$1,$XX[0]#b | 236 add \$1,$XX[0]#b |
241 movzb $TY#b,$TY#d | 237 movzb $TY#b,$TY#d |
242 movzb $XX[0]#b,$XX[0]#d | 238 movzb $XX[0]#b,$XX[0]#d |
243 movzb ($dat,$TY),$TY#d | 239 movzb ($dat,$TY),$TY#d |
244 movzb ($dat,$XX[0]),$TX[0]#d | 240 movzb ($dat,$XX[0]),$TX[0]#d |
245 xorb ($inp),$TY#b | 241 xorb ($inp),$TY#b |
246 lea 1($inp),$inp | 242 lea 1($inp),$inp |
247 movb $TY#b,($out) | 243 movb $TY#b,($out) |
248 lea 1($out),$out | 244 lea 1($out),$out |
249 sub \$1,$len | 245 sub \$1,$len |
250 jnz .Lcloop1 | 246 jnz .Lcloop1 |
251 jmp .Lexit | 247 jmp .Lexit |
| 248 |
| 249 .align 16 |
| 250 .Lexit: |
| 251 sub \$1,$XX[0]#b |
| 252 movl $XX[0]#d,-8($dat) |
| 253 movl $YY#d,-4($dat) |
| 254 |
| 255 mov (%rsp),%r13 |
| 256 mov 8(%rsp),%r12 |
| 257 mov 16(%rsp),%rbx |
| 258 add \$24,%rsp |
| 259 .Lepilogue: |
| 260 ret |
252 .size RC4,.-RC4 | 261 .size RC4,.-RC4 |
253 ___ | 262 ___ |
254 | 263 |
255 $idx="%r8"; | 264 $idx="%r8"; |
256 $ido="%r9"; | 265 $ido="%r9"; |
257 | 266 |
258 $code.=<<___; | 267 $code.=<<___; |
259 .extern OPENSSL_ia32cap_P | 268 .extern OPENSSL_ia32cap_P |
260 .globl RC4_set_key | 269 .globl RC4_set_key |
261 .type RC4_set_key,\@function,3 | 270 .type RC4_set_key,\@function,3 |
(...skipping 64 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
326 | 335 |
327 .align 16 | 336 .align 16 |
328 .Lexit_key: | 337 .Lexit_key: |
329 xor %eax,%eax | 338 xor %eax,%eax |
330 mov %eax,-8($dat) | 339 mov %eax,-8($dat) |
331 mov %eax,-4($dat) | 340 mov %eax,-4($dat) |
332 ret | 341 ret |
333 .size RC4_set_key,.-RC4_set_key | 342 .size RC4_set_key,.-RC4_set_key |
334 | 343 |
335 .globl RC4_options | 344 .globl RC4_options |
336 .type» RC4_options,\@function,0 | 345 .type» RC4_options,\@abi-omnipotent |
337 .align 16 | 346 .align 16 |
338 RC4_options: | 347 RC4_options: |
339 » .picmeup %rax | 348 » lea» .Lopts(%rip),%rax |
340 » lea» .Lopts-.(%rax),%rax | |
341 mov OPENSSL_ia32cap_P(%rip),%edx | 349 mov OPENSSL_ia32cap_P(%rip),%edx |
342 bt \$20,%edx | 350 bt \$20,%edx |
343 jnc .Ldone | 351 jnc .Ldone |
344 add \$12,%rax | 352 add \$12,%rax |
345 bt \$30,%edx | 353 bt \$30,%edx |
346 jnc .Ldone | 354 jnc .Ldone |
347 add \$13,%rax | 355 add \$13,%rax |
348 .Ldone: | 356 .Ldone: |
349 ret | 357 ret |
350 .align 64 | 358 .align 64 |
351 .Lopts: | 359 .Lopts: |
352 .asciz "rc4(8x,int)" | 360 .asciz "rc4(8x,int)" |
353 .asciz "rc4(8x,char)" | 361 .asciz "rc4(8x,char)" |
354 .asciz "rc4(1x,char)" | 362 .asciz "rc4(1x,char)" |
355 .asciz "RC4 for x86_64, CRYPTOGAMS by <appro\@openssl.org>" | 363 .asciz "RC4 for x86_64, CRYPTOGAMS by <appro\@openssl.org>" |
356 .align 64 | 364 .align 64 |
357 .size RC4_options,.-RC4_options | 365 .size RC4_options,.-RC4_options |
358 ___ | 366 ___ |
359 | 367 |
| 368 # EXCEPTION_DISPOSITION handler (EXCEPTION_RECORD *rec,ULONG64 frame, |
| 369 # CONTEXT *context,DISPATCHER_CONTEXT *disp) |
| 370 if ($win64) { |
| 371 $rec="%rcx"; |
| 372 $frame="%rdx"; |
| 373 $context="%r8"; |
| 374 $disp="%r9"; |
| 375 |
| 376 $code.=<<___; |
| 377 .extern __imp_RtlVirtualUnwind |
| 378 .type stream_se_handler,\@abi-omnipotent |
| 379 .align 16 |
| 380 stream_se_handler: |
| 381 push %rsi |
| 382 push %rdi |
| 383 push %rbx |
| 384 push %rbp |
| 385 push %r12 |
| 386 push %r13 |
| 387 push %r14 |
| 388 push %r15 |
| 389 pushfq |
| 390 sub \$64,%rsp |
| 391 |
| 392 mov 120($context),%rax # pull context->Rax |
| 393 mov 248($context),%rbx # pull context->Rip |
| 394 |
| 395 lea .Lprologue(%rip),%r10 |
| 396 cmp %r10,%rbx # context->Rip<prologue label |
| 397 jb .Lin_prologue |
| 398 |
| 399 mov 152($context),%rax # pull context->Rsp |
| 400 |
| 401 lea .Lepilogue(%rip),%r10 |
| 402 cmp %r10,%rbx # context->Rip>=epilogue label |
| 403 jae .Lin_prologue |
| 404 |
| 405 lea 24(%rax),%rax |
| 406 |
| 407 mov -8(%rax),%rbx |
| 408 mov -16(%rax),%r12 |
| 409 mov -24(%rax),%r13 |
| 410 mov %rbx,144($context) # restore context->Rbx |
| 411 mov %r12,216($context) # restore context->R12 |
| 412 mov %r13,224($context) # restore context->R13 |
| 413 |
| 414 .Lin_prologue: |
| 415 mov 8(%rax),%rdi |
| 416 mov 16(%rax),%rsi |
| 417 mov %rax,152($context) # restore context->Rsp |
| 418 mov %rsi,168($context) # restore context->Rsi |
| 419 mov %rdi,176($context) # restore context->Rdi |
| 420 |
| 421 jmp .Lcommon_seh_exit |
| 422 .size stream_se_handler,.-stream_se_handler |
| 423 |
| 424 .type key_se_handler,\@abi-omnipotent |
| 425 .align 16 |
| 426 key_se_handler: |
| 427 push %rsi |
| 428 push %rdi |
| 429 push %rbx |
| 430 push %rbp |
| 431 push %r12 |
| 432 push %r13 |
| 433 push %r14 |
| 434 push %r15 |
| 435 pushfq |
| 436 sub \$64,%rsp |
| 437 |
| 438 mov 152($context),%rax # pull context->Rsp |
| 439 mov 8(%rax),%rdi |
| 440 mov 16(%rax),%rsi |
| 441 mov %rsi,168($context) # restore context->Rsi |
| 442 mov %rdi,176($context) # restore context->Rdi |
| 443 |
| 444 .Lcommon_seh_exit: |
| 445 |
| 446 mov 40($disp),%rdi # disp->ContextRecord |
| 447 mov $context,%rsi # context |
| 448 mov \$154,%ecx # sizeof(CONTEXT) |
| 449 .long 0xa548f3fc # cld; rep movsq |
| 450 |
| 451 mov $disp,%rsi |
| 452 xor %rcx,%rcx # arg1, UNW_FLAG_NHANDLER |
| 453 mov 8(%rsi),%rdx # arg2, disp->ImageBase |
| 454 mov 0(%rsi),%r8 # arg3, disp->ControlPc |
| 455 mov 16(%rsi),%r9 # arg4, disp->FunctionEntry |
| 456 mov 40(%rsi),%r10 # disp->ContextRecord |
| 457 lea 56(%rsi),%r11 # &disp->HandlerData |
| 458 lea 24(%rsi),%r12 # &disp->EstablisherFrame |
| 459 mov %r10,32(%rsp) # arg5 |
| 460 mov %r11,40(%rsp) # arg6 |
| 461 mov %r12,48(%rsp) # arg7 |
| 462 mov %rcx,56(%rsp) # arg8, (NULL) |
| 463 call *__imp_RtlVirtualUnwind(%rip) |
| 464 |
| 465 mov \$1,%eax # ExceptionContinueSearch |
| 466 add \$64,%rsp |
| 467 popfq |
| 468 pop %r15 |
| 469 pop %r14 |
| 470 pop %r13 |
| 471 pop %r12 |
| 472 pop %rbp |
| 473 pop %rbx |
| 474 pop %rdi |
| 475 pop %rsi |
| 476 ret |
| 477 .size key_se_handler,.-key_se_handler |
| 478 |
| 479 .section .pdata |
| 480 .align 4 |
| 481 .rva .LSEH_begin_RC4 |
| 482 .rva .LSEH_end_RC4 |
| 483 .rva .LSEH_info_RC4 |
| 484 |
| 485 .rva .LSEH_begin_RC4_set_key |
| 486 .rva .LSEH_end_RC4_set_key |
| 487 .rva .LSEH_info_RC4_set_key |
| 488 |
| 489 .section .xdata |
| 490 .align 8 |
| 491 .LSEH_info_RC4: |
| 492 .byte 9,0,0,0 |
| 493 .rva stream_se_handler |
| 494 .LSEH_info_RC4_set_key: |
| 495 .byte 9,0,0,0 |
| 496 .rva key_se_handler |
| 497 ___ |
| 498 } |
| 499 |
360 $code =~ s/#([bwd])/$1/gm; | 500 $code =~ s/#([bwd])/$1/gm; |
361 | 501 |
362 $code =~ s/RC4_set_key/private_RC4_set_key/g if ($ENV{FIPSCANLIB} ne ""); | |
363 | |
364 print $code; | 502 print $code; |
365 | 503 |
366 close STDOUT; | 504 close STDOUT; |
OLD | NEW |