]> git.ipfire.org Git - thirdparty/openssl.git/blame - crypto/md5/asm/md5-x86_64.pl
x86_64 assembly pack: make Windows build more robust [from master].
[thirdparty/openssl.git] / crypto / md5 / asm / md5-x86_64.pl
CommitLineData
d37a65bc
AP
1#!/usr/bin/perl -w
2#
3# MD5 optimized for AMD64.
4#
5# Author: Marc Bevand <bevand_m (at) epita.fr>
6# Licence: I hereby disclaim the copyright on this code and place it
7# in the public domain.
8#
9
10use strict;
11
12my $code;
13
14# round1_step() does:
15# dst = x + ((dst + F(x,y,z) + X[k] + T_i) <<< s)
16# %r10d = X[k_next]
17# %r11d = z' (copy of z for the next step)
9c4fe782 18# Each round1_step() takes about 5.3 clocks (9 instructions, 1.7 IPC)
d37a65bc
AP
19sub round1_step
20{
21 my ($pos, $dst, $x, $y, $z, $k_next, $T_i, $s) = @_;
22 $code .= " mov 0*4(%rsi), %r10d /* (NEXT STEP) X[0] */\n" if ($pos == -1);
23 $code .= " mov %edx, %r11d /* (NEXT STEP) z' = %edx */\n" if ($pos == -1);
24 $code .= <<EOF;
25 xor $y, %r11d /* y ^ ... */
26 lea $T_i($dst,%r10d),$dst /* Const + dst + ... */
27 and $x, %r11d /* x & ... */
28 xor $z, %r11d /* z ^ ... */
29 mov $k_next*4(%rsi),%r10d /* (NEXT STEP) X[$k_next] */
30 add %r11d, $dst /* dst += ... */
31 rol \$$s, $dst /* dst <<< s */
32 mov $y, %r11d /* (NEXT STEP) z' = $y */
33 add $x, $dst /* dst += x */
34EOF
35}
36
37# round2_step() does:
38# dst = x + ((dst + G(x,y,z) + X[k] + T_i) <<< s)
39# %r10d = X[k_next]
9c4fe782
AP
40# %r11d = z' (copy of z for the next step)
41# %r12d = z' (copy of z for the next step)
42# Each round2_step() takes about 5.4 clocks (11 instructions, 2.0 IPC)
d37a65bc
AP
43sub round2_step
44{
45 my ($pos, $dst, $x, $y, $z, $k_next, $T_i, $s) = @_;
46 $code .= " mov 1*4(%rsi), %r10d /* (NEXT STEP) X[1] */\n" if ($pos == -1);
9c4fe782
AP
47 $code .= " mov %edx, %r11d /* (NEXT STEP) z' = %edx */\n" if ($pos == -1);
48 $code .= " mov %edx, %r12d /* (NEXT STEP) z' = %edx */\n" if ($pos == -1);
d37a65bc 49 $code .= <<EOF;
9c4fe782 50 not %r11d /* not z */
d37a65bc 51 lea $T_i($dst,%r10d),$dst /* Const + dst + ... */
9c4fe782
AP
52 and $x, %r12d /* x & z */
53 and $y, %r11d /* y & (not z) */
d37a65bc 54 mov $k_next*4(%rsi),%r10d /* (NEXT STEP) X[$k_next] */
9c4fe782
AP
55 or %r11d, %r12d /* (y & (not z)) | (x & z) */
56 mov $y, %r11d /* (NEXT STEP) z' = $y */
57 add %r12d, $dst /* dst += ... */
58 mov $y, %r12d /* (NEXT STEP) z' = $y */
d37a65bc 59 rol \$$s, $dst /* dst <<< s */
d37a65bc
AP
60 add $x, $dst /* dst += x */
61EOF
62}
63
64# round3_step() does:
65# dst = x + ((dst + H(x,y,z) + X[k] + T_i) <<< s)
66# %r10d = X[k_next]
67# %r11d = y' (copy of y for the next step)
9c4fe782 68# Each round3_step() takes about 4.2 clocks (8 instructions, 1.9 IPC)
d37a65bc
AP
69sub round3_step
70{
71 my ($pos, $dst, $x, $y, $z, $k_next, $T_i, $s) = @_;
72 $code .= " mov 5*4(%rsi), %r10d /* (NEXT STEP) X[5] */\n" if ($pos == -1);
73 $code .= " mov %ecx, %r11d /* (NEXT STEP) y' = %ecx */\n" if ($pos == -1);
74 $code .= <<EOF;
75 lea $T_i($dst,%r10d),$dst /* Const + dst + ... */
76 mov $k_next*4(%rsi),%r10d /* (NEXT STEP) X[$k_next] */
77 xor $z, %r11d /* z ^ ... */
78 xor $x, %r11d /* x ^ ... */
79 add %r11d, $dst /* dst += ... */
80 rol \$$s, $dst /* dst <<< s */
81 mov $x, %r11d /* (NEXT STEP) y' = $x */
82 add $x, $dst /* dst += x */
83EOF
84}
85
86# round4_step() does:
87# dst = x + ((dst + I(x,y,z) + X[k] + T_i) <<< s)
88# %r10d = X[k_next]
89# %r11d = not z' (copy of not z for the next step)
9c4fe782 90# Each round4_step() takes about 5.2 clocks (9 instructions, 1.7 IPC)
d37a65bc
AP
91sub round4_step
92{
93 my ($pos, $dst, $x, $y, $z, $k_next, $T_i, $s) = @_;
94 $code .= " mov 0*4(%rsi), %r10d /* (NEXT STEP) X[0] */\n" if ($pos == -1);
95 $code .= " mov \$0xffffffff, %r11d\n" if ($pos == -1);
96 $code .= " xor %edx, %r11d /* (NEXT STEP) not z' = not %edx*/\n"
97 if ($pos == -1);
98 $code .= <<EOF;
99 lea $T_i($dst,%r10d),$dst /* Const + dst + ... */
100 or $x, %r11d /* x | ... */
101 xor $y, %r11d /* y ^ ... */
102 add %r11d, $dst /* dst += ... */
103 mov $k_next*4(%rsi),%r10d /* (NEXT STEP) X[$k_next] */
104 mov \$0xffffffff, %r11d
105 rol \$$s, $dst /* dst <<< s */
106 xor $y, %r11d /* (NEXT STEP) not z' = not $y */
107 add $x, $dst /* dst += x */
108EOF
109}
110
be01f79d
AP
111my $flavour = shift;
112my $output = shift;
113if ($flavour =~ /\./) { $output = $flavour; undef $flavour; }
114
115my $win64=0; $win64=1 if ($flavour =~ /[nm]asm|mingw64/ || $output =~ /\.asm$/);
abe7f8b4
AP
116
117$0 =~ m/(.*[\/\\])[^\/\\]+$/; my $dir=$1; my $xlate;
118( $xlate="${dir}x86_64-xlate.pl" and -f $xlate ) or
119( $xlate="${dir}../../perlasm/x86_64-xlate.pl" and -f $xlate) or
120die "can't locate x86_64-xlate.pl";
121
be01f79d 122no warnings qw(uninitialized);
b286a968
AP
123open OUT,"| \"$^X\" $xlate $flavour $output";
124*STDOUT=*OUT;
d37a65bc
AP
125
126$code .= <<EOF;
127.text
128.align 16
129
c69ed6ea
AP
130.globl md5_block_asm_data_order
131.type md5_block_asm_data_order,\@function,3
132md5_block_asm_data_order:
d37a65bc
AP
133 push %rbp
134 push %rbx
9c4fe782 135 push %r12
d37a65bc
AP
136 push %r14
137 push %r15
be01f79d 138.Lprologue:
d37a65bc
AP
139
140 # rdi = arg #1 (ctx, MD5_CTX pointer)
141 # rsi = arg #2 (ptr, data pointer)
142 # rdx = arg #3 (nbr, number of 16-word blocks to process)
143 mov %rdi, %rbp # rbp = ctx
144 shl \$6, %rdx # rdx = nbr in bytes
145 lea (%rsi,%rdx), %rdi # rdi = end
146 mov 0*4(%rbp), %eax # eax = ctx->A
147 mov 1*4(%rbp), %ebx # ebx = ctx->B
148 mov 2*4(%rbp), %ecx # ecx = ctx->C
149 mov 3*4(%rbp), %edx # edx = ctx->D
d37a65bc
AP
150 # end is 'rdi'
151 # ptr is 'rsi'
152 # A is 'eax'
153 # B is 'ebx'
154 # C is 'ecx'
155 # D is 'edx'
156
157 cmp %rdi, %rsi # cmp end with ptr
8b5bf52a 158 je .Lend # jmp if ptr == end
d37a65bc
AP
159
160 # BEGIN of loop over 16-word blocks
8b5bf52a 161.Lloop: # save old values of A, B, C, D
d37a65bc
AP
162 mov %eax, %r8d
163 mov %ebx, %r9d
164 mov %ecx, %r14d
165 mov %edx, %r15d
166EOF
167round1_step(-1,'%eax','%ebx','%ecx','%edx', '1','0xd76aa478', '7');
168round1_step( 0,'%edx','%eax','%ebx','%ecx', '2','0xe8c7b756','12');
169round1_step( 0,'%ecx','%edx','%eax','%ebx', '3','0x242070db','17');
170round1_step( 0,'%ebx','%ecx','%edx','%eax', '4','0xc1bdceee','22');
171round1_step( 0,'%eax','%ebx','%ecx','%edx', '5','0xf57c0faf', '7');
172round1_step( 0,'%edx','%eax','%ebx','%ecx', '6','0x4787c62a','12');
173round1_step( 0,'%ecx','%edx','%eax','%ebx', '7','0xa8304613','17');
174round1_step( 0,'%ebx','%ecx','%edx','%eax', '8','0xfd469501','22');
175round1_step( 0,'%eax','%ebx','%ecx','%edx', '9','0x698098d8', '7');
176round1_step( 0,'%edx','%eax','%ebx','%ecx','10','0x8b44f7af','12');
177round1_step( 0,'%ecx','%edx','%eax','%ebx','11','0xffff5bb1','17');
178round1_step( 0,'%ebx','%ecx','%edx','%eax','12','0x895cd7be','22');
179round1_step( 0,'%eax','%ebx','%ecx','%edx','13','0x6b901122', '7');
180round1_step( 0,'%edx','%eax','%ebx','%ecx','14','0xfd987193','12');
181round1_step( 0,'%ecx','%edx','%eax','%ebx','15','0xa679438e','17');
182round1_step( 1,'%ebx','%ecx','%edx','%eax', '0','0x49b40821','22');
183
184round2_step(-1,'%eax','%ebx','%ecx','%edx', '6','0xf61e2562', '5');
185round2_step( 0,'%edx','%eax','%ebx','%ecx','11','0xc040b340', '9');
186round2_step( 0,'%ecx','%edx','%eax','%ebx', '0','0x265e5a51','14');
187round2_step( 0,'%ebx','%ecx','%edx','%eax', '5','0xe9b6c7aa','20');
188round2_step( 0,'%eax','%ebx','%ecx','%edx','10','0xd62f105d', '5');
189round2_step( 0,'%edx','%eax','%ebx','%ecx','15', '0x2441453', '9');
190round2_step( 0,'%ecx','%edx','%eax','%ebx', '4','0xd8a1e681','14');
191round2_step( 0,'%ebx','%ecx','%edx','%eax', '9','0xe7d3fbc8','20');
192round2_step( 0,'%eax','%ebx','%ecx','%edx','14','0x21e1cde6', '5');
193round2_step( 0,'%edx','%eax','%ebx','%ecx', '3','0xc33707d6', '9');
194round2_step( 0,'%ecx','%edx','%eax','%ebx', '8','0xf4d50d87','14');
195round2_step( 0,'%ebx','%ecx','%edx','%eax','13','0x455a14ed','20');
196round2_step( 0,'%eax','%ebx','%ecx','%edx', '2','0xa9e3e905', '5');
197round2_step( 0,'%edx','%eax','%ebx','%ecx', '7','0xfcefa3f8', '9');
198round2_step( 0,'%ecx','%edx','%eax','%ebx','12','0x676f02d9','14');
199round2_step( 1,'%ebx','%ecx','%edx','%eax', '0','0x8d2a4c8a','20');
200
201round3_step(-1,'%eax','%ebx','%ecx','%edx', '8','0xfffa3942', '4');
202round3_step( 0,'%edx','%eax','%ebx','%ecx','11','0x8771f681','11');
203round3_step( 0,'%ecx','%edx','%eax','%ebx','14','0x6d9d6122','16');
204round3_step( 0,'%ebx','%ecx','%edx','%eax', '1','0xfde5380c','23');
205round3_step( 0,'%eax','%ebx','%ecx','%edx', '4','0xa4beea44', '4');
206round3_step( 0,'%edx','%eax','%ebx','%ecx', '7','0x4bdecfa9','11');
207round3_step( 0,'%ecx','%edx','%eax','%ebx','10','0xf6bb4b60','16');
208round3_step( 0,'%ebx','%ecx','%edx','%eax','13','0xbebfbc70','23');
209round3_step( 0,'%eax','%ebx','%ecx','%edx', '0','0x289b7ec6', '4');
210round3_step( 0,'%edx','%eax','%ebx','%ecx', '3','0xeaa127fa','11');
211round3_step( 0,'%ecx','%edx','%eax','%ebx', '6','0xd4ef3085','16');
212round3_step( 0,'%ebx','%ecx','%edx','%eax', '9', '0x4881d05','23');
213round3_step( 0,'%eax','%ebx','%ecx','%edx','12','0xd9d4d039', '4');
214round3_step( 0,'%edx','%eax','%ebx','%ecx','15','0xe6db99e5','11');
215round3_step( 0,'%ecx','%edx','%eax','%ebx', '2','0x1fa27cf8','16');
216round3_step( 1,'%ebx','%ecx','%edx','%eax', '0','0xc4ac5665','23');
217
218round4_step(-1,'%eax','%ebx','%ecx','%edx', '7','0xf4292244', '6');
219round4_step( 0,'%edx','%eax','%ebx','%ecx','14','0x432aff97','10');
220round4_step( 0,'%ecx','%edx','%eax','%ebx', '5','0xab9423a7','15');
221round4_step( 0,'%ebx','%ecx','%edx','%eax','12','0xfc93a039','21');
222round4_step( 0,'%eax','%ebx','%ecx','%edx', '3','0x655b59c3', '6');
223round4_step( 0,'%edx','%eax','%ebx','%ecx','10','0x8f0ccc92','10');
224round4_step( 0,'%ecx','%edx','%eax','%ebx', '1','0xffeff47d','15');
225round4_step( 0,'%ebx','%ecx','%edx','%eax', '8','0x85845dd1','21');
226round4_step( 0,'%eax','%ebx','%ecx','%edx','15','0x6fa87e4f', '6');
227round4_step( 0,'%edx','%eax','%ebx','%ecx', '6','0xfe2ce6e0','10');
228round4_step( 0,'%ecx','%edx','%eax','%ebx','13','0xa3014314','15');
229round4_step( 0,'%ebx','%ecx','%edx','%eax', '4','0x4e0811a1','21');
230round4_step( 0,'%eax','%ebx','%ecx','%edx','11','0xf7537e82', '6');
231round4_step( 0,'%edx','%eax','%ebx','%ecx', '2','0xbd3af235','10');
232round4_step( 0,'%ecx','%edx','%eax','%ebx', '9','0x2ad7d2bb','15');
233round4_step( 1,'%ebx','%ecx','%edx','%eax', '0','0xeb86d391','21');
234$code .= <<EOF;
235 # add old values of A, B, C, D
236 add %r8d, %eax
237 add %r9d, %ebx
238 add %r14d, %ecx
239 add %r15d, %edx
240
241 # loop control
242 add \$64, %rsi # ptr += 64
243 cmp %rdi, %rsi # cmp end with ptr
8b5bf52a 244 jb .Lloop # jmp if ptr < end
d37a65bc
AP
245 # END of loop over 16-word blocks
246
8b5bf52a 247.Lend:
d37a65bc
AP
248 mov %eax, 0*4(%rbp) # ctx->A = A
249 mov %ebx, 1*4(%rbp) # ctx->B = B
250 mov %ecx, 2*4(%rbp) # ctx->C = C
251 mov %edx, 3*4(%rbp) # ctx->D = D
252
be01f79d
AP
253 mov (%rsp),%r15
254 mov 8(%rsp),%r14
255 mov 16(%rsp),%r12
256 mov 24(%rsp),%rbx
257 mov 32(%rsp),%rbp
258 add \$40,%rsp
259.Lepilogue:
260 ret
261.size md5_block_asm_data_order,.-md5_block_asm_data_order
262EOF
263
264# EXCEPTION_DISPOSITION handler (EXCEPTION_RECORD *rec,ULONG64 frame,
265# CONTEXT *context,DISPATCHER_CONTEXT *disp)
266if ($win64) {
267my $rec="%rcx";
268my $frame="%rdx";
269my $context="%r8";
270my $disp="%r9";
271
272$code.=<<___;
273.extern __imp_RtlVirtualUnwind
274.type se_handler,\@abi-omnipotent
275.align 16
276se_handler:
277 push %rsi
278 push %rdi
279 push %rbx
280 push %rbp
281 push %r12
282 push %r13
283 push %r14
284 push %r15
285 pushfq
286 sub \$64,%rsp
287
288 mov 120($context),%rax # pull context->Rax
289 mov 248($context),%rbx # pull context->Rip
290
291 lea .Lprologue(%rip),%r10
292 cmp %r10,%rbx # context->Rip<.Lprologue
293 jb .Lin_prologue
294
295 mov 152($context),%rax # pull context->Rsp
296
297 lea .Lepilogue(%rip),%r10
298 cmp %r10,%rbx # context->Rip>=.Lepilogue
299 jae .Lin_prologue
300
301 lea 40(%rax),%rax
302
303 mov -8(%rax),%rbp
304 mov -16(%rax),%rbx
305 mov -24(%rax),%r12
306 mov -32(%rax),%r14
307 mov -40(%rax),%r15
308 mov %rbx,144($context) # restore context->Rbx
309 mov %rbp,160($context) # restore context->Rbp
310 mov %r12,216($context) # restore context->R12
311 mov %r14,232($context) # restore context->R14
312 mov %r15,240($context) # restore context->R15
313
314.Lin_prologue:
315 mov 8(%rax),%rdi
316 mov 16(%rax),%rsi
317 mov %rax,152($context) # restore context->Rsp
318 mov %rsi,168($context) # restore context->Rsi
319 mov %rdi,176($context) # restore context->Rdi
320
321 mov 40($disp),%rdi # disp->ContextRecord
322 mov $context,%rsi # context
323 mov \$154,%ecx # sizeof(CONTEXT)
324 .long 0xa548f3fc # cld; rep movsq
325
326 mov $disp,%rsi
327 xor %rcx,%rcx # arg1, UNW_FLAG_NHANDLER
328 mov 8(%rsi),%rdx # arg2, disp->ImageBase
329 mov 0(%rsi),%r8 # arg3, disp->ControlPc
330 mov 16(%rsi),%r9 # arg4, disp->FunctionEntry
331 mov 40(%rsi),%r10 # disp->ContextRecord
332 lea 56(%rsi),%r11 # &disp->HandlerData
333 lea 24(%rsi),%r12 # &disp->EstablisherFrame
334 mov %r10,32(%rsp) # arg5
335 mov %r11,40(%rsp) # arg6
336 mov %r12,48(%rsp) # arg7
337 mov %rcx,56(%rsp) # arg8, (NULL)
338 call *__imp_RtlVirtualUnwind(%rip)
339
340 mov \$1,%eax # ExceptionContinueSearch
341 add \$64,%rsp
342 popfq
d37a65bc
AP
343 pop %r15
344 pop %r14
be01f79d 345 pop %r13
9c4fe782 346 pop %r12
d37a65bc 347 pop %rbp
be01f79d
AP
348 pop %rbx
349 pop %rdi
350 pop %rsi
d37a65bc 351 ret
be01f79d
AP
352.size se_handler,.-se_handler
353
354.section .pdata
355.align 4
356 .rva .LSEH_begin_md5_block_asm_data_order
357 .rva .LSEH_end_md5_block_asm_data_order
358 .rva .LSEH_info_md5_block_asm_data_order
359
360.section .xdata
361.align 8
362.LSEH_info_md5_block_asm_data_order:
363 .byte 9,0,0,0
364 .rva se_handler
365___
366}
d37a65bc
AP
367
368print $code;
8b5bf52a
AP
369
370close STDOUT;