Mercurial > hg > sv-dependency-builds
comparison src/zlib-1.2.8/contrib/amd64/amd64-match.S @ 43:5ea0608b923f
Current zlib source
author | Chris Cannam |
---|---|
date | Tue, 18 Oct 2016 14:33:52 +0100 |
parents | |
children |
comparison
equal
deleted
inserted
replaced
42:2cd0e3b3e1fd | 43:5ea0608b923f |
---|---|
1 /* | |
2 * match.S -- optimized version of longest_match() | |
3 * based on the similar work by Gilles Vollant, and Brian Raiter, written 1998 | |
4 * | |
5 * This is free software; you can redistribute it and/or modify it | |
6 * under the terms of the BSD License. Use by owners of Che Guevarra | |
7 * parafernalia is prohibited, where possible, and highly discouraged | |
8 * elsewhere. | |
9 */ | |
10 | |
11 #ifndef NO_UNDERLINE | |
12 # define match_init _match_init | |
13 # define longest_match _longest_match | |
14 #endif | |
15 | |
16 #define scanend ebx | |
17 #define scanendw bx | |
18 #define chainlenwmask edx /* high word: current chain len low word: s->wmask */ | |
19 #define curmatch rsi | |
20 #define curmatchd esi | |
21 #define windowbestlen r8 | |
22 #define scanalign r9 | |
23 #define scanalignd r9d | |
24 #define window r10 | |
25 #define bestlen r11 | |
26 #define bestlend r11d | |
27 #define scanstart r12d | |
28 #define scanstartw r12w | |
29 #define scan r13 | |
30 #define nicematch r14d | |
31 #define limit r15 | |
32 #define limitd r15d | |
33 #define prev rcx | |
34 | |
35 /* | |
36 * The 258 is a "magic number, not a parameter -- changing it | |
37 * breaks the hell loose | |
38 */ | |
39 #define MAX_MATCH (258) | |
40 #define MIN_MATCH (3) | |
41 #define MIN_LOOKAHEAD (MAX_MATCH + MIN_MATCH + 1) | |
42 #define MAX_MATCH_8 ((MAX_MATCH + 7) & ~7) | |
43 | |
44 /* stack frame offsets */ | |
45 #define LocalVarsSize (112) | |
46 #define _chainlenwmask ( 8-LocalVarsSize)(%rsp) | |
47 #define _windowbestlen (16-LocalVarsSize)(%rsp) | |
48 #define save_r14 (24-LocalVarsSize)(%rsp) | |
49 #define save_rsi (32-LocalVarsSize)(%rsp) | |
50 #define save_rbx (40-LocalVarsSize)(%rsp) | |
51 #define save_r12 (56-LocalVarsSize)(%rsp) | |
52 #define save_r13 (64-LocalVarsSize)(%rsp) | |
53 #define save_r15 (80-LocalVarsSize)(%rsp) | |
54 | |
55 | |
56 .globl match_init, longest_match | |
57 | |
58 /* | |
59 * On AMD64 the first argument of a function (in our case -- the pointer to | |
60 * deflate_state structure) is passed in %rdi, hence our offsets below are | |
61 * all off of that. | |
62 */ | |
63 | |
64 /* you can check the structure offset by running | |
65 | |
66 #include <stdlib.h> | |
67 #include <stdio.h> | |
68 #include "deflate.h" | |
69 | |
70 void print_depl() | |
71 { | |
72 deflate_state ds; | |
73 deflate_state *s=&ds; | |
74 printf("size pointer=%u\n",(int)sizeof(void*)); | |
75 | |
76 printf("#define dsWSize (%3u)(%%rdi)\n",(int)(((char*)&(s->w_size))-((char*)s))); | |
77 printf("#define dsWMask (%3u)(%%rdi)\n",(int)(((char*)&(s->w_mask))-((char*)s))); | |
78 printf("#define dsWindow (%3u)(%%rdi)\n",(int)(((char*)&(s->window))-((char*)s))); | |
79 printf("#define dsPrev (%3u)(%%rdi)\n",(int)(((char*)&(s->prev))-((char*)s))); | |
80 printf("#define dsMatchLen (%3u)(%%rdi)\n",(int)(((char*)&(s->match_length))-((char*)s))); | |
81 printf("#define dsPrevMatch (%3u)(%%rdi)\n",(int)(((char*)&(s->prev_match))-((char*)s))); | |
82 printf("#define dsStrStart (%3u)(%%rdi)\n",(int)(((char*)&(s->strstart))-((char*)s))); | |
83 printf("#define dsMatchStart (%3u)(%%rdi)\n",(int)(((char*)&(s->match_start))-((char*)s))); | |
84 printf("#define dsLookahead (%3u)(%%rdi)\n",(int)(((char*)&(s->lookahead))-((char*)s))); | |
85 printf("#define dsPrevLen (%3u)(%%rdi)\n",(int)(((char*)&(s->prev_length))-((char*)s))); | |
86 printf("#define dsMaxChainLen (%3u)(%%rdi)\n",(int)(((char*)&(s->max_chain_length))-((char*)s))); | |
87 printf("#define dsGoodMatch (%3u)(%%rdi)\n",(int)(((char*)&(s->good_match))-((char*)s))); | |
88 printf("#define dsNiceMatch (%3u)(%%rdi)\n",(int)(((char*)&(s->nice_match))-((char*)s))); | |
89 } | |
90 | |
91 */ | |
92 | |
93 | |
94 /* | |
95 to compile for XCode 3.2 on MacOSX x86_64 | |
96 - run "gcc -g -c -DXCODE_MAC_X64_STRUCTURE amd64-match.S" | |
97 */ | |
98 | |
99 | |
100 #ifndef CURRENT_LINX_XCODE_MAC_X64_STRUCTURE | |
101 #define dsWSize ( 68)(%rdi) | |
102 #define dsWMask ( 76)(%rdi) | |
103 #define dsWindow ( 80)(%rdi) | |
104 #define dsPrev ( 96)(%rdi) | |
105 #define dsMatchLen (144)(%rdi) | |
106 #define dsPrevMatch (148)(%rdi) | |
107 #define dsStrStart (156)(%rdi) | |
108 #define dsMatchStart (160)(%rdi) | |
109 #define dsLookahead (164)(%rdi) | |
110 #define dsPrevLen (168)(%rdi) | |
111 #define dsMaxChainLen (172)(%rdi) | |
112 #define dsGoodMatch (188)(%rdi) | |
113 #define dsNiceMatch (192)(%rdi) | |
114 | |
115 #else | |
116 | |
117 #ifndef STRUCT_OFFSET | |
118 # define STRUCT_OFFSET (0) | |
119 #endif | |
120 | |
121 | |
122 #define dsWSize ( 56 + STRUCT_OFFSET)(%rdi) | |
123 #define dsWMask ( 64 + STRUCT_OFFSET)(%rdi) | |
124 #define dsWindow ( 72 + STRUCT_OFFSET)(%rdi) | |
125 #define dsPrev ( 88 + STRUCT_OFFSET)(%rdi) | |
126 #define dsMatchLen (136 + STRUCT_OFFSET)(%rdi) | |
127 #define dsPrevMatch (140 + STRUCT_OFFSET)(%rdi) | |
128 #define dsStrStart (148 + STRUCT_OFFSET)(%rdi) | |
129 #define dsMatchStart (152 + STRUCT_OFFSET)(%rdi) | |
130 #define dsLookahead (156 + STRUCT_OFFSET)(%rdi) | |
131 #define dsPrevLen (160 + STRUCT_OFFSET)(%rdi) | |
132 #define dsMaxChainLen (164 + STRUCT_OFFSET)(%rdi) | |
133 #define dsGoodMatch (180 + STRUCT_OFFSET)(%rdi) | |
134 #define dsNiceMatch (184 + STRUCT_OFFSET)(%rdi) | |
135 | |
136 #endif | |
137 | |
138 | |
139 | |
140 | |
141 .text | |
142 | |
143 /* uInt longest_match(deflate_state *deflatestate, IPos curmatch) */ | |
144 | |
145 longest_match: | |
146 /* | |
147 * Retrieve the function arguments. %curmatch will hold cur_match | |
148 * throughout the entire function (passed via rsi on amd64). | |
149 * rdi will hold the pointer to the deflate_state (first arg on amd64) | |
150 */ | |
151 mov %rsi, save_rsi | |
152 mov %rbx, save_rbx | |
153 mov %r12, save_r12 | |
154 mov %r13, save_r13 | |
155 mov %r14, save_r14 | |
156 mov %r15, save_r15 | |
157 | |
158 /* uInt wmask = s->w_mask; */ | |
159 /* unsigned chain_length = s->max_chain_length; */ | |
160 /* if (s->prev_length >= s->good_match) { */ | |
161 /* chain_length >>= 2; */ | |
162 /* } */ | |
163 | |
164 movl dsPrevLen, %eax | |
165 movl dsGoodMatch, %ebx | |
166 cmpl %ebx, %eax | |
167 movl dsWMask, %eax | |
168 movl dsMaxChainLen, %chainlenwmask | |
169 jl LastMatchGood | |
170 shrl $2, %chainlenwmask | |
171 LastMatchGood: | |
172 | |
173 /* chainlen is decremented once beforehand so that the function can */ | |
174 /* use the sign flag instead of the zero flag for the exit test. */ | |
175 /* It is then shifted into the high word, to make room for the wmask */ | |
176 /* value, which it will always accompany. */ | |
177 | |
178 decl %chainlenwmask | |
179 shll $16, %chainlenwmask | |
180 orl %eax, %chainlenwmask | |
181 | |
182 /* if ((uInt)nice_match > s->lookahead) nice_match = s->lookahead; */ | |
183 | |
184 movl dsNiceMatch, %eax | |
185 movl dsLookahead, %ebx | |
186 cmpl %eax, %ebx | |
187 jl LookaheadLess | |
188 movl %eax, %ebx | |
189 LookaheadLess: movl %ebx, %nicematch | |
190 | |
191 /* register Bytef *scan = s->window + s->strstart; */ | |
192 | |
193 mov dsWindow, %window | |
194 movl dsStrStart, %limitd | |
195 lea (%limit, %window), %scan | |
196 | |
197 /* Determine how many bytes the scan ptr is off from being */ | |
198 /* dword-aligned. */ | |
199 | |
200 mov %scan, %scanalign | |
201 negl %scanalignd | |
202 andl $3, %scanalignd | |
203 | |
204 /* IPos limit = s->strstart > (IPos)MAX_DIST(s) ? */ | |
205 /* s->strstart - (IPos)MAX_DIST(s) : NIL; */ | |
206 | |
207 movl dsWSize, %eax | |
208 subl $MIN_LOOKAHEAD, %eax | |
209 xorl %ecx, %ecx | |
210 subl %eax, %limitd | |
211 cmovng %ecx, %limitd | |
212 | |
213 /* int best_len = s->prev_length; */ | |
214 | |
215 movl dsPrevLen, %bestlend | |
216 | |
217 /* Store the sum of s->window + best_len in %windowbestlen locally, and in memory. */ | |
218 | |
219 lea (%window, %bestlen), %windowbestlen | |
220 mov %windowbestlen, _windowbestlen | |
221 | |
222 /* register ush scan_start = *(ushf*)scan; */ | |
223 /* register ush scan_end = *(ushf*)(scan+best_len-1); */ | |
224 /* Posf *prev = s->prev; */ | |
225 | |
226 movzwl (%scan), %scanstart | |
227 movzwl -1(%scan, %bestlen), %scanend | |
228 mov dsPrev, %prev | |
229 | |
230 /* Jump into the main loop. */ | |
231 | |
232 movl %chainlenwmask, _chainlenwmask | |
233 jmp LoopEntry | |
234 | |
235 .balign 16 | |
236 | |
237 /* do { | |
238 * match = s->window + cur_match; | |
239 * if (*(ushf*)(match+best_len-1) != scan_end || | |
240 * *(ushf*)match != scan_start) continue; | |
241 * [...] | |
242 * } while ((cur_match = prev[cur_match & wmask]) > limit | |
243 * && --chain_length != 0); | |
244 * | |
245 * Here is the inner loop of the function. The function will spend the | |
246 * majority of its time in this loop, and majority of that time will | |
247 * be spent in the first ten instructions. | |
248 */ | |
249 LookupLoop: | |
250 andl %chainlenwmask, %curmatchd | |
251 movzwl (%prev, %curmatch, 2), %curmatchd | |
252 cmpl %limitd, %curmatchd | |
253 jbe LeaveNow | |
254 subl $0x00010000, %chainlenwmask | |
255 js LeaveNow | |
256 LoopEntry: cmpw -1(%windowbestlen, %curmatch), %scanendw | |
257 jne LookupLoop | |
258 cmpw %scanstartw, (%window, %curmatch) | |
259 jne LookupLoop | |
260 | |
261 /* Store the current value of chainlen. */ | |
262 movl %chainlenwmask, _chainlenwmask | |
263 | |
264 /* %scan is the string under scrutiny, and %prev to the string we */ | |
265 /* are hoping to match it up with. In actuality, %esi and %edi are */ | |
266 /* both pointed (MAX_MATCH_8 - scanalign) bytes ahead, and %edx is */ | |
267 /* initialized to -(MAX_MATCH_8 - scanalign). */ | |
268 | |
269 mov $(-MAX_MATCH_8), %rdx | |
270 lea (%curmatch, %window), %windowbestlen | |
271 lea MAX_MATCH_8(%windowbestlen, %scanalign), %windowbestlen | |
272 lea MAX_MATCH_8(%scan, %scanalign), %prev | |
273 | |
274 /* the prefetching below makes very little difference... */ | |
275 prefetcht1 (%windowbestlen, %rdx) | |
276 prefetcht1 (%prev, %rdx) | |
277 | |
278 /* | |
279 * Test the strings for equality, 8 bytes at a time. At the end, | |
280 * adjust %rdx so that it is offset to the exact byte that mismatched. | |
281 * | |
282 * It should be confessed that this loop usually does not represent | |
283 * much of the total running time. Replacing it with a more | |
284 * straightforward "rep cmpsb" would not drastically degrade | |
285 * performance -- unrolling it, for example, makes no difference. | |
286 */ | |
287 | |
288 #undef USE_SSE /* works, but is 6-7% slower, than non-SSE... */ | |
289 | |
290 LoopCmps: | |
291 #ifdef USE_SSE | |
292 /* Preload the SSE registers */ | |
293 movdqu (%windowbestlen, %rdx), %xmm1 | |
294 movdqu (%prev, %rdx), %xmm2 | |
295 pcmpeqb %xmm2, %xmm1 | |
296 movdqu 16(%windowbestlen, %rdx), %xmm3 | |
297 movdqu 16(%prev, %rdx), %xmm4 | |
298 pcmpeqb %xmm4, %xmm3 | |
299 movdqu 32(%windowbestlen, %rdx), %xmm5 | |
300 movdqu 32(%prev, %rdx), %xmm6 | |
301 pcmpeqb %xmm6, %xmm5 | |
302 movdqu 48(%windowbestlen, %rdx), %xmm7 | |
303 movdqu 48(%prev, %rdx), %xmm8 | |
304 pcmpeqb %xmm8, %xmm7 | |
305 | |
306 /* Check the comparisions' results */ | |
307 pmovmskb %xmm1, %rax | |
308 notw %ax | |
309 bsfw %ax, %ax | |
310 jnz LeaveLoopCmps | |
311 | |
312 /* this is the only iteration of the loop with a possibility of having | |
313 incremented rdx by 0x108 (each loop iteration add 16*4 = 0x40 | |
314 and (0x40*4)+8=0x108 */ | |
315 add $8, %rdx | |
316 jz LenMaximum | |
317 add $8, %rdx | |
318 | |
319 | |
320 pmovmskb %xmm3, %rax | |
321 notw %ax | |
322 bsfw %ax, %ax | |
323 jnz LeaveLoopCmps | |
324 | |
325 | |
326 add $16, %rdx | |
327 | |
328 | |
329 pmovmskb %xmm5, %rax | |
330 notw %ax | |
331 bsfw %ax, %ax | |
332 jnz LeaveLoopCmps | |
333 | |
334 add $16, %rdx | |
335 | |
336 | |
337 pmovmskb %xmm7, %rax | |
338 notw %ax | |
339 bsfw %ax, %ax | |
340 jnz LeaveLoopCmps | |
341 | |
342 add $16, %rdx | |
343 | |
344 jmp LoopCmps | |
345 LeaveLoopCmps: add %rax, %rdx | |
346 #else | |
347 mov (%windowbestlen, %rdx), %rax | |
348 xor (%prev, %rdx), %rax | |
349 jnz LeaveLoopCmps | |
350 | |
351 mov 8(%windowbestlen, %rdx), %rax | |
352 xor 8(%prev, %rdx), %rax | |
353 jnz LeaveLoopCmps8 | |
354 | |
355 mov 16(%windowbestlen, %rdx), %rax | |
356 xor 16(%prev, %rdx), %rax | |
357 jnz LeaveLoopCmps16 | |
358 | |
359 add $24, %rdx | |
360 jnz LoopCmps | |
361 jmp LenMaximum | |
362 # if 0 | |
363 /* | |
364 * This three-liner is tantalizingly simple, but bsf is a slow instruction, | |
365 * and the complicated alternative down below is quite a bit faster. Sad... | |
366 */ | |
367 | |
368 LeaveLoopCmps: bsf %rax, %rax /* find the first non-zero bit */ | |
369 shrl $3, %eax /* divide by 8 to get the byte */ | |
370 add %rax, %rdx | |
371 # else | |
372 LeaveLoopCmps16: | |
373 add $8, %rdx | |
374 LeaveLoopCmps8: | |
375 add $8, %rdx | |
376 LeaveLoopCmps: testl $0xFFFFFFFF, %eax /* Check the first 4 bytes */ | |
377 jnz Check16 | |
378 add $4, %rdx | |
379 shr $32, %rax | |
380 Check16: testw $0xFFFF, %ax | |
381 jnz LenLower | |
382 add $2, %rdx | |
383 shrl $16, %eax | |
384 LenLower: subb $1, %al | |
385 adc $0, %rdx | |
386 # endif | |
387 #endif | |
388 | |
389 /* Calculate the length of the match. If it is longer than MAX_MATCH, */ | |
390 /* then automatically accept it as the best possible match and leave. */ | |
391 | |
392 lea (%prev, %rdx), %rax | |
393 sub %scan, %rax | |
394 cmpl $MAX_MATCH, %eax | |
395 jge LenMaximum | |
396 | |
397 /* If the length of the match is not longer than the best match we */ | |
398 /* have so far, then forget it and return to the lookup loop. */ | |
399 | |
400 cmpl %bestlend, %eax | |
401 jg LongerMatch | |
402 mov _windowbestlen, %windowbestlen | |
403 mov dsPrev, %prev | |
404 movl _chainlenwmask, %edx | |
405 jmp LookupLoop | |
406 | |
407 /* s->match_start = cur_match; */ | |
408 /* best_len = len; */ | |
409 /* if (len >= nice_match) break; */ | |
410 /* scan_end = *(ushf*)(scan+best_len-1); */ | |
411 | |
412 LongerMatch: | |
413 movl %eax, %bestlend | |
414 movl %curmatchd, dsMatchStart | |
415 cmpl %nicematch, %eax | |
416 jge LeaveNow | |
417 | |
418 lea (%window, %bestlen), %windowbestlen | |
419 mov %windowbestlen, _windowbestlen | |
420 | |
421 movzwl -1(%scan, %rax), %scanend | |
422 mov dsPrev, %prev | |
423 movl _chainlenwmask, %chainlenwmask | |
424 jmp LookupLoop | |
425 | |
426 /* Accept the current string, with the maximum possible length. */ | |
427 | |
428 LenMaximum: | |
429 movl $MAX_MATCH, %bestlend | |
430 movl %curmatchd, dsMatchStart | |
431 | |
432 /* if ((uInt)best_len <= s->lookahead) return (uInt)best_len; */ | |
433 /* return s->lookahead; */ | |
434 | |
435 LeaveNow: | |
436 movl dsLookahead, %eax | |
437 cmpl %eax, %bestlend | |
438 cmovngl %bestlend, %eax | |
439 LookaheadRet: | |
440 | |
441 /* Restore the registers and return from whence we came. */ | |
442 | |
443 mov save_rsi, %rsi | |
444 mov save_rbx, %rbx | |
445 mov save_r12, %r12 | |
446 mov save_r13, %r13 | |
447 mov save_r14, %r14 | |
448 mov save_r15, %r15 | |
449 | |
450 ret | |
451 | |
452 match_init: ret |