File tree Expand file tree Collapse file tree 2 files changed +31
-31
lines changed Expand file tree Collapse file tree 2 files changed +31
-31
lines changed Original file line number Diff line number Diff line change @@ -327,31 +327,31 @@ SYM_FUNC_START(__tfw_strtolower_avx2)
327
327
328
328
.text
329
329
.str2low_len8:
330
- endbr64;
330
+ endbr64
331
331
__STRTOLOWER_SMALL_STR 7
332
332
.str2low_len7:
333
- endbr64;
333
+ endbr64
334
334
__STRTOLOWER_SMALL_STR 6
335
335
.str2low_len6:
336
- endbr64;
336
+ endbr64
337
337
__STRTOLOWER_SMALL_STR 5
338
338
.str2low_len5:
339
- endbr64;
339
+ endbr64
340
340
__STRTOLOWER_SMALL_STR 4
341
341
.str2low_len4:
342
- endbr64;
342
+ endbr64
343
343
__STRTOLOWER_SMALL_STR 3
344
344
.str2low_len3:
345
- endbr64;
345
+ endbr64
346
346
__STRTOLOWER_SMALL_STR 2
347
347
.str2low_len2:
348
- endbr64;
348
+ endbr64
349
349
__STRTOLOWER_SMALL_STR 1
350
350
.str2low_len1:
351
- endbr64;
351
+ endbr64
352
352
__STRTOLOWER_SMALL_STR 0
353
353
.str2low_len0:
354
- endbr64;
354
+ endbr64
355
355
addq $88 , %rsp
356
356
popq %rbx
357
357
popq %r10
@@ -505,7 +505,7 @@ SYM_FUNC_START(__tfw_stricmp_avx2)
505
505
.quad .stricmp_len8
506
506
.text
507
507
.stricmp_len7:
508
- endbr64;
508
+ endbr64
509
509
xorl %eax , %eax
510
510
/*
511
511
* The blocks at the below use complex mixture for the registers,
@@ -562,35 +562,35 @@ SYM_FUNC_START(__tfw_stricmp_avx2)
562
562
orl %edx , %eax
563
563
RET
564
564
.stricmp_len0:
565
- endbr64;
565
+ endbr64
566
566
xorl %eax , %eax
567
567
RET
568
568
.stricmp_len1:
569
- endbr64;
569
+ endbr64
570
570
xorl %edx , %edx
571
571
jmp .stricmp_do_len1
572
572
.stricmp_len2:
573
- endbr64;
573
+ endbr64
574
574
xorl %edx , %edx
575
575
jmp .stricmp_do_len2
576
576
.stricmp_len3:
577
- endbr64;
577
+ endbr64
578
578
xorl %eax , %eax
579
579
jmp .stricmp_do_len3
580
580
.stricmp_len4:
581
- endbr64;
581
+ endbr64
582
582
xorl %eax , %eax
583
583
jmp .stricmp_do_len4
584
584
.stricmp_len5:
585
- endbr64;
585
+ endbr64
586
586
xorl %edx , %edx
587
587
jmp .stricmp_do_len5
588
588
.stricmp_len6:
589
- endbr64;
589
+ endbr64
590
590
xorl %edx , %edx
591
591
jmp .stricmp_do_len6
592
592
.stricmp_len8:
593
- endbr64;
593
+ endbr64
594
594
movzbl 7 (%rdi ), %edx
595
595
movzbl 7 (%rsi ), %eax
596
596
movzbl __tfw_lct(%rdx ), %edx
@@ -871,7 +871,7 @@ SYM_FUNC_START(__tfw_stricmp_avx2_2lc)
871
871
.quad .sic2lc_len8
872
872
.text
873
873
.sic2lc_len7:
874
- endbr64;
874
+ endbr64
875
875
xorl %eax , %eax
876
876
.sic2lc_do_len7:
877
877
movzbl 6 (%rdi ), %edx
@@ -919,35 +919,35 @@ SYM_FUNC_START(__tfw_stricmp_avx2_2lc)
919
919
orl %edx , %eax
920
920
RET
921
921
.sic2lc_len0:
922
- endbr64;
922
+ endbr64
923
923
xorl %eax , %eax
924
924
RET
925
925
.sic2lc_len1:
926
- endbr64;
926
+ endbr64
927
927
xorl %edx , %edx
928
928
jmp .sic2lc_do_len1
929
929
.sic2lc_len2:
930
- endbr64;
930
+ endbr64
931
931
xorl %edx , %edx
932
932
jmp .sic2lc_do_len2
933
933
.sic2lc_len3:
934
- endbr64;
934
+ endbr64
935
935
xorl %eax , %eax
936
936
jmp .sic2lc_do_len3
937
937
.sic2lc_len4:
938
- endbr64;
938
+ endbr64
939
939
xorl %eax , %eax
940
940
jmp .sic2lc_do_len4
941
941
.sic2lc_len5:
942
- endbr64;
942
+ endbr64
943
943
xorl %edx , %edx
944
944
jmp .sic2lc_do_len5
945
945
.sic2lc_len6:
946
- endbr64;
946
+ endbr64
947
947
xorl %edx , %edx
948
948
jmp .sic2lc_do_len6
949
949
.sic2lc_len8:
950
- endbr64;
950
+ endbr64
951
951
movzbl 7 (%rdi ), %eax
952
952
movzbl __tfw_lct(%rax ), %edx
953
953
xorb 7 (%rsi ), %dl
Original file line number Diff line number Diff line change @@ -199,25 +199,25 @@ SYM_FUNC_START(mpi_sub_x86_64)
199
199
jmpq *%rbx
200
200
201
201
.sub_tail3:
202
- endbr64;
202
+ endbr64
203
203
movq (%rdx , %rax , 8 ), %r9
204
204
sbbq (%rsi , %rax , 8 ), %r9
205
205
movq %r9 , (%rdi , %rax , 8 )
206
206
incq %rax
207
207
.sub_tail2:
208
- endbr64;
208
+ endbr64
209
209
movq (%rdx , %rax , 8 ), %r10
210
210
sbbq (%rsi , %rax , 8 ), %r10
211
211
movq %r10 , (%rdi , %rax , 8 )
212
212
incq %rax
213
213
.sub_tail1:
214
- endbr64;
214
+ endbr64
215
215
movq (%rdx , %rax , 8 ), %r11
216
216
sbbq (%rsi , %rax , 8 ), %r11
217
217
movq %r11 , (%rdi , %rax , 8 )
218
218
incq %rax
219
219
.sub_tail0:
220
- endbr64;
220
+ endbr64
221
221
popq %rbx
222
222
223
223
/*
You can’t perform that action at this time.
0 commit comments