zend_ulong ret;
__asm__ (
- ".L0%=:\n\t"
+ ".LL0%=:\n\t"
"movl (%2,%3), %0\n\t"
"xorl (%2), %0\n\t"
- "jne .L1%=\n\t"
+ "jne .LL1%=\n\t"
"addl $0x4, %2\n\t"
"subl $0x4, %1\n\t"
- "ja .L0%=\n\t"
+ "ja .LL0%=\n\t"
"movl $0x1, %0\n\t"
- "jmp .L3%=\n\t"
- ".L1%=:\n\t"
+ "jmp .LL3%=\n\t"
+ ".LL1%=:\n\t"
"cmpl $0x4,%1\n\t"
- "jb .L2%=\n\t"
+ "jb .LL2%=\n\t"
"xorl %0, %0\n\t"
- "jmp .L3%=\n\t"
- ".L2%=:\n\t"
+ "jmp .LL3%=\n\t"
+ ".LL2%=:\n\t"
"negl %1\n\t"
"lea 0x1c(,%1,8), %1\n\t"
"shll %b1, %0\n\t"
"sete %b0\n\t"
"movzbl %b0, %0\n\t"
- ".L3%=:\n"
+ ".LL3%=:\n"
: "=&a"(ret),
"+c"(len),
"+r"(ptr)
__asm__ (
"test %1, %1\n\t"
- "jnz .L1%=\n\t"
+ "jnz .LL1%=\n\t"
"movl $0x1, %0\n\t"
- "jmp .L2%=\n\t"
- ".L1%=:\n\t"
+ "jmp .LL2%=\n\t"
+ ".LL1%=:\n\t"
"cld\n\t"
"rep\n\t"
"cmpsb\n\t"
"sete %b0\n\t"
"movzbl %b0, %0\n\t"
- ".L2%=:\n"
+ ".LL2%=:\n"
: "=a"(ret),
"+c"(len),
"+D"(ptr1),
zend_ulong ret;
__asm__ (
- ".L0%=:\n\t"
+ ".LL0%=:\n\t"
"movq (%2,%3), %0\n\t"
"xorq (%2), %0\n\t"
- "jne .L1%=\n\t"
+ "jne .LL1%=\n\t"
"addq $0x8, %2\n\t"
"subq $0x8, %1\n\t"
- "ja .L0%=\n\t"
+ "ja .LL0%=\n\t"
"movq $0x1, %0\n\t"
- "jmp .L3%=\n\t"
- ".L1%=:\n\t"
+ "jmp .LL3%=\n\t"
+ ".LL1%=:\n\t"
"cmpq $0x8,%1\n\t"
- "jb .L2%=\n\t"
+ "jb .LL2%=\n\t"
"xorq %0, %0\n\t"
- "jmp .L3%=\n\t"
- ".L2%=:\n\t"
+ "jmp .LL3%=\n\t"
+ ".LL2%=:\n\t"
"negq %1\n\t"
"lea 0x3c(,%1,8), %1\n\t"
"shlq %b1, %0\n\t"
"sete %b0\n\t"
"movzbq %b0, %0\n\t"
- ".L3%=:\n"
+ ".LL3%=:\n"
: "=&a"(ret),
"+c"(len),
"+r"(ptr)
__asm__ (
"test %1, %1\n\t"
- "jnz .L1%=\n\t"
+ "jnz .LL1%=\n\t"
"movq $0x1, %0\n\t"
- "jmp .L2%=\n\t"
- ".L1%=:\n\t"
+ "jmp .LL2%=\n\t"
+ ".LL1%=:\n\t"
"cld\n\t"
"rep\n\t"
"cmpsb\n\t"
"sete %b0\n\t"
"movzbq %b0, %0\n\t"
- ".L2%=:\n"
+ ".LL2%=:\n"
: "=a"(ret),
"+c"(len),
"+D"(ptr1),
__asm__ volatile (
".align 16\n\t"
- ".L0%=:\n\t"
+ ".LL0%=:\n\t"
"prefetchnta 0x40(%1)\n\t"
"movdqa (%1), %%xmm0\n\t"
"movdqa 0x10(%1), %%xmm1\n\t"
"movdqa %%xmm3, 0x30(%1,%2)\n\t"
"addl $0x40, %1\n\t"
"subl $0x40, %0\n\t"
- "ja .L0%="
+ "ja .LL0%="
: "+r"(size),
"+r"(src)
: "r"(delta)
__asm__ volatile (
".align 16\n\t"
- ".L0%=:\n\t"
+ ".LL0%=:\n\t"
"prefetchnta 0x40(%1)\n\t"
"movdqa (%1), %%xmm0\n\t"
"movdqa 0x10(%1), %%xmm1\n\t"
"movdqa %%xmm3, 0x30(%1,%2)\n\t"
"addq $0x40, %1\n\t"
"subq $0x40, %0\n\t"
- "ja .L0%="
+ "ja .LL0%="
: "+r"(size),
"+r"(src)
: "r"(delta)