const uint8_t *s = src;
const uint8_t *end;
#ifdef HAVE_MMX
- uint8_t *mm_end;
+ const uint8_t *mm_end;
#endif
end = s + src_size;
#ifdef HAVE_MMX
__asm __volatile(PREFETCH" %0"::"m"(*s):"memory");
+ mm_end = end - 23;
__asm __volatile("movq %0, %%mm7"::"m"(mask32):"memory");
- mm_end = (uint8_t*)((((unsigned long)end)/24)*24);
while(s < mm_end)
{
__asm __volatile(
const uint8_t *s = src;
const uint8_t *end;
#ifdef HAVE_MMX
- uint8_t *mm_end;
+ const uint8_t *mm_end;
#endif
end = s + src_size;
#ifdef HAVE_MMX
__asm __volatile(PREFETCH" %0"::"m"(*s):"memory");
- mm_end = (uint8_t*)((((unsigned long)end)/32)*32);
+ mm_end = end - 31;
while(s < mm_end)
{
__asm __volatile(
register const uint8_t* s=src;
register uint8_t* d=dst;
register const uint8_t *end;
- uint8_t *mm_end;
+ const uint8_t *mm_end;
end = s + src_size;
#ifdef HAVE_MMX
__asm __volatile(PREFETCH" %0"::"m"(*s));
__asm __volatile("movq %0, %%mm4"::"m"(mask15s));
- mm_end = (uint8_t*)((((unsigned long)end)/16)*16);
+ mm_end = end - 15;
while(s<mm_end)
{
__asm __volatile(
__asm __volatile(SFENCE:::"memory");
__asm __volatile(EMMS:::"memory");
#endif
- mm_end = (uint8_t*)((((unsigned long)end)/4)*4);
+ mm_end = end - 3;
while(s < mm_end)
{
register unsigned x= *((uint32_t *)s);
"movq %0, %%mm7\n\t"
"movq %1, %%mm6\n\t"
::"m"(red_16mask),"m"(green_16mask));
- mm_end = (uint8_t*)((((unsigned long)end)/16)*16);
+ mm_end = end - 15;
while(s < mm_end)
{
__asm __volatile(
"movq %0, %%mm7\n\t"
"movq %1, %%mm6\n\t"
::"m"(red_15mask),"m"(green_15mask));
- mm_end = (uint8_t*)((((unsigned long)end)/16)*16);
+ mm_end = end - 15;
while(s < mm_end)
{
__asm __volatile(
"movq %0, %%mm7\n\t"
"movq %1, %%mm6\n\t"
::"m"(red_16mask),"m"(green_16mask));
- mm_end = (uint8_t*)((((unsigned long)end)/16)*16);
+ mm_end = end - 11;
while(s < mm_end)
{
__asm __volatile(
"movq %0, %%mm7\n\t"
"movq %1, %%mm6\n\t"
::"m"(red_15mask),"m"(green_15mask));
- mm_end = (uint8_t*)((((unsigned long)end)/16)*16);
+ mm_end = end - 11;
while(s < mm_end)
{
__asm __volatile(
end = s + src_size/2;
#ifdef HAVE_MMX
__asm __volatile(PREFETCH" %0"::"m"(*s):"memory");
- mm_end = (uint16_t*)((((unsigned long)end)/8)*8);
+ mm_end = end - 7;
while(s < mm_end)
{
__asm __volatile(
end = s + src_size/2;
#ifdef HAVE_MMX
__asm __volatile(PREFETCH" %0"::"m"(*s):"memory");
- mm_end = (uint16_t*)((((unsigned long)end)/8)*8);
+ mm_end = end - 7;
while(s < mm_end)
{
__asm __volatile(
#ifdef HAVE_MMX
__asm __volatile(PREFETCH" %0"::"m"(*s):"memory");
__asm __volatile("pxor %%mm7,%%mm7\n\t":::"memory");
- mm_end = (uint16_t*)((((unsigned long)end)/4)*4);
+ mm_end = end - 3;
while(s < mm_end)
{
__asm __volatile(
#ifdef HAVE_MMX
__asm __volatile(PREFETCH" %0"::"m"(*s):"memory");
__asm __volatile("pxor %%mm7,%%mm7\n\t":::"memory");
- mm_end = (uint16_t*)((((unsigned long)end)/4)*4);
+ mm_end = end - 3;
while(s < mm_end)
{
__asm __volatile(
"addl $8, %%eax \n\t"
"cmpl %2, %%eax \n\t"
" jb 1b \n\t"
- :: "r" (src), "r"(dst), "r" (src_size)
+ :: "r" (src), "r"(dst), "r" (src_size-7)
: "%eax"
);