--- /dev/null Thu Jan 01 00:00:00 1970 +0000
+++ b/genericopenlibs/liboil/src/i386/composite_i386.c Tue Aug 31 16:54:36 2010 +0300
@@ -0,0 +1,1147 @@
+/*
+ * LIBOIL - Library of Optimized Inner Loops
+ * Copyright (c) 2005 David A. Schleef <ds@schleef.org>
+ * All rights reserved.
+ *
+ * Redistribution and use in source and binary forms, with or without
+ * modification, are permitted provided that the following conditions
+ * are met:
+ * 1. Redistributions of source code must retain the above copyright
+ * notice, this list of conditions and the following disclaimer.
+ * 2. Redistributions in binary form must reproduce the above copyright
+ * notice, this list of conditions and the following disclaimer in the
+ * documentation and/or other materials provided with the distribution.
+ *
+ * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
+ * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
+ * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT,
+ * INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR
+ * SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
+ * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT,
+ * STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING
+ * IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
+ * POSSIBILITY OF SUCH DAMAGE.
+ */
+//Portions Copyright (c) 2008-2009 Nokia Corporation and/or its subsidiary(-ies). All rights reserved.
+
+#ifdef HAVE_CONFIG_H
+#include "config.h"
+#endif
+
+#include <liboil/liboil.h>
+#include <liboil/liboilfunction.h>
+
+OIL_DECLARE_CLASS (composite_in_argb);
+OIL_DECLARE_CLASS (composite_in_argb_const_src);
+OIL_DECLARE_CLASS (composite_in_argb_const_mask);
+OIL_DECLARE_CLASS (composite_over_argb);
+OIL_DECLARE_CLASS (composite_over_argb_const_src);
+OIL_DECLARE_CLASS (composite_add_argb);
+OIL_DECLARE_CLASS (composite_add_argb_const_src);
+OIL_DECLARE_CLASS (composite_in_over_argb);
+OIL_DECLARE_CLASS (composite_in_over_argb_const_src);
+OIL_DECLARE_CLASS (composite_in_over_argb_const_mask);
+
+#if 0
+static void
+composite_in_argb_mmx (uint32_t *dest, uint32_t *src, uint8_t *mask, int n)
+{
+ int i;
+
+ for(i=0;i<n;i++){
+ dest[i] = ARGB(
+ COMPOSITE_IN(ARGB_A(src[i]), mask[i]),
+ COMPOSITE_IN(ARGB_R(src[i]), mask[i]),
+ COMPOSITE_IN(ARGB_G(src[i]), mask[i]),
+ COMPOSITE_IN(ARGB_B(src[i]), mask[i]));
+ }
+}
+OIL_DEFINE_IMPL_FULL (composite_in_argb_mmx, composite_in_argb);
+#endif
+
+/*
+ * This macro loads the constants:
+ * mm7 = { 0, 0, 0, 0 }
+ * mm6 = { 128, 128, 128, 128 }
+ * mm5 = { 255, 255, 255, 255 }
+ */
+#define MMX_LOAD_CONSTANTS \
+ " pxor %%mm7, %%mm7\n" \
+ " movl $0x80808080, %%eax\n" \
+ " movd %%eax, %%mm6\n" \
+ " punpcklbw %%mm7, %%mm6\n" \
+ " movl $0xffffffff, %%eax\n" \
+ " movd %%eax, %%mm5\n" \
+ " punpcklbw %%mm7, %%mm5\n"
+
+/*
+ * a = muldiv255(a, b)
+ * a, b are unpacked
+ * destroys both registers
+ * requires mm6 set up as above
+ */
+#define MMX_MULDIV255(a,b) \
+ " pmullw %%" #b ", %%" #a "\n" \
+ " paddw %%mm6, %%" #a "\n" \
+ " movq %%" #a ", %%" #b "\n" \
+ " psrlw $8, %%" #b "\n" \
+ " paddw %%" #b ", %%" #a "\n" \
+ " psrlw $8, %%" #a "\n"
+
+static void
+composite_in_argb_mmx (uint32_t *dest, uint32_t *src, const uint8_t *mask, int n)
+{
+#if !defined(__WINSCW__) && !defined(__WINS__)
+ __asm__ __volatile__ (
+ MMX_LOAD_CONSTANTS
+ "1:\n"
+ " movd (%2), %%mm0\n"
+ " punpcklbw %%mm7, %%mm0\n"
+ " pshufw $0x00, %%mm0, %%mm1\n"
+
+ " movd (%1), %%mm2\n"
+ " punpcklbw %%mm7, %%mm2\n"
+
+ MMX_MULDIV255(mm2, mm1)
+
+ " packuswb %%mm2, %%mm2\n"
+ " movd %%mm2, (%0)\n"
+ " addl $4, %0\n"
+ " addl $4, %1\n"
+ " addl $1, %2\n"
+ " decl %3\n"
+ " jnz 1b\n"
+ " emms\n"
+ :"+r" (dest), "+r" (src), "+r" (mask), "+r" (n)
+ :
+ :"eax");
+#endif
+}
+OIL_DEFINE_IMPL_FULL (composite_in_argb_mmx, composite_in_argb, OIL_IMPL_FLAG_MMX | OIL_IMPL_FLAG_MMXEXT);
+
+/*
+ * This is a different style than the others. Should be moved elsewhere.
+ */
+static void
+composite_in_argb_mmx2 (uint32_t *dest, uint32_t *src, const uint8_t *mask, int n)
+{
+#if !defined(__WINSCW__) && !defined(__WINS__)
+ __asm__ __volatile__ (
+ MMX_LOAD_CONSTANTS
+ "1:\n"
+ " movl (%2), %%eax\n"
+ /* if alpha == 0, write a 0 */
+ " testl $0x000000ff, %%eax\n"
+ " je 2f\n"
+ /* if alpha == 0xff, write src value */
+ " cmp $0xff, %%al\n"
+ " je 3f\n"
+
+ " movd %%eax, %%mm0\n"
+ " punpcklbw %%mm7, %%mm0\n"
+ " pshufw $0x00, %%mm0, %%mm1\n"
+
+ " movd (%1), %%mm2\n"
+ " punpcklbw %%mm7, %%mm2\n"
+
+ MMX_MULDIV255(mm2, mm1)
+
+ " packuswb %%mm2, %%mm2\n"
+ " movd %%mm2, (%0)\n"
+ " jmp 4f\n"
+ "2:\n"
+ " movl $0, (%0)\n"
+ " jmp 4f\n"
+ "3:\n"
+ " movl (%1), %%eax\n"
+ " movl %%eax, (%0)\n"
+ "4:\n"
+ " addl $4, %0\n"
+ " addl $4, %1\n"
+ " addl $1, %2\n"
+ " decl %3\n"
+ " jnz 1b\n"
+ " emms\n"
+ :"+r" (dest), "+r" (src), "+r" (mask), "+r" (n)
+ :
+ :"eax");
+#endif
+}
+OIL_DEFINE_IMPL_FULL (composite_in_argb_mmx2, composite_in_argb, OIL_IMPL_FLAG_MMX | OIL_IMPL_FLAG_MMXEXT);
+
+static void
+composite_in_argb_const_src_mmx (uint32_t *dest, uint32_t *src, const uint8_t *mask, int n)
+{
+#if !defined(__WINSCW__) && !defined(__WINS__)
+ __asm__ __volatile__ (
+ MMX_LOAD_CONSTANTS
+ " movd (%1), %%mm3\n"
+ " punpcklbw %%mm7, %%mm3\n"
+ "1:\n"
+ " movd (%2), %%mm0\n"
+ " punpcklbw %%mm7, %%mm0\n"
+ " pshufw $0x00, %%mm0, %%mm1\n"
+
+ " movq %%mm3, %%mm2\n"
+
+ MMX_MULDIV255(mm2, mm1)
+
+ " packuswb %%mm2, %%mm2\n"
+ " movd %%mm2, (%0)\n"
+ " addl $4, %0\n"
+ " addl $1, %2\n"
+ " decl %3\n"
+ " jnz 1b\n"
+ " emms\n"
+ :"+r" (dest), "+r" (src), "+r" (mask), "+r" (n)
+ :
+ :"eax");
+#endif
+}
+OIL_DEFINE_IMPL_FULL (composite_in_argb_const_src_mmx, composite_in_argb_const_src, OIL_IMPL_FLAG_MMX | OIL_IMPL_FLAG_MMXEXT);
+
+static void
+composite_in_argb_const_mask_mmx (uint32_t *dest, uint32_t *src, const uint8_t *mask, int n)
+{
+#if !defined(__WINSCW__) && !defined(__WINS__)
+ __asm__ __volatile__ (
+ MMX_LOAD_CONSTANTS
+ " movd (%2), %%mm0\n"
+ " punpcklbw %%mm7, %%mm0\n"
+ " pshufw $0x00, %%mm0, %%mm3\n"
+ "1:\n"
+ " movq %%mm3, %%mm1\n"
+ " movd (%1), %%mm2\n"
+ " punpcklbw %%mm7, %%mm2\n"
+
+ MMX_MULDIV255(mm2, mm1)
+
+ " packuswb %%mm2, %%mm2\n"
+ " movd %%mm2, (%0)\n"
+ " addl $4, %0\n"
+ " addl $4, %1\n"
+ " decl %3\n"
+ " jnz 1b\n"
+ " emms\n"
+ :"+r" (dest), "+r" (src), "+r" (mask), "+r" (n)
+ :
+ :"eax");
+#endif
+}
+OIL_DEFINE_IMPL_FULL (composite_in_argb_const_mask_mmx, composite_in_argb_const_mask, OIL_IMPL_FLAG_MMX | OIL_IMPL_FLAG_MMXEXT);
+
+static void
+composite_over_argb_mmx (uint32_t *dest, uint32_t *src, int n)
+{
+#if !defined(__WINSCW__) && !defined(__WINS__)
+ __asm__ __volatile__ (
+ MMX_LOAD_CONSTANTS
+ "1:\n"
+ " movl (%1), %%eax\n"
+ " testl $0xff000000, %%eax\n"
+ " jz 2f\n"
+
+ " movd %%eax, %%mm0\n"
+ " punpcklbw %%mm7, %%mm0\n"
+ " pshufw $0xff, %%mm0, %%mm1\n"
+ " pxor %%mm5, %%mm1\n"
+
+ " movd (%0), %%mm2\n"
+ " punpcklbw %%mm7, %%mm2\n"
+
+ MMX_MULDIV255(mm2, mm1)
+
+ " paddw %%mm0, %%mm2\n"
+ " packuswb %%mm2, %%mm2\n"
+
+ " movd %%mm2, (%0)\n"
+ "2:\n"
+ " addl $4, %0\n"
+ " addl $4, %1\n"
+ " decl %2\n"
+ " jnz 1b\n"
+ " emms\n"
+ :"+r" (dest), "+r" (src), "+r" (n)
+ :
+ :"eax");
+#endif
+}
+OIL_DEFINE_IMPL_FULL (composite_over_argb_mmx, composite_over_argb, OIL_IMPL_FLAG_MMX | OIL_IMPL_FLAG_MMXEXT);
+
+/* unroll 2 */
+static void
+composite_over_argb_mmx_2 (uint32_t *dest, uint32_t *src, int n)
+{
+#if !defined(__WINSCW__) && !defined(__WINS__)
+ __asm__ __volatile__ (
+ MMX_LOAD_CONSTANTS
+
+ " testl $0x1, %2\n"
+ " jz 2f\n"
+
+ " movl (%1), %%eax\n"
+ " testl $0xff000000, %%eax\n"
+ " jz 1f\n"
+
+ " movd %%eax, %%mm0\n"
+ " punpcklbw %%mm7, %%mm0\n"
+ " pshufw $0xff, %%mm0, %%mm1\n"
+ " pxor %%mm5, %%mm1\n"
+
+ " movd (%0), %%mm2\n"
+ " punpcklbw %%mm7, %%mm2\n"
+ " pmullw %%mm1, %%mm2\n"
+ " paddw %%mm6, %%mm2\n"
+ " movq %%mm2, %%mm1\n"
+ " psrlw $8, %%mm1\n"
+ " paddw %%mm1, %%mm2\n"
+ " psrlw $8, %%mm2\n"
+
+ " paddw %%mm0, %%mm2\n"
+ " packuswb %%mm2, %%mm2\n"
+
+ " movd %%mm2, (%0)\n"
+
+ "1:\n"
+ " addl $4, %0\n"
+ " addl $4, %1\n"
+
+ "2:\n"
+ " shr $1, %2\n"
+ " jz 5f\n"
+ "3:\n"
+ " movl (%1), %%eax\n"
+ " orl 4(%1), %%eax\n"
+ " testl $0xff000000, %%eax\n"
+ " jz 4f\n"
+
+ " movd (%1), %%mm0\n"
+ " movd (%0), %%mm2\n"
+
+ " punpcklbw %%mm7, %%mm0\n"
+ " movd 4(%1), %%mm3\n"
+
+ " pshufw $0xff, %%mm0, %%mm1\n"
+ " punpcklbw %%mm7, %%mm2\n"
+
+ " pxor %%mm5, %%mm1\n"
+ " movd 4(%0), %%mm4\n"
+
+ " pmullw %%mm1, %%mm2\n"
+ " punpcklbw %%mm7, %%mm3\n"
+
+ " paddw %%mm6, %%mm2\n"
+ " punpcklbw %%mm7, %%mm4\n"
+
+ " movq %%mm2, %%mm1\n"
+ " pshufw $0xff, %%mm3, %%mm7\n"
+
+ " psrlw $8, %%mm1\n"
+ " pxor %%mm5, %%mm7\n"
+
+ " paddw %%mm1, %%mm2\n"
+ " pmullw %%mm7, %%mm4\n"
+
+ " psrlw $8, %%mm2\n"
+ " paddw %%mm6, %%mm4\n"
+
+ " paddw %%mm0, %%mm2\n"
+ " movq %%mm4, %%mm7\n"
+
+ " packuswb %%mm2, %%mm2\n"
+ " psrlw $8, %%mm7\n"
+
+ " movd %%mm2, (%0)\n"
+ " paddw %%mm7, %%mm4\n"
+
+ " psrlw $8, %%mm4\n"
+ " paddw %%mm3, %%mm4\n"
+ " packuswb %%mm4, %%mm4\n"
+ " movd %%mm4, 4(%0)\n"
+
+ " pxor %%mm7, %%mm7\n"
+ "4:\n"
+ " addl $8, %0\n"
+ " addl $8, %1\n"
+ " decl %2\n"
+ " jnz 3b\n"
+ "5:\n"
+ " emms\n"
+ :"+r" (dest), "+r" (src), "+r" (n)
+ :
+ :"eax");
+#endif
+}
+OIL_DEFINE_IMPL_FULL (composite_over_argb_mmx_2, composite_over_argb, OIL_IMPL_FLAG_MMX | OIL_IMPL_FLAG_MMXEXT);
+
+/* replace pshufw with punpck */
+static void
+composite_over_argb_mmx_3 (uint32_t *dest, uint32_t *src, int n)
+{
+#if !defined(__WINSCW__) && !defined(__WINS__)
+ __asm__ __volatile__ (
+ MMX_LOAD_CONSTANTS
+ "1:\n"
+ " movl (%1), %%eax\n"
+ " testl $0xff000000, %%eax\n"
+ " jz 2f\n"
+
+ " movd %%eax, %%mm0\n"
+ " punpcklbw %%mm7, %%mm0\n"
+ " movq %%mm0, %%mm1\n"
+ " punpckhwd %%mm1, %%mm1\n"
+ " punpckhdq %%mm1, %%mm1\n"
+ " pxor %%mm5, %%mm1\n"
+
+ " movd (%0), %%mm2\n"
+ " punpcklbw %%mm7, %%mm2\n"
+ " pmullw %%mm1, %%mm2\n"
+ " paddw %%mm6, %%mm2\n"
+ " movq %%mm2, %%mm1\n"
+ " psrlw $8, %%mm1\n"
+ " paddw %%mm1, %%mm2\n"
+ " psrlw $8, %%mm2\n"
+
+ " paddw %%mm0, %%mm2\n"
+ " packuswb %%mm2, %%mm2\n"
+ " movd %%mm2, (%0)\n"
+
+ "2:\n"
+ " addl $4, %0\n"
+ " addl $4, %1\n"
+ " decl %2\n"
+ " jnz 1b\n"
+ " emms\n"
+ :"+r" (dest), "+r" (src), "+r" (n)
+ :
+ :"eax");
+#endif
+}
+OIL_DEFINE_IMPL_FULL (composite_over_argb_mmx_3, composite_over_argb, OIL_IMPL_FLAG_MMX);
+
+/* written for gromit */
+static void
+composite_over_argb_mmx_4 (uint32_t *dest, uint32_t *src, int n)
+{
+#if !defined(__WINSCW__) && !defined(__WINS__)
+ __asm__ __volatile__ (" pxor %%mm7, %%mm7\n" // mm7 = { 0, 0, 0, 0 }
+ " movl $0x80808080, %%eax\n"
+ " movd %%eax, %%mm6\n" // mm6 = { 128, 128, 128, 128 }
+ " punpcklbw %%mm7, %%mm6\n"
+ " movl $0xffffffff, %%eax\n" // mm5 = { 255, 255, 255, 255 }
+ " movd %%eax, %%mm5\n"
+ " punpcklbw %%mm7, %%mm5\n"
+ " movl $0x02020202, %%eax\n"
+ " movd %%eax, %%mm4\n"
+ " punpcklbw %%mm7, %%mm4\n"
+ " paddw %%mm5, %%mm4\n" // mm5 = { 257, 257, 257, 257 }
+ "1:\n"
+ " movl (%1), %%eax\n"
+ " testl $0xff000000, %%eax\n"
+ " jz 2f\n"
+
+ " movd %%eax, %%mm0\n"
+ " punpcklbw %%mm7, %%mm0\n"
+ " pshufw $0xff, %%mm0, %%mm1\n"
+ " pxor %%mm5, %%mm1\n"
+
+ " movd (%0), %%mm2\n"
+ " punpcklbw %%mm7, %%mm2\n"
+ " pmullw %%mm1, %%mm2\n"
+ " paddw %%mm6, %%mm2\n"
+ " pmulhuw %%mm4, %%mm2\n"
+
+ " paddw %%mm0, %%mm2\n"
+ " packuswb %%mm2, %%mm2\n"
+
+ " movd %%mm2, (%0)\n"
+ "2:\n"
+ " addl $4, %0\n"
+ " addl $4, %1\n"
+ " subl $1, %2\n"
+ " jnz 1b\n"
+ " emms\n"
+ :"+r" (dest), "+r" (src), "+r" (n)
+ :
+ :"eax");
+#endif
+}
+OIL_DEFINE_IMPL_FULL (composite_over_argb_mmx_4, composite_over_argb, OIL_IMPL_FLAG_MMX | OIL_IMPL_FLAG_MMXEXT);
+
+static void
+composite_over_argb_mmx_5 (uint32_t *dest, uint32_t *src, int n)
+{
+#if !defined(__WINSCW__) && !defined(__WINS__)
+ __asm__ __volatile__ (" pxor %%mm7, %%mm7\n" // mm7 = { 0, 0, 0, 0 }
+ " movl $0x80808080, %%eax\n"
+ " movd %%eax, %%mm6\n" // mm6 = { 128, 128, 128, 128 }
+ " punpcklbw %%mm7, %%mm6\n"
+#if 0
+ " movl $0xffffffff, %%eax\n" // mm5 = { 255, 255, 255, 255 }
+ " movd %%eax, %%mm5\n"
+ " punpcklbw %%mm7, %%mm5\n"
+#else
+ " pcmpeqw %%mm5, %%mm5\n"
+ " psrlw $8, %%mm5\n" // mm5 = { 255, 255, 255, 255 }
+#endif
+ " movl $0x02020202, %%eax\n"
+ " movd %%eax, %%mm4\n"
+ " punpcklbw %%mm7, %%mm4\n"
+ " paddw %%mm5, %%mm4\n" // mm5 = { 257, 257, 257, 257 }
+ "1:\n"
+ " movd (%1), %%mm0\n"
+ " punpcklbw %%mm7, %%mm0\n"
+ " xor %%eax, %%eax\n"
+ " pextrw $3, %%mm0, %%eax\n"
+ " test %%eax, %%eax\n"
+ " jz 2f\n"
+
+ " pshufw $0xff, %%mm0, %%mm1\n"
+ " pxor %%mm5, %%mm1\n"
+
+ " movd (%0), %%mm2\n"
+ " punpcklbw %%mm7, %%mm2\n"
+ " pmullw %%mm1, %%mm2\n"
+ " paddw %%mm6, %%mm2\n"
+ " pmulhuw %%mm4, %%mm2\n"
+
+ " paddw %%mm0, %%mm2\n"
+ " packuswb %%mm2, %%mm2\n"
+
+ " movd %%mm2, (%0)\n"
+ "2:\n"
+ " addl $4, %0\n"
+ " addl $4, %1\n"
+ " subl $1, %2\n"
+ " jnz 1b\n"
+ " emms\n"
+ :"+r" (dest), "+r" (src), "+r" (n)
+ :
+ :"eax");
+#endif
+}
+OIL_DEFINE_IMPL_FULL (composite_over_argb_mmx_5, composite_over_argb, OIL_IMPL_FLAG_MMX | OIL_IMPL_FLAG_MMXEXT);
+
+static void
+composite_over_argb_sse2 (uint32_t *dest, uint32_t *src, int n)
+{
+#if !defined(__WINSCW__) && !defined(__WINS__)
+ __asm__ __volatile__ (" pxor %%xmm7, %%xmm7\n" // mm7 = { 0, 0, 0, 0 }
+ " movl $0x80808080, %%eax\n"
+ " movd %%eax, %%xmm6\n" // mm6 = { 128, 128, 128, 128 }
+ " punpcklbw %%xmm7, %%xmm6\n"
+ " movl $0xffffffff, %%eax\n" // mm5 = { 255, 255, 255, 255 }
+ " movd %%eax, %%xmm5\n"
+ " punpcklbw %%xmm7, %%xmm5\n"
+ " movl $0x02020202, %%eax\n"
+ " movd %%eax, %%xmm4\n"
+ " punpcklbw %%xmm7, %%xmm4\n"
+ " paddw %%xmm5, %%xmm4\n" // mm4 = { 255, 255, 255, 255 }
+ "1:\n"
+ " movl (%1), %%eax\n"
+ " testl $0xff000000, %%eax\n"
+ " jz 2f\n"
+
+ " movd (%1), %%xmm1\n"
+ " punpcklbw %%xmm7, %%xmm1\n"
+ " pshuflw $0xff, %%xmm1, %%xmm0\n"
+ " pxor %%xmm5, %%xmm0\n"
+
+ " movd (%0), %%xmm3\n"
+ " punpcklbw %%xmm7, %%xmm3\n"
+ " pmullw %%xmm0, %%xmm3\n"
+ " paddw %%xmm6, %%xmm3\n"
+ " pmulhuw %%xmm4, %%xmm3\n"
+
+ " paddw %%xmm1, %%xmm3\n"
+ " packuswb %%xmm3, %%xmm3\n"
+ " movd %%xmm3, (%0)\n"
+ "2:\n"
+ " addl $4, %0\n"
+ " addl $4, %1\n"
+ " decl %2\n"
+ " jnz 1b\n"
+ :"+r" (dest), "+r" (src), "+r" (n)
+ :
+ :"eax");
+#endif
+}
+OIL_DEFINE_IMPL_FULL (composite_over_argb_sse2, composite_over_argb, OIL_IMPL_FLAG_SSE2);
+
+/* written for shaun */
+static void
+composite_over_argb_sse2_2 (uint32_t *dest, uint32_t *src, int n)
+{
+#if !defined(__WINSCW__) && !defined(__WINS__)
+ __asm__ __volatile__ (" pxor %%xmm7, %%xmm7\n" // mm7 = { 0, 0, 0, 0 }
+ " movl $0x80808080, %%eax\n"
+ " movd %%eax, %%xmm6\n" // mm6 = { 128, 128, 128, 128 }
+ " punpcklbw %%xmm7, %%xmm6\n"
+ " punpcklwd %%xmm6, %%xmm6\n"
+ " movl $0xffffffff, %%eax\n" // mm5 = { 255, 255, 255, 255 }
+ " movd %%eax, %%xmm5\n"
+ " punpcklbw %%xmm7, %%xmm5\n"
+ " punpcklwd %%xmm5, %%xmm5\n"
+ " movl $0x02020202, %%eax\n"
+ " movd %%eax, %%xmm4\n"
+ " punpcklbw %%xmm7, %%xmm4\n"
+ " paddw %%xmm5, %%xmm4\n" // mm4 = { 257, 257, 257, 257 }
+ " punpcklwd %%xmm4, %%xmm4\n"
+ :
+ :
+ :"eax");
+
+ if (n&1) {
+ __asm__ __volatile__ (
+ " movl (%1), %%eax\n"
+ " testl $0xff000000, %%eax\n"
+ " jz 1f\n"
+
+ " movd (%1), %%xmm1\n"
+ " punpcklbw %%xmm7, %%xmm1\n"
+ " pshuflw $0xff, %%xmm1, %%xmm0\n"
+ " pxor %%xmm5, %%xmm0\n"
+
+ " movd (%0), %%xmm3\n"
+ " punpcklbw %%xmm7, %%xmm3\n"
+ " pmullw %%xmm0, %%xmm3\n"
+ " paddw %%xmm6, %%xmm3\n"
+ " pmulhuw %%xmm4, %%xmm3\n"
+
+ " paddw %%xmm1, %%xmm3\n"
+ " packuswb %%xmm3, %%xmm3\n"
+ " movd %%xmm3, (%0)\n"
+
+ "1:\n"
+ " addl $4, %0\n"
+ " addl $4, %1\n"
+ :"+r" (dest), "+r" (src)
+ :
+ :"eax");
+ }
+ n>>=1;
+
+ if (n>0){
+ __asm__ __volatile__ ("\n"
+ "3:\n"
+#if 0
+ " movl (%1), %%eax\n"
+ " orl 4(%1), %%eax\n"
+ " testl $0xff000000, %%eax\n"
+ " jz 4f\n"
+#endif
+
+ " movq (%1), %%xmm1\n"
+ " punpcklbw %%xmm7, %%xmm1\n"
+ " pshuflw $0xff, %%xmm1, %%xmm0\n"
+ " pshufhw $0xff, %%xmm0, %%xmm0\n"
+ " pxor %%xmm5, %%xmm0\n"
+
+ " movq (%0), %%xmm3\n"
+ " punpcklbw %%xmm7, %%xmm3\n"
+ " pmullw %%xmm0, %%xmm3\n"
+ " paddw %%xmm6, %%xmm3\n"
+ " pmulhuw %%xmm4, %%xmm3\n"
+ " paddw %%xmm1, %%xmm3\n"
+ " packuswb %%xmm3, %%xmm3\n"
+ " movq %%xmm3, (%0)\n"
+ "4:\n"
+ " addl $8, %0\n"
+ " addl $8, %1\n"
+ " subl $1, %2\n"
+ " jnz 3b\n"
+ :"+r" (dest), "+r" (src), "+r" (n)
+ :
+ :"eax");
+ }
+#endif
+}
+OIL_DEFINE_IMPL_FULL (composite_over_argb_sse2_2, composite_over_argb, OIL_IMPL_FLAG_SSE2);
+
+/* written for shaun */
+static void
+composite_over_argb_sse2_3 (uint32_t *dest, uint32_t *src, int n)
+{
+ int begin;
+ int middle;
+ int end;
+#if !defined(__WINSCW__) && !defined(__WINS__)
+ __asm__ __volatile__ (" pxor %%xmm7, %%xmm7\n" // mm7 = { 0, 0, 0, 0 }
+ " movl $0x80808080, %%eax\n"
+ " movd %%eax, %%xmm6\n" // mm6 = { 128, 128, 128, 128 }
+ " punpcklbw %%xmm7, %%xmm6\n"
+ " punpcklwd %%xmm6, %%xmm6\n"
+ " movl $0xffffffff, %%eax\n" // mm5 = { 255, 255, 255, 255 }
+ " movd %%eax, %%xmm5\n"
+ " punpcklbw %%xmm7, %%xmm5\n"
+ " punpcklwd %%xmm5, %%xmm5\n"
+ " movl $0x02020202, %%eax\n"
+ " movd %%eax, %%xmm4\n"
+ " punpcklbw %%xmm7, %%xmm4\n"
+ " paddw %%xmm5, %%xmm4\n" // mm4 = { 257, 257, 257, 257 }
+ " punpcklwd %%xmm4, %%xmm4\n"
+ :
+ :
+ :"eax");
+
+ begin = 0x3 & (4 - (((unsigned long)dest & 0xf) >> 2));
+ if (begin>n) {
+ begin = n;
+ middle = 0;
+ end = 0;
+ } else {
+ middle = (n-begin)>>2;
+ end = n - begin - middle*4;
+ }
+
+ if (begin>0) {
+ __asm__ __volatile__ ("\n"
+ "1:\n"
+ " movl (%1), %%eax\n"
+ " testl $0xff000000, %%eax\n"
+ " jz 2f\n"
+
+ " movd (%1), %%xmm1\n"
+ " punpcklbw %%xmm7, %%xmm1\n"
+ " pshuflw $0xff, %%xmm1, %%xmm0\n"
+ " pxor %%xmm5, %%xmm0\n"
+
+ " movd (%0), %%xmm3\n"
+ " punpcklbw %%xmm7, %%xmm3\n"
+ " pmullw %%xmm0, %%xmm3\n"
+ " paddw %%xmm6, %%xmm3\n"
+ " pmulhuw %%xmm4, %%xmm3\n"
+
+ " paddw %%xmm1, %%xmm3\n"
+ " packuswb %%xmm3, %%xmm3\n"
+ " movd %%xmm3, (%0)\n"
+
+ "2:\n"
+ " addl $4, %0\n"
+ " addl $4, %1\n"
+ " subl $1, %2\n"
+ " jnz 1b\n"
+ :"+r" (dest), "+r" (src), "+r" (begin)
+ :
+ :"eax");
+ }
+
+ if (middle>0){
+ __asm__ __volatile__ ("\n"
+ "1:\n"
+ " movq (%1), %%xmm1\n"
+ " movq 8(%1), %%xmm0\n"
+ " movl (%1), %%eax\n"
+ " orl 4(%1), %%eax\n"
+ " orl 8(%1), %%eax\n"
+ " orl 12(%1), %%eax\n"
+ " test $0xff000000, %%eax\n"
+ " jz 2f\n"
+ " punpcklbw %%xmm7, %%xmm1\n"
+ " punpcklbw %%xmm7, %%xmm0\n"
+ " pshuflw $0xff, %%xmm1, %%xmm1\n"
+ " pshuflw $0xff, %%xmm0, %%xmm0\n"
+ " pshufhw $0xff, %%xmm1, %%xmm1\n"
+ " pshufhw $0xff, %%xmm0, %%xmm0\n"
+
+ " pxor %%xmm5, %%xmm1\n"
+ " pxor %%xmm5, %%xmm0\n"
+
+ " movq (%0), %%xmm3\n"
+ " movq 8(%0), %%xmm2\n"
+ " punpcklbw %%xmm7, %%xmm3\n"
+ " punpcklbw %%xmm7, %%xmm2\n"
+
+ " pmullw %%xmm1, %%xmm3\n"
+ " paddw %%xmm6, %%xmm3\n"
+ " pmulhuw %%xmm4, %%xmm3\n"
+ " pmullw %%xmm0, %%xmm2\n"
+ " paddw %%xmm6, %%xmm2\n"
+ " pmulhuw %%xmm4, %%xmm2\n"
+ " packuswb %%xmm2, %%xmm3\n"
+
+ " movdqu (%1), %%xmm1\n"
+ " paddb %%xmm1, %%xmm3\n"
+ " movdqa %%xmm3, (%0)\n"
+ "2:\n"
+ " addl $16, %0\n"
+ " addl $16, %1\n"
+ " subl $1, %2\n"
+ " jnz 1b\n"
+ :"+r" (dest), "+r" (src), "+r" (middle)
+ :
+ :"eax");
+ }
+ if (end>0) {
+ __asm__ __volatile__ ("\n"
+ "1:\n"
+ " movl (%1), %%eax\n"
+ " testl $0xff000000, %%eax\n"
+ " jz 2f\n"
+
+ " movd (%1), %%xmm1\n"
+ " punpcklbw %%xmm7, %%xmm1\n"
+ " pshuflw $0xff, %%xmm1, %%xmm0\n"
+ " pxor %%xmm5, %%xmm0\n"
+
+ " movd (%0), %%xmm3\n"
+ " punpcklbw %%xmm7, %%xmm3\n"
+ " pmullw %%xmm0, %%xmm3\n"
+ " paddw %%xmm6, %%xmm3\n"
+ " pmulhuw %%xmm4, %%xmm3\n"
+
+ " paddw %%xmm1, %%xmm3\n"
+ " packuswb %%xmm3, %%xmm3\n"
+ " movd %%xmm3, (%0)\n"
+
+ "2:\n"
+ " addl $4, %0\n"
+ " addl $4, %1\n"
+ " subl $1, %2\n"
+ " jnz 1b\n"
+ :"+r" (dest), "+r" (src), "+r" (end)
+ :
+ :"eax");
+ }
+#endif
+}
+OIL_DEFINE_IMPL_FULL (composite_over_argb_sse2_3, composite_over_argb, OIL_IMPL_FLAG_SSE2);
+
+
+static void
+composite_over_argb_const_src_mmx (uint32_t *dest, uint32_t *src, int n)
+{
+#if !defined(__WINSCW__) && !defined(__WINS__)
+ __asm__ __volatile__ (
+ MMX_LOAD_CONSTANTS
+ " movl (%1), %%eax\n"
+ " movd %%eax, %%mm0\n"
+ " punpcklbw %%mm7, %%mm0\n"
+ " pshufw $0xff, %%mm0, %%mm3\n"
+ " pxor %%mm5, %%mm3\n"
+ "1:\n"
+ " movq %%mm3, %%mm1\n"
+ " movd (%0), %%mm2\n"
+ " punpcklbw %%mm7, %%mm2\n"
+
+ MMX_MULDIV255(mm2, mm1)
+
+ " paddw %%mm0, %%mm2\n"
+ " packuswb %%mm2, %%mm2\n"
+
+ " movd %%mm2, (%0)\n"
+ " addl $4, %0\n"
+ " decl %2\n"
+ " jnz 1b\n"
+ " emms\n"
+ :"+r" (dest), "+r" (src), "+r" (n)
+ :
+ :"eax");
+#endif
+}
+OIL_DEFINE_IMPL_FULL (composite_over_argb_const_src_mmx, composite_over_argb_const_src, OIL_IMPL_FLAG_MMX | OIL_IMPL_FLAG_MMXEXT);
+
+static void
+composite_add_argb_mmx (uint32_t *dest, uint32_t *src, int n)
+{
+#if !defined(__WINSCW__) && !defined(__WINS__)
+ __asm__ __volatile__ (
+ "1:\n"
+ " movd (%1), %%mm0\n"
+ " movd (%0), %%mm2\n"
+ " paddusb %%mm0, %%mm2\n"
+ " movd %%mm2, (%0)\n"
+ " addl $4, %0\n"
+ " addl $4, %1\n"
+ " decl %2\n"
+ " jnz 1b\n"
+ " emms\n"
+ :"+r" (dest), "+r" (src), "+r" (n)
+ :
+ :"eax");
+#endif
+}
+OIL_DEFINE_IMPL_FULL (composite_add_argb_mmx, composite_add_argb, OIL_IMPL_FLAG_MMX);
+
+static void
+composite_add_argb_const_src_mmx (uint32_t *dest, uint32_t *src, int n)
+{
+#if !defined(__WINSCW__) && !defined(__WINS__)
+ __asm__ __volatile__ (
+ " movd (%1), %%mm0\n"
+ "1:\n"
+ " movd (%0), %%mm2\n"
+ " paddusb %%mm0, %%mm2\n"
+ " movd %%mm2, (%0)\n"
+ " addl $4, %0\n"
+ " decl %2\n"
+ " jnz 1b\n"
+ " emms\n"
+ :"+r" (dest), "+r" (src), "+r" (n)
+ :
+ :"eax");
+#endif
+}
+OIL_DEFINE_IMPL_FULL (composite_add_argb_const_src_mmx, composite_add_argb_const_src, OIL_IMPL_FLAG_MMX);
+
+static void
+composite_in_over_argb_mmx (uint32_t *dest, uint32_t *src, uint8_t *mask, int n)
+{
+#if !defined(__WINSCW__) && !defined(__WINS__)
+ __asm__ __volatile__ (
+ MMX_LOAD_CONSTANTS
+ "1:\n"
+ " movd (%2), %%mm0\n"
+ " punpcklbw %%mm7, %%mm0\n"
+ " pshufw $0x00, %%mm0, %%mm1\n"
+
+ " movd (%1), %%mm2\n"
+ " punpcklbw %%mm7, %%mm2\n"
+
+ MMX_MULDIV255(mm2, mm1)
+
+ " movd (%0), %%mm0\n"
+ " punpcklbw %%mm7, %%mm0\n"
+
+ " pshufw $0xff, %%mm2, %%mm1\n"
+ " pxor %%mm5, %%mm1\n"
+
+ MMX_MULDIV255(mm0, mm1)
+
+ " paddw %%mm0, %%mm2\n"
+ " packuswb %%mm2, %%mm2\n"
+
+ " movd %%mm2, (%0)\n"
+ " addl $4, %0\n"
+ " addl $4, %1\n"
+ " addl $1, %2\n"
+ " decl %3\n"
+ " jnz 1b\n"
+ " emms\n"
+ :"+r" (dest), "+r" (src), "+r" (mask), "+r" (n)
+ :
+ :"eax");
+#endif
+}
+OIL_DEFINE_IMPL_FULL (composite_in_over_argb_mmx, composite_in_over_argb, OIL_IMPL_FLAG_MMX | OIL_IMPL_FLAG_MMXEXT);
+
+static void
+composite_in_over_argb_const_src_mmx (uint32_t *dest, uint32_t *src, uint8_t *mask, int n)
+{
+#if !defined(__WINSCW__) && !defined(__WINS__)
+ __asm__ __volatile__ (
+ MMX_LOAD_CONSTANTS
+
+ " movd (%1), %%mm3\n"
+ " punpcklbw %%mm7, %%mm3\n"
+ "1:\n"
+ " movd (%2), %%mm0\n"
+ " punpcklbw %%mm7, %%mm0\n"
+ " pshufw $0x00, %%mm0, %%mm1\n"
+
+ " movq %%mm3, %%mm2\n"
+
+ MMX_MULDIV255(mm2, mm1)
+
+ " movd (%0), %%mm0\n"
+ " punpcklbw %%mm7, %%mm0\n"
+
+ " pshufw $0xff, %%mm2, %%mm1\n"
+ " pxor %%mm5, %%mm1\n"
+
+ MMX_MULDIV255(mm0, mm1)
+
+ " paddw %%mm0, %%mm2\n"
+ " packuswb %%mm2, %%mm2\n"
+
+ " movd %%mm2, (%0)\n"
+ " addl $4, %0\n"
+ " addl $1, %2\n"
+ " decl %3\n"
+ " jnz 1b\n"
+ " emms\n"
+ :"+r" (dest), "+r" (src), "+r" (mask), "+r" (n)
+ :
+ :"eax");
+#endif
+}
+OIL_DEFINE_IMPL_FULL (composite_in_over_argb_const_src_mmx, composite_in_over_argb_const_src, OIL_IMPL_FLAG_MMX | OIL_IMPL_FLAG_MMXEXT);
+
+static void
+composite_in_over_argb_const_mask_mmx (uint32_t *dest, uint32_t *src, uint8_t *mask, int n)
+{
+#if !defined(__WINSCW__) && !defined(__WINS__)
+ __asm__ __volatile__ (
+ MMX_LOAD_CONSTANTS
+ " movd (%2), %%mm0\n"
+ " punpcklbw %%mm7, %%mm0\n"
+ " pshufw $0x00, %%mm0, %%mm3\n"
+
+ "1:\n"
+ " movd (%1), %%mm2\n"
+ " punpcklbw %%mm7, %%mm2\n"
+ " movq %%mm3, %%mm1\n"
+
+ MMX_MULDIV255(mm2, mm1)
+
+ " movd (%0), %%mm0\n"
+ " punpcklbw %%mm7, %%mm0\n"
+
+ " pshufw $0xff, %%mm2, %%mm1\n"
+ " pxor %%mm5, %%mm1\n"
+
+ MMX_MULDIV255(mm0, mm1)
+
+ " paddw %%mm0, %%mm2\n"
+ " packuswb %%mm2, %%mm2\n"
+
+ " movd %%mm2, (%0)\n"
+ " addl $4, %0\n"
+ " addl $4, %1\n"
+ " decl %3\n"
+ " jnz 1b\n"
+ " emms\n"
+ :"+r" (dest), "+r" (src), "+r" (mask), "+r" (n)
+ :
+ :"eax");
+#endif
+}
+OIL_DEFINE_IMPL_FULL (composite_in_over_argb_const_mask_mmx, composite_in_over_argb_const_mask, OIL_IMPL_FLAG_MMX | OIL_IMPL_FLAG_MMXEXT);
+
+
+
+#ifdef __SYMBIAN32__
+
+OilFunctionImpl* __oil_function_impl_composite_in_argb_mmx() {
+ return &_oil_function_impl_composite_in_argb_mmx;
+}
+#endif
+
+#ifdef __SYMBIAN32__
+
+OilFunctionImpl* __oil_function_impl_composite_in_argb_mmx, composite_in_argb() {
+ return &_oil_function_impl_composite_in_argb_mmx, composite_in_argb;
+}
+#endif
+
+#ifdef __SYMBIAN32__
+
+OilFunctionImpl* __oil_function_impl_composite_in_argb_mmx2, composite_in_argb() {
+ return &_oil_function_impl_composite_in_argb_mmx2, composite_in_argb;
+}
+#endif
+
+#ifdef __SYMBIAN32__
+
+OilFunctionImpl* __oil_function_impl_composite_in_argb_const_src_mmx, composite_in_argb_const_src() {
+ return &_oil_function_impl_composite_in_argb_const_src_mmx, composite_in_argb_const_src;
+}
+#endif
+
+#ifdef __SYMBIAN32__
+
+OilFunctionImpl* __oil_function_impl_composite_in_argb_const_mask_mmx, composite_in_argb_const_mask() {
+ return &_oil_function_impl_composite_in_argb_const_mask_mmx, composite_in_argb_const_mask;
+}
+#endif
+
+#ifdef __SYMBIAN32__
+
+OilFunctionImpl* __oil_function_impl_composite_over_argb_mmx, composite_over_argb() {
+ return &_oil_function_impl_composite_over_argb_mmx, composite_over_argb;
+}
+#endif
+
+#ifdef __SYMBIAN32__
+
+OilFunctionImpl* __oil_function_impl_composite_over_argb_mmx_2, composite_over_argb() {
+ return &_oil_function_impl_composite_over_argb_mmx_2, composite_over_argb;
+}
+#endif
+
+#ifdef __SYMBIAN32__
+
+OilFunctionImpl* __oil_function_impl_composite_over_argb_mmx_3, composite_over_argb() {
+ return &_oil_function_impl_composite_over_argb_mmx_3, composite_over_argb;
+}
+#endif
+
+#ifdef __SYMBIAN32__
+
+OilFunctionImpl* __oil_function_impl_composite_over_argb_mmx_4, composite_over_argb() {
+ return &_oil_function_impl_composite_over_argb_mmx_4, composite_over_argb;
+}
+#endif
+
+#ifdef __SYMBIAN32__
+
+OilFunctionImpl* __oil_function_impl_composite_over_argb_mmx_5, composite_over_argb() {
+ return &_oil_function_impl_composite_over_argb_mmx_5, composite_over_argb;
+}
+#endif
+
+#ifdef __SYMBIAN32__
+
+OilFunctionImpl* __oil_function_impl_composite_over_argb_sse2, composite_over_argb() {
+ return &_oil_function_impl_composite_over_argb_sse2, composite_over_argb;
+}
+#endif
+
+#ifdef __SYMBIAN32__
+
+OilFunctionImpl* __oil_function_impl_composite_over_argb_sse2_2, composite_over_argb() {
+ return &_oil_function_impl_composite_over_argb_sse2_2, composite_over_argb;
+}
+#endif
+
+#ifdef __SYMBIAN32__
+
+OilFunctionImpl* __oil_function_impl_composite_over_argb_sse2_3, composite_over_argb() {
+ return &_oil_function_impl_composite_over_argb_sse2_3, composite_over_argb;
+}
+#endif
+
+#ifdef __SYMBIAN32__
+
+OilFunctionImpl* __oil_function_impl_composite_over_argb_const_src_mmx, composite_over_argb_const_src() {
+ return &_oil_function_impl_composite_over_argb_const_src_mmx, composite_over_argb_const_src;
+}
+#endif
+
+#ifdef __SYMBIAN32__
+
+OilFunctionImpl* __oil_function_impl_composite_add_argb_mmx, composite_add_argb() {
+ return &_oil_function_impl_composite_add_argb_mmx, composite_add_argb;
+}
+#endif
+
+#ifdef __SYMBIAN32__
+
+OilFunctionImpl* __oil_function_impl_composite_add_argb_const_src_mmx, composite_add_argb_const_src() {
+ return &_oil_function_impl_composite_add_argb_const_src_mmx, composite_add_argb_const_src;
+}
+#endif
+
+#ifdef __SYMBIAN32__
+
+OilFunctionImpl* __oil_function_impl_composite_in_over_argb_mmx, composite_in_over_argb() {
+ return &_oil_function_impl_composite_in_over_argb_mmx, composite_in_over_argb;
+}
+#endif
+
+#ifdef __SYMBIAN32__
+
+OilFunctionImpl* __oil_function_impl_composite_in_over_argb_const_src_mmx, composite_in_over_argb_const_src() {
+ return &_oil_function_impl_composite_in_over_argb_const_src_mmx, composite_in_over_argb_const_src;
+}
+#endif
+
+#ifdef __SYMBIAN32__
+
+OilFunctionImpl* __oil_function_impl_composite_in_over_argb_const_mask_mmx, composite_in_over_argb_const_mask() {
+ return &_oil_function_impl_composite_in_over_argb_const_mask_mmx, composite_in_over_argb_const_mask;
+}
+#endif
+