genericopenlibs/liboil/src/i386/recon8x8_i386.c
changeset 18 47c74d1534e1
--- /dev/null	Thu Jan 01 00:00:00 1970 +0000
+++ b/genericopenlibs/liboil/src/i386/recon8x8_i386.c	Fri Apr 16 16:46:38 2010 +0300
@@ -0,0 +1,184 @@
+/*
+ * LIBOIL - Library of Optimized Inner Loops
+ * Copyright (c) 2003,2004 David A. Schleef <ds@schleef.org>
+ * All rights reserved.
+ *
+ * Redistribution and use in source and binary forms, with or without
+ * modification, are permitted provided that the following conditions
+ * are met:
+ * 1. Redistributions of source code must retain the above copyright
+ *    notice, this list of conditions and the following disclaimer.
+ * 2. Redistributions in binary form must reproduce the above copyright
+ *    notice, this list of conditions and the following disclaimer in the
+ *    documentation and/or other materials provided with the distribution.
+ * 
+ * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
+ * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
+ * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
+ * ARE DISCLAIMED.  IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT,
+ * INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
+ * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR
+ * SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
+ * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT,
+ * STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING
+ * IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
+ * POSSIBILITY OF SUCH DAMAGE.
+ */
+//Portions Copyright (c)  2008-2009 Nokia Corporation and/or its subsidiary(-ies). All rights reserved. 
+
+#ifdef HAVE_CONFIG_H
+#include "config.h"
+#endif
+
+#include <liboil/liboilfunction.h>
+#include "liboil/simdpack/simdpack.h"
+
+OIL_DECLARE_CLASS (recon8x8_intra);
+OIL_DECLARE_CLASS (recon8x8_inter);
+OIL_DECLARE_CLASS (recon8x8_inter2);
+
+const uint8_t c0x80[8] = {
+  0x80, 0x80, 0x80, 0x80, 0x80, 0x80, 0x80, 0x80
+};
+
+static void
+recon8x8_intra_i386_mmx (uint8_t *dest, int ds, int16_t *change)
+{
+#if !defined(__WINSCW__) && !defined(__WINS__)      
+  __asm__ __volatile__ (
+    "  movq        (%3), %%mm0       \n\t" /* Set mm0 to 0x8080808080808080 */
+
+    "  lea         128(%1), %%edi      \n\t" /* Endpoint in input buffer */
+    "1:                                \n\t" 
+    "  movq         (%1), %%mm2        \n\t" /* First four input values */
+
+    "  packsswb    8(%1), %%mm2        \n\t" /* pack with next(high) four values */
+    "  por         %%mm0, %%mm0        \n\t" 
+    "  pxor        %%mm0, %%mm2        \n\t" /* Convert result to unsigned (same as add 128) */
+    "  lea         16(%1), %1          \n\t" /* Step source buffer */
+    "  cmp         %%edi, %1           \n\t" /* are we done */
+
+    "  movq        %%mm2, (%0)         \n\t" /* store results */
+
+    "  lea         (%0, %2), %0        \n\t" /* Step output buffer */
+    "  jc          1b                  \n\t" /* Loop back if we are not done */
+    "  emms                            \n\t"
+      : "+r" (dest)
+      : "r" (change),
+        "r" (ds),
+        "r" (c0x80)
+      : "memory", "edi"
+  );
+#endif  
+}
+
+OIL_DEFINE_IMPL_FULL (recon8x8_intra_i386_mmx, recon8x8_intra, OIL_IMPL_FLAG_MMX);
+
+#if 0
+static void
+recon8x8_inter_i386_mmx (uint8_t *dest, int ds, uint8_t *src, int ss, int16_t *change)
+{
+  /* FIXME doesn't handle ss */
+  __asm__ __volatile__ (
+    "  pxor        %%mm0, %%mm0        \n\t"
+    "  lea         128(%1), %%edi      \n\t"
+
+    "1:                                \n\t"
+    "  movq        (%2), %%mm2         \n\t" /* (+3 misaligned) 8 reference pixels */
+
+    "  movq        (%1), %%mm4         \n\t" /* first 4 changes */
+    "  movq        %%mm2, %%mm3        \n\t"
+    "  movq        8(%1), %%mm5        \n\t" /* last 4 changes */
+    "  punpcklbw   %%mm0, %%mm2        \n\t" /* turn first 4 refs into positive 16-bit #s */
+    "  paddsw      %%mm4, %%mm2        \n\t" /* add in first 4 changes */
+    "  punpckhbw   %%mm0, %%mm3        \n\t" /* turn last 4 refs into positive 16-bit #s */
+    "  paddsw      %%mm5, %%mm3        \n\t" /* add in last 4 changes */
+    "  add         %3, %2              \n\t" /* next row of reference pixels */
+    "  packuswb    %%mm3, %%mm2        \n\t" /* pack result to unsigned 8-bit values */
+    "  lea         16(%1), %1          \n\t" /* next row of changes */
+    "  cmp         %%edi, %1           \n\t" /* are we done? */
+
+    "  movq        %%mm2, (%0)         \n\t" /* store result */
+
+    "  lea         (%0, %3), %0        \n\t" /* next row of output */
+    "  jc          1b                  \n\t"
+    "  emms                            \n\t"
+      : "+r" (dest)
+      : "r" (change),
+        "r" (src),
+        "r" (ds)
+      : "memory", "edi"
+  );
+}
+
+OIL_DEFINE_IMPL_FULL (recon8x8_inter_i386_mmx, recon8x8_inter, OIL_IMPL_FLAG_MMX);
+
+static void
+recon8x8_inter2_i386_mmx (uint8_t *dest, int ds, uint8_t *s1, int ss1, uint8_t *s2, int ss2, int16_t *change)
+{
+  /* FIXME doesn't handle ss1, ss2 */
+  __asm__ __volatile__ (
+    "  pxor        %%mm0, %%mm0        \n\t"
+    "  lea         128(%1), %%edi      \n\t"
+
+    "1:                                \n\t"
+    "  movq        (%2), %%mm2         \n\t" /* (+3 misaligned) 8 reference pixels */
+    "  movq        (%3), %%mm4         \n\t" /* (+3 misaligned) 8 reference pixels */
+
+    "  movq        %%mm2, %%mm3        \n\t"
+    "  punpcklbw   %%mm0, %%mm2        \n\t" /* mm2 = start ref1 as positive 16-bit #s */
+    "  movq        %%mm4, %%mm5        \n\t"
+    "  movq        (%1), %%mm6         \n\t" /* first 4 changes */
+    "  punpckhbw   %%mm0, %%mm3        \n\t" /* mm3 = end ref1 as positive 16-bit #s */
+    "  movq        8(%1), %%mm7        \n\t" /* last 4 changes */
+    "  punpcklbw   %%mm0, %%mm4        \n\t" /* mm4 = start ref2 as positive 16-bit #s */
+    "  punpckhbw   %%mm0, %%mm5        \n\t" /* mm5 = end ref2 as positive 16-bit #s */
+    "  paddw       %%mm4, %%mm2        \n\t" /* mm2 = start (ref1 + ref2) */
+    "  paddw       %%mm5, %%mm3        \n\t" /* mm3 = end (ref1 + ref2) */
+    "  psrlw       $1, %%mm2           \n\t" /* mm2 = start (ref1 + ref2)/2 */
+    "  psrlw       $1, %%mm3           \n\t" /* mm3 = end (ref1 + ref2)/2 */
+    "  paddw       %%mm6, %%mm2        \n\t" /* add changes to start */
+    "  paddw       %%mm7, %%mm3        \n\t" /* add changes to end */
+    "  lea         16(%1), %1          \n\t" /* next row of changes */
+    "  packuswb    %%mm3, %%mm2        \n\t" /* pack start|end to unsigned 8-bit */
+    "  add         %4, %2              \n\t" /* next row of reference pixels */
+    "  add         %4, %3              \n\t" /* next row of reference pixels */
+    "  movq        %%mm2, (%0)         \n\t" /* store result */
+    "  add         %4, %0              \n\t" /* next row of output */
+    "  cmp         %%edi, %1           \n\t" /* are we done? */
+    "  jc          1b                  \n\t"
+    "  emms                            \n\t"
+      : "+r" (dest)
+      : "r" (change),
+        "r" (s1),
+        "r" (s2),
+        "m" (ds)
+      : "memory", "edi"
+  );
+}
+
+OIL_DEFINE_IMPL_FULL (recon8x8_inter2_i386_mmx, recon8x8_inter2, OIL_IMPL_FLAG_MMX);
+#endif
+
+
+#ifdef	__SYMBIAN32__
+ 
+OilFunctionImpl* __oil_function_impl_recon8x8_intra_i386_mmx, recon8x8_intra() {
+		return &_oil_function_impl_recon8x8_intra_i386_mmx, recon8x8_intra;
+}
+#endif
+
+#ifdef	__SYMBIAN32__
+ 
+OilFunctionImpl* __oil_function_impl_recon8x8_inter_i386_mmx, recon8x8_inter() {
+		return &_oil_function_impl_recon8x8_inter_i386_mmx, recon8x8_inter;
+}
+#endif
+
+#ifdef	__SYMBIAN32__
+ 
+OilFunctionImpl* __oil_function_impl_recon8x8_inter2_i386_mmx, recon8x8_inter2() {
+		return &_oil_function_impl_recon8x8_inter2_i386_mmx, recon8x8_inter2;
+}
+#endif
+