changeset 841:e0f7001ef072 libavutil

TOMI: 16- and 32-bit intreadwrite functions
author mru
date Fri, 19 Feb 2010 03:20:01 +0000
parents c34612e84b06
children 935e766af9a1
files intreadwrite.h tomi/intreadwrite.h
diffstat 2 files changed, 150 insertions(+), 0 deletions(-) [+]
line wrap: on
line diff
--- a/intreadwrite.h	Thu Feb 18 13:44:38 2010 +0000
+++ b/intreadwrite.h	Fri Feb 19 03:20:01 2010 +0000
@@ -60,6 +60,8 @@
 #   include "mips/intreadwrite.h"
 #elif ARCH_PPC
 #   include "ppc/intreadwrite.h"
+#elif ARCH_TOMI
+#   include "tomi/intreadwrite.h"
 #elif ARCH_X86
 #   include "x86/intreadwrite.h"
 #endif
--- /dev/null	Thu Jan 01 00:00:00 1970 +0000
+++ b/tomi/intreadwrite.h	Fri Feb 19 03:20:01 2010 +0000
@@ -0,0 +1,148 @@
+/*
+ * Copyright (c) 2010 Mans Rullgard <mans@mansr.com>
+ *
+ * This file is part of FFmpeg.
+ *
+ * FFmpeg is free software; you can redistribute it and/or
+ * modify it under the terms of the GNU Lesser General Public
+ * License as published by the Free Software Foundation; either
+ * version 2.1 of the License, or (at your option) any later version.
+ *
+ * FFmpeg is distributed in the hope that it will be useful,
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
+ * Lesser General Public License for more details.
+ *
+ * You should have received a copy of the GNU Lesser General Public
+ * License along with FFmpeg; if not, write to the Free Software
+ * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
+ */
+
+#ifndef AVUTIL_TOMI_INTREADWRITE_H
+#define AVUTIL_TOMI_INTREADWRITE_H
+
+#include <stdint.h>
+#include "config.h"
+
+#define AV_RB16 AV_RB16
+static av_always_inline uint16_t AV_RB16(const void *p)
+{
+    uint16_t v;
+    __asm__ ("loadacc,   (%1+) \n\t"
+             "rol8             \n\t"
+             "storeacc,  %0    \n\t"
+             "loadacc,   (%1+) \n\t"
+             "add,       %0    \n\t"
+             : "=r"(v), "+a"(p));
+    return v;
+}
+
+#define AV_WB16 AV_WB16
+static av_always_inline void AV_WB16(void *p, uint16_t v)
+{
+    __asm__ volatile ("loadacc,   %1    \n\t"
+                      "lsr8             \n\t"
+                      "storeacc,  (%0+) \n\t"
+                      "loadacc,   %1    \n\t"
+                      "storeacc,  (%0+) \n\t"
+                      : "+&a"(p) : "r"(v));
+}
+
+#define AV_RL16 AV_RL16
+static av_always_inline uint16_t AV_RL16(const void *p)
+{
+    uint16_t v;
+    __asm__ ("loadacc,   (%1+) \n\t"
+             "storeacc,  %0    \n\t"
+             "loadacc,   (%1+) \n\t"
+             "rol8             \n\t"
+             "add,       %0    \n\t"
+             : "=r"(v), "+a"(p));
+    return v;
+}
+
+#define AV_WL16 AV_WL16
+static av_always_inline void AV_WL16(void *p, uint16_t v)
+{
+    __asm__ volatile ("loadacc,   %1    \n\t"
+                      "storeacc,  (%0+) \n\t"
+                      "lsr8             \n\t"
+                      "storeacc,  (%0+) \n\t"
+                      : "+&a"(p) : "r"(v));
+}
+
+#define AV_RB32 AV_RB32
+static av_always_inline uint32_t AV_RB32(const void *p)
+{
+    uint32_t v;
+    __asm__ ("loadacc,   (%1+) \n\t"
+             "rol8             \n\t"
+             "rol8             \n\t"
+             "rol8             \n\t"
+             "storeacc,  %0    \n\t"
+             "loadacc,   (%1+) \n\t"
+             "rol8             \n\t"
+             "rol8             \n\t"
+             "add,       %0    \n\t"
+             "loadacc,   (%1+) \n\t"
+             "rol8             \n\t"
+             "add,       %0    \n\t"
+             "loadacc,   (%1+) \n\t"
+             "add,       %0    \n\t"
+             : "=r"(v), "+a"(p));
+    return v;
+}
+
+#define AV_WB32 AV_WB32
+static av_always_inline void AV_WB32(void *p, uint32_t v)
+{
+    __asm__ volatile ("loadacc,   #4    \n\t"
+                      "add,       %0    \n\t"
+                      "loadacc,   %1    \n\t"
+                      "storeacc,  (-%0) \n\t"
+                      "lsr8             \n\t"
+                      "storeacc,  (-%0) \n\t"
+                      "lsr8             \n\t"
+                      "storeacc,  (-%0) \n\t"
+                      "lsr8             \n\t"
+                      "storeacc,  (-%0) \n\t"
+                      : "+&a"(p) : "r"(v));
+}
+
+#define AV_RL32 AV_RL32
+static av_always_inline uint32_t AV_RL32(const void *p)
+{
+    uint32_t v;
+    __asm__ ("loadacc,   (%1+) \n\t"
+             "storeacc,  %0    \n\t"
+             "loadacc,   (%1+) \n\t"
+             "rol8             \n\t"
+             "add,       %0    \n\t"
+             "loadacc,   (%1+) \n\t"
+             "rol8             \n\t"
+             "rol8             \n\t"
+             "add,       %0    \n\t"
+             "loadacc,   (%1+) \n\t"
+             "rol8             \n\t"
+             "rol8             \n\t"
+             "rol8             \n\t"
+             "add,       %0    \n\t"
+             : "=r"(v), "+a"(p));
+    return v;
+}
+
+#define AV_WL32 AV_WL32
+static av_always_inline void AV_WL32(void *p, uint32_t v)
+{
+    __asm__ volatile ("loadacc,   %1    \n\t"
+                      "storeacc,  (%0+) \n\t"
+                      "lsr8             \n\t"
+                      "storeacc,  (%0+) \n\t"
+                      "lsr8             \n\t"
+                      "storeacc,  (%0+) \n\t"
+                      "lsr8             \n\t"
+                      "storeacc,  (%0+) \n\t"
+                      : "+&a"(p) : "r"(v));
+}
+
+#endif /* AVUTIL_TOMI_INTREADWRITE_H */