/* * Copyright (c) 2010 Mans Rullgard * * This file is part of Libav. * * Libav is free software; you can redistribute it and/or * modify it under the terms of the GNU Lesser General Public * License as published by the Free Software Foundation; either * version 2.1 of the License, or (at your option) any later version. * * Libav is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * Lesser General Public License for more details. * * You should have received a copy of the GNU Lesser General Public * License along with Libav; if not, write to the Free Software * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA */ #ifndef AVUTIL_TOMI_INTREADWRITE_H #define AVUTIL_TOMI_INTREADWRITE_H #include #include "config.h" #include "libavutil/attributes.h" #define AV_RB16 AV_RB16 static av_always_inline uint16_t AV_RB16(const void *p) { uint16_t v; __asm__ ("loadacc, (%1+) \n\t" "rol8 \n\t" "storeacc, %0 \n\t" "loadacc, (%1+) \n\t" "add, %0 \n\t" : "=r"(v), "+a"(p)); return v; } #define AV_WB16 AV_WB16 static av_always_inline void AV_WB16(void *p, uint16_t v) { __asm__ volatile ("loadacc, %1 \n\t" "lsr8 \n\t" "storeacc, (%0+) \n\t" "loadacc, %1 \n\t" "storeacc, (%0+) \n\t" : "+&a"(p) : "r"(v)); } #define AV_RL16 AV_RL16 static av_always_inline uint16_t AV_RL16(const void *p) { uint16_t v; __asm__ ("loadacc, (%1+) \n\t" "storeacc, %0 \n\t" "loadacc, (%1+) \n\t" "rol8 \n\t" "add, %0 \n\t" : "=r"(v), "+a"(p)); return v; } #define AV_WL16 AV_WL16 static av_always_inline void AV_WL16(void *p, uint16_t v) { __asm__ volatile ("loadacc, %1 \n\t" "storeacc, (%0+) \n\t" "lsr8 \n\t" "storeacc, (%0+) \n\t" : "+&a"(p) : "r"(v)); } #define AV_RB32 AV_RB32 static av_always_inline uint32_t AV_RB32(const void *p) { uint32_t v; __asm__ ("loadacc, (%1+) \n\t" "rol8 \n\t" "rol8 \n\t" "rol8 \n\t" "storeacc, %0 \n\t" "loadacc, (%1+) \n\t" "rol8 \n\t" "rol8 \n\t" "add, %0 \n\t" "loadacc, (%1+) \n\t" "rol8 \n\t" "add, %0 \n\t" "loadacc, (%1+) \n\t" "add, %0 \n\t" : "=r"(v), "+a"(p)); return v; } #define AV_WB32 AV_WB32 static av_always_inline void AV_WB32(void *p, uint32_t v) { __asm__ volatile ("loadacc, #4 \n\t" "add, %0 \n\t" "loadacc, %1 \n\t" "storeacc, (-%0) \n\t" "lsr8 \n\t" "storeacc, (-%0) \n\t" "lsr8 \n\t" "storeacc, (-%0) \n\t" "lsr8 \n\t" "storeacc, (-%0) \n\t" : "+&a"(p) : "r"(v)); } #define AV_RL32 AV_RL32 static av_always_inline uint32_t AV_RL32(const void *p) { uint32_t v; __asm__ ("loadacc, (%1+) \n\t" "storeacc, %0 \n\t" "loadacc, (%1+) \n\t" "rol8 \n\t" "add, %0 \n\t" "loadacc, (%1+) \n\t" "rol8 \n\t" "rol8 \n\t" "add, %0 \n\t" "loadacc, (%1+) \n\t" "rol8 \n\t" "rol8 \n\t" "rol8 \n\t" "add, %0 \n\t" : "=r"(v), "+a"(p)); return v; } #define AV_WL32 AV_WL32 static av_always_inline void AV_WL32(void *p, uint32_t v) { __asm__ volatile ("loadacc, %1 \n\t" "storeacc, (%0+) \n\t" "lsr8 \n\t" "storeacc, (%0+) \n\t" "lsr8 \n\t" "storeacc, (%0+) \n\t" "lsr8 \n\t" "storeacc, (%0+) \n\t" : "+&a"(p) : "r"(v)); } #endif /* AVUTIL_TOMI_INTREADWRITE_H */