Commit a7831d50 authored by Mans Rullgard's avatar Mans Rullgard

arm: use HAVE*_INLINE/EXTERNAL macros for conditional compilation

These macros reflect the actual capabilities required here.
Signed-off-by: 's avatarMans Rullgard <mans@mansr.com>
parent 7f2b3dca
...@@ -23,7 +23,7 @@ ...@@ -23,7 +23,7 @@
#include "config.h" #include "config.h"
#if HAVE_NEON && HAVE_INLINE_ASM #if HAVE_NEON_INLINE
#define VMUL2 VMUL2 #define VMUL2 VMUL2
static inline float *VMUL2(float *dst, const float *v, unsigned idx, static inline float *VMUL2(float *dst, const float *v, unsigned idx,
...@@ -138,6 +138,6 @@ static inline float *VMUL4S(float *dst, const float *v, unsigned idx, ...@@ -138,6 +138,6 @@ static inline float *VMUL4S(float *dst, const float *v, unsigned idx,
return dst; return dst;
} }
#endif /* HAVE_NEON && HAVE_INLINE_ASM */ #endif /* HAVE_NEON_INLINE */
#endif /* AVCODEC_ARM_AAC_H */ #endif /* AVCODEC_ARM_AAC_H */
...@@ -25,7 +25,7 @@ ...@@ -25,7 +25,7 @@
#include "config.h" #include "config.h"
#include "libavutil/intmath.h" #include "libavutil/intmath.h"
#if HAVE_ARMV6 && HAVE_INLINE_ASM && AV_GCC_VERSION_AT_LEAST(4,4) #if HAVE_ARMV6_INLINE && AV_GCC_VERSION_AT_LEAST(4,4)
#define decode_blockcodes decode_blockcodes #define decode_blockcodes decode_blockcodes
static inline int decode_blockcodes(int code1, int code2, int levels, static inline int decode_blockcodes(int code1, int code2, int levels,
...@@ -79,7 +79,7 @@ static inline int decode_blockcodes(int code1, int code2, int levels, ...@@ -79,7 +79,7 @@ static inline int decode_blockcodes(int code1, int code2, int levels,
#endif #endif
#if HAVE_NEON && HAVE_INLINE_ASM && HAVE_ASM_MOD_Y #if HAVE_NEON_INLINE && HAVE_ASM_MOD_Y
#define int8x8_fmul_int32 int8x8_fmul_int32 #define int8x8_fmul_int32 int8x8_fmul_int32
static inline void int8x8_fmul_int32(float *dst, const int8_t *src, int scale) static inline void int8x8_fmul_int32(float *dst, const int8_t *src, int scale)
......
...@@ -22,7 +22,7 @@ ...@@ -22,7 +22,7 @@
#include "config.h" #include "config.h"
#include "libavutil/arm/asm.S" #include "libavutil/arm/asm.S"
#if HAVE_ARMV5TE #if HAVE_ARMV5TE_EXTERNAL
function ff_prefetch_arm, export=1 function ff_prefetch_arm, export=1
subs r2, r2, #1 subs r2, r2, #1
pld [r0] pld [r0]
......
...@@ -28,7 +28,7 @@ ...@@ -28,7 +28,7 @@
#if HAVE_INLINE_ASM #if HAVE_INLINE_ASM
#if HAVE_ARMV6 #if HAVE_ARMV6_INLINE
#define MULH MULH #define MULH MULH
static inline av_const int MULH(int a, int b) static inline av_const int MULH(int a, int b)
{ {
...@@ -50,7 +50,7 @@ static av_always_inline av_const int FASTDIV(int a, int b) ...@@ -50,7 +50,7 @@ static av_always_inline av_const int FASTDIV(int a, int b)
return r; return r;
} }
#else /* HAVE_ARMV6 */ #else /* HAVE_ARMV6_INLINE */
#define FASTDIV FASTDIV #define FASTDIV FASTDIV
static av_always_inline av_const int FASTDIV(int a, int b) static av_always_inline av_const int FASTDIV(int a, int b)
...@@ -64,7 +64,7 @@ static av_always_inline av_const int FASTDIV(int a, int b) ...@@ -64,7 +64,7 @@ static av_always_inline av_const int FASTDIV(int a, int b)
#define MLS64(d, a, b) MAC64(d, -(a), b) #define MLS64(d, a, b) MAC64(d, -(a), b)
#if HAVE_ARMV5TE #if HAVE_ARMV5TE_INLINE
/* signed 16x16 -> 32 multiply add accumulate */ /* signed 16x16 -> 32 multiply add accumulate */
# define MAC16(rt, ra, rb) \ # define MAC16(rt, ra, rb) \
......
...@@ -37,7 +37,7 @@ ...@@ -37,7 +37,7 @@
# define U(x) # define U(x)
#endif #endif
#if HAVE_ARMV6 && HAVE_INLINE_ASM #if HAVE_ARMV6_INLINE
#define vp56_rac_get_prob vp56_rac_get_prob_armv6 #define vp56_rac_get_prob vp56_rac_get_prob_armv6
static inline int vp56_rac_get_prob_armv6(VP56RangeCoder *c, int pr) static inline int vp56_rac_get_prob_armv6(VP56RangeCoder *c, int pr)
......
...@@ -21,7 +21,7 @@ ...@@ -21,7 +21,7 @@
#include "config.h" #include "config.h"
#if HAVE_ARMV6 #if HAVE_ARMV6_EXTERNAL
#define decode_block_coeffs_internal ff_decode_block_coeffs_armv6 #define decode_block_coeffs_internal ff_decode_block_coeffs_armv6
int ff_decode_block_coeffs_armv6(VP56RangeCoder *rc, DCTELEM block[16], int ff_decode_block_coeffs_armv6(VP56RangeCoder *rc, DCTELEM block[16],
uint8_t probs[8][3][NUM_DCT_TOKENS-1], uint8_t probs[8][3][NUM_DCT_TOKENS-1],
......
...@@ -118,7 +118,7 @@ function ff_vp8_luma_dc_wht_armv6, export=1 ...@@ -118,7 +118,7 @@ function ff_vp8_luma_dc_wht_armv6, export=1
usub16 r4, r4, r6 @ block[0,1][3] usub16 r4, r4, r6 @ block[0,1][3]
usub16 r5, r5, r12 @ block[2,3][3] usub16 r5, r5, r12 @ block[2,3][3]
#if HAVE_ARMV6T2 #if HAVE_ARMV6T2_EXTERNAL
sbfx r6, r8, #3, #13 sbfx r6, r8, #3, #13
sbfx r12, r7, #3, #13 sbfx r12, r7, #3, #13
sbfx r1, r9, #3, #13 sbfx r1, r9, #3, #13
...@@ -151,7 +151,7 @@ function ff_vp8_luma_dc_wht_armv6, export=1 ...@@ -151,7 +151,7 @@ function ff_vp8_luma_dc_wht_armv6, export=1
strh r4, [r0], #32 strh r4, [r0], #32
asr r10, r5, #19 @ block[3][3] asr r10, r5, #19 @ block[3][3]
#if HAVE_ARMV6T2 #if HAVE_ARMV6T2_EXTERNAL
sbfx r2, r2, #3, #13 sbfx r2, r2, #3, #13
sbfx lr, lr, #3, #13 sbfx lr, lr, #3, #13
sbfx r3, r3, #3, #13 sbfx r3, r3, #3, #13
...@@ -284,7 +284,7 @@ function ff_vp8_idct_add_armv6, export=1 ...@@ -284,7 +284,7 @@ function ff_vp8_idct_add_armv6, export=1
sxth r12, r12 sxth r12, r12
ldr r9, [r0, r2] ldr r9, [r0, r2]
sxth r1, r1 sxth r1, r1
#if HAVE_ARMV6T2 #if HAVE_ARMV6T2_EXTERNAL
sbfx r7, r7, #3, #13 sbfx r7, r7, #3, #13
sbfx r10, r10, #3, #13 sbfx r10, r10, #3, #13
#else #else
......
...@@ -89,7 +89,7 @@ ELF .size \name, . - \name ...@@ -89,7 +89,7 @@ ELF .size \name, . - \name
\name: \name:
.endm .endm
#if !HAVE_ARMV6T2 #if !HAVE_ARMV6T2_EXTERNAL
.macro movw rd, val .macro movw rd, val
mov \rd, \val & 255 mov \rd, \val & 255
orr \rd, \val & ~255 orr \rd, \val & ~255
...@@ -97,7 +97,7 @@ ELF .size \name, . - \name ...@@ -97,7 +97,7 @@ ELF .size \name, . - \name
#endif #endif
.macro mov32 rd, val .macro mov32 rd, val
#if HAVE_ARMV6T2 #if HAVE_ARMV6T2_EXTERNAL
movw \rd, #(\val) & 0xffff movw \rd, #(\val) & 0xffff
.if (\val) >> 16 .if (\val) >> 16
movt \rd, #(\val) >> 16 movt \rd, #(\val) >> 16
...@@ -146,7 +146,7 @@ T ldr \rd, [\rd] ...@@ -146,7 +146,7 @@ T ldr \rd, [\rd]
.macro movrel rd, val .macro movrel rd, val
#if CONFIG_PIC #if CONFIG_PIC
ldpic \rd, \val ldpic \rd, \val
#elif HAVE_ARMV6T2 && !defined(__APPLE__) #elif HAVE_ARMV6T2_EXTERNAL && !defined(__APPLE__)
movw \rd, #:lower16:\val movw \rd, #:lower16:\val
movt \rd, #:upper16:\val movt \rd, #:upper16:\val
#else #else
......
...@@ -35,7 +35,7 @@ static av_always_inline av_const uint32_t av_bswap32(uint32_t x) ...@@ -35,7 +35,7 @@ static av_always_inline av_const uint32_t av_bswap32(uint32_t x)
#elif HAVE_INLINE_ASM #elif HAVE_INLINE_ASM
#if HAVE_ARMV6 #if HAVE_ARMV6_INLINE
#define av_bswap16 av_bswap16 #define av_bswap16 av_bswap16
static av_always_inline av_const unsigned av_bswap16(unsigned x) static av_always_inline av_const unsigned av_bswap16(unsigned x)
{ {
...@@ -48,7 +48,7 @@ static av_always_inline av_const unsigned av_bswap16(unsigned x) ...@@ -48,7 +48,7 @@ static av_always_inline av_const unsigned av_bswap16(unsigned x)
#define av_bswap32 av_bswap32 #define av_bswap32 av_bswap32
static av_always_inline av_const uint32_t av_bswap32(uint32_t x) static av_always_inline av_const uint32_t av_bswap32(uint32_t x)
{ {
#if HAVE_ARMV6 #if HAVE_ARMV6_INLINE
__asm__("rev %0, %0" : "+r"(x)); __asm__("rev %0, %0" : "+r"(x));
#else #else
uint32_t t; uint32_t t;
...@@ -57,7 +57,7 @@ static av_always_inline av_const uint32_t av_bswap32(uint32_t x) ...@@ -57,7 +57,7 @@ static av_always_inline av_const uint32_t av_bswap32(uint32_t x)
"mov %0, %0, ror #8 \n\t" "mov %0, %0, ror #8 \n\t"
"eor %0, %0, %1, lsr #8 \n\t" "eor %0, %0, %1, lsr #8 \n\t"
: "+r"(x), "=&r"(t)); : "+r"(x), "=&r"(t));
#endif /* HAVE_ARMV6 */ #endif /* HAVE_ARMV6_INLINE */
return x; return x;
} }
#endif /* !AV_GCC_VERSION_AT_LEAST(4,5) */ #endif /* !AV_GCC_VERSION_AT_LEAST(4,5) */
......
...@@ -28,7 +28,7 @@ ...@@ -28,7 +28,7 @@
#if HAVE_INLINE_ASM #if HAVE_INLINE_ASM
#if HAVE_ARMV6 #if HAVE_ARMV6_INLINE
#define av_clip_uint8 av_clip_uint8_arm #define av_clip_uint8 av_clip_uint8_arm
static av_always_inline av_const unsigned av_clip_uint8_arm(int a) static av_always_inline av_const unsigned av_clip_uint8_arm(int a)
...@@ -86,7 +86,7 @@ static av_always_inline int av_sat_dadd32_arm(int a, int b) ...@@ -86,7 +86,7 @@ static av_always_inline int av_sat_dadd32_arm(int a, int b)
return r; return r;
} }
#endif /* HAVE_ARMV6 */ #endif /* HAVE_ARMV6_INLINE */
#if HAVE_ASM_MOD_Q #if HAVE_ASM_MOD_Q
......
Markdown is supported
0% or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment