Commit ee155011 authored by Diego Biurrun's avatar Diego Biurrun

whitespace cosmetics, prettyprinting

Originally committed as revision 19751 to svn://svn.ffmpeg.org/ffmpeg/trunk
parent f4258b13
...@@ -56,7 +56,7 @@ ...@@ -56,7 +56,7 @@
#endif #endif
#ifndef INT16_MIN #ifndef INT16_MIN
#define INT16_MIN (-0x7fff-1) #define INT16_MIN (-0x7fff - 1)
#endif #endif
#ifndef INT16_MAX #ifndef INT16_MAX
...@@ -64,7 +64,7 @@ ...@@ -64,7 +64,7 @@
#endif #endif
#ifndef INT32_MIN #ifndef INT32_MIN
#define INT32_MIN (-0x7fffffff-1) #define INT32_MIN (-0x7fffffff - 1)
#endif #endif
#ifndef INT32_MAX #ifndef INT32_MAX
...@@ -76,7 +76,7 @@ ...@@ -76,7 +76,7 @@
#endif #endif
#ifndef INT64_MIN #ifndef INT64_MIN
#define INT64_MIN (-0x7fffffffffffffffLL-1) #define INT64_MIN (-0x7fffffffffffffffLL - 1)
#endif #endif
#ifndef INT64_MAX #ifndef INT64_MAX
...@@ -96,7 +96,7 @@ ...@@ -96,7 +96,7 @@
#endif #endif
#ifndef offsetof #ifndef offsetof
# define offsetof(T,F) ((unsigned int)((char *)&((T *)0)->F)) # define offsetof(T, F) ((unsigned int)((char *)&((T *)0)->F))
#endif #endif
/* Use to export labels from asm. */ /* Use to export labels from asm. */
...@@ -129,11 +129,11 @@ extern const uint32_t ff_inverse[256]; ...@@ -129,11 +129,11 @@ extern const uint32_t ff_inverse[256];
#if ARCH_X86 #if ARCH_X86
# define FASTDIV(a,b) \ # define FASTDIV(a,b) \
({\ ({\
int ret,dmy;\ int ret, dmy;\
__asm__ volatile(\ __asm__ volatile(\
"mull %3"\ "mull %3"\
:"=d"(ret),"=a"(dmy)\ :"=d"(ret), "=a"(dmy)\
:"1"(a),"g"(ff_inverse[b])\ :"1"(a), "g"(ff_inverse[b])\
);\ );\
ret;\ ret;\
}) })
...@@ -152,14 +152,14 @@ static inline av_const int FASTDIV(int a, int b) ...@@ -152,14 +152,14 @@ static inline av_const int FASTDIV(int a, int b)
static inline av_const int FASTDIV(int a, int b) static inline av_const int FASTDIV(int a, int b)
{ {
int r, t; int r, t;
__asm__ volatile ("umull %1, %0, %2, %3" __asm__ volatile("umull %1, %0, %2, %3"
: "=&r"(r), "=&r"(t) : "r"(a), "r"(ff_inverse[b])); : "=&r"(r), "=&r"(t) : "r"(a), "r"(ff_inverse[b]));
return r; return r;
} }
#elif CONFIG_FASTDIV #elif CONFIG_FASTDIV
# define FASTDIV(a,b) ((uint32_t)((((uint64_t)a)*ff_inverse[b])>>32)) # define FASTDIV(a,b) ((uint32_t)((((uint64_t)a) * ff_inverse[b]) >> 32))
#else #else
# define FASTDIV(a,b) ((a)/(b)) # define FASTDIV(a,b) ((a) / (b))
#endif #endif
extern const uint8_t ff_sqrt_tab[256]; extern const uint8_t ff_sqrt_tab[256];
...@@ -168,20 +168,20 @@ static inline av_const unsigned int ff_sqrt(unsigned int a) ...@@ -168,20 +168,20 @@ static inline av_const unsigned int ff_sqrt(unsigned int a)
{ {
unsigned int b; unsigned int b;
if(a<255) return (ff_sqrt_tab[a+1]-1)>>4; if (a < 255) return (ff_sqrt_tab[a + 1] - 1) >> 4;
else if(a<(1<<12)) b= ff_sqrt_tab[a>>4 ]>>2; else if (a < (1 << 12)) b = ff_sqrt_tab[a >> 4] >> 2;
#if !CONFIG_SMALL #if !CONFIG_SMALL
else if(a<(1<<14)) b= ff_sqrt_tab[a>>6 ]>>1; else if (a < (1 << 14)) b = ff_sqrt_tab[a >> 6] >> 1;
else if(a<(1<<16)) b= ff_sqrt_tab[a>>8 ] ; else if (a < (1 << 16)) b = ff_sqrt_tab[a >> 8] ;
#endif #endif
else{ else {
int s= av_log2_16bit(a>>16)>>1; int s = av_log2_16bit(a >> 16) >> 1;
unsigned int c= a>>(s+2); unsigned int c = a >> (s + 2);
b= ff_sqrt_tab[c>>(s+8)]; b = ff_sqrt_tab[c >> (s + 8)];
b= FASTDIV(c,b) + (b<<s); b = FASTDIV(c,b) + (b << s);
} }
return b - (a<b*b); return b - (a < b * b);
} }
#if ARCH_X86 #if ARCH_X86
...@@ -194,14 +194,14 @@ static inline av_const unsigned int ff_sqrt(unsigned int a) ...@@ -194,14 +194,14 @@ static inline av_const unsigned int ff_sqrt(unsigned int a)
); );
#else #else
#define MASK_ABS(mask, level)\ #define MASK_ABS(mask, level)\
mask= level>>31;\ mask = level >> 31;\
level= (level^mask)-mask; level = (level ^ mask) - mask;
#endif #endif
#if HAVE_CMOV #if HAVE_CMOV
#define COPY3_IF_LT(x,y,a,b,c,d)\ #define COPY3_IF_LT(x, y, a, b, c, d)\
__asm__ volatile (\ __asm__ volatile(\
"cmpl %0, %3 \n\t"\ "cmpl %0, %3 \n\t"\
"cmovl %3, %0 \n\t"\ "cmovl %3, %0 \n\t"\
"cmovl %4, %1 \n\t"\ "cmovl %4, %1 \n\t"\
"cmovl %5, %2 \n\t"\ "cmovl %5, %2 \n\t"\
...@@ -209,11 +209,11 @@ __asm__ volatile (\ ...@@ -209,11 +209,11 @@ __asm__ volatile (\
: "r" (y), "r" (b), "r" (d)\ : "r" (y), "r" (b), "r" (d)\
); );
#else #else
#define COPY3_IF_LT(x,y,a,b,c,d)\ #define COPY3_IF_LT(x, y, a, b, c, d)\
if((y)<(x)){\ if ((y) < (x)) {\
(x)=(y);\ (x) = (y);\
(a)=(b);\ (a) = (b);\
(c)=(d);\ (c) = (d);\
} }
#endif #endif
...@@ -251,8 +251,8 @@ if((y)<(x)){\ ...@@ -251,8 +251,8 @@ if((y)<(x)){\
#define CHECKED_ALLOC(p, size)\ #define CHECKED_ALLOC(p, size)\
{\ {\
p= av_malloc(size);\ p = av_malloc(size);\
if(p==NULL && (size)!=0){\ if (p == NULL && (size) != 0) {\
av_log(NULL, AV_LOG_ERROR, "Cannot allocate memory.");\ av_log(NULL, AV_LOG_ERROR, "Cannot allocate memory.");\
goto fail;\ goto fail;\
}\ }\
...@@ -260,8 +260,8 @@ if((y)<(x)){\ ...@@ -260,8 +260,8 @@ if((y)<(x)){\
#define CHECKED_ALLOCZ(p, size)\ #define CHECKED_ALLOCZ(p, size)\
{\ {\
p= av_mallocz(size);\ p = av_mallocz(size);\
if(p==NULL && (size)!=0){\ if (p == NULL && (size) != 0) {\
av_log(NULL, AV_LOG_ERROR, "Cannot allocate memory.");\ av_log(NULL, AV_LOG_ERROR, "Cannot allocate memory.");\
goto fail;\ goto fail;\
}\ }\
......
Markdown is supported
0% or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment