Commit 21c41e76 authored by Michael Niedermayer's avatar Michael Niedermayer

Merge commit 'cec5ce49'

* commit 'cec5ce49':
  cdxl: remove an unused variable
  c93: use the AVFrame API properly.
  bethsoftvid: use the AVFrame API properly.
  avs: use the AVFrame API properly.

Conflicts:
	libavcodec/bethsoftvideo.c
Merged-by: 's avatarMichael Niedermayer <michaelni@gmx.at>
parents cc4a6435 cec5ce49
...@@ -25,7 +25,7 @@ ...@@ -25,7 +25,7 @@
typedef struct { typedef struct {
AVFrame picture; AVFrame *frame;
} AvsContext; } AvsContext;
typedef enum { typedef enum {
...@@ -52,7 +52,7 @@ avs_decode_frame(AVCodecContext * avctx, ...@@ -52,7 +52,7 @@ avs_decode_frame(AVCodecContext * avctx,
int buf_size = avpkt->size; int buf_size = avpkt->size;
AvsContext *const avs = avctx->priv_data; AvsContext *const avs = avctx->priv_data;
AVFrame *picture = data; AVFrame *picture = data;
AVFrame *const p = &avs->picture; AVFrame *const p = avs->frame;
const uint8_t *table, *vect; const uint8_t *table, *vect;
uint8_t *out; uint8_t *out;
int i, j, x, y, stride, ret, vect_w = 3, vect_h = 3; int i, j, x, y, stride, ret, vect_w = 3, vect_h = 3;
...@@ -65,8 +65,8 @@ avs_decode_frame(AVCodecContext * avctx, ...@@ -65,8 +65,8 @@ avs_decode_frame(AVCodecContext * avctx,
p->pict_type = AV_PICTURE_TYPE_P; p->pict_type = AV_PICTURE_TYPE_P;
p->key_frame = 0; p->key_frame = 0;
out = avs->picture.data[0]; out = p->data[0];
stride = avs->picture.linesize[0]; stride = p->linesize[0];
if (buf_end - buf < 4) if (buf_end - buf < 4)
return AVERROR_INVALIDDATA; return AVERROR_INVALIDDATA;
...@@ -76,7 +76,7 @@ avs_decode_frame(AVCodecContext * avctx, ...@@ -76,7 +76,7 @@ avs_decode_frame(AVCodecContext * avctx,
if (type == AVS_PALETTE) { if (type == AVS_PALETTE) {
int first, last; int first, last;
uint32_t *pal = (uint32_t *) avs->picture.data[1]; uint32_t *pal = (uint32_t *) p->data[1];
first = AV_RL16(buf); first = AV_RL16(buf);
last = first + AV_RL16(buf + 2); last = first + AV_RL16(buf + 2);
...@@ -149,7 +149,7 @@ avs_decode_frame(AVCodecContext * avctx, ...@@ -149,7 +149,7 @@ avs_decode_frame(AVCodecContext * avctx,
align_get_bits(&change_map); align_get_bits(&change_map);
} }
if ((ret = av_frame_ref(picture, &avs->picture)) < 0) if ((ret = av_frame_ref(picture, p)) < 0)
return ret; return ret;
*got_frame = 1; *got_frame = 1;
...@@ -159,16 +159,21 @@ avs_decode_frame(AVCodecContext * avctx, ...@@ -159,16 +159,21 @@ avs_decode_frame(AVCodecContext * avctx,
static av_cold int avs_decode_init(AVCodecContext * avctx) static av_cold int avs_decode_init(AVCodecContext * avctx)
{ {
AvsContext *s = avctx->priv_data; AvsContext *s = avctx->priv_data;
s->frame = av_frame_alloc();
if (!s->frame)
return AVERROR(ENOMEM);
avctx->pix_fmt = AV_PIX_FMT_PAL8; avctx->pix_fmt = AV_PIX_FMT_PAL8;
ff_set_dimensions(avctx, 318, 198); ff_set_dimensions(avctx, 318, 198);
avcodec_get_frame_defaults(&s->picture);
return 0; return 0;
} }
static av_cold int avs_decode_end(AVCodecContext *avctx) static av_cold int avs_decode_end(AVCodecContext *avctx)
{ {
AvsContext *s = avctx->priv_data; AvsContext *s = avctx->priv_data;
av_frame_unref(&s->picture); av_frame_free(&s->frame);
return 0; return 0;
} }
......
...@@ -34,21 +34,25 @@ ...@@ -34,21 +34,25 @@
#include "internal.h" #include "internal.h"
typedef struct BethsoftvidContext { typedef struct BethsoftvidContext {
AVFrame frame; AVFrame *frame;
GetByteContext g; GetByteContext g;
} BethsoftvidContext; } BethsoftvidContext;
static av_cold int bethsoftvid_decode_init(AVCodecContext *avctx) static av_cold int bethsoftvid_decode_init(AVCodecContext *avctx)
{ {
BethsoftvidContext *vid = avctx->priv_data; BethsoftvidContext *vid = avctx->priv_data;
avcodec_get_frame_defaults(&vid->frame);
avctx->pix_fmt = AV_PIX_FMT_PAL8; avctx->pix_fmt = AV_PIX_FMT_PAL8;
vid->frame = av_frame_alloc();
if (!vid->frame)
return AVERROR(ENOMEM);
return 0; return 0;
} }
static int set_palette(BethsoftvidContext *ctx) static int set_palette(BethsoftvidContext *ctx)
{ {
uint32_t *palette = (uint32_t *)ctx->frame.data[1]; uint32_t *palette = (uint32_t *)ctx->frame->data[1];
int a; int a;
if (bytestream2_get_bytes_left(&ctx->g) < 256*3) if (bytestream2_get_bytes_left(&ctx->g) < 256*3)
...@@ -58,7 +62,7 @@ static int set_palette(BethsoftvidContext *ctx) ...@@ -58,7 +62,7 @@ static int set_palette(BethsoftvidContext *ctx)
palette[a] = 0xFFU << 24 | bytestream2_get_be24u(&ctx->g) * 4; palette[a] = 0xFFU << 24 | bytestream2_get_be24u(&ctx->g) * 4;
palette[a] |= palette[a] >> 6 & 0x30303; palette[a] |= palette[a] >> 6 & 0x30303;
} }
ctx->frame.palette_has_changed = 1; ctx->frame->palette_has_changed = 1;
return 0; return 0;
} }
...@@ -75,9 +79,9 @@ static int bethsoftvid_decode_frame(AVCodecContext *avctx, ...@@ -75,9 +79,9 @@ static int bethsoftvid_decode_frame(AVCodecContext *avctx,
int code, ret; int code, ret;
int yoffset; int yoffset;
if ((ret = ff_reget_buffer(avctx, &vid->frame)) < 0) if ((ret = ff_reget_buffer(avctx, vid->frame)) < 0)
return ret; return ret;
wrap_to_next_line = vid->frame.linesize[0] - avctx->width; wrap_to_next_line = vid->frame->linesize[0] - avctx->width;
if (avpkt->side_data_elems > 0 && if (avpkt->side_data_elems > 0 &&
avpkt->side_data[0].type == AV_PKT_DATA_PALETTE) { avpkt->side_data[0].type == AV_PKT_DATA_PALETTE) {
...@@ -88,8 +92,8 @@ static int bethsoftvid_decode_frame(AVCodecContext *avctx, ...@@ -88,8 +92,8 @@ static int bethsoftvid_decode_frame(AVCodecContext *avctx,
} }
bytestream2_init(&vid->g, avpkt->data, avpkt->size); bytestream2_init(&vid->g, avpkt->data, avpkt->size);
dst = vid->frame.data[0]; dst = vid->frame->data[0];
frame_end = vid->frame.data[0] + vid->frame.linesize[0] * avctx->height; frame_end = vid->frame->data[0] + vid->frame->linesize[0] * avctx->height;
switch(block_type = bytestream2_get_byte(&vid->g)){ switch(block_type = bytestream2_get_byte(&vid->g)){
case PALETTE_BLOCK: { case PALETTE_BLOCK: {
...@@ -104,7 +108,7 @@ static int bethsoftvid_decode_frame(AVCodecContext *avctx, ...@@ -104,7 +108,7 @@ static int bethsoftvid_decode_frame(AVCodecContext *avctx,
yoffset = bytestream2_get_le16(&vid->g); yoffset = bytestream2_get_le16(&vid->g);
if(yoffset >= avctx->height) if(yoffset >= avctx->height)
return AVERROR_INVALIDDATA; return AVERROR_INVALIDDATA;
dst += vid->frame.linesize[0] * yoffset; dst += vid->frame->linesize[0] * yoffset;
} }
// main code // main code
...@@ -134,7 +138,7 @@ static int bethsoftvid_decode_frame(AVCodecContext *avctx, ...@@ -134,7 +138,7 @@ static int bethsoftvid_decode_frame(AVCodecContext *avctx,
} }
end: end:
if ((ret = av_frame_ref(data, &vid->frame)) < 0) if ((ret = av_frame_ref(data, vid->frame)) < 0)
return ret; return ret;
*got_frame = 1; *got_frame = 1;
...@@ -145,7 +149,7 @@ static int bethsoftvid_decode_frame(AVCodecContext *avctx, ...@@ -145,7 +149,7 @@ static int bethsoftvid_decode_frame(AVCodecContext *avctx,
static av_cold int bethsoftvid_decode_end(AVCodecContext *avctx) static av_cold int bethsoftvid_decode_end(AVCodecContext *avctx)
{ {
BethsoftvidContext * vid = avctx->priv_data; BethsoftvidContext * vid = avctx->priv_data;
av_frame_unref(&vid->frame); av_frame_free(&vid->frame);
return 0; return 0;
} }
......
...@@ -24,7 +24,7 @@ ...@@ -24,7 +24,7 @@
#include "internal.h" #include "internal.h"
typedef struct { typedef struct {
AVFrame pictures[2]; AVFrame *pictures[2];
int currentpic; int currentpic;
} C93DecoderContext; } C93DecoderContext;
...@@ -46,21 +46,27 @@ typedef enum { ...@@ -46,21 +46,27 @@ typedef enum {
#define C93_HAS_PALETTE 0x01 #define C93_HAS_PALETTE 0x01
#define C93_FIRST_FRAME 0x02 #define C93_FIRST_FRAME 0x02
static av_cold int decode_init(AVCodecContext *avctx) static av_cold int decode_end(AVCodecContext *avctx)
{ {
C93DecoderContext *s = avctx->priv_data; C93DecoderContext * const c93 = avctx->priv_data;
avctx->pix_fmt = AV_PIX_FMT_PAL8;
avcodec_get_frame_defaults(&s->pictures[0]); av_frame_free(&c93->pictures[0]);
avcodec_get_frame_defaults(&s->pictures[1]); av_frame_free(&c93->pictures[1]);
return 0; return 0;
} }
static av_cold int decode_end(AVCodecContext *avctx) static av_cold int decode_init(AVCodecContext *avctx)
{ {
C93DecoderContext * const c93 = avctx->priv_data; C93DecoderContext *s = avctx->priv_data;
avctx->pix_fmt = AV_PIX_FMT_PAL8;
av_frame_unref(&c93->pictures[0]); s->pictures[0] = av_frame_alloc();
av_frame_unref(&c93->pictures[1]); s->pictures[1] = av_frame_alloc();
if (!s->pictures[0] || !s->pictures[1]) {
decode_end(avctx);
return AVERROR(ENOMEM);
}
return 0; return 0;
} }
...@@ -121,8 +127,8 @@ static int decode_frame(AVCodecContext *avctx, void *data, ...@@ -121,8 +127,8 @@ static int decode_frame(AVCodecContext *avctx, void *data,
const uint8_t *buf = avpkt->data; const uint8_t *buf = avpkt->data;
int buf_size = avpkt->size; int buf_size = avpkt->size;
C93DecoderContext * const c93 = avctx->priv_data; C93DecoderContext * const c93 = avctx->priv_data;
AVFrame * const newpic = &c93->pictures[c93->currentpic]; AVFrame * const newpic = c93->pictures[c93->currentpic];
AVFrame * const oldpic = &c93->pictures[c93->currentpic^1]; AVFrame * const oldpic = c93->pictures[c93->currentpic^1];
GetByteContext gb; GetByteContext gb;
uint8_t *out; uint8_t *out;
int stride, ret, i, x, y, b, bt = 0; int stride, ret, i, x, y, b, bt = 0;
......
Markdown is supported
0% or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment