Commit 2852740e authored by Rémi Denis-Courmont's avatar Rémi Denis-Courmont Committed by Anton Khirnov

vdpau: store picture data in picture's rather than codec's context

Signed-off-by: 's avatarAnton Khirnov <anton@khirnov.net>
parent 549294fb
...@@ -680,7 +680,7 @@ SKIPHEADERS-$(CONFIG_LIBSCHROEDINGER) += libschroedinger.h ...@@ -680,7 +680,7 @@ SKIPHEADERS-$(CONFIG_LIBSCHROEDINGER) += libschroedinger.h
SKIPHEADERS-$(CONFIG_MPEG_XVMC_DECODER) += xvmc.h SKIPHEADERS-$(CONFIG_MPEG_XVMC_DECODER) += xvmc.h
SKIPHEADERS-$(CONFIG_VAAPI) += vaapi_internal.h SKIPHEADERS-$(CONFIG_VAAPI) += vaapi_internal.h
SKIPHEADERS-$(CONFIG_VDA) += vda.h SKIPHEADERS-$(CONFIG_VDA) += vda.h
SKIPHEADERS-$(CONFIG_VDPAU) += vdpau.h SKIPHEADERS-$(CONFIG_VDPAU) += vdpau.h vdpau_internal.h
EXAMPLES = api EXAMPLES = api
......
...@@ -38,13 +38,15 @@ ...@@ -38,13 +38,15 @@
* @{ * @{
*/ */
int ff_vdpau_common_start_frame(AVCodecContext *avctx, int ff_vdpau_common_start_frame(Picture *pic,
av_unused const uint8_t *buffer, av_unused const uint8_t *buffer,
av_unused uint32_t size) av_unused uint32_t size)
{ {
AVVDPAUContext *hwctx = avctx->hwaccel_context; struct vdpau_picture_context *pic_ctx = pic->hwaccel_picture_private;
hwctx->bitstream_buffers_used = 0; pic_ctx->bitstream_buffers_allocated = 0;
pic_ctx->bitstream_buffers_used = 0;
pic_ctx->bitstream_buffers = NULL;
return 0; return 0;
} }
...@@ -55,31 +57,32 @@ int ff_vdpau_mpeg_end_frame(AVCodecContext *avctx) ...@@ -55,31 +57,32 @@ int ff_vdpau_mpeg_end_frame(AVCodecContext *avctx)
{ {
AVVDPAUContext *hwctx = avctx->hwaccel_context; AVVDPAUContext *hwctx = avctx->hwaccel_context;
MpegEncContext *s = avctx->priv_data; MpegEncContext *s = avctx->priv_data;
VdpVideoSurface surf = ff_vdpau_get_surface_id(s->current_picture_ptr); Picture *pic = s->current_picture_ptr;
struct vdpau_picture_context *pic_ctx = pic->hwaccel_picture_private;
VdpVideoSurface surf = ff_vdpau_get_surface_id(pic);
hwctx->render(hwctx->decoder, surf, (void *)&hwctx->info, hwctx->render(hwctx->decoder, surf, (void *)&pic_ctx->info,
hwctx->bitstream_buffers_used, hwctx->bitstream_buffers); pic_ctx->bitstream_buffers_used, pic_ctx->bitstream_buffers);
ff_mpeg_draw_horiz_band(s, 0, s->avctx->height); ff_mpeg_draw_horiz_band(s, 0, s->avctx->height);
hwctx->bitstream_buffers_used = 0; av_freep(&pic_ctx->bitstream_buffers);
return 0; return 0;
} }
#endif #endif
int ff_vdpau_add_buffer(AVCodecContext *avctx, int ff_vdpau_add_buffer(Picture *pic, const uint8_t *buf, uint32_t size)
const uint8_t *buf, uint32_t size)
{ {
AVVDPAUContext *hwctx = avctx->hwaccel_context; struct vdpau_picture_context *pic_ctx = pic->hwaccel_picture_private;
VdpBitstreamBuffer *buffers = hwctx->bitstream_buffers; VdpBitstreamBuffer *buffers = pic_ctx->bitstream_buffers;
buffers = av_fast_realloc(buffers, &hwctx->bitstream_buffers_allocated, buffers = av_fast_realloc(buffers, &pic_ctx->bitstream_buffers_allocated,
(hwctx->bitstream_buffers_used + 1) * sizeof(*buffers)); (pic_ctx->bitstream_buffers_used + 1) * sizeof(*buffers));
if (!buffers) if (!buffers)
return AVERROR(ENOMEM); return AVERROR(ENOMEM);
hwctx->bitstream_buffers = buffers; pic_ctx->bitstream_buffers = buffers;
buffers += hwctx->bitstream_buffers_used++; buffers += pic_ctx->bitstream_buffers_used++;
buffers->struct_version = VDP_BITSTREAM_BUFFER_VERSION; buffers->struct_version = VDP_BITSTREAM_BUFFER_VERSION;
buffers->bitstream = buf; buffers->bitstream = buf;
......
...@@ -66,8 +66,8 @@ static void vdpau_h264_set_rf(VdpReferenceFrameH264 *rf, Picture *pic, ...@@ -66,8 +66,8 @@ static void vdpau_h264_set_rf(VdpReferenceFrameH264 *rf, Picture *pic,
static void vdpau_h264_set_reference_frames(AVCodecContext *avctx) static void vdpau_h264_set_reference_frames(AVCodecContext *avctx)
{ {
H264Context * const h = avctx->priv_data; H264Context * const h = avctx->priv_data;
AVVDPAUContext *hwctx = avctx->hwaccel_context; struct vdpau_picture_context *pic_ctx = h->cur_pic_ptr->hwaccel_picture_private;
VdpPictureInfoH264 *info = &hwctx->info.h264; VdpPictureInfoH264 *info = &pic_ctx->info.h264;
int list; int list;
VdpReferenceFrameH264 *rf = &info->referenceFrames[0]; VdpReferenceFrameH264 *rf = &info->referenceFrames[0];
...@@ -118,9 +118,9 @@ static int vdpau_h264_start_frame(AVCodecContext *avctx, ...@@ -118,9 +118,9 @@ static int vdpau_h264_start_frame(AVCodecContext *avctx,
const uint8_t *buffer, uint32_t size) const uint8_t *buffer, uint32_t size)
{ {
H264Context * const h = avctx->priv_data; H264Context * const h = avctx->priv_data;
AVVDPAUContext *hwctx = avctx->hwaccel_context;
VdpPictureInfoH264 *info = &hwctx->info.h264;
Picture *pic = h->cur_pic_ptr; Picture *pic = h->cur_pic_ptr;
struct vdpau_picture_context *pic_ctx = pic->hwaccel_picture_private;
VdpPictureInfoH264 *info = &pic_ctx->info.h264;
/* init VdpPictureInfoH264 */ /* init VdpPictureInfoH264 */
info->slice_count = 0; info->slice_count = 0;
...@@ -161,7 +161,7 @@ static int vdpau_h264_start_frame(AVCodecContext *avctx, ...@@ -161,7 +161,7 @@ static int vdpau_h264_start_frame(AVCodecContext *avctx,
vdpau_h264_set_reference_frames(avctx); vdpau_h264_set_reference_frames(avctx);
return ff_vdpau_common_start_frame(avctx, buffer, size); return ff_vdpau_common_start_frame(pic, buffer, size);
} }
static const uint8_t start_code_prefix[3] = { 0x00, 0x00, 0x01 }; static const uint8_t start_code_prefix[3] = { 0x00, 0x00, 0x01 };
...@@ -169,18 +169,20 @@ static const uint8_t start_code_prefix[3] = { 0x00, 0x00, 0x01 }; ...@@ -169,18 +169,20 @@ static const uint8_t start_code_prefix[3] = { 0x00, 0x00, 0x01 };
static int vdpau_h264_decode_slice(AVCodecContext *avctx, static int vdpau_h264_decode_slice(AVCodecContext *avctx,
const uint8_t *buffer, uint32_t size) const uint8_t *buffer, uint32_t size)
{ {
AVVDPAUContext *hwctx = avctx->hwaccel_context; H264Context *h = avctx->priv_data;
Picture *pic = h->cur_pic_ptr;
struct vdpau_picture_context *pic_ctx = pic->hwaccel_picture_private;
int val; int val;
val = ff_vdpau_add_buffer(avctx, start_code_prefix, 3); val = ff_vdpau_add_buffer(pic, start_code_prefix, 3);
if (val) if (val)
return val; return val;
val = ff_vdpau_add_buffer(avctx, buffer, size); val = ff_vdpau_add_buffer(pic, buffer, size);
if (val) if (val)
return val; return val;
hwctx->info.h264.slice_count++; pic_ctx->info.h264.slice_count++;
return 0; return 0;
} }
...@@ -188,13 +190,15 @@ static int vdpau_h264_end_frame(AVCodecContext *avctx) ...@@ -188,13 +190,15 @@ static int vdpau_h264_end_frame(AVCodecContext *avctx)
{ {
AVVDPAUContext *hwctx = avctx->hwaccel_context; AVVDPAUContext *hwctx = avctx->hwaccel_context;
H264Context *h = avctx->priv_data; H264Context *h = avctx->priv_data;
VdpVideoSurface surf = ff_vdpau_get_surface_id(h->cur_pic_ptr); Picture *pic = h->cur_pic_ptr;
struct vdpau_picture_context *pic_ctx = pic->hwaccel_picture_private;
VdpVideoSurface surf = ff_vdpau_get_surface_id(pic);
hwctx->render(hwctx->decoder, surf, (void *)&hwctx->info, hwctx->render(hwctx->decoder, surf, (void *)&pic_ctx->info,
hwctx->bitstream_buffers_used, hwctx->bitstream_buffers); pic_ctx->bitstream_buffers_used, pic_ctx->bitstream_buffers);
ff_h264_draw_horiz_band(h, 0, h->avctx->height); ff_h264_draw_horiz_band(h, 0, h->avctx->height);
hwctx->bitstream_buffers_used = 0; av_freep(&pic_ctx->bitstream_buffers);
return 0; return 0;
} }
...@@ -207,4 +211,5 @@ AVHWAccel ff_h264_vdpau_hwaccel = { ...@@ -207,4 +211,5 @@ AVHWAccel ff_h264_vdpau_hwaccel = {
.start_frame = vdpau_h264_start_frame, .start_frame = vdpau_h264_start_frame,
.end_frame = vdpau_h264_end_frame, .end_frame = vdpau_h264_end_frame,
.decode_slice = vdpau_h264_decode_slice, .decode_slice = vdpau_h264_decode_slice,
.priv_data_size = sizeof(struct vdpau_picture_context),
}; };
...@@ -25,6 +25,7 @@ ...@@ -25,6 +25,7 @@
#define AVCODEC_VDPAU_INTERNAL_H #define AVCODEC_VDPAU_INTERNAL_H
#include <stdint.h> #include <stdint.h>
#include <vdpau/vdpau.h>
#include "h264.h" #include "h264.h"
#include "mpegvideo.h" #include "mpegvideo.h"
...@@ -34,10 +35,31 @@ static inline uintptr_t ff_vdpau_get_surface_id(Picture *pic) ...@@ -34,10 +35,31 @@ static inline uintptr_t ff_vdpau_get_surface_id(Picture *pic)
return (uintptr_t)pic->f.data[3]; return (uintptr_t)pic->f.data[3];
} }
int ff_vdpau_common_start_frame(AVCodecContext *avctx, struct vdpau_picture_context {
/**
* VDPAU picture information.
*/
union AVVDPAUPictureInfo info;
/**
* Allocated size of the bitstream_buffers table.
*/
int bitstream_buffers_allocated;
/**
* Useful bitstream buffers in the bitstream buffers table.
*/
int bitstream_buffers_used;
/**
* Table of bitstream buffers.
*/
VdpBitstreamBuffer *bitstream_buffers;
};
int ff_vdpau_common_start_frame(Picture *pic,
const uint8_t *buffer, uint32_t size); const uint8_t *buffer, uint32_t size);
int ff_vdpau_mpeg_end_frame(AVCodecContext *avctx); int ff_vdpau_mpeg_end_frame(AVCodecContext *avctx);
int ff_vdpau_add_buffer(AVCodecContext *avctx, int ff_vdpau_add_buffer(Picture *pic, const uint8_t *buf, uint32_t buf_size);
const uint8_t *buf, uint32_t buf_size);
#endif /* AVCODEC_VDPAU_INTERNAL_H */ #endif /* AVCODEC_VDPAU_INTERNAL_H */
...@@ -31,8 +31,9 @@ static int vdpau_mpeg_start_frame(AVCodecContext *avctx, ...@@ -31,8 +31,9 @@ static int vdpau_mpeg_start_frame(AVCodecContext *avctx,
const uint8_t *buffer, uint32_t size) const uint8_t *buffer, uint32_t size)
{ {
MpegEncContext * const s = avctx->priv_data; MpegEncContext * const s = avctx->priv_data;
AVVDPAUContext *hwctx = avctx->hwaccel_context; Picture *pic = s->current_picture_ptr;
VdpPictureInfoMPEG1Or2 *info = &hwctx->info.mpeg; struct vdpau_picture_context *pic_ctx = pic->hwaccel_picture_private;
VdpPictureInfoMPEG1Or2 *info = &pic_ctx->info.mpeg;
VdpVideoSurface ref; VdpVideoSurface ref;
int i; int i;
...@@ -44,11 +45,11 @@ static int vdpau_mpeg_start_frame(AVCodecContext *avctx, ...@@ -44,11 +45,11 @@ static int vdpau_mpeg_start_frame(AVCodecContext *avctx,
case AV_PICTURE_TYPE_B: case AV_PICTURE_TYPE_B:
ref = ff_vdpau_get_surface_id(&s->next_picture); ref = ff_vdpau_get_surface_id(&s->next_picture);
assert(ref != VDP_INVALID_HANDLE); assert(ref != VDP_INVALID_HANDLE);
hwctx->info.mpeg.backward_reference = ref; info->backward_reference = ref;
/* fall through to forward prediction */ /* fall through to forward prediction */
case AV_PICTURE_TYPE_P: case AV_PICTURE_TYPE_P:
ref = ff_vdpau_get_surface_id(&s->last_picture); ref = ff_vdpau_get_surface_id(&s->last_picture);
hwctx->info.mpeg.forward_reference = ref; info->forward_reference = ref;
} }
info->slice_count = 0; info->slice_count = 0;
...@@ -74,20 +75,22 @@ static int vdpau_mpeg_start_frame(AVCodecContext *avctx, ...@@ -74,20 +75,22 @@ static int vdpau_mpeg_start_frame(AVCodecContext *avctx,
info->non_intra_quantizer_matrix[i] = s->inter_matrix[i]; info->non_intra_quantizer_matrix[i] = s->inter_matrix[i];
} }
return ff_vdpau_common_start_frame(avctx, buffer, size); return ff_vdpau_common_start_frame(pic, buffer, size);
} }
static int vdpau_mpeg_decode_slice(AVCodecContext *avctx, static int vdpau_mpeg_decode_slice(AVCodecContext *avctx,
const uint8_t *buffer, uint32_t size) const uint8_t *buffer, uint32_t size)
{ {
AVVDPAUContext *hwctx = avctx->hwaccel_context; MpegEncContext * const s = avctx->priv_data;
Picture *pic = s->current_picture_ptr;
struct vdpau_picture_context *pic_ctx = pic->hwaccel_picture_private;
int val; int val;
val = ff_vdpau_add_buffer(avctx, buffer, size); val = ff_vdpau_add_buffer(pic, buffer, size);
if (val < 0) if (val < 0)
return val; return val;
hwctx->info.mpeg.slice_count++; pic_ctx->info.mpeg.slice_count++;
return 0; return 0;
} }
...@@ -100,6 +103,7 @@ AVHWAccel ff_mpeg1_vdpau_hwaccel = { ...@@ -100,6 +103,7 @@ AVHWAccel ff_mpeg1_vdpau_hwaccel = {
.start_frame = vdpau_mpeg_start_frame, .start_frame = vdpau_mpeg_start_frame,
.end_frame = ff_vdpau_mpeg_end_frame, .end_frame = ff_vdpau_mpeg_end_frame,
.decode_slice = vdpau_mpeg_decode_slice, .decode_slice = vdpau_mpeg_decode_slice,
.priv_data_size = sizeof(struct vdpau_picture_context),
}; };
#endif #endif
...@@ -112,5 +116,6 @@ AVHWAccel ff_mpeg2_vdpau_hwaccel = { ...@@ -112,5 +116,6 @@ AVHWAccel ff_mpeg2_vdpau_hwaccel = {
.start_frame = vdpau_mpeg_start_frame, .start_frame = vdpau_mpeg_start_frame,
.end_frame = ff_vdpau_mpeg_end_frame, .end_frame = ff_vdpau_mpeg_end_frame,
.decode_slice = vdpau_mpeg_decode_slice, .decode_slice = vdpau_mpeg_decode_slice,
.priv_data_size = sizeof(struct vdpau_picture_context),
}; };
#endif #endif
...@@ -31,8 +31,9 @@ static int vdpau_mpeg4_start_frame(AVCodecContext *avctx, ...@@ -31,8 +31,9 @@ static int vdpau_mpeg4_start_frame(AVCodecContext *avctx,
const uint8_t *buffer, uint32_t size) const uint8_t *buffer, uint32_t size)
{ {
MpegEncContext * const s = avctx->priv_data; MpegEncContext * const s = avctx->priv_data;
AVVDPAUContext *hwctx = avctx->hwaccel_context; Picture *pic = s->current_picture_ptr;
VdpPictureInfoMPEG4Part2 *info = &hwctx->info.mpeg4; struct vdpau_picture_context *pic_ctx = pic->hwaccel_picture_private;
VdpPictureInfoMPEG4Part2 *info = &pic_ctx->info.mpeg4;
VdpVideoSurface ref; VdpVideoSurface ref;
int i; int i;
...@@ -74,8 +75,8 @@ static int vdpau_mpeg4_start_frame(AVCodecContext *avctx, ...@@ -74,8 +75,8 @@ static int vdpau_mpeg4_start_frame(AVCodecContext *avctx,
info->non_intra_quantizer_matrix[i] = s->inter_matrix[i]; info->non_intra_quantizer_matrix[i] = s->inter_matrix[i];
} }
ff_vdpau_common_start_frame(avctx, buffer, size); ff_vdpau_common_start_frame(pic, buffer, size);
return ff_vdpau_add_buffer(avctx, buffer, size); return ff_vdpau_add_buffer(pic, buffer, size);
} }
static int vdpau_mpeg4_decode_slice(av_unused AVCodecContext *avctx, static int vdpau_mpeg4_decode_slice(av_unused AVCodecContext *avctx,
...@@ -94,6 +95,7 @@ AVHWAccel ff_h263_vdpau_hwaccel = { ...@@ -94,6 +95,7 @@ AVHWAccel ff_h263_vdpau_hwaccel = {
.start_frame = vdpau_mpeg4_start_frame, .start_frame = vdpau_mpeg4_start_frame,
.end_frame = ff_vdpau_mpeg_end_frame, .end_frame = ff_vdpau_mpeg_end_frame,
.decode_slice = vdpau_mpeg4_decode_slice, .decode_slice = vdpau_mpeg4_decode_slice,
.priv_data_size = sizeof(struct vdpau_picture_context),
}; };
#endif #endif
...@@ -106,5 +108,6 @@ AVHWAccel ff_mpeg4_vdpau_hwaccel = { ...@@ -106,5 +108,6 @@ AVHWAccel ff_mpeg4_vdpau_hwaccel = {
.start_frame = vdpau_mpeg4_start_frame, .start_frame = vdpau_mpeg4_start_frame,
.end_frame = ff_vdpau_mpeg_end_frame, .end_frame = ff_vdpau_mpeg_end_frame,
.decode_slice = vdpau_mpeg4_decode_slice, .decode_slice = vdpau_mpeg4_decode_slice,
.priv_data_size = sizeof(struct vdpau_picture_context),
}; };
#endif #endif
...@@ -32,9 +32,10 @@ static int vdpau_vc1_start_frame(AVCodecContext *avctx, ...@@ -32,9 +32,10 @@ static int vdpau_vc1_start_frame(AVCodecContext *avctx,
const uint8_t *buffer, uint32_t size) const uint8_t *buffer, uint32_t size)
{ {
VC1Context * const v = avctx->priv_data; VC1Context * const v = avctx->priv_data;
AVVDPAUContext *hwctx = avctx->hwaccel_context;
MpegEncContext * const s = &v->s; MpegEncContext * const s = &v->s;
VdpPictureInfoVC1 *info = &hwctx->info.vc1; Picture *pic = s->current_picture_ptr;
struct vdpau_picture_context *pic_ctx = pic->hwaccel_picture_private;
VdpPictureInfoVC1 *info = &pic_ctx->info.vc1;
VdpVideoSurface ref; VdpVideoSurface ref;
/* fill LvPictureInfoVC1 struct */ /* fill LvPictureInfoVC1 struct */
...@@ -88,20 +89,23 @@ static int vdpau_vc1_start_frame(AVCodecContext *avctx, ...@@ -88,20 +89,23 @@ static int vdpau_vc1_start_frame(AVCodecContext *avctx,
info->deblockEnable = v->postprocflag & 1; info->deblockEnable = v->postprocflag & 1;
info->pquant = v->pq; info->pquant = v->pq;
return ff_vdpau_common_start_frame(avctx, buffer, size); return ff_vdpau_common_start_frame(pic, buffer, size);
} }
static int vdpau_vc1_decode_slice(AVCodecContext *avctx, static int vdpau_vc1_decode_slice(AVCodecContext *avctx,
const uint8_t *buffer, uint32_t size) const uint8_t *buffer, uint32_t size)
{ {
AVVDPAUContext *hwctx = avctx->hwaccel_context; VC1Context * const v = avctx->priv_data;
MpegEncContext * const s = &v->s;
Picture *pic = s->current_picture_ptr;
struct vdpau_picture_context *pic_ctx = pic->hwaccel_picture_private;
int val; int val;
val = ff_vdpau_add_buffer(avctx, buffer, size); val = ff_vdpau_add_buffer(pic, buffer, size);
if (val < 0) if (val < 0)
return val; return val;
hwctx->info.vc1.slice_count++; pic_ctx->info.vc1.slice_count++;
return 0; return 0;
} }
...@@ -114,6 +118,7 @@ AVHWAccel ff_wmv3_vdpau_hwaccel = { ...@@ -114,6 +118,7 @@ AVHWAccel ff_wmv3_vdpau_hwaccel = {
.start_frame = vdpau_vc1_start_frame, .start_frame = vdpau_vc1_start_frame,
.end_frame = ff_vdpau_mpeg_end_frame, .end_frame = ff_vdpau_mpeg_end_frame,
.decode_slice = vdpau_vc1_decode_slice, .decode_slice = vdpau_vc1_decode_slice,
.priv_data_size = sizeof(struct vdpau_picture_context),
}; };
#endif #endif
...@@ -125,4 +130,5 @@ AVHWAccel ff_vc1_vdpau_hwaccel = { ...@@ -125,4 +130,5 @@ AVHWAccel ff_vc1_vdpau_hwaccel = {
.start_frame = vdpau_vc1_start_frame, .start_frame = vdpau_vc1_start_frame,
.end_frame = ff_vdpau_mpeg_end_frame, .end_frame = ff_vdpau_mpeg_end_frame,
.decode_slice = vdpau_vc1_decode_slice, .decode_slice = vdpau_vc1_decode_slice,
.priv_data_size = sizeof(struct vdpau_picture_context),
}; };
Markdown is supported
0% or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment