Commit 6faf0a21 authored by Michael Niedermayer's avatar Michael Niedermayer

Merge remote-tracking branch 'qatar/master'

* qatar/master: (53 commits)
  probe: Restore identification of files with very large id3 tags and no extension.
  probe: Remove id3 tag presence as a criteria to do file extension checking.
  mpegts: MP4 SL support
  mpegts: MP4 OD support
  mpegts: Add support for Sections in PMT
  mpegts: Replace the MP4 descriptor parser with a recursive parser.
  mpegts: Add support for multiple mp4 descriptors
  mpegts: Parse mpeg2 SL descriptors.
  isom: Add MPEG4SYSTEMS dummy object type indication.
  aacdec: allow output reconfiguration on channel changes
  nellymoserenc: take float input samples instead of int16
  nellymoserdec: use dsp functions for overlap and windowing
  nellymoserdec: do not fail if there is extra data in the packet
  nellymoserdec: fail if output buffer is too small
  nellymoserdec: remove pointless buffer size check.
  lavf: add init_put_byte() to the list of visible symbols.
  seek-test: free options dictionary after use
  snow: do not draw_edge if emu_edge is set
  tools/pktdumper: update to recent avformat api
  seek-test: update to recent avformat api
  ...

Conflicts:
	doc/APIchanges
	libavcodec/mpegaudiodec.c
	libavcodec/nellymoserdec.c
	libavcodec/snow.c
	libavcodec/version.h
	libavcodec/wmadec.c
	libavformat/avformat.h
	libavformat/mpegts.c
	libavformat/mxfdec.c
	libavformat/utils.c
	libavformat/wtv.c
Merged-by: 's avatarMichael Niedermayer <michaelni@gmx.at>
parents ed1aa892 61856d06
...@@ -16,6 +16,10 @@ API changes, most recent first: ...@@ -16,6 +16,10 @@ API changes, most recent first:
2011-10-20 - b35e9e1 - lavu 51.22.0 2011-10-20 - b35e9e1 - lavu 51.22.0
Add av_strtok() to avstring.h. Add av_strtok() to avstring.h.
2011-xx-xx - xxxxxxx - lavc 53.15.0
Remove avcodec_parse_frame.
Deprecate AVCodecContext.parse_only and CODEC_CAP_PARSE_ONLY.
2011-10-xx - xxxxxxx - lavf 53.10.0 2011-10-xx - xxxxxxx - lavf 53.10.0
Add avformat_new_stream(). Deprecate av_new_stream(). Add avformat_new_stream(). Deprecate av_new_stream().
......
This diff is collapsed.
...@@ -383,6 +383,8 @@ enum CodecID { ...@@ -383,6 +383,8 @@ enum CodecID {
CODEC_ID_MPEG2TS= 0x20000, /**< _FAKE_ codec to indicate a raw MPEG-2 TS CODEC_ID_MPEG2TS= 0x20000, /**< _FAKE_ codec to indicate a raw MPEG-2 TS
* stream (only used by libavformat) */ * stream (only used by libavformat) */
CODEC_ID_MPEG4SYSTEMS = 0x20001, /**< _FAKE_ codec to indicate a MPEG-4 Systems
* stream (only used by libavformat) */
CODEC_ID_FFMETADATA=0x21000, ///< Dummy codec for streams containing only metadata information. CODEC_ID_FFMETADATA=0x21000, ///< Dummy codec for streams containing only metadata information.
}; };
...@@ -682,8 +684,10 @@ typedef struct RcOverride{ ...@@ -682,8 +684,10 @@ typedef struct RcOverride{
* assume the buffer was allocated by avcodec_default_get_buffer. * assume the buffer was allocated by avcodec_default_get_buffer.
*/ */
#define CODEC_CAP_DR1 0x0002 #define CODEC_CAP_DR1 0x0002
#if FF_API_PARSE_FRAME
/* If 'parse_only' field is true, then avcodec_parse_frame() can be used. */ /* If 'parse_only' field is true, then avcodec_parse_frame() can be used. */
#define CODEC_CAP_PARSE_ONLY 0x0004 #define CODEC_CAP_PARSE_ONLY 0x0004
#endif
#define CODEC_CAP_TRUNCATED 0x0008 #define CODEC_CAP_TRUNCATED 0x0008
/* Codec can export data for HW decoding (XvMC). */ /* Codec can export data for HW decoding (XvMC). */
#define CODEC_CAP_HWACCEL 0x0010 #define CODEC_CAP_HWACCEL 0x0010
...@@ -1590,9 +1594,15 @@ typedef struct AVCodecContext { ...@@ -1590,9 +1594,15 @@ typedef struct AVCodecContext {
*/ */
int block_align; int block_align;
int parse_only; /* - decoding only: If true, only parsing is done #if FF_API_PARSE_FRAME
(function avcodec_parse_frame()). The frame /**
data is returned. Only MPEG codecs support this now. */ * If true, only parsing is done. The frame data is returned.
* Only MPEG audio decoders support this now.
* - encoding: unused
* - decoding: Set by user
*/
attribute_deprecated int parse_only;
#endif
/** /**
* 0-> h263 quant 1-> mpeg quant * 0-> h263 quant 1-> mpeg quant
...@@ -4047,10 +4057,6 @@ int avcodec_decode_subtitle2(AVCodecContext *avctx, AVSubtitle *sub, ...@@ -4047,10 +4057,6 @@ int avcodec_decode_subtitle2(AVCodecContext *avctx, AVSubtitle *sub,
*/ */
void avsubtitle_free(AVSubtitle *sub); void avsubtitle_free(AVSubtitle *sub);
int avcodec_parse_frame(AVCodecContext *avctx, uint8_t **pdata,
int *data_size_ptr,
uint8_t *buf, int buf_size);
/** /**
* Encode an audio frame from samples into buf. * Encode an audio frame from samples into buf.
* *
......
This diff is collapsed.
...@@ -30,7 +30,9 @@ AVCodec ff_mp1float_decoder = { ...@@ -30,7 +30,9 @@ AVCodec ff_mp1float_decoder = {
.priv_data_size = sizeof(MPADecodeContext), .priv_data_size = sizeof(MPADecodeContext),
.init = decode_init, .init = decode_init,
.decode = decode_frame, .decode = decode_frame,
#if FF_API_PARSE_FRAME
.capabilities = CODEC_CAP_PARSE_ONLY, .capabilities = CODEC_CAP_PARSE_ONLY,
#endif
.flush = flush, .flush = flush,
.long_name = NULL_IF_CONFIG_SMALL("MP1 (MPEG audio layer 1)"), .long_name = NULL_IF_CONFIG_SMALL("MP1 (MPEG audio layer 1)"),
}; };
...@@ -43,7 +45,9 @@ AVCodec ff_mp2float_decoder = { ...@@ -43,7 +45,9 @@ AVCodec ff_mp2float_decoder = {
.priv_data_size = sizeof(MPADecodeContext), .priv_data_size = sizeof(MPADecodeContext),
.init = decode_init, .init = decode_init,
.decode = decode_frame, .decode = decode_frame,
#if FF_API_PARSE_FRAME
.capabilities = CODEC_CAP_PARSE_ONLY, .capabilities = CODEC_CAP_PARSE_ONLY,
#endif
.flush = flush, .flush = flush,
.long_name = NULL_IF_CONFIG_SMALL("MP2 (MPEG audio layer 2)"), .long_name = NULL_IF_CONFIG_SMALL("MP2 (MPEG audio layer 2)"),
}; };
...@@ -56,7 +60,9 @@ AVCodec ff_mp3float_decoder = { ...@@ -56,7 +60,9 @@ AVCodec ff_mp3float_decoder = {
.priv_data_size = sizeof(MPADecodeContext), .priv_data_size = sizeof(MPADecodeContext),
.init = decode_init, .init = decode_init,
.decode = decode_frame, .decode = decode_frame,
#if FF_API_PARSE_FRAME
.capabilities = CODEC_CAP_PARSE_ONLY, .capabilities = CODEC_CAP_PARSE_ONLY,
#endif
.flush = flush, .flush = flush,
.long_name = NULL_IF_CONFIG_SMALL("MP3 (MPEG audio layer 3)"), .long_name = NULL_IF_CONFIG_SMALL("MP3 (MPEG audio layer 3)"),
}; };
...@@ -69,7 +75,9 @@ AVCodec ff_mp3adufloat_decoder = { ...@@ -69,7 +75,9 @@ AVCodec ff_mp3adufloat_decoder = {
.priv_data_size = sizeof(MPADecodeContext), .priv_data_size = sizeof(MPADecodeContext),
.init = decode_init, .init = decode_init,
.decode = decode_frame_adu, .decode = decode_frame_adu,
#if FF_API_PARSE_FRAME
.capabilities = CODEC_CAP_PARSE_ONLY, .capabilities = CODEC_CAP_PARSE_ONLY,
#endif
.flush = flush, .flush = flush,
.long_name = NULL_IF_CONFIG_SMALL("ADU (Application Data Unit) MP3 (MPEG audio layer 3)"), .long_name = NULL_IF_CONFIG_SMALL("ADU (Application Data Unit) MP3 (MPEG audio layer 3)"),
}; };
......
...@@ -48,7 +48,7 @@ ...@@ -48,7 +48,7 @@
typedef struct NellyMoserDecodeContext { typedef struct NellyMoserDecodeContext {
AVCodecContext* avctx; AVCodecContext* avctx;
float *float_buf; float *float_buf;
float state[NELLY_BUF_LEN]; DECLARE_ALIGNED(16, float, state)[NELLY_BUF_LEN];
AVLFG random_state; AVLFG random_state;
GetBitContext gb; GetBitContext gb;
float scale_bias; float scale_bias;
...@@ -58,23 +58,6 @@ typedef struct NellyMoserDecodeContext { ...@@ -58,23 +58,6 @@ typedef struct NellyMoserDecodeContext {
DECLARE_ALIGNED(32, float, imdct_out)[NELLY_BUF_LEN * 2]; DECLARE_ALIGNED(32, float, imdct_out)[NELLY_BUF_LEN * 2];
} NellyMoserDecodeContext; } NellyMoserDecodeContext;
static void overlap_and_window(NellyMoserDecodeContext *s, float *state, float *audio, float *a_in)
{
int bot, top;
bot = 0;
top = NELLY_BUF_LEN-1;
while (bot < NELLY_BUF_LEN) {
audio[bot] = a_in [bot]*ff_sine_128[bot]
+state[bot]*ff_sine_128[top];
bot++;
top--;
}
memcpy(state, a_in + NELLY_BUF_LEN, sizeof(float)*NELLY_BUF_LEN);
}
static void nelly_decode_block(NellyMoserDecodeContext *s, static void nelly_decode_block(NellyMoserDecodeContext *s,
const unsigned char block[NELLY_BLOCK_LEN], const unsigned char block[NELLY_BLOCK_LEN],
float audio[NELLY_SAMPLES]) float audio[NELLY_SAMPLES])
...@@ -125,7 +108,9 @@ static void nelly_decode_block(NellyMoserDecodeContext *s, ...@@ -125,7 +108,9 @@ static void nelly_decode_block(NellyMoserDecodeContext *s,
s->imdct_ctx.imdct_calc(&s->imdct_ctx, s->imdct_out, aptr); s->imdct_ctx.imdct_calc(&s->imdct_ctx, s->imdct_out, aptr);
/* XXX: overlapping and windowing should be part of a more /* XXX: overlapping and windowing should be part of a more
generic imdct function */ generic imdct function */
overlap_and_window(s, s->state, aptr, s->imdct_out); s->dsp.vector_fmul_reverse(s->state, s->state, ff_sine_128, NELLY_BUF_LEN);
s->dsp.vector_fmul_add(aptr, s->imdct_out, ff_sine_128, s->state, NELLY_BUF_LEN);
memcpy(s->state, s->imdct_out + NELLY_BUF_LEN, sizeof(float)*NELLY_BUF_LEN);
} }
} }
...@@ -172,20 +157,21 @@ static int decode_tag(AVCodecContext * avctx, ...@@ -172,20 +157,21 @@ static int decode_tag(AVCodecContext * avctx,
float *samples_flt = data; float *samples_flt = data;
*data_size = 0; *data_size = 0;
if (buf_size < avctx->block_align) {
return buf_size;
}
if (buf_size % NELLY_BLOCK_LEN) {
av_log(avctx, AV_LOG_ERROR, "Tag size %d.\n", buf_size);
return buf_size;
}
block_size = NELLY_SAMPLES * av_get_bytes_per_sample(avctx->sample_fmt); block_size = NELLY_SAMPLES * av_get_bytes_per_sample(avctx->sample_fmt);
blocks = FFMIN(buf_size / NELLY_BLOCK_LEN, data_max / block_size); blocks = buf_size / NELLY_BLOCK_LEN;
if (blocks <= 0) { if (blocks <= 0) {
av_log(avctx, AV_LOG_ERROR, "Packet is too small\n");
return AVERROR_INVALIDDATA;
}
if (data_max < blocks * block_size) {
av_log(avctx, AV_LOG_ERROR, "Output buffer is too small\n"); av_log(avctx, AV_LOG_ERROR, "Output buffer is too small\n");
return AVERROR(EINVAL); return AVERROR(EINVAL);
} }
if (buf_size % NELLY_BLOCK_LEN) {
av_log(avctx, AV_LOG_WARNING, "Leftover bytes: %d.\n",
buf_size % NELLY_BLOCK_LEN);
}
/* Normal numbers of blocks for sample rates: /* Normal numbers of blocks for sample rates:
* 8000 Hz - 1 * 8000 Hz - 1
* 11025 Hz - 2 * 11025 Hz - 2
......
...@@ -146,7 +146,7 @@ static av_cold int encode_init(AVCodecContext *avctx) ...@@ -146,7 +146,7 @@ static av_cold int encode_init(AVCodecContext *avctx)
avctx->frame_size = NELLY_SAMPLES; avctx->frame_size = NELLY_SAMPLES;
s->avctx = avctx; s->avctx = avctx;
ff_mdct_init(&s->mdct_ctx, 8, 0, 1.0); ff_mdct_init(&s->mdct_ctx, 8, 0, 32768.0);
dsputil_init(&s->dsp, avctx); dsputil_init(&s->dsp, avctx);
/* Generate overlap window */ /* Generate overlap window */
...@@ -352,17 +352,15 @@ static void encode_block(NellyMoserEncodeContext *s, unsigned char *output, int ...@@ -352,17 +352,15 @@ static void encode_block(NellyMoserEncodeContext *s, unsigned char *output, int
static int encode_frame(AVCodecContext *avctx, uint8_t *frame, int buf_size, void *data) static int encode_frame(AVCodecContext *avctx, uint8_t *frame, int buf_size, void *data)
{ {
NellyMoserEncodeContext *s = avctx->priv_data; NellyMoserEncodeContext *s = avctx->priv_data;
const int16_t *samples = data; const float *samples = data;
int i; int i;
if (s->last_frame) if (s->last_frame)
return 0; return 0;
if (data) { if (data) {
for (i = 0; i < avctx->frame_size; i++) { memcpy(s->buf[s->bufsel], samples, avctx->frame_size * sizeof(*samples));
s->buf[s->bufsel][i] = samples[i]; for (i = avctx->frame_size; i < NELLY_SAMPLES; i++) {
}
for (; i < NELLY_SAMPLES; i++) {
s->buf[s->bufsel][i] = 0; s->buf[s->bufsel][i] = 0;
} }
s->bufsel = 1 - s->bufsel; s->bufsel = 1 - s->bufsel;
...@@ -393,5 +391,5 @@ AVCodec ff_nellymoser_encoder = { ...@@ -393,5 +391,5 @@ AVCodec ff_nellymoser_encoder = {
.close = encode_end, .close = encode_end,
.capabilities = CODEC_CAP_SMALL_LAST_FRAME | CODEC_CAP_DELAY, .capabilities = CODEC_CAP_SMALL_LAST_FRAME | CODEC_CAP_DELAY,
.long_name = NULL_IF_CONFIG_SMALL("Nellymoser Asao"), .long_name = NULL_IF_CONFIG_SMALL("Nellymoser Asao"),
.sample_fmts = (const enum AVSampleFormat[]){AV_SAMPLE_FMT_S16,AV_SAMPLE_FMT_NONE}, .sample_fmts = (const enum AVSampleFormat[]){AV_SAMPLE_FMT_FLT,AV_SAMPLE_FMT_NONE},
}; };
...@@ -1665,7 +1665,7 @@ static int frame_start(SnowContext *s){ ...@@ -1665,7 +1665,7 @@ static int frame_start(SnowContext *s){
int w= s->avctx->width; //FIXME round up to x16 ? int w= s->avctx->width; //FIXME round up to x16 ?
int h= s->avctx->height; int h= s->avctx->height;
if(s->current_picture.data[0] && !(s->avctx->flags&CODEC_FLAG_EMU_EDGE)){ if (s->current_picture.data[0] && !(s->avctx->flags&CODEC_FLAG_EMU_EDGE)) {
s->dsp.draw_edges(s->current_picture.data[0], s->dsp.draw_edges(s->current_picture.data[0],
s->current_picture.linesize[0], w , h , s->current_picture.linesize[0], w , h ,
EDGE_WIDTH , EDGE_WIDTH , EDGE_TOP | EDGE_BOTTOM); EDGE_WIDTH , EDGE_WIDTH , EDGE_TOP | EDGE_BOTTOM);
......
...@@ -21,7 +21,7 @@ ...@@ -21,7 +21,7 @@
#define AVCODEC_VERSION_H #define AVCODEC_VERSION_H
#define LIBAVCODEC_VERSION_MAJOR 53 #define LIBAVCODEC_VERSION_MAJOR 53
#define LIBAVCODEC_VERSION_MINOR 23 #define LIBAVCODEC_VERSION_MINOR 24
#define LIBAVCODEC_VERSION_MICRO 0 #define LIBAVCODEC_VERSION_MICRO 0
#define LIBAVCODEC_VERSION_INT AV_VERSION_INT(LIBAVCODEC_VERSION_MAJOR, \ #define LIBAVCODEC_VERSION_INT AV_VERSION_INT(LIBAVCODEC_VERSION_MAJOR, \
...@@ -101,5 +101,8 @@ ...@@ -101,5 +101,8 @@
#ifndef FF_API_GET_ALPHA_INFO #ifndef FF_API_GET_ALPHA_INFO
#define FF_API_GET_ALPHA_INFO (LIBAVCODEC_VERSION_MAJOR < 54) #define FF_API_GET_ALPHA_INFO (LIBAVCODEC_VERSION_MAJOR < 54)
#endif #endif
#ifndef FF_API_PARSE_FRAME
#define FF_API_PARSE_FRAME (LIBAVCODEC_VERSION_MAJOR < 54)
#endif
#endif /* AVCODEC_VERSION_H */ #endif /* AVCODEC_VERSION_H */
...@@ -816,7 +816,7 @@ static int wma_decode_superframe(AVCodecContext *avctx, ...@@ -816,7 +816,7 @@ static int wma_decode_superframe(AVCodecContext *avctx,
const uint8_t *buf = avpkt->data; const uint8_t *buf = avpkt->data;
int buf_size = avpkt->size; int buf_size = avpkt->size;
WMACodecContext *s = avctx->priv_data; WMACodecContext *s = avctx->priv_data;
int nb_frames, bit_offset, i, pos, len; int nb_frames, bit_offset, i, pos, len, out_size;
uint8_t *q; uint8_t *q;
int16_t *samples; int16_t *samples;
...@@ -838,13 +838,19 @@ static int wma_decode_superframe(AVCodecContext *avctx, ...@@ -838,13 +838,19 @@ static int wma_decode_superframe(AVCodecContext *avctx,
if (s->use_bit_reservoir) { if (s->use_bit_reservoir) {
/* read super frame header */ /* read super frame header */
skip_bits(&s->gb, 4); /* super frame index */ skip_bits(&s->gb, 4); /* super frame index */
nb_frames = get_bits(&s->gb, 4) - 1; nb_frames = get_bits(&s->gb, 4) - (s->last_superframe_len <= 0);
} else {
nb_frames = 1;
}
if((nb_frames+1) * s->nb_channels * s->frame_len * sizeof(int16_t) > *data_size){ out_size = nb_frames * s->frame_len * s->nb_channels *
av_get_bytes_per_sample(avctx->sample_fmt);
if (*data_size < out_size) {
av_log(s->avctx, AV_LOG_ERROR, "Insufficient output space\n"); av_log(s->avctx, AV_LOG_ERROR, "Insufficient output space\n");
goto fail; goto fail;
} }
if (s->use_bit_reservoir) {
bit_offset = get_bits(&s->gb, s->byte_offset_bits + 3); bit_offset = get_bits(&s->gb, s->byte_offset_bits + 3);
if (s->last_superframe_len > 0) { if (s->last_superframe_len > 0) {
...@@ -873,6 +879,7 @@ static int wma_decode_superframe(AVCodecContext *avctx, ...@@ -873,6 +879,7 @@ static int wma_decode_superframe(AVCodecContext *avctx,
if (wma_decode_frame(s, samples) < 0) if (wma_decode_frame(s, samples) < 0)
goto fail; goto fail;
samples += s->nb_channels * s->frame_len; samples += s->nb_channels * s->frame_len;
nb_frames--;
} }
/* read each frame starting from bit_offset */ /* read each frame starting from bit_offset */
...@@ -901,10 +908,6 @@ static int wma_decode_superframe(AVCodecContext *avctx, ...@@ -901,10 +908,6 @@ static int wma_decode_superframe(AVCodecContext *avctx,
s->last_superframe_len = len; s->last_superframe_len = len;
memcpy(s->last_superframe, buf + pos, len); memcpy(s->last_superframe, buf + pos, len);
} else { } else {
if(s->nb_channels * s->frame_len * sizeof(int16_t) > *data_size){
av_log(s->avctx, AV_LOG_ERROR, "Insufficient output space\n");
goto fail;
}
/* single frame decode */ /* single frame decode */
if (wma_decode_frame(s, samples) < 0) if (wma_decode_frame(s, samples) < 0)
goto fail; goto fail;
...@@ -912,7 +915,7 @@ static int wma_decode_superframe(AVCodecContext *avctx, ...@@ -912,7 +915,7 @@ static int wma_decode_superframe(AVCodecContext *avctx,
} }
//av_log(NULL, AV_LOG_ERROR, "%d %d %d %d outbytes:%d eaten:%d\n", s->frame_len_bits, s->block_len_bits, s->frame_len, s->block_len, (int8_t *)samples - (int8_t *)data, s->block_align); //av_log(NULL, AV_LOG_ERROR, "%d %d %d %d outbytes:%d eaten:%d\n", s->frame_len_bits, s->block_len_bits, s->frame_len, s->block_len, (int8_t *)samples - (int8_t *)data, s->block_align);
*data_size = (int8_t *)samples - (int8_t *)data; *data_size = out_size;
return buf_size; return buf_size;
fail: fail:
/* when error, we reset the bit reservoir */ /* when error, we reset the bit reservoir */
......
...@@ -86,12 +86,14 @@ ...@@ -86,12 +86,14 @@
* subframe in order to reconstruct the output samples. * subframe in order to reconstruct the output samples.
*/ */
#include "libavutil/intreadwrite.h"
#include "avcodec.h" #include "avcodec.h"
#include "internal.h" #include "internal.h"
#include "get_bits.h" #include "get_bits.h"
#include "put_bits.h" #include "put_bits.h"
#include "wmaprodata.h" #include "wmaprodata.h"
#include "dsputil.h" #include "dsputil.h"
#include "fmtconvert.h"
#include "sinewin.h" #include "sinewin.h"
#include "wma.h" #include "wma.h"
...@@ -166,6 +168,7 @@ typedef struct WMAProDecodeCtx { ...@@ -166,6 +168,7 @@ typedef struct WMAProDecodeCtx {
/* generic decoder variables */ /* generic decoder variables */
AVCodecContext* avctx; ///< codec context for av_log AVCodecContext* avctx; ///< codec context for av_log
DSPContext dsp; ///< accelerated DSP functions DSPContext dsp; ///< accelerated DSP functions
FmtConvertContext fmt_conv;
uint8_t frame_data[MAX_FRAMESIZE + uint8_t frame_data[MAX_FRAMESIZE +
FF_INPUT_BUFFER_PADDING_SIZE];///< compressed frame data FF_INPUT_BUFFER_PADDING_SIZE];///< compressed frame data
PutBitContext pb; ///< context for filling the frame_data buffer PutBitContext pb; ///< context for filling the frame_data buffer
...@@ -279,6 +282,7 @@ static av_cold int decode_init(AVCodecContext *avctx) ...@@ -279,6 +282,7 @@ static av_cold int decode_init(AVCodecContext *avctx)
s->avctx = avctx; s->avctx = avctx;
dsputil_init(&s->dsp, avctx); dsputil_init(&s->dsp, avctx);
ff_fmt_convert_init(&s->fmt_conv, avctx);
init_put_bits(&s->pb, s->frame_data, MAX_FRAMESIZE); init_put_bits(&s->pb, s->frame_data, MAX_FRAMESIZE);
avctx->sample_fmt = AV_SAMPLE_FMT_FLT; avctx->sample_fmt = AV_SAMPLE_FMT_FLT;
...@@ -767,7 +771,7 @@ static int decode_coeffs(WMAProDecodeCtx *s, int c) ...@@ -767,7 +771,7 @@ static int decode_coeffs(WMAProDecodeCtx *s, int c)
/* Integers 0..15 as single-precision floats. The table saves a /* Integers 0..15 as single-precision floats. The table saves a
costly int to float conversion, and storing the values as costly int to float conversion, and storing the values as
integers allows fast sign-flipping. */ integers allows fast sign-flipping. */
static const int fval_tab[16] = { static const uint32_t fval_tab[16] = {
0x00000000, 0x3f800000, 0x40000000, 0x40400000, 0x00000000, 0x3f800000, 0x40000000, 0x40400000,
0x40800000, 0x40a00000, 0x40c00000, 0x40e00000, 0x40800000, 0x40a00000, 0x40c00000, 0x40e00000,
0x41000000, 0x41100000, 0x41200000, 0x41300000, 0x41000000, 0x41100000, 0x41200000, 0x41300000,
...@@ -799,7 +803,7 @@ static int decode_coeffs(WMAProDecodeCtx *s, int c) ...@@ -799,7 +803,7 @@ static int decode_coeffs(WMAProDecodeCtx *s, int c)
4 vector coded large values) */ 4 vector coded large values) */
while ((s->transmit_num_vec_coeffs || !rl_mode) && while ((s->transmit_num_vec_coeffs || !rl_mode) &&
(cur_coeff + 3 < ci->num_vec_coeffs)) { (cur_coeff + 3 < ci->num_vec_coeffs)) {
int vals[4]; uint32_t vals[4];
int i; int i;
unsigned int idx; unsigned int idx;
...@@ -809,15 +813,15 @@ static int decode_coeffs(WMAProDecodeCtx *s, int c) ...@@ -809,15 +813,15 @@ static int decode_coeffs(WMAProDecodeCtx *s, int c)
for (i = 0; i < 4; i += 2) { for (i = 0; i < 4; i += 2) {
idx = get_vlc2(&s->gb, vec2_vlc.table, VLCBITS, VEC2MAXDEPTH); idx = get_vlc2(&s->gb, vec2_vlc.table, VLCBITS, VEC2MAXDEPTH);
if (idx == HUFF_VEC2_SIZE - 1) { if (idx == HUFF_VEC2_SIZE - 1) {
int v0, v1; uint32_t v0, v1;
v0 = get_vlc2(&s->gb, vec1_vlc.table, VLCBITS, VEC1MAXDEPTH); v0 = get_vlc2(&s->gb, vec1_vlc.table, VLCBITS, VEC1MAXDEPTH);
if (v0 == HUFF_VEC1_SIZE - 1) if (v0 == HUFF_VEC1_SIZE - 1)
v0 += ff_wma_get_large_val(&s->gb); v0 += ff_wma_get_large_val(&s->gb);
v1 = get_vlc2(&s->gb, vec1_vlc.table, VLCBITS, VEC1MAXDEPTH); v1 = get_vlc2(&s->gb, vec1_vlc.table, VLCBITS, VEC1MAXDEPTH);
if (v1 == HUFF_VEC1_SIZE - 1) if (v1 == HUFF_VEC1_SIZE - 1)
v1 += ff_wma_get_large_val(&s->gb); v1 += ff_wma_get_large_val(&s->gb);
((float*)vals)[i ] = v0; vals[i ] = ((av_alias32){ .f32 = v0 }).u32;
((float*)vals)[i+1] = v1; vals[i+1] = ((av_alias32){ .f32 = v1 }).u32;
} else { } else {
vals[i] = fval_tab[symbol_to_vec2[idx] >> 4 ]; vals[i] = fval_tab[symbol_to_vec2[idx] >> 4 ];
vals[i+1] = fval_tab[symbol_to_vec2[idx] & 0xF]; vals[i+1] = fval_tab[symbol_to_vec2[idx] & 0xF];
...@@ -833,8 +837,8 @@ static int decode_coeffs(WMAProDecodeCtx *s, int c) ...@@ -833,8 +837,8 @@ static int decode_coeffs(WMAProDecodeCtx *s, int c)
/** decode sign */ /** decode sign */
for (i = 0; i < 4; i++) { for (i = 0; i < 4; i++) {
if (vals[i]) { if (vals[i]) {
int sign = get_bits1(&s->gb) - 1; uint32_t sign = get_bits1(&s->gb) - 1;
*(uint32_t*)&ci->coeffs[cur_coeff] = vals[i] ^ sign<<31; AV_WN32A(&ci->coeffs[cur_coeff], vals[i] ^ sign << 31);
num_zeros = 0; num_zeros = 0;
} else { } else {
ci->coeffs[cur_coeff] = 0; ci->coeffs[cur_coeff] = 0;
...@@ -1281,6 +1285,7 @@ static int decode_frame(WMAProDecodeCtx *s) ...@@ -1281,6 +1285,7 @@ static int decode_frame(WMAProDecodeCtx *s)
int more_frames = 0; int more_frames = 0;
int len = 0; int len = 0;
int i; int i;
const float *out_ptr[WMAPRO_MAX_CHANNELS];
/** check for potential output buffer overflow */ /** check for potential output buffer overflow */
if (s->num_channels * s->samples_per_frame > s->samples_end - s->samples) { if (s->num_channels * s->samples_per_frame > s->samples_end - s->samples) {
...@@ -1356,18 +1361,12 @@ static int decode_frame(WMAProDecodeCtx *s) ...@@ -1356,18 +1361,12 @@ static int decode_frame(WMAProDecodeCtx *s)
} }
/** interleave samples and write them to the output buffer */ /** interleave samples and write them to the output buffer */
for (i = 0; i < s->num_channels; i++) { for (i = 0; i < s->num_channels; i++)
float* ptr = s->samples + i; out_ptr[i] = s->channel[i].out;
int incr = s->num_channels; s->fmt_conv.float_interleave(s->samples, out_ptr, s->samples_per_frame,
float* iptr = s->channel[i].out; s->num_channels);
float* iend = iptr + s->samples_per_frame;
// FIXME should create/use a DSP function here
while (iptr < iend) {
*ptr = *iptr++;
ptr += incr;
}
for (i = 0; i < s->num_channels; i++) {
/** reuse second half of the IMDCT output for the next frame */ /** reuse second half of the IMDCT output for the next frame */
memcpy(&s->channel[i].out[0], memcpy(&s->channel[i].out[0],
&s->channel[i].out[s->samples_per_frame], &s->channel[i].out[s->samples_per_frame],
......
...@@ -1730,7 +1730,7 @@ static int synth_superframe(AVCodecContext *ctx, ...@@ -1730,7 +1730,7 @@ static int synth_superframe(AVCodecContext *ctx,
{ {
WMAVoiceContext *s = ctx->priv_data; WMAVoiceContext *s = ctx->priv_data;
GetBitContext *gb = &s->gb, s_gb; GetBitContext *gb = &s->gb, s_gb;
int n, res, n_samples = 480; int n, res, out_size, n_samples = 480;
double lsps[MAX_FRAMES][MAX_LSPS]; double lsps[MAX_FRAMES][MAX_LSPS];
const double *mean_lsf = s->lsps == 16 ? const double *mean_lsf = s->lsps == 16 ?
wmavoice_mean_lsf16[s->lsp_def_mode] : wmavoice_mean_lsf10[s->lsp_def_mode]; wmavoice_mean_lsf16[s->lsp_def_mode] : wmavoice_mean_lsf10[s->lsp_def_mode];
...@@ -1748,7 +1748,10 @@ static int synth_superframe(AVCodecContext *ctx, ...@@ -1748,7 +1748,10 @@ static int synth_superframe(AVCodecContext *ctx,
s->sframe_cache_size = 0; s->sframe_cache_size = 0;
} }
if ((res = check_bits_for_superframe(gb, s)) == 1) return 1; if ((res = check_bits_for_superframe(gb, s)) == 1) {
*data_size = 0;
return 1;
}
/* First bit is speech/music bit, it differentiates between WMAVoice /* First bit is speech/music bit, it differentiates between WMAVoice
* speech samples (the actual codec) and WMAVoice music samples, which * speech samples (the actual codec) and WMAVoice music samples, which
...@@ -1789,6 +1792,14 @@ static int synth_superframe(AVCodecContext *ctx, ...@@ -1789,6 +1792,14 @@ static int synth_superframe(AVCodecContext *ctx,
stabilize_lsps(lsps[n], s->lsps); stabilize_lsps(lsps[n], s->lsps);
} }
out_size = n_samples * av_get_bytes_per_sample(ctx->sample_fmt);
if (*data_size < out_size) {
av_log(ctx, AV_LOG_ERROR,
"Output buffer too small (%d given - %zu needed)\n",
*data_size, out_size);
return -1;
}
/* Parse frames, optionally preceeded by per-frame (independent) LSPs. */ /* Parse frames, optionally preceeded by per-frame (independent) LSPs. */
for (n = 0; n < 3; n++) { for (n = 0; n < 3; n++) {
if (!s->has_residual_lsps) { if (!s->has_residual_lsps) {
...@@ -1808,9 +1819,11 @@ static int synth_superframe(AVCodecContext *ctx, ...@@ -1808,9 +1819,11 @@ static int synth_superframe(AVCodecContext *ctx,
&samples[n * MAX_FRAMESIZE], &samples[n * MAX_FRAMESIZE],
lsps[n], n == 0 ? s->prev_lsps : lsps[n - 1], lsps[n], n == 0 ? s->prev_lsps : lsps[n - 1],
&excitation[s->history_nsamples + n * MAX_FRAMESIZE], &excitation[s->history_nsamples + n * MAX_FRAMESIZE],
&synth[s->lsps + n * MAX_FRAMESIZE]))) &synth[s->lsps + n * MAX_FRAMESIZE]))) {
*data_size = 0;
return res; return res;
} }
}
/* Statistics? FIXME - we don't check for length, a slight overrun /* Statistics? FIXME - we don't check for length, a slight overrun
* will be caught by internal buffer padding, and anything else * will be caught by internal buffer padding, and anything else
...@@ -1821,7 +1834,7 @@ static int synth_superframe(AVCodecContext *ctx, ...@@ -1821,7 +1834,7 @@ static int synth_superframe(AVCodecContext *ctx,
} }
/* Specify nr. of output samples */ /* Specify nr. of output samples */
*data_size = n_samples * sizeof(float); *data_size = out_size;
/* Update history */ /* Update history */
memcpy(s->prev_lsps, lsps[2], memcpy(s->prev_lsps, lsps[2],
...@@ -1915,22 +1928,16 @@ static int wmavoice_decode_packet(AVCodecContext *ctx, void *data, ...@@ -1915,22 +1928,16 @@ static int wmavoice_decode_packet(AVCodecContext *ctx, void *data,
GetBitContext *gb = &s->gb; GetBitContext *gb = &s->gb;
int size, res, pos; int size, res, pos;
if (*data_size < 480 * sizeof(float)) {
av_log(ctx, AV_LOG_ERROR,
"Output buffer too small (%d given - %zu needed)\n",
*data_size, 480 * sizeof(float));
return -1;
}
*data_size = 0;
/* Packets are sometimes a multiple of ctx->block_align, with a packet /* Packets are sometimes a multiple of ctx->block_align, with a packet
* header at each ctx->block_align bytes. However, FFmpeg's ASF demuxer * header at each ctx->block_align bytes. However, FFmpeg's ASF demuxer
* feeds us ASF packets, which may concatenate multiple "codec" packets * feeds us ASF packets, which may concatenate multiple "codec" packets
* in a single "muxer" packet, so we artificially emulate that by * in a single "muxer" packet, so we artificially emulate that by
* capping the packet size at ctx->block_align. */ * capping the packet size at ctx->block_align. */
for (size = avpkt->size; size > ctx->block_align; size -= ctx->block_align); for (size = avpkt->size; size > ctx->block_align; size -= ctx->block_align);
if (!size) if (!size) {
*data_size = 0;
return 0; return 0;
}
init_get_bits(&s->gb, avpkt->data, size << 3); init_get_bits(&s->gb, avpkt->data, size << 3);
/* size == ctx->block_align is used to indicate whether we are dealing with /* size == ctx->block_align is used to indicate whether we are dealing with
......
...@@ -1297,7 +1297,7 @@ static int asf_read_seek(AVFormatContext *s, int stream_index, int64_t pts, int ...@@ -1297,7 +1297,7 @@ static int asf_read_seek(AVFormatContext *s, int stream_index, int64_t pts, int
} }
} }
/* no index or seeking by index failed */ /* no index or seeking by index failed */
if(av_seek_frame_binary(s, stream_index, pts, flags)<0) if (ff_seek_frame_binary(s, stream_index, pts, flags) < 0)
return -1; return -1;
asf_reset_header(s); asf_reset_header(s);
return 0; return 0;
......
...@@ -523,8 +523,10 @@ typedef struct AVStream { ...@@ -523,8 +523,10 @@ typedef struct AVStream {
AVRational r_frame_rate; AVRational r_frame_rate;
void *priv_data; void *priv_data;
#if FF_API_REORDER_PRIVATE
/* internal data used in av_find_stream_info() */ /* internal data used in av_find_stream_info() */
int64_t first_dts; int64_t first_dts;
#endif
/** /**
* encoding: pts generation when outputting stream * encoding: pts generation when outputting stream
...@@ -539,7 +541,9 @@ typedef struct AVStream { ...@@ -539,7 +541,9 @@ typedef struct AVStream {
* encoding: set by libavformat in av_write_header * encoding: set by libavformat in av_write_header
*/ */
AVRational time_base; AVRational time_base;
#if FF_API_REORDER_PRIVATE
int pts_wrap_bits; /**< number of bits in pts (used for wrapping control) */ int pts_wrap_bits; /**< number of bits in pts (used for wrapping control) */
#endif
#if FF_API_STREAM_COPY #if FF_API_STREAM_COPY
/* ffmpeg.c private use */ /* ffmpeg.c private use */
attribute_deprecated int stream_copy; /**< If set, just copy stream. */ attribute_deprecated int stream_copy; /**< If set, just copy stream. */
...@@ -572,6 +576,7 @@ typedef struct AVStream { ...@@ -572,6 +576,7 @@ typedef struct AVStream {
*/ */
int64_t duration; int64_t duration;
#if FF_API_REORDER_PRIVATE
/* av_read_frame() support */ /* av_read_frame() support */
enum AVStreamParseType need_parsing; enum AVStreamParseType need_parsing;
struct AVCodecParserContext *parser; struct AVCodecParserContext *parser;
...@@ -584,14 +589,17 @@ typedef struct AVStream { ...@@ -584,14 +589,17 @@ typedef struct AVStream {
support seeking natively. */ support seeking natively. */
int nb_index_entries; int nb_index_entries;
unsigned int index_entries_allocated_size; unsigned int index_entries_allocated_size;
#endif
int64_t nb_frames; ///< number of frames in this stream if known or 0 int64_t nb_frames; ///< number of frames in this stream if known or 0
int disposition; /**< AV_DISPOSITION_* bit field */ int disposition; /**< AV_DISPOSITION_* bit field */
#if FF_API_REORDER_PRIVATE
AVProbeData probe_data; AVProbeData probe_data;
#define MAX_REORDER_DELAY 16 #define MAX_REORDER_DELAY 16
int64_t pts_buffer[MAX_REORDER_DELAY+1]; int64_t pts_buffer[MAX_REORDER_DELAY+1];
#endif
/** /**
* sample aspect ratio (0 if unknown) * sample aspect ratio (0 if unknown)
...@@ -602,6 +610,7 @@ typedef struct AVStream { ...@@ -602,6 +610,7 @@ typedef struct AVStream {
AVDictionary *metadata; AVDictionary *metadata;
#if FF_API_REORDER_PRIVATE
/* Intended mostly for av_read_frame() support. Not supposed to be used by */ /* Intended mostly for av_read_frame() support. Not supposed to be used by */
/* external applications; try to use something else if at all possible. */ /* external applications; try to use something else if at all possible. */
const uint8_t *cur_ptr; const uint8_t *cur_ptr;
...@@ -630,12 +639,21 @@ typedef struct AVStream { ...@@ -630,12 +639,21 @@ typedef struct AVStream {
* used internally, NOT PART OF PUBLIC API, dont read or write from outside of libav* * used internally, NOT PART OF PUBLIC API, dont read or write from outside of libav*
*/ */
struct AVPacketList *last_in_packet_buffer; struct AVPacketList *last_in_packet_buffer;
#endif
/** /**
* Average framerate * Average framerate
*/ */
AVRational avg_frame_rate; AVRational avg_frame_rate;
/*****************************************************************
* All fields below this line are not part of the public API. They
* may not be used outside of libavformat and can be changed and
* removed at will.
* New public fields should be added right above.
*****************************************************************
*/
/** /**
* Number of frames that have been demuxed during av_find_stream_info() * Number of frames that have been demuxed during av_find_stream_info()
*/ */
...@@ -665,6 +683,49 @@ typedef struct AVStream { ...@@ -665,6 +683,49 @@ typedef struct AVStream {
* NOT PART OF PUBLIC API * NOT PART OF PUBLIC API
*/ */
int request_probe; int request_probe;
#if !FF_API_REORDER_PRIVATE
const uint8_t *cur_ptr;
int cur_len;
AVPacket cur_pkt;
// Timestamp generation support:
/**
* Timestamp corresponding to the last dts sync point.
*
* Initialized when AVCodecParserContext.dts_sync_point >= 0 and
* a DTS is received from the underlying container. Otherwise set to
* AV_NOPTS_VALUE by default.
*/
int64_t reference_dts;
int64_t first_dts;
int64_t cur_dts;
int last_IP_duration;
int64_t last_IP_pts;
/**
* Number of packets to buffer for codec probing
*/
#define MAX_PROBE_PACKETS 2500
int probe_packets;
/**
* last packet in packet_buffer for this stream when muxing.
*/
struct AVPacketList *last_in_packet_buffer;
AVProbeData probe_data;
#define MAX_REORDER_DELAY 16
int64_t pts_buffer[MAX_REORDER_DELAY+1];
/* av_read_frame() support */
enum AVStreamParseType need_parsing;
struct AVCodecParserContext *parser;
AVIndexEntry *index_entries; /**< Only used if the format does not
support seeking natively. */
int nb_index_entries;
unsigned int index_entries_allocated_size;
int pts_wrap_bits; /**< number of bits in pts (used for wrapping control) */
#endif
} AVStream; } AVStream;
#define AV_PROGRAM_RUNNING 1 #define AV_PROGRAM_RUNNING 1
...@@ -724,6 +785,7 @@ typedef struct AVFormatContext { ...@@ -724,6 +785,7 @@ typedef struct AVFormatContext {
#endif #endif
int ctx_flags; /**< Format-specific flags, see AVFMTCTX_xx */ int ctx_flags; /**< Format-specific flags, see AVFMTCTX_xx */
#if FF_API_REORDER_PRIVATE
/* private data for pts handling (do not modify directly). */ /* private data for pts handling (do not modify directly). */
/** /**
* This buffer is only needed when packets were already buffered but * This buffer is only needed when packets were already buffered but
...@@ -731,6 +793,7 @@ typedef struct AVFormatContext { ...@@ -731,6 +793,7 @@ typedef struct AVFormatContext {
* streams. * streams.
*/ */
struct AVPacketList *packet_buffer; struct AVPacketList *packet_buffer;
#endif
/** /**
* Decoding: position of the first frame of the component, in * Decoding: position of the first frame of the component, in
...@@ -761,11 +824,13 @@ typedef struct AVFormatContext { ...@@ -761,11 +824,13 @@ typedef struct AVFormatContext {
*/ */
int bit_rate; int bit_rate;
#if FF_API_REORDER_PRIVATE
/* av_read_frame() support */ /* av_read_frame() support */
AVStream *cur_st; AVStream *cur_st;
/* av_seek_frame() support */ /* av_seek_frame() support */
int64_t data_offset; /**< offset of the first packet */ int64_t data_offset; /**< offset of the first packet */
#endif
#if FF_API_MUXRATE #if FF_API_MUXRATE
/** /**
...@@ -876,6 +941,7 @@ typedef struct AVFormatContext { ...@@ -876,6 +941,7 @@ typedef struct AVFormatContext {
int debug; int debug;
#define FF_FDEBUG_TS 0x0001 #define FF_FDEBUG_TS 0x0001
#if FF_API_REORDER_PRIVATE
/** /**
* Raw packets from the demuxer, prior to parsing and decoding. * Raw packets from the demuxer, prior to parsing and decoding.
* This buffer is used for buffering packets until the codec can * This buffer is used for buffering packets until the codec can
...@@ -886,15 +952,18 @@ typedef struct AVFormatContext { ...@@ -886,15 +952,18 @@ typedef struct AVFormatContext {
struct AVPacketList *raw_packet_buffer_end; struct AVPacketList *raw_packet_buffer_end;
struct AVPacketList *packet_buffer_end; struct AVPacketList *packet_buffer_end;
#endif
AVDictionary *metadata; AVDictionary *metadata;
#if FF_API_REORDER_PRIVATE
/** /**
* Remaining size available for raw_packet_buffer, in bytes. * Remaining size available for raw_packet_buffer, in bytes.
* NOT PART OF PUBLIC API * NOT PART OF PUBLIC API
*/ */
#define RAW_PACKET_BUFFER_SIZE 2500000 #define RAW_PACKET_BUFFER_SIZE 2500000
int raw_packet_buffer_remaining_size; int raw_packet_buffer_remaining_size;
#endif
/** /**
* Start time of the stream in real world time, in microseconds * Start time of the stream in real world time, in microseconds
...@@ -923,6 +992,43 @@ typedef struct AVFormatContext { ...@@ -923,6 +992,43 @@ typedef struct AVFormatContext {
* This will be moved into demuxer private options. Thus no API/ABI compatibility * This will be moved into demuxer private options. Thus no API/ABI compatibility
*/ */
int ts_id; int ts_id;
/*****************************************************************
* All fields below this line are not part of the public API. They
* may not be used outside of libavformat and can be changed and
* removed at will.
* New public fields should be added right above.
*****************************************************************
*/
#if !FF_API_REORDER_PRIVATE
/**
* Raw packets from the demuxer, prior to parsing and decoding.
* This buffer is used for buffering packets until the codec can
* be identified, as parsing cannot be done without knowing the
* codec.
*/
struct AVPacketList *raw_packet_buffer;
struct AVPacketList *raw_packet_buffer_end;
/**
* Remaining size available for raw_packet_buffer, in bytes.
*/
#define RAW_PACKET_BUFFER_SIZE 2500000
int raw_packet_buffer_remaining_size;
/**
* This buffer is only needed when packets were already buffered but
* not decoded, for example to get the codec parameters in MPEG
* streams.
*/
struct AVPacketList *packet_buffer;
struct AVPacketList *packet_buffer_end;
/* av_read_frame() support */
AVStream *cur_st;
/* av_seek_frame() support */
int64_t data_offset; /**< offset of the first packet */
#endif
} AVFormatContext; } AVFormatContext;
typedef struct AVPacketList { typedef struct AVPacketList {
...@@ -1479,40 +1585,20 @@ int av_index_search_timestamp(AVStream *st, int64_t timestamp, int flags); ...@@ -1479,40 +1585,20 @@ int av_index_search_timestamp(AVStream *st, int64_t timestamp, int flags);
int av_add_index_entry(AVStream *st, int64_t pos, int64_t timestamp, int av_add_index_entry(AVStream *st, int64_t pos, int64_t timestamp,
int size, int distance, int flags); int size, int distance, int flags);
/** #if FF_API_SEEK_PUBLIC
* Perform a binary search using av_index_search_timestamp() and attribute_deprecated
* AVInputFormat.read_timestamp().
* This is not supposed to be called directly by a user application,
* but by demuxers.
* @param target_ts target timestamp in the time base of the given stream
* @param stream_index stream number
*/
int av_seek_frame_binary(AVFormatContext *s, int stream_index, int av_seek_frame_binary(AVFormatContext *s, int stream_index,
int64_t target_ts, int flags); int64_t target_ts, int flags);
attribute_deprecated
/**
* Update cur_dts of all streams based on the given timestamp and AVStream.
*
* Stream ref_st unchanged, others set cur_dts in their native time base.
* Only needed for timestamp wrapping or if (dts not set and pts!=dts).
* @param timestamp new dts expressed in time_base of param ref_st
* @param ref_st reference stream giving time_base of param timestamp
*/
void av_update_cur_dts(AVFormatContext *s, AVStream *ref_st, int64_t timestamp); void av_update_cur_dts(AVFormatContext *s, AVStream *ref_st, int64_t timestamp);
attribute_deprecated
/**
* Perform a binary search using read_timestamp().
* This is not supposed to be called directly by a user application,
* but by demuxers.
* @param target_ts target timestamp in the time base of the given stream
* @param stream_index stream number
*/
int64_t av_gen_search(AVFormatContext *s, int stream_index, int64_t av_gen_search(AVFormatContext *s, int stream_index,
int64_t target_ts, int64_t pos_min, int64_t target_ts, int64_t pos_min,
int64_t pos_max, int64_t pos_limit, int64_t pos_max, int64_t pos_limit,
int64_t ts_min, int64_t ts_max, int64_t ts_min, int64_t ts_max,
int flags, int64_t *ts_ret, int flags, int64_t *ts_ret,
int64_t (*read_timestamp)(struct AVFormatContext *, int , int64_t *, int64_t )); int64_t (*read_timestamp)(struct AVFormatContext *, int , int64_t *, int64_t ));
#endif
/** /**
* media file output * media file output
......
...@@ -251,4 +251,37 @@ enum CodecID ff_guess_image2_codec(const char *filename); ...@@ -251,4 +251,37 @@ enum CodecID ff_guess_image2_codec(const char *filename);
*/ */
int64_t ff_iso8601_to_unix_time(const char *datestr); int64_t ff_iso8601_to_unix_time(const char *datestr);
/**
* Perform a binary search using av_index_search_timestamp() and
* AVInputFormat.read_timestamp().
*
* @param target_ts target timestamp in the time base of the given stream
* @param stream_index stream number
*/
int ff_seek_frame_binary(AVFormatContext *s, int stream_index,
int64_t target_ts, int flags);
/**
* Update cur_dts of all streams based on the given timestamp and AVStream.
*
* Stream ref_st unchanged, others set cur_dts in their native time base.
* Only needed for timestamp wrapping or if (dts not set and pts!=dts).
* @param timestamp new dts expressed in time_base of param ref_st
* @param ref_st reference stream giving time_base of param timestamp
*/
void ff_update_cur_dts(AVFormatContext *s, AVStream *ref_st, int64_t timestamp);
/**
* Perform a binary search using read_timestamp().
*
* @param target_ts target timestamp in the time base of the given stream
* @param stream_index stream number
*/
int64_t ff_gen_search(AVFormatContext *s, int stream_index,
int64_t target_ts, int64_t pos_min,
int64_t pos_max, int64_t pos_limit,
int64_t ts_min, int64_t ts_max,
int flags, int64_t *ts_ret,
int64_t (*read_timestamp)(struct AVFormatContext *, int , int64_t *, int64_t ));
#endif /* AVFORMAT_INTERNAL_H */ #endif /* AVFORMAT_INTERNAL_H */
...@@ -61,6 +61,8 @@ const AVCodecTag ff_mp4_obj_type[] = { ...@@ -61,6 +61,8 @@ const AVCodecTag ff_mp4_obj_type[] = {
{ CODEC_ID_VORBIS , 0xDD }, /* non standard, gpac uses it */ { CODEC_ID_VORBIS , 0xDD }, /* non standard, gpac uses it */
{ CODEC_ID_DVD_SUBTITLE, 0xE0 }, /* non standard, see unsupported-embedded-subs-2.mp4 */ { CODEC_ID_DVD_SUBTITLE, 0xE0 }, /* non standard, see unsupported-embedded-subs-2.mp4 */
{ CODEC_ID_QCELP , 0xE1 }, { CODEC_ID_QCELP , 0xE1 },
{ CODEC_ID_MPEG4SYSTEMS, 0x01 },
{ CODEC_ID_MPEG4SYSTEMS, 0x02 },
{ CODEC_ID_NONE , 0 }, { CODEC_ID_NONE , 0 },
}; };
......
...@@ -150,10 +150,12 @@ int ff_mp4_read_descr(AVFormatContext *fc, AVIOContext *pb, int *tag); ...@@ -150,10 +150,12 @@ int ff_mp4_read_descr(AVFormatContext *fc, AVIOContext *pb, int *tag);
int ff_mp4_read_dec_config_descr(AVFormatContext *fc, AVStream *st, AVIOContext *pb); int ff_mp4_read_dec_config_descr(AVFormatContext *fc, AVStream *st, AVIOContext *pb);
void ff_mp4_parse_es_descr(AVIOContext *pb, int *es_id); void ff_mp4_parse_es_descr(AVIOContext *pb, int *es_id);
#define MP4ODescrTag 0x01
#define MP4IODescrTag 0x02 #define MP4IODescrTag 0x02
#define MP4ESDescrTag 0x03 #define MP4ESDescrTag 0x03
#define MP4DecConfigDescrTag 0x04 #define MP4DecConfigDescrTag 0x04
#define MP4DecSpecificDescrTag 0x05 #define MP4DecSpecificDescrTag 0x05
#define MP4SLDescrTag 0x06
int ff_mov_read_esds(AVFormatContext *fc, AVIOContext *pb, MOVAtom atom); int ff_mov_read_esds(AVFormatContext *fc, AVIOContext *pb, MOVAtom atom);
enum CodecID ff_mov_get_lpcm_codec_id(int bps, int flags); enum CodecID ff_mov_get_lpcm_codec_id(int bps, int flags);
......
...@@ -2071,7 +2071,7 @@ static int matroska_read_seek(AVFormatContext *s, int stream_index, ...@@ -2071,7 +2071,7 @@ static int matroska_read_seek(AVFormatContext *s, int stream_index,
matroska->skip_to_keyframe = !(flags & AVSEEK_FLAG_ANY); matroska->skip_to_keyframe = !(flags & AVSEEK_FLAG_ANY);
matroska->skip_to_timecode = st->index_entries[index].timestamp; matroska->skip_to_timecode = st->index_entries[index].timestamp;
matroska->done = 0; matroska->done = 0;
av_update_cur_dts(s, st, st->index_entries[index].timestamp); ff_update_cur_dts(s, st, st->index_entries[index].timestamp);
return 0; return 0;
} }
......
This diff is collapsed.
...@@ -39,6 +39,7 @@ ...@@ -39,6 +39,7 @@
/* table ids */ /* table ids */
#define PAT_TID 0x00 #define PAT_TID 0x00
#define PMT_TID 0x02 #define PMT_TID 0x02
#define M4OD_TID 0x05
#define SDT_TID 0x42 #define SDT_TID 0x42
#define STREAM_TYPE_VIDEO_MPEG1 0x01 #define STREAM_TYPE_VIDEO_MPEG1 0x01
...@@ -64,6 +65,30 @@ int ff_mpegts_parse_packet(MpegTSContext *ts, AVPacket *pkt, ...@@ -64,6 +65,30 @@ int ff_mpegts_parse_packet(MpegTSContext *ts, AVPacket *pkt,
const uint8_t *buf, int len); const uint8_t *buf, int len);
void ff_mpegts_parse_close(MpegTSContext *ts); void ff_mpegts_parse_close(MpegTSContext *ts);
typedef struct {
int use_au_start;
int use_au_end;
int use_rand_acc_pt;
int use_padding;
int use_timestamps;
int use_idle;
int timestamp_res;
int timestamp_len;
int ocr_len;
int au_len;
int inst_bitrate_len;
int degr_prior_len;
int au_seq_num_len;
int packet_seq_num_len;
} SLConfigDescr;
typedef struct {
int es_id;
int dec_config_descr_len;
uint8_t *dec_config_descr;
SLConfigDescr sl;
} Mp4Descr;
/** /**
* Parse an MPEG-2 descriptor * Parse an MPEG-2 descriptor
* @param[in] fc Format context (used for logging only) * @param[in] fc Format context (used for logging only)
...@@ -79,7 +104,7 @@ void ff_mpegts_parse_close(MpegTSContext *ts); ...@@ -79,7 +104,7 @@ void ff_mpegts_parse_close(MpegTSContext *ts);
*/ */
int ff_parse_mpeg2_descriptor(AVFormatContext *fc, AVStream *st, int stream_type, int ff_parse_mpeg2_descriptor(AVFormatContext *fc, AVStream *st, int stream_type,
const uint8_t **pp, const uint8_t *desc_list_end, const uint8_t **pp, const uint8_t *desc_list_end,
int mp4_dec_config_descr_len, int mp4_es_id, int pid, Mp4Descr *mp4_descr, int mp4_descr_count, int pid,
uint8_t *mp4_dec_config_descr); MpegTSContext *ts);
#endif /* AVFORMAT_MPEGTS_H */ #endif /* AVFORMAT_MPEGTS_H */
...@@ -49,6 +49,7 @@ ...@@ -49,6 +49,7 @@
#include "libavutil/mathematics.h" #include "libavutil/mathematics.h"
#include "libavcodec/bytestream.h" #include "libavcodec/bytestream.h"
#include "avformat.h" #include "avformat.h"
#include "internal.h"
#include "mxf.h" #include "mxf.h"
typedef enum { typedef enum {
...@@ -1141,7 +1142,7 @@ static int mxf_read_seek(AVFormatContext *s, int stream_index, int64_t sample_ti ...@@ -1141,7 +1142,7 @@ static int mxf_read_seek(AVFormatContext *s, int stream_index, int64_t sample_ti
seconds = av_rescale(sample_time, st->time_base.num, st->time_base.den); seconds = av_rescale(sample_time, st->time_base.num, st->time_base.den);
if (avio_seek(s->pb, (s->bit_rate * seconds) >> 3, SEEK_SET) < 0) if (avio_seek(s->pb, (s->bit_rate * seconds) >> 3, SEEK_SET) < 0)
return -1; return -1;
av_update_cur_dts(s, st, sample_time); ff_update_cur_dts(s, st, sample_time);
return 0; return 0;
} }
......
...@@ -874,7 +874,7 @@ static int read_seek(AVFormatContext *s, int stream_index, int64_t pts, int flag ...@@ -874,7 +874,7 @@ static int read_seek(AVFormatContext *s, int stream_index, int64_t pts, int flag
(void **) next_node); (void **) next_node);
av_log(s, AV_LOG_DEBUG, "%"PRIu64"-%"PRIu64" %"PRId64"-%"PRId64"\n", next_node[0]->pos, next_node[1]->pos, av_log(s, AV_LOG_DEBUG, "%"PRIu64"-%"PRIu64" %"PRId64"-%"PRId64"\n", next_node[0]->pos, next_node[1]->pos,
next_node[0]->ts , next_node[1]->ts); next_node[0]->ts , next_node[1]->ts);
pos= av_gen_search(s, -1, dummy.ts, next_node[0]->pos, next_node[1]->pos, next_node[1]->pos, pos = ff_gen_search(s, -1, dummy.ts, next_node[0]->pos, next_node[1]->pos, next_node[1]->pos,
next_node[0]->ts , next_node[1]->ts, AVSEEK_FLAG_BACKWARD, &ts, nut_read_timestamp); next_node[0]->ts , next_node[1]->ts, AVSEEK_FLAG_BACKWARD, &ts, nut_read_timestamp);
if(!(flags & AVSEEK_FLAG_BACKWARD)){ if(!(flags & AVSEEK_FLAG_BACKWARD)){
...@@ -882,7 +882,7 @@ static int read_seek(AVFormatContext *s, int stream_index, int64_t pts, int flag ...@@ -882,7 +882,7 @@ static int read_seek(AVFormatContext *s, int stream_index, int64_t pts, int flag
next_node[1]= &nopts_sp; next_node[1]= &nopts_sp;
av_tree_find(nut->syncpoints, &dummy, (void *) ff_nut_sp_pos_cmp, av_tree_find(nut->syncpoints, &dummy, (void *) ff_nut_sp_pos_cmp,
(void **) next_node); (void **) next_node);
pos2= av_gen_search(s, -2, dummy.pos, next_node[0]->pos , next_node[1]->pos, next_node[1]->pos, pos2 = ff_gen_search(s, -2, dummy.pos, next_node[0]->pos , next_node[1]->pos, next_node[1]->pos,
next_node[0]->back_ptr, next_node[1]->back_ptr, flags, &ts, nut_read_timestamp); next_node[0]->back_ptr, next_node[1]->back_ptr, flags, &ts, nut_read_timestamp);
if(pos2>=0) if(pos2>=0)
pos= pos2; pos= pos2;
......
...@@ -32,6 +32,7 @@ ...@@ -32,6 +32,7 @@
#include <stdio.h> #include <stdio.h>
#include "oggdec.h" #include "oggdec.h"
#include "avformat.h" #include "avformat.h"
#include "internal.h"
#include "vorbiscomment.h" #include "vorbiscomment.h"
#define MAX_PAGE_SIZE 65307 #define MAX_PAGE_SIZE 65307
...@@ -661,7 +662,7 @@ static int ogg_read_seek(AVFormatContext *s, int stream_index, ...@@ -661,7 +662,7 @@ static int ogg_read_seek(AVFormatContext *s, int stream_index,
&& !(flags & AVSEEK_FLAG_ANY)) && !(flags & AVSEEK_FLAG_ANY))
os->keyframe_seek = 1; os->keyframe_seek = 1;
ret = av_seek_frame_binary(s, stream_index, timestamp, flags); ret = ff_seek_frame_binary(s, stream_index, timestamp, flags);
os = ogg->streams + stream_index; os = ogg->streams + stream_index;
if (ret < 0) if (ret < 0)
os->keyframe_seek = 0; os->keyframe_seek = 0;
......
...@@ -64,10 +64,10 @@ int main(int argc, char **argv) ...@@ -64,10 +64,10 @@ int main(int argc, char **argv)
AVFormatContext *ic = NULL; AVFormatContext *ic = NULL;
int i, ret, stream_id; int i, ret, stream_id;
int64_t timestamp; int64_t timestamp;
AVFormatParameters params, *ap= &params; AVDictionary *format_opts = NULL;
memset(ap, 0, sizeof(params));
ap->channels=1; av_dict_set(&format_opts, "channels", "1", 0);
ap->sample_rate= 22050; av_dict_set(&format_opts, "sample_rate", "22050", 0);
/* initialize libavcodec, and register all codecs and formats */ /* initialize libavcodec, and register all codecs and formats */
av_register_all(); av_register_all();
...@@ -80,7 +80,8 @@ int main(int argc, char **argv) ...@@ -80,7 +80,8 @@ int main(int argc, char **argv)
filename = argv[1]; filename = argv[1];
ret = av_open_input_file(&ic, filename, NULL, 0, ap); ret = avformat_open_input(&ic, filename, NULL, &format_opts);
av_dict_free(&format_opts);
if (ret < 0) { if (ret < 0) {
fprintf(stderr, "cannot open %s\n", filename); fprintf(stderr, "cannot open %s\n", filename);
exit(1); exit(1);
......
...@@ -340,6 +340,7 @@ AVInputFormat *av_probe_input_format3(AVProbeData *pd, int is_opened, int *score ...@@ -340,6 +340,7 @@ AVInputFormat *av_probe_input_format3(AVProbeData *pd, int is_opened, int *score
fmt = NULL; fmt = NULL;
} }
*score_ret= score_max; *score_ret= score_max;
return fmt; return fmt;
} }
...@@ -1427,7 +1428,15 @@ void ff_read_frame_flush(AVFormatContext *s) ...@@ -1427,7 +1428,15 @@ void ff_read_frame_flush(AVFormatContext *s)
} }
} }
void av_update_cur_dts(AVFormatContext *s, AVStream *ref_st, int64_t timestamp){ #if FF_API_SEEK_PUBLIC
void av_update_cur_dts(AVFormatContext *s, AVStream *ref_st, int64_t timestamp)
{
return ff_update_cur_dts(s, ref_st, timestamp);
}
#endif
void ff_update_cur_dts(AVFormatContext *s, AVStream *ref_st, int64_t timestamp)
{
int i; int i;
for(i = 0; i < s->nb_streams; i++) { for(i = 0; i < s->nb_streams; i++) {
...@@ -1547,7 +1556,14 @@ int av_index_search_timestamp(AVStream *st, int64_t wanted_timestamp, ...@@ -1547,7 +1556,14 @@ int av_index_search_timestamp(AVStream *st, int64_t wanted_timestamp,
wanted_timestamp, flags); wanted_timestamp, flags);
} }
#if FF_API_SEEK_PUBLIC
int av_seek_frame_binary(AVFormatContext *s, int stream_index, int64_t target_ts, int flags){ int av_seek_frame_binary(AVFormatContext *s, int stream_index, int64_t target_ts, int flags){
return ff_seek_frame_binary(s, stream_index, target_ts, flags);
}
#endif
int ff_seek_frame_binary(AVFormatContext *s, int stream_index, int64_t target_ts, int flags)
{
AVInputFormat *avif= s->iformat; AVInputFormat *avif= s->iformat;
int64_t av_uninit(pos_min), av_uninit(pos_max), pos, pos_limit; int64_t av_uninit(pos_min), av_uninit(pos_max), pos, pos_limit;
int64_t ts_min, ts_max, ts; int64_t ts_min, ts_max, ts;
...@@ -1594,7 +1610,7 @@ int av_seek_frame_binary(AVFormatContext *s, int stream_index, int64_t target_ts ...@@ -1594,7 +1610,7 @@ int av_seek_frame_binary(AVFormatContext *s, int stream_index, int64_t target_ts
} }
} }
pos= av_gen_search(s, stream_index, target_ts, pos_min, pos_max, pos_limit, ts_min, ts_max, flags, &ts, avif->read_timestamp); pos= ff_gen_search(s, stream_index, target_ts, pos_min, pos_max, pos_limit, ts_min, ts_max, flags, &ts, avif->read_timestamp);
if(pos<0) if(pos<0)
return -1; return -1;
...@@ -1603,12 +1619,28 @@ int av_seek_frame_binary(AVFormatContext *s, int stream_index, int64_t target_ts ...@@ -1603,12 +1619,28 @@ int av_seek_frame_binary(AVFormatContext *s, int stream_index, int64_t target_ts
return ret; return ret;
ff_read_frame_flush(s); ff_read_frame_flush(s);
av_update_cur_dts(s, st, ts); ff_update_cur_dts(s, st, ts);
return 0; return 0;
} }
int64_t av_gen_search(AVFormatContext *s, int stream_index, int64_t target_ts, int64_t pos_min, int64_t pos_max, int64_t pos_limit, int64_t ts_min, int64_t ts_max, int flags, int64_t *ts_ret, int64_t (*read_timestamp)(struct AVFormatContext *, int , int64_t *, int64_t )){ #if FF_API_SEEK_PUBLIC
int64_t av_gen_search(AVFormatContext *s, int stream_index, int64_t target_ts,
int64_t pos_min, int64_t pos_max, int64_t pos_limit,
int64_t ts_min, int64_t ts_max, int flags, int64_t *ts_ret,
int64_t (*read_timestamp)(struct AVFormatContext *, int , int64_t *, int64_t ))
{
return ff_gen_search(s, stream_index, target_ts, pos_min, pos_max,
pos_limit, ts_min, ts_max, flags, ts_ret,
read_timestamp);
}
#endif
int64_t ff_gen_search(AVFormatContext *s, int stream_index, int64_t target_ts,
int64_t pos_min, int64_t pos_max, int64_t pos_limit,
int64_t ts_min, int64_t ts_max, int flags, int64_t *ts_ret,
int64_t (*read_timestamp)(struct AVFormatContext *, int , int64_t *, int64_t ))
{
int64_t pos, ts; int64_t pos, ts;
int64_t start_pos, filesize; int64_t start_pos, filesize;
int no_change; int no_change;
...@@ -1775,7 +1807,7 @@ static int seek_frame_generic(AVFormatContext *s, ...@@ -1775,7 +1807,7 @@ static int seek_frame_generic(AVFormatContext *s,
ie= &st->index_entries[st->nb_index_entries-1]; ie= &st->index_entries[st->nb_index_entries-1];
if ((ret = avio_seek(s->pb, ie->pos, SEEK_SET)) < 0) if ((ret = avio_seek(s->pb, ie->pos, SEEK_SET)) < 0)
return ret; return ret;
av_update_cur_dts(s, st, ie->timestamp); ff_update_cur_dts(s, st, ie->timestamp);
}else{ }else{
if ((ret = avio_seek(s->pb, s->data_offset, SEEK_SET)) < 0) if ((ret = avio_seek(s->pb, s->data_offset, SEEK_SET)) < 0)
return ret; return ret;
...@@ -1812,7 +1844,7 @@ static int seek_frame_generic(AVFormatContext *s, ...@@ -1812,7 +1844,7 @@ static int seek_frame_generic(AVFormatContext *s,
ie = &st->index_entries[index]; ie = &st->index_entries[index];
if ((ret = avio_seek(s->pb, ie->pos, SEEK_SET)) < 0) if ((ret = avio_seek(s->pb, ie->pos, SEEK_SET)) < 0)
return ret; return ret;
av_update_cur_dts(s, st, ie->timestamp); ff_update_cur_dts(s, st, ie->timestamp);
return 0; return 0;
} }
...@@ -1853,7 +1885,7 @@ int av_seek_frame(AVFormatContext *s, int stream_index, int64_t timestamp, int f ...@@ -1853,7 +1885,7 @@ int av_seek_frame(AVFormatContext *s, int stream_index, int64_t timestamp, int f
if (s->iformat->read_timestamp && !(s->iformat->flags & AVFMT_NOBINSEARCH)) { if (s->iformat->read_timestamp && !(s->iformat->flags & AVFMT_NOBINSEARCH)) {
ff_read_frame_flush(s); ff_read_frame_flush(s);
return av_seek_frame_binary(s, stream_index, timestamp, flags); return ff_seek_frame_binary(s, stream_index, timestamp, flags);
} else if (!(s->iformat->flags & AVFMT_NOGENSEARCH)) { } else if (!(s->iformat->flags & AVFMT_NOGENSEARCH)) {
ff_read_frame_flush(s); ff_read_frame_flush(s);
return seek_frame_generic(s, stream_index, timestamp, flags); return seek_frame_generic(s, stream_index, timestamp, flags);
......
...@@ -107,5 +107,11 @@ ...@@ -107,5 +107,11 @@
#ifndef FF_API_STREAM_COPY #ifndef FF_API_STREAM_COPY
#define FF_API_STREAM_COPY (LIBAVFORMAT_VERSION_MAJOR < 54) #define FF_API_STREAM_COPY (LIBAVFORMAT_VERSION_MAJOR < 54)
#endif #endif
#ifndef FF_API_SEEK_PUBLIC
#define FF_API_SEEK_PUBLIC (LIBAVFORMAT_VERSION_MAJOR < 54)
#endif
#ifndef FF_API_REORDER_PRIVATE
#define FF_API_REORDER_PRIVATE (LIBAVFORMAT_VERSION_MAJOR < 54)
#endif
#endif /* AVFORMAT_VERSION_H */ #endif /* AVFORMAT_VERSION_H */
...@@ -773,7 +773,7 @@ static int parse_chunks(AVFormatContext *s, int mode, int64_t seekts, int *len_p ...@@ -773,7 +773,7 @@ static int parse_chunks(AVFormatContext *s, int mode, int64_t seekts, int *len_p
buf_size = FFMIN(len - consumed, sizeof(buf)); buf_size = FFMIN(len - consumed, sizeof(buf));
avio_read(pb, buf, buf_size); avio_read(pb, buf, buf_size);
consumed += buf_size; consumed += buf_size;
ff_parse_mpeg2_descriptor(s, st, 0, &pbuf, buf + buf_size, 0, 0, 0, 0); ff_parse_mpeg2_descriptor(s, st, 0, &pbuf, buf + buf_size, NULL, 0, 0, NULL);
} }
} else if (!ff_guidcmp(g, EVENTID_AudioTypeSpanningEvent)) { } else if (!ff_guidcmp(g, EVENTID_AudioTypeSpanningEvent)) {
int stream_index = ff_find_stream_index(s, sid); int stream_index = ff_find_stream_index(s, sid);
......
...@@ -44,7 +44,7 @@ int main(int argc, char **argv) ...@@ -44,7 +44,7 @@ int main(int argc, char **argv)
{ {
char fntemplate[PATH_MAX]; char fntemplate[PATH_MAX];
char pktfilename[PATH_MAX]; char pktfilename[PATH_MAX];
AVFormatContext *fctx; AVFormatContext *fctx = NULL;
AVPacket pkt; AVPacket pkt;
int64_t pktnum = 0; int64_t pktnum = 0;
int64_t maxpkts = 0; int64_t maxpkts = 0;
...@@ -83,9 +83,9 @@ int main(int argc, char **argv) ...@@ -83,9 +83,9 @@ int main(int argc, char **argv)
// register all file formats // register all file formats
av_register_all(); av_register_all();
err = av_open_input_file(&fctx, argv[1], NULL, 0, NULL); err = avformat_open_input(&fctx, argv[1], NULL, NULL);
if (err < 0) { if (err < 0) {
fprintf(stderr, "av_open_input_file: error %d\n", err); fprintf(stderr, "cannot open input: error %d\n", err);
return 1; return 1;
} }
......
Markdown is supported
0% or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment