Go to the documentation of this file.
54 if (!
s->last_frame || !
s->last2_frame)
61 const uint8_t *buf,
const uint8_t *buf_end)
63 unsigned char *dst =
frame->data[0];
66 for (
i=0;
i <
s->avctx->height && buf_end - buf >=
s->avctx->width;
i++) {
67 memcpy(dst, buf,
s->avctx->width);
68 dst +=
frame->linesize[0];
69 buf +=
s->avctx->width;
73 static void cmv_motcomp(
unsigned char *dst, ptrdiff_t dst_stride,
74 const unsigned char *
src, ptrdiff_t src_stride,
76 int xoffset,
int yoffset,
83 if (
i+xoffset>=0 &&
i+xoffset<
width &&
84 j+yoffset>=0 && j+yoffset<
height) {
85 dst[j*dst_stride +
i] =
src[(j+yoffset)*src_stride +
i+xoffset];
87 dst[j*dst_stride +
i] = 0;
93 const uint8_t *buf_end)
95 const uint8_t *raw = buf + (
s->avctx->width*
s->avctx->height/16);
99 for(y=0; y<
s->avctx->height/4; y++)
100 for(x=0; x<
s->avctx->width/4 && buf_end - buf >
i; x++) {
102 unsigned char *dst =
frame->data[0] + (y*4)*
frame->linesize[0] + x*4;
103 if (raw+16<buf_end && *raw==0xFF) {
106 memcpy(dst +
frame->linesize[0], raw+4, 4);
107 memcpy(dst + 2 *
frame->linesize[0], raw+8, 4);
108 memcpy(dst + 3 *
frame->linesize[0], raw+12, 4);
110 }
else if(raw<buf_end) {
111 int xoffset = (*raw & 0xF) - 7;
112 int yoffset = ((*raw >> 4)) - 7;
113 if (
s->last2_frame->data[0])
115 s->last2_frame->data[0],
s->last2_frame->linesize[0],
116 x*4, y*4, xoffset, yoffset,
s->avctx->width,
s->avctx->height);
120 int xoffset = (buf[
i] & 0xF) - 7;
121 int yoffset = ((buf[
i] >> 4)) - 7;
122 if (
s->last_frame->data[0])
124 s->last_frame->data[0],
s->last_frame->linesize[0],
125 x*4, y*4, xoffset, yoffset,
s->avctx->width,
s->avctx->height);
133 int pal_start, pal_count,
i,
ret, fps;
135 if(buf_end - buf < 16) {
143 if (
s->width !=
s->avctx->width ||
144 s->height !=
s->avctx->height) {
169 #define EA_PREAMBLE_SIZE 8
170 #define MVIh_TAG MKTAG('M', 'V', 'I', 'h')
175 const uint8_t *buf = avpkt->
data;
176 int buf_size = avpkt->
size;
178 const uint8_t *buf_end = buf + buf_size;
205 frame->key_frame = 0;
208 frame->key_frame = 1;
#define AV_LOG_WARNING
Something somehow does not look correct.
#define FF_CODEC_CAP_INIT_CLEANUP
The codec allows calling the close function for deallocation even if the init function returned a fai...
Filter the word “frame” indicates either a video frame or a group of audio as stored in an AVFrame structure Format for each input and each output the list of supported formats For video that means pixel format For audio that means channel sample they are references to shared objects When the negotiation mechanism computes the intersection of the formats supported at each end of a all references to both lists are replaced with a reference to the intersection And when a single format is eventually chosen for a link amongst the remaining all references to the list are updated That means that if a filter requires that its input and output have the same format amongst a supported all it has to do is use a reference to the same list of formats query_formats can leave some formats unset and return AVERROR(EAGAIN) to cause the negotiation mechanism toagain later. That can be used by filters with complex requirements to use the format negotiated on one link to set the formats supported on another. Frame references ownership and permissions
static av_cold int cmv_decode_init(AVCodecContext *avctx)
void av_frame_free(AVFrame **frame)
Free the frame and any dynamically allocated objects in it, e.g.
This structure describes decoded (raw) audio or video data.
static av_cold int cmv_decode_end(AVCodecContext *avctx)
int ff_set_dimensions(AVCodecContext *s, int width, int height)
Check that the provided frame dimensions are valid and set them on the codec context.
static int cmv_decode_frame(AVCodecContext *avctx, AVFrame *frame, int *got_frame, AVPacket *avpkt)
AVCodec p
The public AVCodec.
AVFrame * av_frame_alloc(void)
Allocate an AVFrame and set its fields to default values.
#define FF_CODEC_DECODE_CB(func)
#define AV_GET_BUFFER_FLAG_REF
The decoder will keep a reference to the frame and may reuse it later.
int(* init)(AVBSFContext *ctx)
uint64_t_TMPL AV_WL64 unsigned int_TMPL AV_WL32 unsigned int_TMPL AV_WL24 unsigned int_TMPL AV_RL16
const FFCodec ff_eacmv_decoder
#define CODEC_LONG_NAME(str)
Rational number (pair of numerator and denominator).
@ AV_PICTURE_TYPE_I
Intra.
static int cmv_process_header(CmvContext *s, const uint8_t *buf, const uint8_t *buf_end)
int ff_get_buffer(AVCodecContext *avctx, AVFrame *frame, int flags)
Get a buffer for a frame.
#define AV_CODEC_CAP_DR1
Codec uses get_buffer() or get_encode_buffer() for allocating buffers and supports custom allocators.
int av_frame_ref(AVFrame *dst, const AVFrame *src)
Set up a new reference to the data described by the source frame.
unsigned int palette[AVPALETTE_COUNT]
uint64_t_TMPL AV_WL64 unsigned int_TMPL AV_WL32 unsigned int_TMPL AV_WL24 unsigned int_TMPL AV_WL16 uint64_t_TMPL AV_WB64 unsigned int_TMPL AV_RB32
static void cmv_decode_inter(CmvContext *s, AVFrame *frame, const uint8_t *buf, const uint8_t *buf_end)
static void cmv_motcomp(unsigned char *dst, ptrdiff_t dst_stride, const unsigned char *src, ptrdiff_t src_stride, int x, int y, int xoffset, int yoffset, int width, int height)
#define i(width, name, range_min, range_max)
void av_frame_move_ref(AVFrame *dst, AVFrame *src)
Move everything contained in src to dst and reset src.
void av_frame_unref(AVFrame *frame)
Unreference all the buffers referenced by frame and reset the frame fields.
const char * name
Name of the codec implementation.
enum AVPixelFormat pix_fmt
Pixel format, see AV_PIX_FMT_xxx.
static void cmv_decode_intra(CmvContext *s, AVFrame *frame, const uint8_t *buf, const uint8_t *buf_end)
@ AV_PIX_FMT_PAL8
8 bits with AV_PIX_FMT_RGB32 palette
these buffered frames must be flushed immediately if a new input produces new the filter must not call request_frame to get more It must just process the frame or queue it The task of requesting more frames is left to the filter s request_frame method or the application If a filter has several the filter must be ready for frames arriving randomly on any input any filter with several inputs will most likely require some kind of queuing mechanism It is perfectly acceptable to have a limited queue and to drop frames when the inputs are too unbalanced request_frame For filters that do not use the this method is called when a frame is wanted on an output For a it should directly call filter_frame on the corresponding output For a if there are queued frames already one of these frames should be pushed If the filter should request a frame on one of its repeatedly until at least one frame has been pushed Return or at least make progress towards producing a frame
uint64_t_TMPL AV_WL64 unsigned int_TMPL AV_RL32
main external API structure.
@ AV_PICTURE_TYPE_P
Predicted.
This structure stores compressed data.
AVFrame * last2_frame
second-last
#define AVERROR_INVALIDDATA
Invalid data found when processing input.
int av_image_check_size(unsigned int w, unsigned int h, int log_offset, void *log_ctx)
Check if the given dimension of an image is valid, meaning that all bytes of the image can be address...
uint64_t_TMPL AV_WL64 unsigned int_TMPL AV_WL32 unsigned int_TMPL AV_WL24 unsigned int_TMPL AV_WL16 uint64_t_TMPL AV_WB64 unsigned int_TMPL AV_WB32 unsigned int_TMPL AV_RB24