Go to the documentation of this file.
50 const uint8_t *buf = avpkt->
data;
53 int min_stride = 2 * avctx->
width;
59 min_stride = (avctx->
width + 15) / 16 * 40;
75 for (y = 0; y < avctx->
height; y++) {
84 uint16_t *luma = (uint16_t*)&
frame->data[0][y*
frame->linesize[0]];
85 uint16_t *
cb = (uint16_t*)&
frame->data[1][y*
frame->linesize[1]];
86 uint16_t *
cr = (uint16_t*)&
frame->data[2][y*
frame->linesize[2]];
88 const uint8_t *buf_src = buf + src_y*
stride + 40*
block;
89 for (x = 0; x < 16 && x + 16*
block < avctx->
width; x++) {
90 int xd = x + 16*
block;
92 luma [xd] = (4*buf_src[2*x + 0]) + ((buf_src[32 + (x>>1)]>>4)&3);
94 luma [xd] = (4*buf_src[2*x + 0]) + (buf_src[32 + (x>>1)] &3);
95 cb[xd>>1] = (4*buf_src[2*x + 1]) + ((buf_src[32 + (x>>1)]>>2)&3);
96 cr[xd>>1] = (4*buf_src[2*x + 3]) + (buf_src[32 + (x>>1)]>>6);
static av_cold int m101_decode_init(AVCodecContext *avctx)
static double cb(void *priv, double x, double y)
This structure describes decoded (raw) audio or video data.
#define AV_FRAME_FLAG_TOP_FIELD_FIRST
A flag to mark frames where the top field is displayed first if the content is interlaced.
AVCodec p
The public AVCodec.
#define AV_LOG_ERROR
Something went wrong and cannot losslessly be recovered.
#define FF_CODEC_DECODE_CB(func)
#define CODEC_LONG_NAME(str)
@ AV_PIX_FMT_YUYV422
packed YUV 4:2:2, 16bpp, Y0 Cb Y1 Cr
#define AV_PIX_FMT_YUV422P10
int ff_get_buffer(AVCodecContext *avctx, AVFrame *frame, int flags)
Get a buffer for a frame.
#define AV_CODEC_CAP_DR1
Codec uses get_buffer() or get_encode_buffer() for allocating buffers and supports custom allocators.
const FFCodec ff_m101_decoder
static int m101_decode_frame(AVCodecContext *avctx, AVFrame *frame, int *got_frame, AVPacket *avpkt)
uint8_t * extradata
some codecs need / can use extradata like Huffman tables.
const char * name
Name of the codec implementation.
enum AVPixelFormat pix_fmt
Pixel format, see AV_PIX_FMT_xxx.
#define AV_FRAME_FLAG_INTERLACED
A flag to mark frames whose content is interlaced.
these buffered frames must be flushed immediately if a new input produces new the filter must not call request_frame to get more It must just process the frame or queue it The task of requesting more frames is left to the filter s request_frame method or the application If a filter has several the filter must be ready for frames arriving randomly on any input any filter with several inputs will most likely require some kind of queuing mechanism It is perfectly acceptable to have a limited queue and to drop frames when the inputs are too unbalanced request_frame For filters that do not use the this method is called when a frame is wanted on an output For a it should directly call filter_frame on the corresponding output For a if there are queued frames already one of these frames should be pushed If the filter should request a frame on one of its repeatedly until at least one frame has been pushed Return or at least make progress towards producing a frame
uint64_t_TMPL AV_WL64 unsigned int_TMPL AV_RL32
main external API structure.
#define avpriv_request_sample(...)
This structure stores compressed data.
static double cr(void *priv, double x, double y)
int width
picture width / height.
The exact code depends on how similar the blocks are and how related they are to the block
#define AVERROR_INVALIDDATA
Invalid data found when processing input.