Go to the documentation of this file.
24 #include <linux/videodev2.h>
25 #include <sys/ioctl.h>
46 return V4L2_TYPE_IS_OUTPUT(
ctx->type) ?
58 return V4L2_TYPE_IS_MULTIPLANAR(
fmt->type) ?
fmt->fmt.pix_mp.width :
fmt->fmt.pix.width;
63 return V4L2_TYPE_IS_MULTIPLANAR(
fmt->type) ?
fmt->fmt.pix_mp.height :
fmt->fmt.pix.height;
68 struct v4l2_format *fmt1 = &
ctx->format;
69 int ret = V4L2_TYPE_IS_MULTIPLANAR(
ctx->type) ?
70 fmt1->fmt.pix_mp.width != fmt2->fmt.pix_mp.width ||
71 fmt1->fmt.pix_mp.height != fmt2->fmt.pix_mp.height
73 fmt1->fmt.pix.width != fmt2->fmt.pix.width ||
74 fmt1->fmt.pix.height != fmt2->fmt.pix.height;
87 return ctx->type == V4L2_BUF_TYPE_VIDEO_CAPTURE_MPLANE ||
88 ctx->type == V4L2_BUF_TYPE_VIDEO_OUTPUT_MPLANE ||
89 ctx->type == V4L2_BUF_TYPE_VIDEO_CAPTURE ||
90 ctx->type == V4L2_BUF_TYPE_VIDEO_OUTPUT;
96 const int SZ_4K = 0x1000;
109 ctx->format.type =
ctx->type;
111 if (
fmt->update_avfmt)
112 ctx->av_pix_fmt =
fmt->av_fmt;
114 if (V4L2_TYPE_IS_MULTIPLANAR(
ctx->type)) {
116 ctx->format.fmt.pix_mp.height =
ctx->height;
117 ctx->format.fmt.pix_mp.width =
ctx->width;
118 if (
fmt->update_v4l2) {
119 ctx->format.fmt.pix_mp.pixelformat =
fmt->v4l2_fmt;
122 ctx->format.fmt.pix_mp.plane_fmt[0].sizeimage =
126 ctx->format.fmt.pix.height =
ctx->height;
127 ctx->format.fmt.pix.width =
ctx->width;
128 if (
fmt->update_v4l2) {
129 ctx->format.fmt.pix.pixelformat =
fmt->v4l2_fmt;
132 ctx->format.fmt.pix.sizeimage =
145 struct v4l2_format cap_fmt =
s->capture.format;
146 struct v4l2_format out_fmt =
s->output.format;
147 struct v4l2_event evt = { 0 };
150 ret = ioctl(
s->fd, VIDIOC_DQEVENT, &evt);
156 if (evt.type != V4L2_EVENT_SOURCE_CHANGE)
159 ret = ioctl(
s->fd, VIDIOC_G_FMT, &out_fmt);
165 ret = ioctl(
s->fd, VIDIOC_G_FMT, &cap_fmt);
183 if (full_reinit ||
reinit)
218 struct v4l2_decoder_cmd cmd = {
219 .cmd = V4L2_DEC_CMD_STOP,
236 struct v4l2_encoder_cmd cmd = {
237 .cmd = V4L2_ENC_CMD_STOP,
254 struct v4l2_plane
planes[VIDEO_MAX_PLANES];
255 struct v4l2_buffer
buf = { 0 };
257 struct pollfd pfd = {
258 .events = POLLIN | POLLRDNORM | POLLPRI | POLLOUT | POLLWRNORM,
265 for (
i = 0;
i <
ctx->num_buffers;
i++) {
274 if (V4L2_TYPE_IS_OUTPUT(
ctx->type))
275 pfd.events = POLLOUT | POLLWRNORM;
279 pfd.events = POLLIN | POLLRDNORM | POLLPRI;
283 ret = poll(&pfd, 1, timeout);
292 if (pfd.revents & POLLERR) {
296 for (
i = 0;
i <
ctx->num_buffers;
i++) {
308 if (pfd.revents & POLLPRI) {
324 if (pfd.revents & (POLLIN | POLLRDNORM | POLLOUT | POLLWRNORM)) {
326 if (!V4L2_TYPE_IS_OUTPUT(
ctx->type)) {
328 if (pfd.revents & (POLLIN | POLLRDNORM))
334 if (pfd.revents & (POLLOUT | POLLWRNORM))
339 memset(&
buf, 0,
sizeof(
buf));
340 buf.memory = V4L2_MEMORY_MMAP;
342 if (V4L2_TYPE_IS_MULTIPLANAR(
ctx->type)) {
344 buf.length = VIDEO_MAX_PLANES;
350 if (errno != EAGAIN) {
359 avbuf = &
ctx->buffers[
buf.index];
362 if (V4L2_TYPE_IS_MULTIPLANAR(
ctx->type)) {
378 if (V4L2_TYPE_IS_OUTPUT(
ctx->type)) {
383 for (
i = 0;
i <
ctx->num_buffers;
i++) {
385 return &
ctx->buffers[
i];
393 struct v4l2_requestbuffers req = {
394 .memory = V4L2_MEMORY_MMAP,
400 for (
i = 0;
i <
ctx->num_buffers;
i++) {
403 for (j = 0; j <
buffer->num_planes; j++) {
404 struct V4L2Plane_info *p = &
buffer->plane_info[j];
405 if (p->mm_addr && p->length)
406 if (munmap(p->mm_addr, p->length) < 0)
416 struct v4l2_format *
fmt = &
ctx->format;
424 if (V4L2_TYPE_IS_MULTIPLANAR(
ctx->type))
425 fmt->fmt.pix_mp.pixelformat = v4l2_fmt;
427 fmt->fmt.pix.pixelformat = v4l2_fmt;
441 struct v4l2_fmtdesc fdesc;
444 memset(&fdesc, 0,
sizeof(fdesc));
445 fdesc.type =
ctx->type;
475 struct v4l2_fmtdesc fdesc;
485 memset(&fdesc, 0,
sizeof(fdesc));
486 fdesc.type =
ctx->type;
493 if (fdesc.pixelformat == v4l2_fmt)
519 ctx->streamon = (cmd == VIDIOC_STREAMON);
624 fmt.update_avfmt = 1;
664 struct v4l2_requestbuffers req;
672 ret = ioctl(
s->fd, VIDIOC_G_FMT, &
ctx->format);
676 memset(&req, 0,
sizeof(req));
677 req.count =
ctx->num_buffers;
678 req.memory = V4L2_MEMORY_MMAP;
679 req.type =
ctx->type;
680 ret = ioctl(
s->fd, VIDIOC_REQBUFS, &req);
684 ctx->num_buffers = req.count;
691 for (
i = 0;
i < req.count;
i++) {
706 V4L2_TYPE_IS_MULTIPLANAR(
ctx->type) ?
ctx->format.fmt.pix_mp.plane_fmt[0].sizeimage :
ctx->format.fmt.pix.sizeimage,
707 V4L2_TYPE_IS_MULTIPLANAR(
ctx->type) ?
ctx->format.fmt.pix_mp.plane_fmt[0].bytesperline :
ctx->format.fmt.pix.bytesperline);
int ff_v4l2_context_init(V4L2Context *ctx)
Initializes a V4L2Context.
#define AV_LOG_WARNING
Something somehow does not look correct.
AVPixelFormat
Pixel format.
Filter the word “frame” indicates either a video frame or a group of audio as stored in an AVFrame structure Format for each input and each output the list of supported formats For video that means pixel format For audio that means channel sample they are references to shared objects When the negotiation mechanism computes the intersection of the formats supported at each end of a all references to both lists are replaced with a reference to the intersection And when a single format is eventually chosen for a link amongst the remaining all references to the list are updated That means that if a filter requires that its input and output have the same format amongst a supported all it has to do is use a reference to the same list of formats query_formats can leave some formats unset and return AVERROR(EAGAIN) to cause the negotiation mechanism toagain later. That can be used by filters with complex requirements to use the format negotiated on one link to set the formats supported on another. Frame references ownership and permissions
int av_codec_is_decoder(const AVCodec *codec)
static unsigned int v4l2_get_width(struct v4l2_format *fmt)
#define AVERROR_EOF
End of file.
int ff_v4l2_m2m_codec_reinit(V4L2m2mContext *s)
Reinitializes the V4L2m2mContext when the driver cannot continue processing with the capture paramete...
filter_frame For filters that do not use the this method is called when a frame is pushed to the filter s input It can be called at any time except in a reentrant way If the input frame is enough to produce output
int ff_v4l2_buffer_buf_to_avpkt(AVPacket *pkt, V4L2Buffer *avbuf)
Extracts the data from a V4L2Buffer to an AVPacket.
This structure describes decoded (raw) audio or video data.
int ff_v4l2_context_dequeue_packet(V4L2Context *ctx, AVPacket *pkt)
Dequeues a buffer from a V4L2Context to an AVPacket.
void ff_v4l2_context_release(V4L2Context *ctx)
Releases a V4L2Context.
static int v4l2_stop_encode(V4L2Context *ctx)
int ff_v4l2_context_dequeue_frame(V4L2Context *ctx, AVFrame *frame)
Dequeues a buffer from a V4L2Context to an AVFrame.
static V4L2m2mContext * ctx_to_m2mctx(V4L2Context *ctx)
static int v4l2_handle_event(V4L2Context *ctx)
returns 1 if reinit was successful, negative if it failed returns 0 if reinit was not executed
static int v4l2_get_framesize_compressed(V4L2Context *ctx, int width, int height)
int ff_v4l2_buffer_buf_to_avframe(AVFrame *frame, V4L2Buffer *avbuf)
Extracts the data from a V4L2Buffer to an AVFrame.
it s the only field you need to keep assuming you have a context There is some magic you don t need to care about around this just let it vf type
static void v4l2_save_to_context(V4L2Context *ctx, struct v4l2_format_update *fmt)
static int v4l2_release_buffers(V4L2Context *ctx)
#define AV_LOG_ERROR
Something went wrong and cannot losslessly be recovered.
int ff_v4l2_buffer_avframe_to_buf(const AVFrame *frame, V4L2Buffer *out)
Extracts the data from an AVFrame to a V4L2Buffer.
int ff_v4l2_context_set_format(V4L2Context *ctx)
Sets the V4L2Context format in the v4l2 driver.
V4L2Buffer (wrapper for v4l2_buffer management)
static V4L2Buffer * v4l2_getfree_v4l2buf(V4L2Context *ctx)
enum AVPixelFormat ff_v4l2_format_v4l2_to_avfmt(uint32_t v4l2_fmt, enum AVCodecID avcodec)
static AVCodecContext * logger(V4L2Context *ctx)
#define AV_LOG_DEBUG
Stuff which is only useful for libav* developers.
int ff_v4l2_buffer_enqueue(V4L2Buffer *avbuf)
Enqueues a V4L2Buffer.
static int v4l2_stop_decode(V4L2Context *ctx)
#define AVERROR_PATCHWELCOME
Not yet implemented in FFmpeg, patches welcome.
#define container_of(ptr, type, member)
int ff_v4l2_context_enqueue_packet(V4L2Context *ctx, const AVPacket *pkt)
Enqueues a buffer to a V4L2Context from an AVPacket.
static V4L2Buffer * v4l2_dequeue_v4l2buf(V4L2Context *ctx, int timeout)
int ff_v4l2_buffer_avpkt_to_buf(const AVPacket *pkt, V4L2Buffer *out)
Extracts the data from an AVPacket to a V4L2Buffer.
static const struct @314 planes[]
static int v4l2_get_raw_format(V4L2Context *ctx, enum AVPixelFormat *p)
#define av_err2str(errnum)
Convenience macro, the return value should be used only directly in function arguments but never stan...
int ff_v4l2_m2m_codec_full_reinit(V4L2m2mContext *s)
Reinitializes the V4L2m2mContext when the driver cannot continue processing with the any of the curre...
int ff_v4l2_context_set_status(V4L2Context *ctx, uint32_t cmd)
Sets the status of a V4L2Context.
static void reinit(Jpeg2000EncoderContext *s)
uint32_t ff_v4l2_format_avcodec_to_v4l2(enum AVCodecID avcodec)
static int v4l2_try_raw_format(V4L2Context *ctx, enum AVPixelFormat pixfmt)
#define i(width, name, range_min, range_max)
struct v4l2_plane planes[VIDEO_MAX_PLANES]
void * av_mallocz(size_t size)
Allocate a memory block with alignment suitable for all memory accesses (including vectors if availab...
static int v4l2_get_coded_format(V4L2Context *ctx, uint32_t *p)
int ff_v4l2_context_enqueue_frame(V4L2Context *ctx, const AVFrame *frame)
Enqueues a buffer to a V4L2Context from an AVFrame.
enum AVPixelFormat pixfmt
these buffered frames must be flushed immediately if a new input produces new the filter must not call request_frame to get more It must just process the frame or queue it The task of requesting more frames is left to the filter s request_frame method or the application If a filter has several the filter must be ready for frames arriving randomly on any input any filter with several inputs will most likely require some kind of queuing mechanism It is perfectly acceptable to have a limited queue and to drop frames when the inputs are too unbalanced request_frame For filters that do not use the this method is called when a frame is wanted on an output For a it should directly call filter_frame on the corresponding output For a if there are queued frames already one of these frames should be pushed If the filter should request a frame on one of its repeatedly until at least one frame has been pushed Return or at least make progress towards producing a frame
main external API structure.
static unsigned int v4l2_get_height(struct v4l2_format *fmt)
the frame and frame reference mechanism is intended to as much as expensive copies of that data while still allowing the filters to produce correct results The data is stored in buffers represented by AVFrame structures Several references can point to the same frame buffer
static unsigned int v4l2_resolution_changed(V4L2Context *ctx, struct v4l2_format *fmt2)
int ff_set_dimensions(AVCodecContext *s, int width, int height)
Check that the provided frame dimensions are valid and set them on the codec context.
uint32_t ff_v4l2_format_avfmt_to_v4l2(enum AVPixelFormat avfmt)
This structure stores compressed data.
int ff_v4l2_context_get_format(V4L2Context *ctx)
Queries the driver for a valid v4l2 format and copies it to the context.
int ff_v4l2_buffer_initialize(V4L2Buffer *avbuf, int index)
Initializes a V4L2Buffer.
#define av_fourcc2str(fourcc)
static int v4l2_type_supported(V4L2Context *ctx)
enum V4L2Buffer_status status