Go to the documentation of this file.
66 unsigned int consumed;
69 uint8_t crc_header[11] = { 0 };
73 uint32_t audio_sample;
77 uint32_t levels[2][8];
97 frame->blocks = 4 * ((
data[1] >> 4) & 0x03) + 4;
100 frame->allocation = (
data[1] >> 1) & 0x01;
101 frame->subbands =
data[1] & 0x01 ? 8 : 4;
115 crc_header[0] =
data[1];
116 crc_header[1] =
data[2];
120 if (
len * 8 < consumed +
frame->subbands)
124 for (sb = 0; sb <
frame->subbands - 1; sb++)
125 frame->joint |= ((
data[4] >> (7 - sb)) & 0x01) << sb;
126 if (
frame->subbands == 4)
127 crc_header[crc_pos / 8] =
data[4] & 0xf0;
129 crc_header[crc_pos / 8] =
data[4];
131 consumed +=
frame->subbands;
132 crc_pos +=
frame->subbands;
135 if (
len * 8 < consumed + (4 *
frame->subbands *
frame->channels))
138 for (ch = 0; ch <
frame->channels; ch++) {
139 for (sb = 0; sb <
frame->subbands; sb++) {
141 frame->scale_factor[ch][sb] =
142 (
data[consumed >> 3] >> (4 - (consumed & 0x7))) & 0x0F;
143 crc_header[crc_pos >> 3] |=
144 frame->scale_factor[ch][sb] << (4 - (crc_pos & 0x7));
156 for (ch = 0; ch <
frame->channels; ch++) {
157 for (sb = 0; sb <
frame->subbands; sb++)
158 levels[ch][sb] = (1 <<
bits[ch][sb]) - 1;
162 for (ch = 0; ch <
frame->channels; ch++) {
163 for (sb = 0; sb <
frame->subbands; sb++) {
166 if (levels[ch][sb] == 0) {
176 if (consumed >
len * 8)
179 if ((
data[consumed >> 3] >> (7 - (consumed & 0x7))) & 0x01)
180 audio_sample |= 1 << (
bits[ch][sb] -
bit - 1);
186 (((((uint64_t) audio_sample << 1) | 1) <<
shift) /
187 levels[ch][sb]) - (1 <<
shift);
194 for (sb = 0; sb <
frame->subbands; sb++) {
195 if (
frame->joint & (0x01 << sb)) {
207 if ((consumed & 0x7) != 0)
208 consumed += 8 - (consumed & 0x7);
210 return consumed >> 3;
221 for (
i = 0;
i < 8;
i++) {
226 memcpy(v + 80, v, 9 *
sizeof(*v));
238 for (idx = 0,
i = 0;
i < 4;
i++, idx += 5) {
264 for (
i = 0;
i < 16;
i++) {
269 memcpy(v + 160, v, 9 *
sizeof(*v));
285 for (idx = 0,
i = 0;
i < 8;
i++, idx += 5) {
308 switch (
frame->subbands) {
310 for (ch = 0; ch <
frame->channels; ch++)
316 for (ch = 0; ch <
frame->channels; ch++)
332 memset(sbc->
dsp.
V, 0,
sizeof(sbc->
dsp.
V));
333 for (ch = 0; ch < 2; ch++)
340 void *
data,
int *got_frame_ptr,
345 int ret, frame_length;
351 if (frame_length <= 0)
381 .supported_samplerates = (
const int[]) { 16000, 32000, 44100, 48000, 0 },
#define FF_CODEC_CAP_INIT_THREADSAFE
The codec does not modify any global variables in the init function, allowing to call the init functi...
Filter the word “frame” indicates either a video frame or a group of audio as stored in an AVFrame structure Format for each input and each output the list of supported formats For video that means pixel format For audio that means channel sample they are references to shared objects When the negotiation mechanism computes the intersection of the formats supported at each end of a all references to both lists are replaced with a reference to the intersection And when a single format is eventually chosen for a link amongst the remaining all references to the list are updated That means that if a filter requires that its input and output have the same format amongst a supported all it has to do is use a reference to the same list of formats query_formats can leave some formats unset and return AVERROR(EAGAIN) to cause the negotiation mechanism toagain later. That can be used by filters with complex requirements to use the format negotiated on one link to set the formats supported on another. Frame references ownership and permissions
#define AV_CH_LAYOUT_MONO
This structure describes decoded (raw) audio or video data.
static int sbc_decode_init(AVCodecContext *avctx)
#define SBCDEC_FIXED_EXTRA_BITS
static void sbc_synthesize_eight(struct sbc_decoder_state *state, struct sbc_frame *frame, int ch, int blk, AVFrame *output_frame)
#define bit(string, value)
#define AV_CH_LAYOUT_STEREO
const int32_t ff_sbc_proto_4_40m1[]
#define FF_ARRAY_ELEMS(a)
static void decode(AVCodecContext *dec_ctx, AVPacket *pkt, AVFrame *frame, FILE *outfile)
static int sbc_unpack_frame(const uint8_t *data, struct sbc_frame *frame, size_t len)
const int32_t ff_synmatrix8[16][8]
const int32_t ff_synmatrix4[8][4]
struct sbc_decoder_state dsp
Describe the class of an AVClass context structure.
static int sbc_decode_frame(AVCodecContext *avctx, void *data, int *got_frame_ptr, AVPacket *avpkt)
const int32_t ff_sbc_proto_8_80m1[]
#define AV_CODEC_CAP_CHANNEL_CONF
Codec should fill in channel configuration and samplerate instead of container.
int ff_get_buffer(AVCodecContext *avctx, AVFrame *frame, int flags)
Get a buffer for a frame.
#define AV_CODEC_CAP_DR1
Codec uses get_buffer() or get_encode_buffer() for allocating buffers and supports custom allocators.
#define NULL_IF_CONFIG_SMALL(x)
Return NULL if CONFIG_SMALL is true, otherwise the argument without modification.
const AVCodec ff_sbc_decoder
enum AVSampleFormat sample_fmt
audio sample format
static void sbc_synthesize_audio(struct sbc_decoder_state *state, struct sbc_frame *frame, AVFrame *output_frame)
const int32_t ff_sbc_proto_4_40m0[]
const AVCRC * av_crc_get_table(AVCRCId crc_id)
Get an initialized standard CRC table.
static int output_frame(H264Context *h, AVFrame *dst, H264Picture *srcp)
it s the only field you need to keep assuming you have a context There is some magic you don t need to care about around this just let it vf offset
@ AV_SAMPLE_FMT_S16P
signed 16 bits, planar
int channels
number of audio channels
#define DECLARE_ALIGNED(n, t, v)
#define i(width, name, range_min, range_max)
AVSampleFormat
Audio sample formats.
#define xf(width, name, var, range_min, range_max, subs,...)
const char * name
Name of the codec implementation.
uint8_t ff_sbc_crc8(const AVCRC *ctx, const uint8_t *data, size_t len)
these buffered frames must be flushed immediately if a new input produces new the filter must not call request_frame to get more It must just process the frame or queue it The task of requesting more frames is left to the filter s request_frame method or the application If a filter has several the filter must be ready for frames arriving randomly on any input any filter with several inputs will most likely require some kind of queuing mechanism It is perfectly acceptable to have a limited queue and to drop frames when the inputs are too unbalanced request_frame For filters that do not use the this method is called when a frame is wanted on an output For a it should directly call filter_frame on the corresponding output For a if there are queued frames already one of these frames should be pushed If the filter should request a frame on one of its repeatedly until at least one frame has been pushed Return or at least make progress towards producing a frame
main external API structure.
static void sbc_synthesize_four(struct sbc_decoder_state *state, struct sbc_frame *frame, int ch, int blk, AVFrame *output_frame)
void ff_sbc_calculate_bits(const struct sbc_frame *frame, int(*bits)[8])
static int shift(int a, int b)
This structure stores compressed data.
static const uint16_t channel_layouts[7]
const int32_t ff_sbc_proto_8_80m0[]