Go to the documentation of this file.
31 if ((ret = (x)) < 0) \
71 for (
int i = 0;
i < 4;
i++) {
134 for (
int i = 0;
i < 4;
i++) {
137 const int j =
op->swizzle.in[
i];
148 for (
int i = 0;
i < 4;
i++) {
151 const int j =
op->swizzle.in[
i];
229 return (1 <<
p) == x ?
p : 0;
255 for (
int i = 0;
i < 4;
i++) {
265 if (
scale.factor.den)
267 return scale.factor.den;
277 for (
int i = 0;
i < 4;
i++) {
278 bool const_row =
c->m[
i][4].den == 1;
279 for (
int j = 0; j < 4; j++) {
280 const_row &=
c->m[
i][j].num == 0 ||
286 for (
int j = 0; j < 5; j++)
287 c->m[
i][j] =
Q(
i == j);
306 uint32_t nonzero = 0;
307 for (
int i = 0;
i < 4;
i++) {
308 for (
int j = 0; j < 4; j++) {
317 for (
int i = 0;
i < 4;
i++) {
320 for (
int j = 0; j < 4; j++) {
323 c.m[
i][
i] =
c.m[
i][j];
349 for (
int n = 1; n < ops->
num_ops; n++) {
365 for (
int n = 0; n < ops->
num_ops; n++) {
386 if (!
op->rw.packed) {
389 for (
int i = 0;
i <
op->rw.elems;
i++) {
391 swiz.
in[
i] = 3 - (
i - nb_planes);
395 const int idx = nb_planes++;
401 if (nb_planes < op->rw.elems) {
402 op->rw.elems = nb_planes;
437 if (next->
op ==
op->op) {
444 if (!
op->shift.amount) {
451 for (
int i = 0;
i < 4;
i++) {
457 op->clear.value[
i].num == 0)
476 for (
int i = 0;
i < 4;
i++) {
487 for (
int i = 0;
i < 4;
i++) {
490 if (
op->swizzle.in[
i] !=
i)
503 for (
int i = 0;
i < 4;
i++)
512 const int src =
op->swizzle.in[
dst];
513 if (
src >
dst && src < prev->rw.elems) {
515 for (
int i =
dst;
i < 4;
i++) {
516 if (
op->swizzle.in[
i] ==
dst)
518 else if (
op->swizzle.in[
i] ==
src)
528 const int src =
op->swizzle.in[
dst];
529 if (
src >
dst && src < next->rw.elems) {
540 if (
op->type ==
op->convert.to) {
562 op->convert.expand =
true;
569 for (
int i = 0;
i < 4;
i++) {
583 for (
int i = 0;
i < 4;
i++) {
597 for (
int i = 0;
i < 4;
i++) {
598 if (
op->dither.y_offset[
i] < 0)
601 op->dither.y_offset[
i] = -1;
629 for (
int i = 0;
i < 4;
i++) {
630 for (
int j = 0; j < 5; j++) {
632 for (
int k = 0; k < 4; k++)
636 op->lin.m[
i][j] = sum;
645 for (
int j = 0; j < 4; j++) {
649 for (
int i = 0;
i < 4;
i++)
650 op->lin.m[
i][j] =
Q(
i == j);
651 op->lin.mask &= ~col;
656 for (
int i = 0;
i < 4;
i++) {
660 for (
int j = 0; j < 5; j++)
661 op->lin.m[
i][j] =
Q(
i == j);
662 op->lin.mask &= ~row;
700 if (
op->scale.factor.num == 1 &&
op->scale.factor.den == 1) {
710 av_reduce(&
op->scale.factor.num, &
op->scale.factor.den,
p, q, INT_MAX);
719 op->shift.amount =
FFABS(factor2);
740 for (
int n = 0; n < ops->
num_ops - 1; n++) {
757 for (
int n = 0; n < ops->
num_ops - 1; n++) {
788 int size, uint8_t clear_val,
796 (!
read->rw.packed &&
read->rw.elems > 1))
800 uint32_t
mask[4] = {0};
801 for (
int i = 0;
i <
read->rw.elems;
i++)
802 mask[
i] = 0x01010101 *
i * read_size + 0x03020100;
804 for (
int opidx = 1; opidx < ops->
num_ops; opidx++) {
809 for (
int i = 0;
i < 4;
i++)
815 for (
int i = 0;
i < 4;
i++) {
824 for (
int i = 0;
i < 4;
i++) {
827 if (
op->clear.value[
i].num != 0 || !clear_val)
829 mask[
i] = 0x1010101ul * clear_val;
834 if (!
op->convert.expand)
836 for (
int i = 0;
i < 4;
i++) {
838 case 1:
mask[
i] = 0x01010101 * (
mask[
i] & 0xFF);
break;
839 case 2:
mask[
i] = 0x00010001 * (
mask[
i] & 0xFFFF);
break;
846 if (
op->rw.frac ||
op->rw.filter ||
847 (!
op->rw.packed &&
op->rw.elems > 1))
855 const int write_chunk =
op->rw.elems * write_size;
857 for (
int n = 0; n < num_groups; n++) {
859 const int base_out = n * write_chunk;
860 for (
int i = 0;
i <
op->rw.elems;
i++) {
861 const int offset = base_out +
i * write_size;
862 for (
int b = 0;
b < write_size;
b++) {
863 const uint8_t idx =
mask[
i] >> (
b * 8);
864 if (idx != clear_val)
942 for (idx = 0; idx < ops1->
num_ops; idx++) {
943 op = &ops1->
ops[idx];
954 const SwsOp *prev = &ops1->
ops[idx - 1];
968 for (
int i = 0;
i < 4;
i++) {
970 const int o = nb_planes++;
983 for (
int i = 0;
i < nb_planes;
i++) {
1005 .rw.elems = nb_planes,
1013 .rw.elems = nb_planes,
void ff_sws_op_list_free(SwsOpList **p_ops)
AVPixelFormat
Pixel format.
Filter the word “frame” indicates either a video frame or a group of audio as stored in an AVFrame structure Format for each input and each output the list of supported formats For video that means pixel format For audio that means channel sample they are references to shared objects When the negotiation mechanism computes the intersection of the formats supported at each end of a all references to both lists are replaced with a reference to the intersection And when a single format is eventually chosen for a link amongst the remaining all references to the list are updated That means that if a filter requires that its input and output have the same format amongst a supported all it has to do is use a reference to the same list of formats query_formats can leave some formats unset and return AVERROR(EAGAIN) to cause the negotiation mechanism toagain later. That can be used by filters with complex requirements to use the format negotiated on one link to set the formats supported on another. Frame references ownership and permissions
SwsOpList * ff_sws_op_list_duplicate(const SwsOpList *ops)
Returns a duplicate of ops, or NULL on OOM.
SwsComps comps_src
Source component metadata associated with pixel values from each corresponding component (in plane/me...
const SwsOp * ff_sws_op_list_input(const SwsOpList *ops)
Returns the input operation for a given op list, or NULL if there is none (e.g.
const AVPixFmtDescriptor * av_pix_fmt_desc_get(enum AVPixelFormat pix_fmt)
uint32_t ff_sws_linear_mask(const SwsLinearOp c)
AVRational m[4][5]
Generalized 5x5 affine transformation: [ Out.x ] = [ A B C D E ] [ Out.y ] = [ F G H I J ] * [ x y z ...
int ff_sws_op_list_append(SwsOpList *ops, SwsOp *op)
These will take over ownership of op and set it to {0}, even on failure.
static void read_bytes(const uint8_t *src, float *dst, int src_stride, int dst_stride, int width, int height, float scale)
static void get_input_size(const SwsOpList *ops, SwsFormat *fmt)
int ff_sws_op_list_optimize(SwsOpList *ops)
Fuse compatible and eliminate redundant operations, as well as replacing some operations with more ef...
static bool extract_constant_rows(SwsLinearOp *c, SwsComps prev, SwsClearOp *out_clear)
int ff_sws_pixel_type_size(SwsPixelType type)
SwsPixelType
Copyright (C) 2025 Niklas Haas.
#define AV_PIX_FMT_YUVA444P16
#define SWS_COMP_TEST(mask, X)
bool ff_sws_pixel_type_is_int(SwsPixelType type)
it s the only field you need to keep assuming you have a context There is some magic you don t need to care about around this just let it vf type
#define AV_PIX_FMT_GRAY16
int av_reduce(int *dst_num, int *dst_den, int64_t num, int64_t den, int64_t max)
Reduce a fraction.
#define SWS_OP_NEEDED(op, idx)
#define AV_PIX_FMT_YUV444P16
#define SWS_SWIZZLE(X, Y, Z, W)
static int read_chunk(AVFormatContext *s)
static int op(uint8_t **dst, const uint8_t *dst_end, GetByteContext *gb, int pixel, int count, int *x, int width, int linesize)
Perform decode operation.
#define av_assert0(cond)
assert() equivalent, that is always enabled.
static int exact_log2_q(const AVRational x)
static bool extract_scalar(const SwsLinearOp *c, SwsComps comps, SwsComps prev, SwsScaleOp *out_scale)
If a linear operation can be reduced to a scalar multiplication, returns the corresponding scaling fa...
static AVRational ff_sws_pixel_expand(SwsPixelType from, SwsPixelType to)
static bool op_commute_filter(SwsOp *op, SwsOp *prev)
Try to commute a filter op with the previous operation.
#define AV_PIX_FMT_GRAYF32
#define FFABS(a)
Absolute value, Note, INT_MIN / INT64_MIN result in undefined behavior as they are not representable ...
SwsFilterWeights * kernel
Rational number (pair of numerator and denominator).
#define av_unreachable(msg)
Asserts that are used as compiler optimization hints depending upon ASSERT_LEVEL and NBDEBUG.
int ff_sws_op_list_subpass(SwsOpList *ops1, SwsOpList **out_rest)
Eliminate SWS_OP_FILTER_* operations by merging them with prior SWS_OP_READ operations.
void ff_sws_op_list_remove_at(SwsOpList *ops, int index, int count)
#define RET(x)
Copyright (C) 2025 Niklas Haas.
@ AV_PIX_FMT_GRAY8
Y , 8bpp.
Undefined Behavior In the C some operations are like signed integer dereferencing freed accessing outside allocated Undefined Behavior must not occur in a C it is not safe even if the output of undefined operations is unused The unsafety may seem nit picking but Optimizing compilers have in fact optimized code on the assumption that no undefined Behavior occurs Optimizing code based on wrong assumptions can and has in some cases lead to effects beyond the output of computations The signed integer overflow problem in speed critical code Code which is highly optimized and works with signed integers sometimes has the problem that often the output of the computation does not c
void ff_sws_apply_op_q(const SwsOp *op, AVRational x[4])
Apply an operation to an AVRational.
uint8_t pattern[4]
Packed bits are assumed to be LSB-aligned within the underlying integer type; i.e.
uint8_t ptrdiff_t const uint8_t ptrdiff_t int intptr_t intptr_t int int16_t * dst
#define i(width, name, range_min, range_max)
#define AV_PIX_FMT_GBRPF32
int ff_sws_op_list_insert_at(SwsOpList *ops, int index, SwsOp *op)
void ff_sws_op_list_update_comps(SwsOpList *ops)
Infer + propagate known information about components.
void * av_refstruct_ref(void *obj)
Create a new reference to an object managed via this API, i.e.
@ AV_PIX_FMT_YUVA444P
planar YUV 4:4:4 32bpp, (1 Cr & Cb sample per 1x1 Y & A samples)
it s the only field you need to keep assuming you have a context There is some magic you don t need to care about around this just let it vf offset
SwsComps comps
Metadata about the operation's input/output components.
static enum AVPixelFormat get_planar_fmt(SwsPixelType type, int nb_planes)
Determine a suitable intermediate buffer format for a given combination of pixel types and number of ...
SwsCompMask ff_sws_comp_mask_swizzle(const SwsCompMask mask, const SwsSwizzleOp swiz)
static bool extract_swizzle(SwsLinearOp *op, SwsComps prev, SwsSwizzleOp *out_swiz)
#define av_assert1(cond)
assert() equivalent, that does not lie in speed critical code.
static void write_bytes(const float *src, uint8_t *dst, int src_stride, int dst_stride, int width, int height, int depth, float scale)
static int av_cmp_q(AVRational a, AVRational b)
Compare two rationals.
#define FFSWAP(type, a, b)
static bool op_commute_swizzle(SwsOp *op, SwsOp *next)
Try to commute a swizzle op with the next operation.
SwsOpType filter
Filter kernel to apply to each plane while sampling.
#define AV_PIX_FMT_GBRAPF32
int ff_sws_solve_shuffle(const SwsOpList *const ops, uint8_t shuffle[], int size, uint8_t clear_val, int *read_bytes, int *write_bytes)
"Solve" an op list into a fixed shuffle mask, with an optional ability to also directly clear the out...
IDirect3DDxgiInterfaceAccess _COM_Outptr_ void ** p
AVRational av_mul_q(AVRational b, AVRational c)
Multiply two rationals.
@ AV_PIX_FMT_YUV444P
planar YUV 4:4:4, 24bpp, (1 Cr & Cb sample per 1x1 Y samples)
uint8_t elems
Examples: rgba = 4x u8 packed yuv444p = 3x u8 rgb565 = 1x u16 <- use SWS_OP_UNPACK to unpack monow = ...
static void scale(int *out, const int *in, const int w, const int h, const int shift)
AVRational av_add_q(AVRational b, AVRational c)
Add two rationals.
static bool op_commute_clear(SwsOp *op, SwsOp *next)
Try to commute a clear op with the next operation.
Helper struct for representing a list of operations.
static uint64_t shuffle(uint64_t in, const uint8_t *shuffle, int shuffle_len)
static uint32_t BS_FUNC() read(BSCTX *bc, unsigned int n)
Return n bits from the buffer, n has to be in the 0-32 range.
static int exact_log2(const int x)