summaryrefslogtreecommitdiff
path: root/libavfilter
diff options
context:
space:
mode:
authorPaul B Mahol <onemda@gmail.com>2016-08-27 16:04:07 +0200
committerPaul B Mahol <onemda@gmail.com>2016-08-27 16:04:07 +0200
commitf242d74d170e020d433267b6cd4c7fe400883939 (patch)
treec8411b33baa0032c8e4cbf8dbc9c11d5be152eaa /libavfilter
parentb2c6a11fb60429c9698b9d1e8538d32d741e68c5 (diff)
avfilter/vf_convolution: add >8 bit depth support
Diffstat (limited to 'libavfilter')
-rw-r--r--libavfilter/vf_convolution.c190
1 files changed, 165 insertions, 25 deletions
diff --git a/libavfilter/vf_convolution.c b/libavfilter/vf_convolution.c
index 2380cdca07..2cfc3ba62d 100644
--- a/libavfilter/vf_convolution.c
+++ b/libavfilter/vf_convolution.c
@@ -35,6 +35,8 @@ typedef struct ConvolutionContext {
float rdiv[4];
float bias[4];
+ int size[4];
+ int depth;
int bstride;
uint8_t *buffer;
int nb_planes;
@@ -81,46 +83,46 @@ static const int same5x5[25] = {0, 0, 0, 0, 0,
static int query_formats(AVFilterContext *ctx)
{
static const enum AVPixelFormat pix_fmts[] = {
- AV_PIX_FMT_YUV444P, AV_PIX_FMT_YUV422P,
- AV_PIX_FMT_YUV420P, AV_PIX_FMT_YUV411P,
+ AV_PIX_FMT_YUVA444P, AV_PIX_FMT_YUV444P, AV_PIX_FMT_YUV440P,
AV_PIX_FMT_YUVJ444P, AV_PIX_FMT_YUVJ440P,
- AV_PIX_FMT_YUVJ422P, AV_PIX_FMT_YUVJ420P, AV_PIX_FMT_YUVJ411P,
- AV_PIX_FMT_YUV410P, AV_PIX_FMT_YUV440P,
- AV_PIX_FMT_GBRP, AV_PIX_FMT_GBRAP,
- AV_PIX_FMT_GRAY8,
+ AV_PIX_FMT_YUVA422P, AV_PIX_FMT_YUV422P, AV_PIX_FMT_YUVA420P, AV_PIX_FMT_YUV420P,
+ AV_PIX_FMT_YUVJ422P, AV_PIX_FMT_YUVJ420P,
+ AV_PIX_FMT_YUVJ411P, AV_PIX_FMT_YUV411P, AV_PIX_FMT_YUV410P,
+ AV_PIX_FMT_YUV420P9, AV_PIX_FMT_YUV422P9, AV_PIX_FMT_YUV444P9,
+ AV_PIX_FMT_YUV420P10, AV_PIX_FMT_YUV422P10, AV_PIX_FMT_YUV444P10,
+ AV_PIX_FMT_YUV420P12, AV_PIX_FMT_YUV422P12, AV_PIX_FMT_YUV444P12, AV_PIX_FMT_YUV440P12,
+ AV_PIX_FMT_YUV420P14, AV_PIX_FMT_YUV422P14, AV_PIX_FMT_YUV444P14,
+ AV_PIX_FMT_YUV420P16, AV_PIX_FMT_YUV422P16, AV_PIX_FMT_YUV444P16,
+ AV_PIX_FMT_YUVA420P9, AV_PIX_FMT_YUVA422P9, AV_PIX_FMT_YUVA444P9,
+ AV_PIX_FMT_YUVA420P10, AV_PIX_FMT_YUVA422P10, AV_PIX_FMT_YUVA444P10,
+ AV_PIX_FMT_YUVA420P16, AV_PIX_FMT_YUVA422P16, AV_PIX_FMT_YUVA444P16,
+ AV_PIX_FMT_GBRP, AV_PIX_FMT_GBRP9, AV_PIX_FMT_GBRP10,
+ AV_PIX_FMT_GBRP12, AV_PIX_FMT_GBRP14, AV_PIX_FMT_GBRP16,
+ AV_PIX_FMT_GBRAP, AV_PIX_FMT_GBRAP12, AV_PIX_FMT_GBRAP16,
+ AV_PIX_FMT_GRAY8, AV_PIX_FMT_GRAY16,
AV_PIX_FMT_NONE
};
return ff_set_common_formats(ctx, ff_make_format_list(pix_fmts));
}
-static int config_input(AVFilterLink *inlink)
+static inline void line_copy8(uint8_t *line, const uint8_t *srcp, int width, int mergin)
{
- ConvolutionContext *s = inlink->dst->priv;
- const AVPixFmtDescriptor *desc = av_pix_fmt_desc_get(inlink->format);
- int ret;
-
- if ((ret = av_image_fill_linesizes(s->planewidth, inlink->format, inlink->w)) < 0)
- return ret;
-
- s->planeheight[1] = s->planeheight[2] = AV_CEIL_RSHIFT(inlink->h, desc->log2_chroma_h);
- s->planeheight[0] = s->planeheight[3] = inlink->h;
-
- s->nb_planes = av_pix_fmt_count_planes(inlink->format);
+ int i;
- s->bstride = s->planewidth[0] + 32;
- s->buffer = av_malloc(5 * s->bstride);
- if (!s->buffer)
- return AVERROR(ENOMEM);
+ memcpy(line, srcp, width);
- return 0;
+ for (i = mergin; i > 0; i--) {
+ line[-i] = line[i];
+ line[width - 1 + i] = line[width - 1 - i];
+ }
}
-static inline void line_copy8(uint8_t *line, const uint8_t *srcp, int width, int mergin)
+static inline void line_copy16(uint16_t *line, const uint16_t *srcp, int width, int mergin)
{
int i;
- memcpy(line, srcp, width);
+ memcpy(line, srcp, width * 2);
for (i = mergin; i > 0; i--) {
line[-i] = line[i];
@@ -128,6 +130,110 @@ static inline void line_copy8(uint8_t *line, const uint8_t *srcp, int width, int
}
}
+static void filter16_3x3(ConvolutionContext *s, AVFrame *in, AVFrame *out, int plane)
+{
+ const uint16_t *src = (const uint16_t *)in->data[plane];
+ uint16_t *dst = (uint16_t *)out->data[plane];
+ const int peak = (1 << s->depth) - 1;
+ const int stride = in->linesize[plane] / 2;
+ const int bstride = s->bstride;
+ const int height = s->planeheight[plane];
+ const int width = s->planewidth[plane];
+ uint16_t *p0 = (uint16_t *)s->buffer + 16;
+ uint16_t *p1 = p0 + bstride;
+ uint16_t *p2 = p1 + bstride;
+ uint16_t *orig = p0, *end = p2;
+ const int *matrix = s->matrix[plane];
+ const float rdiv = s->rdiv[plane];
+ const float bias = s->bias[plane];
+ int y, x;
+
+ line_copy16(p0, src + stride, width, 1);
+ line_copy16(p1, src, width, 1);
+
+ for (y = 0; y < height; y++) {
+ src += stride * (y < height - 1 ? 1 : -1);
+ line_copy16(p2, src, width, 1);
+
+ for (x = 0; x < width; x++) {
+ int sum = p0[x - 1] * matrix[0] +
+ p0[x] * matrix[1] +
+ p0[x + 1] * matrix[2] +
+ p1[x - 1] * matrix[3] +
+ p1[x] * matrix[4] +
+ p1[x + 1] * matrix[5] +
+ p2[x - 1] * matrix[6] +
+ p2[x] * matrix[7] +
+ p2[x + 1] * matrix[8];
+ sum = (int)(sum * rdiv + bias + 0.5f);
+ dst[x] = av_clip(sum, 0, peak);
+ }
+
+ p0 = p1;
+ p1 = p2;
+ p2 = (p2 == end) ? orig: p2 + bstride;
+ dst += out->linesize[plane] / 2;
+ }
+}
+
+static void filter16_5x5(ConvolutionContext *s, AVFrame *in, AVFrame *out, int plane)
+{
+ const uint16_t *src = (const uint16_t *)in->data[plane];
+ uint16_t *dst = (uint16_t *)out->data[plane];
+ const int peak = (1 << s->depth) - 1;
+ const int stride = in->linesize[plane] / 2;
+ const int bstride = s->bstride;
+ const int height = s->planeheight[plane];
+ const int width = s->planewidth[plane];
+ uint16_t *p0 = (uint16_t *)s->buffer + 16;
+ uint16_t *p1 = p0 + bstride;
+ uint16_t *p2 = p1 + bstride;
+ uint16_t *p3 = p2 + bstride;
+ uint16_t *p4 = p3 + bstride;
+ uint16_t *orig = p0, *end = p4;
+ const int *matrix = s->matrix[plane];
+ float rdiv = s->rdiv[plane];
+ float bias = s->bias[plane];
+ int y, x, i;
+
+ line_copy16(p0, src + 2 * stride, width, 2);
+ line_copy16(p1, src + stride, width, 2);
+ line_copy16(p2, src, width, 2);
+ src += stride;
+ line_copy16(p3, src, width, 2);
+
+
+ for (y = 0; y < height; y++) {
+ uint16_t *array[] = {
+ p0 - 2, p0 - 1, p0, p0 + 1, p0 + 2,
+ p1 - 2, p1 - 1, p1, p1 + 1, p1 + 2,
+ p2 - 2, p2 - 1, p2, p2 + 1, p2 + 2,
+ p3 - 2, p3 - 1, p3, p3 + 1, p3 + 2,
+ p4 - 2, p4 - 1, p4, p4 + 1, p4 + 2
+ };
+
+ src += stride * (y < height - 2 ? 1 : -1);
+ line_copy16(p4, src, width, 2);
+
+ for (x = 0; x < width; x++) {
+ int sum = 0;
+
+ for (i = 0; i < 25; i++) {
+ sum += *(array[i] + x) * matrix[i];
+ }
+ sum = (int)(sum * rdiv + bias + 0.5f);
+ dst[x] = av_clip(sum, 0, peak);
+ }
+
+ p0 = p1;
+ p1 = p2;
+ p2 = p3;
+ p3 = p4;
+ p4 = (p4 == end) ? orig: p4 + bstride;
+ dst += out->linesize[plane] / 2;
+ }
+}
+
static void filter_3x3(ConvolutionContext *s, AVFrame *in, AVFrame *out, int plane)
{
const uint8_t *src = in->data[plane];
@@ -230,6 +336,38 @@ static void filter_5x5(ConvolutionContext *s, AVFrame *in, AVFrame *out, int pla
}
}
+static int config_input(AVFilterLink *inlink)
+{
+ ConvolutionContext *s = inlink->dst->priv;
+ const AVPixFmtDescriptor *desc = av_pix_fmt_desc_get(inlink->format);
+ int ret, p;
+
+ s->depth = desc->comp[0].depth;
+ if ((ret = av_image_fill_linesizes(s->planewidth, inlink->format, inlink->w)) < 0)
+ return ret;
+
+ s->planeheight[1] = s->planeheight[2] = AV_CEIL_RSHIFT(inlink->h, desc->log2_chroma_h);
+ s->planeheight[0] = s->planeheight[3] = inlink->h;
+
+ s->nb_planes = av_pix_fmt_count_planes(inlink->format);
+
+ s->bstride = s->planewidth[0] + 32;
+ s->buffer = av_malloc_array(5 * s->bstride, (s->depth + 7) / 8);
+ if (!s->buffer)
+ return AVERROR(ENOMEM);
+
+ if (s->depth > 8) {
+ for (p = 0; p < s->nb_planes; p++) {
+ if (s->size[p] == 3)
+ s->filter[p] = filter16_3x3;
+ else if (s->size[p] == 5)
+ s->filter[p] = filter16_5x5;
+ }
+ }
+
+ return 0;
+}
+
static int filter_frame(AVFilterLink *inlink, AVFrame *in)
{
ConvolutionContext *s = inlink->dst->priv;
@@ -280,11 +418,13 @@ static av_cold int init(AVFilterContext *ctx)
}
if (s->matrix_length[i] == 9) {
+ s->size[i] = 3;
if (!memcmp(matrix, same3x3, sizeof(same3x3)))
s->copy[i] = 1;
else
s->filter[i] = filter_3x3;
} else if (s->matrix_length[i] == 25) {
+ s->size[i] = 5;
if (!memcmp(matrix, same5x5, sizeof(same5x5)))
s->copy[i] = 1;
else