Subversion Repositories Kolibri OS

Rev

Blame | Last modification | View Log | RSS feed

  1. /*
  2.  * Copyright (c) 2010 Brandon Mintern
  3.  * Copyright (c) 2007 Bobby Bingham
  4.  *
  5.  * This file is part of FFmpeg.
  6.  *
  7.  * FFmpeg is free software; you can redistribute it and/or
  8.  * modify it under the terms of the GNU Lesser General Public
  9.  * License as published by the Free Software Foundation; either
  10.  * version 2.1 of the License, or (at your option) any later version.
  11.  *
  12.  * FFmpeg is distributed in the hope that it will be useful,
  13.  * but WITHOUT ANY WARRANTY; without even the implied warranty of
  14.  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
  15.  * Lesser General Public License for more details.
  16.  *
  17.  * You should have received a copy of the GNU Lesser General Public
  18.  * License along with FFmpeg; if not, write to the Free Software
  19.  * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
  20.  */
  21.  
  22. /**
  23.  * @file
  24.  * video fade filter
  25.  * based heavily on vf_negate.c by Bobby Bingham
  26.  */
  27.  
  28. #include "libavutil/avassert.h"
  29. #include "libavutil/avstring.h"
  30. #include "libavutil/common.h"
  31. #include "libavutil/eval.h"
  32. #include "libavutil/opt.h"
  33. #include "libavutil/pixdesc.h"
  34. #include "avfilter.h"
  35. #include "drawutils.h"
  36. #include "formats.h"
  37. #include "internal.h"
  38. #include "video.h"
  39.  
  40. #define R 0
  41. #define G 1
  42. #define B 2
  43. #define A 3
  44.  
  45. #define Y 0
  46. #define U 1
  47. #define V 2
  48.  
  49. #define FADE_IN  0
  50. #define FADE_OUT 1
  51.  
  52. typedef struct FadeContext {
  53.     const AVClass *class;
  54.     int type;
  55.     int factor, fade_per_frame;
  56.     int start_frame, nb_frames;
  57.     int hsub, vsub, bpp;
  58.     unsigned int black_level, black_level_scaled;
  59.     uint8_t is_packed_rgb;
  60.     uint8_t rgba_map[4];
  61.     int alpha;
  62.     uint64_t start_time, duration;
  63.     enum {VF_FADE_WAITING=0, VF_FADE_FADING, VF_FADE_DONE} fade_state;
  64.     uint8_t color_rgba[4];  ///< fade color
  65.     int black_fade;         ///< if color_rgba is black
  66. } FadeContext;
  67.  
  68. static av_cold int init(AVFilterContext *ctx)
  69. {
  70.     FadeContext *s = ctx->priv;
  71.  
  72.     s->fade_per_frame = (1 << 16) / s->nb_frames;
  73.     s->fade_state = VF_FADE_WAITING;
  74.  
  75.     if (s->duration != 0) {
  76.         // If duration (seconds) is non-zero, assume that we are not fading based on frames
  77.         s->nb_frames = 0; // Mostly to clean up logging
  78.     }
  79.  
  80.     // Choose what to log. If both time-based and frame-based options, both lines will be in the log
  81.     if (s->start_frame || s->nb_frames) {
  82.         av_log(ctx, AV_LOG_VERBOSE,
  83.                "type:%s start_frame:%d nb_frames:%d alpha:%d\n",
  84.                s->type == FADE_IN ? "in" : "out", s->start_frame,
  85.                s->nb_frames,s->alpha);
  86.     }
  87.     if (s->start_time || s->duration) {
  88.         av_log(ctx, AV_LOG_VERBOSE,
  89.                "type:%s start_time:%f duration:%f alpha:%d\n",
  90.                s->type == FADE_IN ? "in" : "out", (s->start_time / (double)AV_TIME_BASE),
  91.                (s->duration / (double)AV_TIME_BASE),s->alpha);
  92.     }
  93.  
  94.     s->black_fade = !memcmp(s->color_rgba, "\x00\x00\x00\xff", 4);
  95.     return 0;
  96. }
  97.  
  98. static int query_formats(AVFilterContext *ctx)
  99. {
  100.     const FadeContext *s = ctx->priv;
  101.     static const enum AVPixelFormat pix_fmts[] = {
  102.         AV_PIX_FMT_YUV444P,  AV_PIX_FMT_YUV422P,  AV_PIX_FMT_YUV420P,
  103.         AV_PIX_FMT_YUV411P,  AV_PIX_FMT_YUV410P,
  104.         AV_PIX_FMT_YUVJ444P, AV_PIX_FMT_YUVJ422P, AV_PIX_FMT_YUVJ420P,
  105.         AV_PIX_FMT_YUV440P,  AV_PIX_FMT_YUVJ440P,
  106.         AV_PIX_FMT_YUVA420P, AV_PIX_FMT_YUVA422P, AV_PIX_FMT_YUVA444P,
  107.         AV_PIX_FMT_RGB24,    AV_PIX_FMT_BGR24,
  108.         AV_PIX_FMT_ARGB,     AV_PIX_FMT_ABGR,
  109.         AV_PIX_FMT_RGBA,     AV_PIX_FMT_BGRA,
  110.         AV_PIX_FMT_NONE
  111.     };
  112.     static const enum AVPixelFormat pix_fmts_rgb[] = {
  113.         AV_PIX_FMT_RGB24,    AV_PIX_FMT_BGR24,
  114.         AV_PIX_FMT_ARGB,     AV_PIX_FMT_ABGR,
  115.         AV_PIX_FMT_RGBA,     AV_PIX_FMT_BGRA,
  116.         AV_PIX_FMT_NONE
  117.     };
  118.     static const enum AVPixelFormat pix_fmts_alpha[] = {
  119.         AV_PIX_FMT_YUVA420P, AV_PIX_FMT_YUVA422P, AV_PIX_FMT_YUVA444P,
  120.         AV_PIX_FMT_ARGB,     AV_PIX_FMT_ABGR,
  121.         AV_PIX_FMT_RGBA,     AV_PIX_FMT_BGRA,
  122.         AV_PIX_FMT_NONE
  123.     };
  124.     static const enum AVPixelFormat pix_fmts_rgba[] = {
  125.         AV_PIX_FMT_ARGB,     AV_PIX_FMT_ABGR,
  126.         AV_PIX_FMT_RGBA,     AV_PIX_FMT_BGRA,
  127.         AV_PIX_FMT_NONE
  128.     };
  129.     AVFilterFormats *fmts_list;
  130.  
  131.     if (s->alpha) {
  132.         if (s->black_fade)
  133.             fmts_list = ff_make_format_list(pix_fmts_alpha);
  134.         else
  135.             fmts_list = ff_make_format_list(pix_fmts_rgba);
  136.     } else {
  137.         if (s->black_fade)
  138.             fmts_list = ff_make_format_list(pix_fmts);
  139.         else
  140.             fmts_list = ff_make_format_list(pix_fmts_rgb);
  141.     }
  142.     if (!fmts_list)
  143.         return AVERROR(ENOMEM);
  144.     return ff_set_common_formats(ctx, fmts_list);
  145. }
  146.  
  147. const static enum AVPixelFormat studio_level_pix_fmts[] = {
  148.     AV_PIX_FMT_YUV444P,  AV_PIX_FMT_YUV422P,  AV_PIX_FMT_YUV420P,
  149.     AV_PIX_FMT_YUV411P,  AV_PIX_FMT_YUV410P,
  150.     AV_PIX_FMT_YUV440P,
  151.     AV_PIX_FMT_NONE
  152. };
  153.  
  154. static int config_props(AVFilterLink *inlink)
  155. {
  156.     FadeContext *s = inlink->dst->priv;
  157.     const AVPixFmtDescriptor *pixdesc = av_pix_fmt_desc_get(inlink->format);
  158.  
  159.     s->hsub = pixdesc->log2_chroma_w;
  160.     s->vsub = pixdesc->log2_chroma_h;
  161.  
  162.     s->bpp = pixdesc->flags & AV_PIX_FMT_FLAG_PLANAR ?
  163.              1 :
  164.              av_get_bits_per_pixel(pixdesc) >> 3;
  165.     s->alpha &= !!(pixdesc->flags & AV_PIX_FMT_FLAG_ALPHA);
  166.     s->is_packed_rgb = ff_fill_rgba_map(s->rgba_map, inlink->format) >= 0;
  167.  
  168.     /* use CCIR601/709 black level for studio-level pixel non-alpha components */
  169.     s->black_level =
  170.             ff_fmt_is_in(inlink->format, studio_level_pix_fmts) && !s->alpha ? 16 : 0;
  171.     /* 32768 = 1 << 15, it is an integer representation
  172.      * of 0.5 and is for rounding. */
  173.     s->black_level_scaled = (s->black_level << 16) + 32768;
  174.     return 0;
  175. }
  176.  
  177. static av_always_inline void filter_rgb(FadeContext *s, const AVFrame *frame,
  178.                                         int slice_start, int slice_end,
  179.                                         int do_alpha, int step)
  180. {
  181.     int i, j;
  182.     const uint8_t r_idx  = s->rgba_map[R];
  183.     const uint8_t g_idx  = s->rgba_map[G];
  184.     const uint8_t b_idx  = s->rgba_map[B];
  185.     const uint8_t a_idx  = s->rgba_map[A];
  186.     const uint8_t *c = s->color_rgba;
  187.  
  188.     for (i = slice_start; i < slice_end; i++) {
  189.         uint8_t *p = frame->data[0] + i * frame->linesize[0];
  190.         for (j = 0; j < frame->width; j++) {
  191. #define INTERP(c_name, c_idx) av_clip_uint8(((c[c_idx]<<16) + ((int)p[c_name] - (int)c[c_idx]) * s->factor + (1<<15)) >> 16)
  192.             p[r_idx] = INTERP(r_idx, 0);
  193.             p[g_idx] = INTERP(g_idx, 1);
  194.             p[b_idx] = INTERP(b_idx, 2);
  195.             if (do_alpha)
  196.                 p[a_idx] = INTERP(a_idx, 3);
  197.             p += step;
  198.         }
  199.     }
  200. }
  201.  
  202. static int filter_slice_rgb(AVFilterContext *ctx, void *arg, int jobnr,
  203.                             int nb_jobs)
  204. {
  205.     FadeContext *s = ctx->priv;
  206.     AVFrame *frame = arg;
  207.     int slice_start = (frame->height *  jobnr   ) / nb_jobs;
  208.     int slice_end   = (frame->height * (jobnr+1)) / nb_jobs;
  209.  
  210.     if      (s->alpha)    filter_rgb(s, frame, slice_start, slice_end, 1, 4);
  211.     else if (s->bpp == 3) filter_rgb(s, frame, slice_start, slice_end, 0, 3);
  212.     else if (s->bpp == 4) filter_rgb(s, frame, slice_start, slice_end, 0, 4);
  213.     else                  av_assert0(0);
  214.  
  215.     return 0;
  216. }
  217.  
  218. static int filter_slice_luma(AVFilterContext *ctx, void *arg, int jobnr,
  219.                              int nb_jobs)
  220. {
  221.     FadeContext *s = ctx->priv;
  222.     AVFrame *frame = arg;
  223.     int slice_start = (frame->height *  jobnr   ) / nb_jobs;
  224.     int slice_end   = (frame->height * (jobnr+1)) / nb_jobs;
  225.     int i, j;
  226.  
  227.     for (i = slice_start; i < slice_end; i++) {
  228.         uint8_t *p = frame->data[0] + i * frame->linesize[0];
  229.         for (j = 0; j < frame->width * s->bpp; j++) {
  230.             /* s->factor is using 16 lower-order bits for decimal
  231.              * places. 32768 = 1 << 15, it is an integer representation
  232.              * of 0.5 and is for rounding. */
  233.             *p = ((*p - s->black_level) * s->factor + s->black_level_scaled) >> 16;
  234.             p++;
  235.         }
  236.     }
  237.  
  238.     return 0;
  239. }
  240.  
  241. static int filter_slice_chroma(AVFilterContext *ctx, void *arg, int jobnr,
  242.                                int nb_jobs)
  243. {
  244.     FadeContext *s = ctx->priv;
  245.     AVFrame *frame = arg;
  246.     int i, j, plane;
  247.     const int width = FF_CEIL_RSHIFT(frame->width, s->hsub);
  248.     const int height= FF_CEIL_RSHIFT(frame->height, s->vsub);
  249.     int slice_start = (height *  jobnr   ) / nb_jobs;
  250.     int slice_end   = (height * (jobnr+1)) / nb_jobs;
  251.  
  252.     for (plane = 1; plane < 3; plane++) {
  253.         for (i = slice_start; i < slice_end; i++) {
  254.             uint8_t *p = frame->data[plane] + i * frame->linesize[plane];
  255.             for (j = 0; j < width; j++) {
  256.                 /* 8421367 = ((128 << 1) + 1) << 15. It is an integer
  257.                  * representation of 128.5. The .5 is for rounding
  258.                  * purposes. */
  259.                 *p = ((*p - 128) * s->factor + 8421367) >> 16;
  260.                 p++;
  261.             }
  262.         }
  263.     }
  264.  
  265.     return 0;
  266. }
  267.  
  268. static int filter_slice_alpha(AVFilterContext *ctx, void *arg, int jobnr,
  269.                               int nb_jobs)
  270. {
  271.     FadeContext *s = ctx->priv;
  272.     AVFrame *frame = arg;
  273.     int plane = s->is_packed_rgb ? 0 : A;
  274.     int slice_start = (frame->height *  jobnr   ) / nb_jobs;
  275.     int slice_end   = (frame->height * (jobnr+1)) / nb_jobs;
  276.     int i, j;
  277.  
  278.     for (i = slice_start; i < slice_end; i++) {
  279.         uint8_t *p = frame->data[plane] + i * frame->linesize[plane] + s->is_packed_rgb*s->rgba_map[A];
  280.         int step = s->is_packed_rgb ? 4 : 1;
  281.         for (j = 0; j < frame->width; j++) {
  282.             /* s->factor is using 16 lower-order bits for decimal
  283.              * places. 32768 = 1 << 15, it is an integer representation
  284.              * of 0.5 and is for rounding. */
  285.             *p = ((*p - s->black_level) * s->factor + s->black_level_scaled) >> 16;
  286.             p += step;
  287.         }
  288.     }
  289.  
  290.     return 0;
  291. }
  292.  
  293. static int filter_frame(AVFilterLink *inlink, AVFrame *frame)
  294. {
  295.     AVFilterContext *ctx = inlink->dst;
  296.     FadeContext *s       = ctx->priv;
  297.     double frame_timestamp = frame->pts == AV_NOPTS_VALUE ? -1 : frame->pts * av_q2d(inlink->time_base);
  298.  
  299.     // Calculate Fade assuming this is a Fade In
  300.     if (s->fade_state == VF_FADE_WAITING) {
  301.         s->factor=0;
  302.         if (frame_timestamp >= s->start_time/(double)AV_TIME_BASE
  303.             && inlink->frame_count >= s->start_frame) {
  304.             // Time to start fading
  305.             s->fade_state = VF_FADE_FADING;
  306.  
  307.             // Save start time in case we are starting based on frames and fading based on time
  308.             if (s->start_time == 0 && s->start_frame != 0) {
  309.                 s->start_time = frame_timestamp*(double)AV_TIME_BASE;
  310.             }
  311.  
  312.             // Save start frame in case we are starting based on time and fading based on frames
  313.             if (s->start_time != 0 && s->start_frame == 0) {
  314.                 s->start_frame = inlink->frame_count;
  315.             }
  316.         }
  317.     }
  318.     if (s->fade_state == VF_FADE_FADING) {
  319.         if (s->duration == 0) {
  320.             // Fading based on frame count
  321.             s->factor = (inlink->frame_count - s->start_frame) * s->fade_per_frame;
  322.             if (inlink->frame_count > s->start_frame + s->nb_frames) {
  323.                 s->fade_state = VF_FADE_DONE;
  324.             }
  325.  
  326.         } else {
  327.             // Fading based on duration
  328.             s->factor = (frame_timestamp - s->start_time/(double)AV_TIME_BASE)
  329.                             * (float) UINT16_MAX / (s->duration/(double)AV_TIME_BASE);
  330.             if (frame_timestamp > s->start_time/(double)AV_TIME_BASE
  331.                                   + s->duration/(double)AV_TIME_BASE) {
  332.                 s->fade_state = VF_FADE_DONE;
  333.             }
  334.         }
  335.     }
  336.     if (s->fade_state == VF_FADE_DONE) {
  337.         s->factor=UINT16_MAX;
  338.     }
  339.  
  340.     s->factor = av_clip_uint16(s->factor);
  341.  
  342.     // Invert fade_factor if Fading Out
  343.     if (s->type == FADE_OUT) {
  344.         s->factor=UINT16_MAX-s->factor;
  345.     }
  346.  
  347.     if (s->factor < UINT16_MAX) {
  348.         if (s->alpha) {
  349.             ctx->internal->execute(ctx, filter_slice_alpha, frame, NULL,
  350.                                 FFMIN(frame->height, ctx->graph->nb_threads));
  351.         } else if (s->is_packed_rgb && !s->black_fade) {
  352.             ctx->internal->execute(ctx, filter_slice_rgb, frame, NULL,
  353.                                    FFMIN(frame->height, ctx->graph->nb_threads));
  354.         } else {
  355.             /* luma, or rgb plane in case of black */
  356.             ctx->internal->execute(ctx, filter_slice_luma, frame, NULL,
  357.                                 FFMIN(frame->height, ctx->graph->nb_threads));
  358.  
  359.             if (frame->data[1] && frame->data[2]) {
  360.                 /* chroma planes */
  361.                 ctx->internal->execute(ctx, filter_slice_chroma, frame, NULL,
  362.                                     FFMIN(frame->height, ctx->graph->nb_threads));
  363.             }
  364.         }
  365.     }
  366.  
  367.     return ff_filter_frame(inlink->dst->outputs[0], frame);
  368. }
  369.  
  370.  
  371. #define OFFSET(x) offsetof(FadeContext, x)
  372. #define FLAGS AV_OPT_FLAG_VIDEO_PARAM|AV_OPT_FLAG_FILTERING_PARAM
  373.  
  374. static const AVOption fade_options[] = {
  375.     { "type", "'in' or 'out' for fade-in/fade-out", OFFSET(type), AV_OPT_TYPE_INT, { .i64 = FADE_IN }, FADE_IN, FADE_OUT, FLAGS, "type" },
  376.     { "t",    "'in' or 'out' for fade-in/fade-out", OFFSET(type), AV_OPT_TYPE_INT, { .i64 = FADE_IN }, FADE_IN, FADE_OUT, FLAGS, "type" },
  377.         { "in",  "fade-in",  0, AV_OPT_TYPE_CONST, { .i64 = FADE_IN },  .unit = "type" },
  378.         { "out", "fade-out", 0, AV_OPT_TYPE_CONST, { .i64 = FADE_OUT }, .unit = "type" },
  379.     { "start_frame", "Number of the first frame to which to apply the effect.",
  380.                                                     OFFSET(start_frame), AV_OPT_TYPE_INT, { .i64 = 0 }, 0, INT_MAX, FLAGS },
  381.     { "s",           "Number of the first frame to which to apply the effect.",
  382.                                                     OFFSET(start_frame), AV_OPT_TYPE_INT, { .i64 = 0 }, 0, INT_MAX, FLAGS },
  383.     { "nb_frames",   "Number of frames to which the effect should be applied.",
  384.                                                     OFFSET(nb_frames),   AV_OPT_TYPE_INT, { .i64 = 25 }, 0, INT_MAX, FLAGS },
  385.     { "n",           "Number of frames to which the effect should be applied.",
  386.                                                     OFFSET(nb_frames),   AV_OPT_TYPE_INT, { .i64 = 25 }, 0, INT_MAX, FLAGS },
  387.     { "alpha",       "fade alpha if it is available on the input", OFFSET(alpha),       AV_OPT_TYPE_INT, {.i64 = 0    }, 0,       1, FLAGS },
  388.     { "start_time",  "Number of seconds of the beginning of the effect.",
  389.                                                     OFFSET(start_time),  AV_OPT_TYPE_DURATION, {.i64 = 0. }, 0, INT32_MAX, FLAGS },
  390.     { "st",          "Number of seconds of the beginning of the effect.",
  391.                                                     OFFSET(start_time),  AV_OPT_TYPE_DURATION, {.i64 = 0. }, 0, INT32_MAX, FLAGS },
  392.     { "duration",    "Duration of the effect in seconds.",
  393.                                                     OFFSET(duration),    AV_OPT_TYPE_DURATION, {.i64 = 0. }, 0, INT32_MAX, FLAGS },
  394.     { "d",           "Duration of the effect in seconds.",
  395.                                                     OFFSET(duration),    AV_OPT_TYPE_DURATION, {.i64 = 0. }, 0, INT32_MAX, FLAGS },
  396.     { "color",       "set color",                   OFFSET(color_rgba),  AV_OPT_TYPE_COLOR,    {.str = "black"}, CHAR_MIN, CHAR_MAX, FLAGS },
  397.     { "c",           "set color",                   OFFSET(color_rgba),  AV_OPT_TYPE_COLOR,    {.str = "black"}, CHAR_MIN, CHAR_MAX, FLAGS },
  398.     { NULL }
  399. };
  400.  
  401. AVFILTER_DEFINE_CLASS(fade);
  402.  
  403. static const AVFilterPad avfilter_vf_fade_inputs[] = {
  404.     {
  405.         .name           = "default",
  406.         .type           = AVMEDIA_TYPE_VIDEO,
  407.         .config_props   = config_props,
  408.         .filter_frame   = filter_frame,
  409.         .needs_writable = 1,
  410.     },
  411.     { NULL }
  412. };
  413.  
  414. static const AVFilterPad avfilter_vf_fade_outputs[] = {
  415.     {
  416.         .name = "default",
  417.         .type = AVMEDIA_TYPE_VIDEO,
  418.     },
  419.     { NULL }
  420. };
  421.  
  422. AVFilter ff_vf_fade = {
  423.     .name          = "fade",
  424.     .description   = NULL_IF_CONFIG_SMALL("Fade in/out input video."),
  425.     .init          = init,
  426.     .priv_size     = sizeof(FadeContext),
  427.     .priv_class    = &fade_class,
  428.     .query_formats = query_formats,
  429.     .inputs        = avfilter_vf_fade_inputs,
  430.     .outputs       = avfilter_vf_fade_outputs,
  431.     .flags         = AVFILTER_FLAG_SLICE_THREADS,
  432. };
  433.