If it helps, here is the diff between af_compand.c in ffmpeg and af_compand.c in libav:
--- /home/andy/dev/ffmpeg/libavfilter/af_compand.c 2014-02-02 12:07:39.789987354 -0500 +++ /home/andy/dev/libav/libavfilter/af_compand.c 2014-02-08 16:50:46.428038489 -0500 @@ -3,8 +3,9 @@ * Copyright (c) 1999 Nick Bailey * Copyright (c) 2007 Rob Sykes <r...@users.sourceforge.net> * Copyright (c) 2013 Paul B Mahol + * Copyright (c) 2014 Andrew Kelley * - * This file is part of FFmpeg. + * This file is part of libav. * * FFmpeg is free software; you can redistribute it and/or * modify it under the terms of the GNU Lesser General Public @@ -22,12 +23,20 @@ * */ +/** + * @file + * audio compand filter + */ + +#include "libavutil/mem.h" #include "libavutil/avassert.h" -#include "libavutil/avstring.h" +#include "libavutil/mathematics.h" +#include "libavutil/channel_layout.h" #include "libavutil/opt.h" -#include "libavutil/samplefmt.h" -#include "avfilter.h" +#include "libavutil/common.h" #include "audio.h" +#include "avfilter.h" +#include "formats.h" #include "internal.h" typedef struct ChanParam { @@ -62,7 +71,7 @@ } CompandContext; #define OFFSET(x) offsetof(CompandContext, x) -#define A AV_OPT_FLAG_AUDIO_PARAM|AV_OPT_FLAG_FILTERING_PARAM +#define A AV_OPT_FLAG_AUDIO_PARAM static const AVOption compand_options[] = { { "attacks", "set time over which increase of volume is determined", OFFSET(attacks), AV_OPT_TYPE_STRING, {.str=NULL}, 0, 0, A }, @@ -75,7 +84,12 @@ { NULL } }; -AVFILTER_DEFINE_CLASS(compand); +static const AVClass compand_class = { + .class_name = "compand filter", + .item_name = av_default_item_name, + .option = compand_options, + .version = LIBAVUTIL_VERSION_INT, +}; static av_cold int init(AVFilterContext *ctx) { @@ -171,11 +185,26 @@ return exp(out_log); } +/** + * Clip a double value into the amin-amax range. + * @param a value to clip + * @param amin minimum value of the clip range + * @param amax maximum value of the clip range + * @return clipped value + */ +static av_always_inline av_const double av_clipd_c(double a, double amin, double amax) +{ + av_assert2(amin <= amax); + if (a < amin) return amin; + else if (a > amax) return amax; + else return a; +} + static int compand_nodelay(AVFilterContext *ctx, AVFrame *frame) { CompandContext *s = ctx->priv; AVFilterLink *inlink = ctx->inputs[0]; - const int channels = inlink->channels; + const int channels = av_get_channel_layout_nb_channels(inlink->channel_layout); const int nb_samples = frame->nb_samples; AVFrame *out_frame; int chan, i; @@ -197,7 +226,7 @@ for (i = 0; i < nb_samples; i++) { update_volume(cp, fabs(src[i])); - dst[i] = av_clipd(src[i] * get_volume(s, cp->volume), -1, 1); + dst[i] = av_clipd_c(src[i] * get_volume(s, cp->volume), -1, 1); } } @@ -213,12 +242,12 @@ { CompandContext *s = ctx->priv; AVFilterLink *inlink = ctx->inputs[0]; - const int channels = inlink->channels; + const int channels = av_get_channel_layout_nb_channels(inlink->channel_layout); const int nb_samples = frame->nb_samples; int chan, i, av_uninit(dindex), oindex, av_uninit(count); AVFrame *out_frame = NULL; - av_assert1(channels > 0); /* would corrupt delay_count and delay_index */ + av_assert2(channels > 0); /* would corrupt delay_count and delay_index */ for (chan = 0; chan < channels; chan++) { const double *src = (double *)frame->extended_data[chan]; @@ -243,7 +272,7 @@ } dst = (double *)out_frame->extended_data[chan]; - dst[oindex++] = av_clipd(dbuf[dindex] * get_volume(s, cp->volume), -1, 1); + dst[oindex++] = av_clipd_c(dbuf[dindex] * get_volume(s, cp->volume), -1, 1); } else { count++; } @@ -264,7 +293,7 @@ { AVFilterContext *ctx = outlink->src; CompandContext *s = ctx->priv; - const int channels = outlink->channels; + const int channels = av_get_channel_layout_nb_channels(outlink->channel_layout); int chan, i, dindex; AVFrame *frame = NULL; @@ -281,7 +310,7 @@ dindex = s->delay_index; for (i = 0; i < frame->nb_samples; i++) { - dst[i] = av_clipd(dbuf[dindex] * get_volume(s, cp->volume), -1, 1); + dst[i] = av_clipd_c(dbuf[dindex] * get_volume(s, cp->volume), -1, 1); dindex = MOD(dindex + 1, s->delay_samples); } } @@ -291,10 +320,56 @@ return ff_filter_frame(outlink, frame); } +static char *av_strtok(char *s, const char *delim, char **saveptr) +{ + char *tok; + + if (!s && !(s = *saveptr)) + return NULL; + + /* skip leading delimiters */ + s += strspn(s, delim); + + /* s now points to the first non delimiter char, or to the end of the string */ + if (!*s) { + *saveptr = NULL; + return NULL; + } + tok = s++; + + /* skip non delimiters */ + s += strcspn(s, delim); + if (*s) { + *s = 0; + *saveptr = s+1; + } else { + *saveptr = NULL; + } + + return tok; +} + + +static int av_samples_alloc_array_and_samples(uint8_t ***audio_data, int *linesize, int nb_channels, + int nb_samples, enum AVSampleFormat sample_fmt, int align) +{ + int ret, nb_planes = av_sample_fmt_is_planar(sample_fmt) ? nb_channels : 1; + + *audio_data = av_mallocz(nb_planes * sizeof(**audio_data)); + if (!*audio_data) + return AVERROR(ENOMEM); + ret = av_samples_alloc(*audio_data, linesize, nb_channels, + nb_samples, sample_fmt, align); + if (ret < 0) + av_freep(audio_data); + return ret; +} + static int config_output(AVFilterLink *outlink) { AVFilterContext *ctx = outlink->src; CompandContext *s = ctx->priv; + const int channels = av_get_channel_layout_nb_channels(outlink->channel_layout); const int sample_rate = outlink->sample_rate; double radius = s->curve_dB * M_LN10 / 20; int nb_attacks, nb_decays, nb_points; @@ -306,14 +381,14 @@ count_items(s->decays, &nb_decays); count_items(s->points, &nb_points); - if ((nb_attacks > outlink->channels) || (nb_decays > outlink->channels)) { + if ((nb_attacks > channels) || (nb_decays > channels)) { av_log(ctx, AV_LOG_ERROR, "Number of attacks/decays bigger than number of channels.\n"); return AVERROR(EINVAL); } uninit(ctx); - s->channels = av_mallocz_array(outlink->channels, sizeof(*s->channels)); + s->channels = av_mallocz_array(channels, sizeof(*s->channels)); s->segments = av_mallocz_array((nb_points + 4) * 2, sizeof(*s->segments)); if (!s->channels || !s->segments) @@ -434,7 +509,7 @@ s->in_min_lin = exp(s->segments[1].x); s->out_min_lin = exp(s->segments[1].y); - for (i = 0; i < outlink->channels; i++) { + for (i = 0; i < channels; i++) { ChanParam *cp = &s->channels[i]; if (cp->attack > 1.0 / sample_rate) @@ -452,12 +527,11 @@ if (s->delay_samples > 0) { int ret; if ((ret = av_samples_alloc_array_and_samples(&s->delayptrs, NULL, - outlink->channels, + channels, s->delay_samples, outlink->format, 0)) < 0) return ret; s->compand = compand_delay; - outlink->flags |= FF_LINK_FLAG_REQUEST_LOOP; } else { s->compand = compand_nodelay; } @@ -480,7 +554,7 @@ ret = ff_request_frame(ctx->inputs[0]); - if (ret == AVERROR_EOF && !ctx->is_disabled && s->delay_count) + if (ret == AVERROR_EOF && s->delay_count) ret = compand_drain(outlink); return ret; @@ -505,6 +579,7 @@ { NULL } }; + AVFilter ff_af_compand = { .name = "compand", .description = NULL_IF_CONFIG_SMALL("Compress or expand audio dynamic range."), @@ -512,7 +587,6 @@ .priv_size = sizeof(CompandContext), .priv_class = &compand_class, .init = init, - .uninit = uninit, .inputs = compand_inputs, .outputs = compand_outputs, }; On Sat, Feb 8, 2014 at 8:21 PM, Andrew Kelley <superjo...@gmail.com> wrote: > This patch adds the `compand` audio filter from ffmpeg master branch > (currently at 7f0f47b3df9d92eea0d1de94df29971b49f10777) adapted to > work with libav. > > My downstream library is successfully using this filter to allow the > user to turn up the volume gain beyond 100%: > > https://github.com/andrewrk/libgroove/commit/b67b17e091d17bc9eec70f89e61fde90bbee0830 > > Some things to consider about this filter: > > In ffmpeg, an AVFilterLink has the concept of FF_LINK_FLAG_REQUEST_LOOP. > "A filter must set this flag on an output link if it may return 0 in > request_frame() without filtering a frame". > > Libav does not have this concept, so I removed the flag. The change > appears to have no effect; however perhaps someone knows better than I do. > > The filter makes use of `av_strtok` to parse the input parameters, and > it appears that libav has removed that function, so I inlined it into > the filter. Same goes for `av_clipd_c`. Same for > `av_samples_alloc_array_and_samples` except it looks like ffmpeg added > that function recently. > --- > doc/filters.texi | 74 ++++++ > libavfilter/Makefile | 1 + > libavfilter/af_compand.c | 592 > +++++++++++++++++++++++++++++++++++++++++++++++ > libavfilter/allfilters.c | 1 + > 4 files changed, 668 insertions(+) > create mode 100644 libavfilter/af_compand.c > > diff --git a/doc/filters.texi b/doc/filters.texi > index 8c83b4e..8863384 100644 > --- a/doc/filters.texi > +++ b/doc/filters.texi > @@ -467,6 +467,80 @@ To fix a 5.1 WAV improperly encoded in AAC's native > channel order > avconv -i in.wav -filter 'channelmap=1|2|0|5|3|4:channel_layout=5.1' > out.wav > @end example > > +@section compand > +Compress or expand audio dynamic range. > + > +A description of the accepted options follows. > + > +@table @option > + > +@item attacks > +@item decays > +Set list of times in seconds for each channel over which the > instantaneous level > +of the input signal is averaged to determine its volume. @var{attacks} > refers to > +increase of volume and @var{decays} refers to decrease of volume. For most > +situations, the attack time (response to the audio getting louder) should > be > +shorter than the decay time because the human ear is more sensitive to > sudden > +loud audio than sudden soft audio. Typical value for attack is 0.3 > seconds and > +for decay 0.8 seconds. > + > +@item points > +Set list of points for transfer function, specified in dB relative to > maximum > +possible signal amplitude. Each key points list need to be defined using > the > +following syntax: @code{x0/y0 x1/y1 x2/y2 ....} > + > +The input values must be in strictly increasing order but the transfer > function > +does not have to be monotonically rising. The point @code{0/0} is assumed > but > +may be overridden (by @code{0/out-dBn}). Typical values for the transfer > +function are @code{-70/-70 -60/-20}. > + > +@item soft-knee > +Set amount for which the points at where adjacent line segments on the > transfer > +function meet will be rounded. Defaults is 0.01. > + > +@item gain > +Set additional gain in dB to be applied at all points on the transfer > function > +and allows easy adjustment of the overall gain. Default is 0. > + > +@item volume > +Set initial volume in dB to be assumed for each channel when filtering > starts. > +This permits the user to supply a nominal level initially, so that, for > +example, a very large gain is not applied to initial signal levels before > the > +companding has begun to operate. A typical value for audio which is > initially > +quiet is -90 dB. Default is 0. > + > +@item delay > +Set delay in seconds. Default is 0. The input audio is analysed > immediately, > +but audio is delayed before being fed to the volume adjuster. Specifying a > +delay approximately equal to the attack/decay times allows the filter to > +effectively operate in predictive rather than reactive mode. > + > +@end table > + > +@subsection Examples > + > +@itemize > +@item > +Make music with both quiet and loud passages suitable for listening in a > noisy > +environment: > +@example > +compand=.3 .3:1 1:-90/-60 -60/-40 -40/-30 -20/-20:6:0:-90:0.2 > +@end example > + > +@item > +Noise-gate for when the noise is at a lower level than the signal: > +@example > +compand=.1 .1:.2 .2:-900/-900 -50.1/-900 -50/-50:.01:0:-90:.1 > +@end example > + > +@item > +Here is another noise-gate, this time for when the noise is at a higher > level > +than the signal (making it, in some ways, similar to squelch): > +@example > +compand=.1 .1:.1 .1:-45.1/-45.1 -45/-900 0/-900:.01:45:-90:.1 > +@end example > +@end itemize > + > @section join > Join multiple input streams into one multi-channel stream. > > diff --git a/libavfilter/Makefile b/libavfilter/Makefile > index 92c1561..2badb3e 100644 > --- a/libavfilter/Makefile > +++ b/libavfilter/Makefile > @@ -37,6 +37,7 @@ OBJS-$(CONFIG_CHANNELSPLIT_FILTER) += > af_channelsplit.o > OBJS-$(CONFIG_JOIN_FILTER) += af_join.o > OBJS-$(CONFIG_RESAMPLE_FILTER) += af_resample.o > OBJS-$(CONFIG_VOLUME_FILTER) += af_volume.o > +OBJS-$(CONFIG_COMPAND_FILTER) += af_compand.o > > OBJS-$(CONFIG_ANULLSRC_FILTER) += asrc_anullsrc.o > > diff --git a/libavfilter/af_compand.c b/libavfilter/af_compand.c > new file mode 100644 > index 0000000..55d75af > --- /dev/null > +++ b/libavfilter/af_compand.c > @@ -0,0 +1,592 @@ > +/* > + * Copyright (c) 1999 Chris Bagwell > + * Copyright (c) 1999 Nick Bailey > + * Copyright (c) 2007 Rob Sykes <r...@users.sourceforge.net> > + * Copyright (c) 2013 Paul B Mahol > + * Copyright (c) 2014 Andrew Kelley > + * > + * This file is part of libav. > + * > + * FFmpeg is free software; you can redistribute it and/or > + * modify it under the terms of the GNU Lesser General Public > + * License as published by the Free Software Foundation; either > + * version 2.1 of the License, or (at your option) any later version. > + * > + * FFmpeg is distributed in the hope that it will be useful, > + * but WITHOUT ANY WARRANTY; without even the implied warranty of > + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU > + * Lesser General Public License for more details. > + * > + * You should have received a copy of the GNU Lesser General Public > + * License along with FFmpeg; if not, write to the Free Software > + * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA > 02110-1301 USA > + * > + */ > + > +/** > + * @file > + * audio compand filter > + */ > + > +#include "libavutil/mem.h" > +#include "libavutil/avassert.h" > +#include "libavutil/mathematics.h" > +#include "libavutil/channel_layout.h" > +#include "libavutil/opt.h" > +#include "libavutil/common.h" > +#include "audio.h" > +#include "avfilter.h" > +#include "formats.h" > +#include "internal.h" > + > +typedef struct ChanParam { > + double attack; > + double decay; > + double volume; > +} ChanParam; > + > +typedef struct CompandSegment { > + double x, y; > + double a, b; > +} CompandSegment; > + > +typedef struct CompandContext { > + const AVClass *class; > + char *attacks, *decays, *points; > + CompandSegment *segments; > + ChanParam *channels; > + double in_min_lin; > + double out_min_lin; > + double curve_dB; > + double gain_dB; > + double initial_volume; > + double delay; > + uint8_t **delayptrs; > + int delay_samples; > + int delay_count; > + int delay_index; > + int64_t pts; > + > + int (*compand)(AVFilterContext *ctx, AVFrame *frame); > +} CompandContext; > + > +#define OFFSET(x) offsetof(CompandContext, x) > +#define A AV_OPT_FLAG_AUDIO_PARAM > + > +static const AVOption compand_options[] = { > + { "attacks", "set time over which increase of volume is determined", > OFFSET(attacks), AV_OPT_TYPE_STRING, {.str=NULL}, 0, 0, A }, > + { "decays", "set time over which decrease of volume is determined", > OFFSET(decays), AV_OPT_TYPE_STRING, {.str=NULL}, 0, 0, A }, > + { "points", "set points of transfer function", OFFSET(points), > AV_OPT_TYPE_STRING, {.str=NULL}, 0, 0, A }, > + { "soft-knee", "set soft-knee", OFFSET(curve_dB), AV_OPT_TYPE_DOUBLE, > {.dbl=0.01}, 0.01, 900, A }, > + { "gain", "set output gain", OFFSET(gain_dB), AV_OPT_TYPE_DOUBLE, > {.dbl=0}, -900, 900, A }, > + { "volume", "set initial volume", OFFSET(initial_volume), > AV_OPT_TYPE_DOUBLE, {.dbl=0}, -900, 0, A }, > + { "delay", "set delay for samples before sending them to volume > adjuster", OFFSET(delay), AV_OPT_TYPE_DOUBLE, {.dbl=0}, 0, 20, A }, > + { NULL } > +}; > + > +static const AVClass compand_class = { > + .class_name = "compand filter", > + .item_name = av_default_item_name, > + .option = compand_options, > + .version = LIBAVUTIL_VERSION_INT, > +}; > + > +static av_cold int init(AVFilterContext *ctx) > +{ > + CompandContext *s = ctx->priv; > + > + if (!s->attacks || !s->decays || !s->points) { > + av_log(ctx, AV_LOG_ERROR, "Missing attacks and/or decays and/or > points.\n"); > + return AVERROR(EINVAL); > + } > + > + return 0; > +} > + > +static av_cold void uninit(AVFilterContext *ctx) > +{ > + CompandContext *s = ctx->priv; > + > + av_freep(&s->channels); > + av_freep(&s->segments); > + if (s->delayptrs) > + av_freep(&s->delayptrs[0]); > + av_freep(&s->delayptrs); > +} > + > +static int query_formats(AVFilterContext *ctx) > +{ > + AVFilterChannelLayouts *layouts; > + AVFilterFormats *formats; > + static const enum AVSampleFormat sample_fmts[] = { > + AV_SAMPLE_FMT_DBLP, > + AV_SAMPLE_FMT_NONE > + }; > + > + layouts = ff_all_channel_layouts(); > + if (!layouts) > + return AVERROR(ENOMEM); > + ff_set_common_channel_layouts(ctx, layouts); > + > + formats = ff_make_format_list(sample_fmts); > + if (!formats) > + return AVERROR(ENOMEM); > + ff_set_common_formats(ctx, formats); > + > + formats = ff_all_samplerates(); > + if (!formats) > + return AVERROR(ENOMEM); > + ff_set_common_samplerates(ctx, formats); > + > + return 0; > +} > + > +static void count_items(char *item_str, int *nb_items) > +{ > + char *p; > + > + *nb_items = 1; > + for (p = item_str; *p; p++) { > + if (*p == ' ') > + (*nb_items)++; > + } > + > +} > + > +static void update_volume(ChanParam *cp, double in) > +{ > + double delta = in - cp->volume; > + > + if (delta > 0.0) > + cp->volume += delta * cp->attack; > + else > + cp->volume += delta * cp->decay; > +} > + > +static double get_volume(CompandContext *s, double in_lin) > +{ > + CompandSegment *cs; > + double in_log, out_log; > + int i; > + > + if (in_lin < s->in_min_lin) > + return s->out_min_lin; > + > + in_log = log(in_lin); > + > + for (i = 1;; i++) > + if (in_log <= s->segments[i + 1].x) > + break; > + > + cs = &s->segments[i]; > + in_log -= cs->x; > + out_log = cs->y + in_log * (cs->a * in_log + cs->b); > + > + return exp(out_log); > +} > + > +/** > + * Clip a double value into the amin-amax range. > + * @param a value to clip > + * @param amin minimum value of the clip range > + * @param amax maximum value of the clip range > + * @return clipped value > + */ > +static av_always_inline av_const double av_clipd_c(double a, double amin, > double amax) > +{ > + av_assert2(amin <= amax); > + if (a < amin) return amin; > + else if (a > amax) return amax; > + else return a; > +} > + > +static int compand_nodelay(AVFilterContext *ctx, AVFrame *frame) > +{ > + CompandContext *s = ctx->priv; > + AVFilterLink *inlink = ctx->inputs[0]; > + const int channels = > av_get_channel_layout_nb_channels(inlink->channel_layout); > + const int nb_samples = frame->nb_samples; > + AVFrame *out_frame; > + int chan, i; > + > + if (av_frame_is_writable(frame)) { > + out_frame = frame; > + } else { > + out_frame = ff_get_audio_buffer(inlink, nb_samples); > + if (!out_frame) > + return AVERROR(ENOMEM); > + av_frame_copy_props(out_frame, frame); > + } > + > + for (chan = 0; chan < channels; chan++) { > + const double *src = (double *)frame->extended_data[chan]; > + double *dst = (double *)out_frame->extended_data[chan]; > + ChanParam *cp = &s->channels[chan]; > + > + for (i = 0; i < nb_samples; i++) { > + update_volume(cp, fabs(src[i])); > + > + dst[i] = av_clipd_c(src[i] * get_volume(s, cp->volume), -1, > 1); > + } > + } > + > + if (frame != out_frame) > + av_frame_free(&frame); > + > + return ff_filter_frame(ctx->outputs[0], out_frame); > +} > + > +#define MOD(a, b) (((a) >= (b)) ? (a) - (b) : (a)) > + > +static int compand_delay(AVFilterContext *ctx, AVFrame *frame) > +{ > + CompandContext *s = ctx->priv; > + AVFilterLink *inlink = ctx->inputs[0]; > + const int channels = > av_get_channel_layout_nb_channels(inlink->channel_layout); > + const int nb_samples = frame->nb_samples; > + int chan, i, av_uninit(dindex), oindex, av_uninit(count); > + AVFrame *out_frame = NULL; > + > + av_assert2(channels > 0); /* would corrupt delay_count and > delay_index */ > + > + for (chan = 0; chan < channels; chan++) { > + const double *src = (double *)frame->extended_data[chan]; > + double *dbuf = (double *)s->delayptrs[chan]; > + ChanParam *cp = &s->channels[chan]; > + double *dst; > + > + count = s->delay_count; > + dindex = s->delay_index; > + for (i = 0, oindex = 0; i < nb_samples; i++) { > + const double in = src[i]; > + update_volume(cp, fabs(in)); > + > + if (count >= s->delay_samples) { > + if (!out_frame) { > + out_frame = ff_get_audio_buffer(inlink, nb_samples - > i); > + if (!out_frame) > + return AVERROR(ENOMEM); > + av_frame_copy_props(out_frame, frame); > + out_frame->pts = s->pts; > + s->pts += av_rescale_q(nb_samples - i, > (AVRational){1, inlink->sample_rate}, inlink->time_base); > + } > + > + dst = (double *)out_frame->extended_data[chan]; > + dst[oindex++] = av_clipd_c(dbuf[dindex] * get_volume(s, > cp->volume), -1, 1); > + } else { > + count++; > + } > + > + dbuf[dindex] = in; > + dindex = MOD(dindex + 1, s->delay_samples); > + } > + } > + > + s->delay_count = count; > + s->delay_index = dindex; > + > + av_frame_free(&frame); > + return out_frame ? ff_filter_frame(ctx->outputs[0], out_frame) : 0; > +} > + > +static int compand_drain(AVFilterLink *outlink) > +{ > + AVFilterContext *ctx = outlink->src; > + CompandContext *s = ctx->priv; > + const int channels = > av_get_channel_layout_nb_channels(outlink->channel_layout); > + int chan, i, dindex; > + AVFrame *frame = NULL; > + > + frame = ff_get_audio_buffer(outlink, FFMIN(2048, s->delay_count)); > + if (!frame) > + return AVERROR(ENOMEM); > + frame->pts = s->pts; > + s->pts += av_rescale_q(frame->nb_samples, (AVRational){1, > outlink->sample_rate}, outlink->time_base); > + > + for (chan = 0; chan < channels; chan++) { > + double *dbuf = (double *)s->delayptrs[chan]; > + double *dst = (double *)frame->extended_data[chan]; > + ChanParam *cp = &s->channels[chan]; > + > + dindex = s->delay_index; > + for (i = 0; i < frame->nb_samples; i++) { > + dst[i] = av_clipd_c(dbuf[dindex] * get_volume(s, cp->volume), > -1, 1); > + dindex = MOD(dindex + 1, s->delay_samples); > + } > + } > + s->delay_count -= frame->nb_samples; > + s->delay_index = dindex; > + > + return ff_filter_frame(outlink, frame); > +} > + > +static char *av_strtok(char *s, const char *delim, char **saveptr) > +{ > + char *tok; > + > + if (!s && !(s = *saveptr)) > + return NULL; > + > + /* skip leading delimiters */ > + s += strspn(s, delim); > + > + /* s now points to the first non delimiter char, or to the end of the > string */ > + if (!*s) { > + *saveptr = NULL; > + return NULL; > + } > + tok = s++; > + > + /* skip non delimiters */ > + s += strcspn(s, delim); > + if (*s) { > + *s = 0; > + *saveptr = s+1; > + } else { > + *saveptr = NULL; > + } > + > + return tok; > +} > + > + > +static int av_samples_alloc_array_and_samples(uint8_t ***audio_data, int > *linesize, int nb_channels, > + int nb_samples, enum > AVSampleFormat sample_fmt, int align) > +{ > + int ret, nb_planes = av_sample_fmt_is_planar(sample_fmt) ? > nb_channels : 1; > + > + *audio_data = av_mallocz(nb_planes * sizeof(**audio_data)); > + if (!*audio_data) > + return AVERROR(ENOMEM); > + ret = av_samples_alloc(*audio_data, linesize, nb_channels, > + nb_samples, sample_fmt, align); > + if (ret < 0) > + av_freep(audio_data); > + return ret; > +} > + > +static int config_output(AVFilterLink *outlink) > +{ > + AVFilterContext *ctx = outlink->src; > + CompandContext *s = ctx->priv; > + const int channels = > av_get_channel_layout_nb_channels(outlink->channel_layout); > + const int sample_rate = outlink->sample_rate; > + double radius = s->curve_dB * M_LN10 / 20; > + int nb_attacks, nb_decays, nb_points; > + char *p, *saveptr = NULL; > + int new_nb_items, num; > + int i; > + > + count_items(s->attacks, &nb_attacks); > + count_items(s->decays, &nb_decays); > + count_items(s->points, &nb_points); > + > + if ((nb_attacks > channels) || (nb_decays > channels)) { > + av_log(ctx, AV_LOG_ERROR, "Number of attacks/decays bigger than > number of channels.\n"); > + return AVERROR(EINVAL); > + } > + > + uninit(ctx); > + > + s->channels = av_mallocz_array(channels, sizeof(*s->channels)); > + s->segments = av_mallocz_array((nb_points + 4) * 2, > sizeof(*s->segments)); > + > + if (!s->channels || !s->segments) > + return AVERROR(ENOMEM); > + > + p = s->attacks; > + for (i = 0, new_nb_items = 0; i < nb_attacks; i++) { > + char *tstr = av_strtok(p, " ", &saveptr); > + p = NULL; > + new_nb_items += sscanf(tstr, "%lf", &s->channels[i].attack) == 1; > + if (s->channels[i].attack < 0) > + return AVERROR(EINVAL); > + } > + nb_attacks = new_nb_items; > + > + p = s->decays; > + for (i = 0, new_nb_items = 0; i < nb_decays; i++) { > + char *tstr = av_strtok(p, " ", &saveptr); > + p = NULL; > + new_nb_items += sscanf(tstr, "%lf", &s->channels[i].decay) == 1; > + if (s->channels[i].decay < 0) > + return AVERROR(EINVAL); > + } > + nb_decays = new_nb_items; > + > + if (nb_attacks != nb_decays) { > + av_log(ctx, AV_LOG_ERROR, "Number of attacks %d differs from > number of decays %d.\n", nb_attacks, nb_decays); > + return AVERROR(EINVAL); > + } > + > +#define S(x) s->segments[2 * ((x) + 1)] > + p = s->points; > + for (i = 0, new_nb_items = 0; i < nb_points; i++) { > + char *tstr = av_strtok(p, " ", &saveptr); > + p = NULL; > + if (sscanf(tstr, "%lf/%lf", &S(i).x, &S(i).y) != 2) { > + av_log(ctx, AV_LOG_ERROR, "Invalid and/or missing > input/output value.\n"); > + return AVERROR(EINVAL); > + } > + if (i && S(i - 1).x > S(i).x) { > + av_log(ctx, AV_LOG_ERROR, "Transfer function input values > must be increasing.\n"); > + return AVERROR(EINVAL); > + } > + S(i).y -= S(i).x; > + av_log(ctx, AV_LOG_DEBUG, "%d: x=%f y=%f\n", i, S(i).x, S(i).y); > + new_nb_items++; > + } > + num = new_nb_items; > + > + /* Add 0,0 if necessary */ > + if (num == 0 || S(num - 1).x) > + num++; > + > +#undef S > +#define S(x) s->segments[2 * (x)] > + /* Add a tail off segment at the start */ > + S(0).x = S(1).x - 2 * s->curve_dB; > + S(0).y = S(1).y; > + num++; > + > + /* Join adjacent colinear segments */ > + for (i = 2; i < num; i++) { > + double g1 = (S(i - 1).y - S(i - 2).y) * (S(i - 0).x - S(i - 1).x); > + double g2 = (S(i - 0).y - S(i - 1).y) * (S(i - 1).x - S(i - 2).x); > + int j; > + > + if (fabs(g1 - g2)) > + continue; > + num--; > + for (j = --i; j < num; j++) > + S(j) = S(j + 1); > + } > + > + for (i = 0; !i || s->segments[i - 2].x; i += 2) { > + s->segments[i].y += s->gain_dB; > + s->segments[i].x *= M_LN10 / 20; > + s->segments[i].y *= M_LN10 / 20; > + } > + > +#define L(x) s->segments[i - (x)] > + for (i = 4; s->segments[i - 2].x; i += 2) { > + double x, y, cx, cy, in1, in2, out1, out2, theta, len, r; > + > + L(4).a = 0; > + L(4).b = (L(2).y - L(4).y) / (L(2).x - L(4).x); > + > + L(2).a = 0; > + L(2).b = (L(0).y - L(2).y) / (L(0).x - L(2).x); > + > + theta = atan2(L(2).y - L(4).y, L(2).x - L(4).x); > + len = sqrt(pow(L(2).x - L(4).x, 2.) + pow(L(2).y - L(4).y, 2.)); > + r = FFMIN(radius, len); > + L(3).x = L(2).x - r * cos(theta); > + L(3).y = L(2).y - r * sin(theta); > + > + theta = atan2(L(0).y - L(2).y, L(0).x - L(2).x); > + len = sqrt(pow(L(0).x - L(2).x, 2.) + pow(L(0).y - L(2).y, 2.)); > + r = FFMIN(radius, len / 2); > + x = L(2).x + r * cos(theta); > + y = L(2).y + r * sin(theta); > + > + cx = (L(3).x + L(2).x + x) / 3; > + cy = (L(3).y + L(2).y + y) / 3; > + > + L(2).x = x; > + L(2).y = y; > + > + in1 = cx - L(3).x; > + out1 = cy - L(3).y; > + in2 = L(2).x - L(3).x; > + out2 = L(2).y - L(3).y; > + L(3).a = (out2 / in2 - out1 / in1) / (in2-in1); > + L(3).b = out1 / in1 - L(3).a * in1; > + } > + L(3).x = 0; > + L(3).y = L(2).y; > + > + s->in_min_lin = exp(s->segments[1].x); > + s->out_min_lin = exp(s->segments[1].y); > + > + for (i = 0; i < channels; i++) { > + ChanParam *cp = &s->channels[i]; > + > + if (cp->attack > 1.0 / sample_rate) > + cp->attack = 1.0 - exp(-1.0 / (sample_rate * cp->attack)); > + else > + cp->attack = 1.0; > + if (cp->decay > 1.0 / sample_rate) > + cp->decay = 1.0 - exp(-1.0 / (sample_rate * cp->decay)); > + else > + cp->decay = 1.0; > + cp->volume = pow(10.0, s->initial_volume / 20); > + } > + > + s->delay_samples = s->delay * sample_rate; > + if (s->delay_samples > 0) { > + int ret; > + if ((ret = av_samples_alloc_array_and_samples(&s->delayptrs, NULL, > + channels, > + s->delay_samples, > + outlink->format, > 0)) < 0) > + return ret; > + s->compand = compand_delay; > + } else { > + s->compand = compand_nodelay; > + } > + return 0; > +} > + > +static int filter_frame(AVFilterLink *inlink, AVFrame *frame) > +{ > + AVFilterContext *ctx = inlink->dst; > + CompandContext *s = ctx->priv; > + > + return s->compand(ctx, frame); > +} > + > +static int request_frame(AVFilterLink *outlink) > +{ > + AVFilterContext *ctx = outlink->src; > + CompandContext *s = ctx->priv; > + int ret; > + > + ret = ff_request_frame(ctx->inputs[0]); > + > + if (ret == AVERROR_EOF && s->delay_count) > + ret = compand_drain(outlink); > + > + return ret; > +} > + > +static const AVFilterPad compand_inputs[] = { > + { > + .name = "default", > + .type = AVMEDIA_TYPE_AUDIO, > + .filter_frame = filter_frame, > + }, > + { NULL } > +}; > + > +static const AVFilterPad compand_outputs[] = { > + { > + .name = "default", > + .request_frame = request_frame, > + .config_props = config_output, > + .type = AVMEDIA_TYPE_AUDIO, > + }, > + { NULL } > +}; > + > + > +AVFilter ff_af_compand = { > + .name = "compand", > + .description = NULL_IF_CONFIG_SMALL("Compress or expand audio > dynamic range."), > + .query_formats = query_formats, > + .priv_size = sizeof(CompandContext), > + .priv_class = &compand_class, > + .init = init, > + .inputs = compand_inputs, > + .outputs = compand_outputs, > +}; > diff --git a/libavfilter/allfilters.c b/libavfilter/allfilters.c > index 9702a0a..e47a22e 100644 > --- a/libavfilter/allfilters.c > +++ b/libavfilter/allfilters.c > @@ -54,6 +54,7 @@ void avfilter_register_all(void) > REGISTER_FILTER(ATRIM, atrim, af); > REGISTER_FILTER(CHANNELMAP, channelmap, af); > REGISTER_FILTER(CHANNELSPLIT, channelsplit, af); > + REGISTER_FILTER(COMPAND, compand, af); > REGISTER_FILTER(JOIN, join, af); > REGISTER_FILTER(RESAMPLE, resample, af); > REGISTER_FILTER(VOLUME, volume, af); > -- > 1.8.3.2 > > _______________________________________________ libav-devel mailing list libav-devel@libav.org https://lists.libav.org/mailman/listinfo/libav-devel