/* * Copyright (c) 2012 Pavel Koshevoy * * This file is part of FFmpeg. * * FFmpeg is free software; you can redistribute it and/or * modify it under the terms of the GNU Lesser General Public * License as published by the Free Software Foundation; either * version 2.1 of the License, or (at your option) any later version. * * FFmpeg is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * Lesser General Public License for more details. * * You should have received a copy of the GNU Lesser General Public * License along with FFmpeg; if not, write to the Free Software * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA */ /** * @file * tempo scaling audio filter -- an implementation of WSOLA algorithm * * Based on MIT licensed yaeAudioTempoFilter.h and yaeAudioFragment.h * from Apprentice Video player by Pavel Koshevoy. * https://sourceforge.net/projects/apprenticevideo/ * * An explanation of SOLA algorithm is available at * http://www.surina.net/article/time-and-pitch-scaling.html * * WSOLA is very similar to SOLA, only one major difference exists between * these algorithms. SOLA shifts audio fragments along the output stream, * where as WSOLA shifts audio fragments along the input stream. * * The advantage of WSOLA algorithm is that the overlap region size is * always the same, therefore the blending function is constant and * can be precomputed. */ #include #include "libavcodec/avfft.h" #include "libavutil/avassert.h" #include "libavutil/avstring.h" #include "libavutil/channel_layout.h" #include "libavutil/eval.h" #include "libavutil/opt.h" #include "libavutil/samplefmt.h" #include "avfilter.h" #include "audio.h" #include "internal.h" /** * A fragment of audio waveform */ typedef struct AudioFragment { // index of the first sample of this fragment in the overall waveform; // 0: input sample position // 1: output sample position int64_t position[2]; // original packed multi-channel samples: uint8_t *data; // number of samples in this fragment: int nsamples; // rDFT transform of the down-mixed mono fragment, used for // fast waveform alignment via correlation in frequency domain: FFTSample *xdat; } AudioFragment; /** * Filter state machine states */ typedef enum { YAE_LOAD_FRAGMENT, YAE_ADJUST_POSITION, YAE_RELOAD_FRAGMENT, YAE_OUTPUT_OVERLAP_ADD, YAE_FLUSH_OUTPUT, } FilterState; /** * Filter state machine */ typedef struct ATempoContext { const AVClass *class; // ring-buffer of input samples, necessary because some times // input fragment position may be adjusted backwards: uint8_t *buffer; // ring-buffer maximum capacity, expressed in sample rate time base: int ring; // ring-buffer house keeping: int size; int head; int tail; // 0: input sample position corresponding to the ring buffer tail // 1: output sample position int64_t position[2]; // first input timestamp, all other timestamps are offset by this one int64_t start_pts; // sample format: enum AVSampleFormat format; // number of channels: int channels; // row of bytes to skip from one sample to next, across multple channels; // stride = (number-of-channels * bits-per-sample-per-channel) / 8 int stride; // fragment window size, power-of-two integer: int window; // Hann window coefficients, for feathering // (blending) the overlapping fragment region: float *hann; // tempo scaling factor: double tempo; // a snapshot of previous fragment input and output position values // captured when the tempo scale factor was set most recently: int64_t origin[2]; // current/previous fragment ring-buffer: AudioFragment frag[2]; // current fragment index: uint64_t nfrag; // current state: FilterState state; // for fast correlation calculation in frequency domain: RDFTContext *real_to_complex; RDFTContext *complex_to_real; FFTSample *correlation; // for managing AVFilterPad.request_frame and AVFilterPad.filter_frame AVFrame *dst_buffer; uint8_t *dst; uint8_t *dst_end; uint64_t nsamples_in; uint64_t nsamples_out; } ATempoContext; #define YAE_ATEMPO_MIN 0.5 #define YAE_ATEMPO_MAX 100.0 #define OFFSET(x) offsetof(ATempoContext, x) static const AVOption atempo_options[] = { { "tempo", "set tempo scale factor", OFFSET(tempo), AV_OPT_TYPE_DOUBLE, { .dbl = 1.0 }, YAE_ATEMPO_MIN, YAE_ATEMPO_MAX, AV_OPT_FLAG_AUDIO_PARAM | AV_OPT_FLAG_FILTERING_PARAM | AV_OPT_FLAG_RUNTIME_PARAM }, { NULL } }; AVFILTER_DEFINE_CLASS(atempo); inline static AudioFragment *yae_curr_frag(ATempoContext *atempo) { return &atempo->frag[atempo->nfrag % 2]; } inline static AudioFragment *yae_prev_frag(ATempoContext *atempo) { return &atempo->frag[(atempo->nfrag + 1) % 2]; } /** * Reset filter to initial state, do not deallocate existing local buffers. */ static void yae_clear(ATempoContext *atempo) { atempo->size = 0; atempo->head = 0; atempo->tail = 0; atempo->nfrag = 0; atempo->state = YAE_LOAD_FRAGMENT; atempo->start_pts = AV_NOPTS_VALUE; atempo->position[0] = 0; atempo->position[1] = 0; atempo->origin[0] = 0; atempo->origin[1] = 0; atempo->frag[0].position[0] = 0; atempo->frag[0].position[1] = 0; atempo->frag[0].nsamples = 0; atempo->frag[1].position[0] = 0; atempo->frag[1].position[1] = 0; atempo->frag[1].nsamples = 0; // shift left position of 1st fragment by half a window // so that no re-normalization would be required for // the left half of the 1st fragment: atempo->frag[0].position[0] = -(int64_t)(atempo->window / 2); atempo->frag[0].position[1] = -(int64_t)(atempo->window / 2); av_frame_free(&atempo->dst_buffer); atempo->dst = NULL; atempo->dst_end = NULL; atempo->nsamples_in = 0; atempo->nsamples_out = 0; } /** * Reset filter to initial state and deallocate all buffers. */ static void yae_release_buffers(ATempoContext *atempo) { yae_clear(atempo); av_freep(&atempo->frag[0].data); av_freep(&atempo->frag[1].data); av_freep(&atempo->frag[0].xdat); av_freep(&atempo->frag[1].xdat); av_freep(&atempo->buffer); av_freep(&atempo->hann); av_freep(&atempo->correlation); av_rdft_end(atempo->real_to_complex); atempo->real_to_complex = NULL; av_rdft_end(atempo->complex_to_real); atempo->complex_to_real = NULL; } /* av_realloc is not aligned enough; fortunately, the data does not need to * be preserved */ #define RE_MALLOC_OR_FAIL(field, field_size) \ do { \ av_freep(&field); \ field = av_malloc(field_size); \ if (!field) { \ yae_release_buffers(atempo); \ return AVERROR(ENOMEM); \ } \ } while (0) /** * Prepare filter for processing audio data of given format, * sample rate and number of channels. */ static int yae_reset(ATempoContext *atempo, enum AVSampleFormat format, int sample_rate, int channels) { const int sample_size = av_get_bytes_per_sample(format); uint32_t nlevels = 0; uint32_t pot; int i; atempo->format = format; atempo->channels = channels; atempo->stride = sample_size * channels; // pick a segment window size: atempo->window = sample_rate / 24; // adjust window size to be a power-of-two integer: nlevels = av_log2(atempo->window); pot = 1 << nlevels; av_assert0(pot <= atempo->window); if (pot < atempo->window) { atempo->window = pot * 2; nlevels++; } // initialize audio fragment buffers: RE_MALLOC_OR_FAIL(atempo->frag[0].data, atempo->window * atempo->stride); RE_MALLOC_OR_FAIL(atempo->frag[1].data, atempo->window * atempo->stride); RE_MALLOC_OR_FAIL(atempo->frag[0].xdat, atempo->window * sizeof(FFTComplex)); RE_MALLOC_OR_FAIL(atempo->frag[1].xdat, atempo->window * sizeof(FFTComplex)); // initialize rDFT contexts: av_rdft_end(atempo->real_to_complex); atempo->real_to_complex = NULL; av_rdft_end(atempo->complex_to_real); atempo->complex_to_real = NULL; atempo->real_to_complex = av_rdft_init(nlevels + 1, DFT_R2C); if (!atempo->real_to_complex) { yae_release_buffers(atempo); return AVERROR(ENOMEM); } atempo->complex_to_real = av_rdft_init(nlevels + 1, IDFT_C2R); if (!atempo->complex_to_real) { yae_release_buffers(atempo); return AVERROR(ENOMEM); } RE_MALLOC_OR_FAIL(atempo->correlation, atempo->window * sizeof(FFTComplex)); atempo->ring = atempo->window * 3; RE_MALLOC_OR_FAIL(atempo->buffer, atempo->ring * atempo->stride); // initialize the Hann window function: RE_MALLOC_OR_FAIL(atempo->hann, atempo->window * sizeof(float)); for (i = 0; i < atempo->window; i++) { double t = (double)i / (double)(atempo->window - 1); double h = 0.5 * (1.0 - cos(2.0 * M_PI * t)); atempo->hann[i] = (float)h; } yae_clear(atempo); return 0; } static int yae_update(AVFilterContext *ctx) { const AudioFragment *prev; ATempoContext *atempo = ctx->priv; prev = yae_prev_frag(atempo); atempo->origin[0] = prev->position[0] + atempo->window / 2; atempo->origin[1] = prev->position[1] + atempo->window / 2; return 0; } /** * A helper macro for initializing complex data buffer with scalar data * of a given type. */ #define yae_init_xdat(scalar_type, scalar_max) \ do { \ const uint8_t *src_end = src + \ frag->nsamples * atempo->channels * sizeof(scalar_type); \ \ FFTSample *xdat = frag->xdat; \ scalar_type tmp; \ \ if (atempo->channels == 1) { \ for (; src < src_end; xdat++) { \ tmp = *(const scalar_type *)src; \ src += sizeof(scalar_type); \ \ *xdat = (FFTSample)tmp; \ } \ } else { \ FFTSample s, max, ti, si; \ int i; \ \ for (; src < src_end; xdat++) { \ tmp = *(const scalar_type *)src; \ src += sizeof(scalar_type); \ \ max = (FFTSample)tmp; \ s = FFMIN((FFTSample)scalar_max, \ (FFTSample)fabsf(max)); \ \ for (i = 1; i < atempo->channels; i++) { \ tmp = *(const scalar_type *)src; \ src += sizeof(scalar_type); \ \ ti = (FFTSample)tmp; \ si = FFMIN((FFTSample)scalar_max, \ (FFTSample)fabsf(ti)); \ \ if (s < si) { \ s = si; \ max = ti; \ } \ } \ \ *xdat = max; \ } \ } \ } while (0) /** * Initialize complex data buffer of a given audio fragment * with down-mixed mono data of appropriate scalar type. */ static void yae_downmix(ATempoContext *atempo, AudioFragment *frag) { // shortcuts: const uint8_t *src = frag->data; // init complex data buffer used for FFT and Correlation: memset(frag->xdat, 0, sizeof(FFTComplex) * atempo->window); if (atempo->format == AV_SAMPLE_FMT_U8) { yae_init_xdat(uint8_t, 127); } else if (atempo->format == AV_SAMPLE_FMT_S16) { yae_init_xdat(int16_t, 32767); } else if (atempo->format == AV_SAMPLE_FMT_S32) { yae_init_xdat(int, 2147483647); } else if (atempo->format == AV_SAMPLE_FMT_FLT) { yae_init_xdat(float, 1); } else if (atempo->format == AV_SAMPLE_FMT_DBL) { yae_init_xdat(double, 1); } } /** * Populate the internal data buffer on as-needed basis. * * @return * 0 if requested data was already available or was successfully loaded, * AVERROR(EAGAIN) if more input data is required. */ static int yae_load_data(ATempoContext *atempo, const uint8_t **src_ref, const uint8_t *src_end, int64_t stop_here) { // shortcut: const uint8_t *src = *src_ref; const int read_size = stop_here - atempo->position[0]; if (stop_here <= atempo->position[0]) { return 0; } // samples are not expected to be skipped, unless tempo is greater than 2: av_assert0(read_size <= atempo->ring || atempo->tempo > 2.0); while (atempo->position[0] < stop_here && src < src_end) { int src_samples = (src_end - src) / atempo->stride; // load data piece-wise, in order to avoid complicating the logic: int nsamples = FFMIN(read_size, src_samples); int na; int nb; nsamples = FFMIN(nsamples, atempo->ring); na = FFMIN(nsamples, atempo->ring - atempo->tail); nb = FFMIN(nsamples - na, atempo->ring); if (na) { uint8_t *a = atempo->buffer + atempo->tail * atempo->stride; memcpy(a, src, na * atempo->stride); src += na * atempo->stride; atempo->position[0] += na; atempo->size = FFMIN(atempo->size + na, atempo->ring); atempo->tail = (atempo->tail + na) % atempo->ring; atempo->head = atempo->size < atempo->ring ? atempo->tail - atempo->size : atempo->tail; } if (nb) { uint8_t *b = atempo->buffer; memcpy(b, src, nb * atempo->stride); src += nb * atempo->stride; atempo->position[0] += nb; atempo->size = FFMIN(atempo->size + nb, atempo->ring); atempo->tail = (atempo->tail + nb) % atempo->ring; atempo->head = atempo->size < atempo->ring ? atempo->tail - atempo->size : atempo->tail; } } // pass back the updated source buffer pointer: *src_ref = src; // sanity check: av_assert0(atempo->position[0] <= stop_here); return atempo->position[0] == stop_here ? 0 : AVERROR(EAGAIN); } /** * Populate current audio fragment data buffer. * * @return * 0 when the fragment is ready, * AVERROR(EAGAIN) if more input data is required. */ static int yae_load_frag(ATempoContext *atempo, const uint8_t **src_ref, const uint8_t *src_end) { // shortcuts: AudioFragment *frag = yae_curr_frag(atempo); uint8_t *dst; int64_t missing, start, zeros; uint32_t nsamples; const uint8_t *a, *b; int i0, i1, n0, n1, na, nb; int64_t stop_here = frag->position[0] + atempo->window; if (src_ref && yae_load_data(atempo, src_ref, src_end, stop_here) != 0) { return AVERROR(EAGAIN); } // calculate the number of samples we don't have: missing = stop_here > atempo->position[0] ? stop_here - atempo->position[0] : 0; nsamples = missing < (int64_t)atempo->window ? (uint32_t)(atempo->window - missing) : 0; // setup the output buffer: frag->nsamples = nsamples; dst = frag->data; start = atempo->position[0] - atempo->size; zeros = 0; if (frag->position[0] < start) { // what we don't have we substitute with zeros: zeros = FFMIN(start - frag->position[0], (int64_t)nsamples); av_assert0(zeros != nsamples); memset(dst, 0, zeros * atempo->stride); dst += zeros * atempo->stride; } if (zeros == nsamples) { return 0; } // get the remaining data from the ring buffer: na = (atempo->head < atempo->tail ? atempo->tail - atempo->head : atempo->ring - atempo->head); nb = atempo->head < atempo->tail ? 0 : atempo->tail; // sanity check: av_assert0(nsamples <= zeros + na + nb); a = atempo->buffer + atempo->head * atempo->stride; b = atempo->buffer; i0 = frag->position[0] + zeros - start; i1 = i0 < na ? 0 : i0 - na; n0 = i0 < na ? FFMIN(na - i0, (int)(nsamples - zeros)) : 0; n1 = nsamples - zeros - n0; if (n0) { memcpy(dst, a + i0 * atempo->stride, n0 * atempo->stride); dst += n0 * atempo->stride; } if (n1) { memcpy(dst, b + i1 * atempo->stride, n1 * atempo->stride); } return 0; } /** * Prepare for loading next audio fragment. */ static void yae_advance_to_next_frag(ATempoContext *atempo) { const double fragment_step = atempo->tempo * (double)(atempo->window / 2); const AudioFragment *prev; AudioFragment *frag; atempo->nfrag++; prev = yae_prev_frag(atempo); frag = yae_curr_frag(atempo); frag->position[0] = prev->position[0] + (int64_t)fragment_step; frag->position[1] = prev->position[1] + atempo->window / 2; frag->nsamples = 0; } /** * Calculate cross-correlation via rDFT. * * Multiply two vectors of complex numbers (result of real_to_complex rDFT) * and transform back via complex_to_real rDFT. */ static void yae_xcorr_via_rdft(FFTSample *xcorr, RDFTContext *complex_to_real, const FFTComplex *xa, const FFTComplex *xb, const int window) { FFTComplex *xc = (FFTComplex *)xcorr; int i; // NOTE: first element requires special care -- Given Y = rDFT(X), // Im(Y[0]) and Im(Y[N/2]) are always zero, therefore av_rdft_calc // stores Re(Y[N/2]) in place of Im(Y[0]). xc->re = xa->re * xb->re; xc->im = xa->im * xb->im; xa++; xb++; xc++; for (i = 1; i < window; i++, xa++, xb++, xc++) { xc->re = (xa->re * xb->re + xa->im * xb->im); xc->im = (xa->im * xb->re - xa->re * xb->im); } // apply inverse rDFT: av_rdft_calc(complex_to_real, xcorr); } /** * Calculate alignment offset for given fragment * relative to the previous fragment. * * @return alignment offset of current fragment relative to previous. */ static int yae_align(AudioFragment *frag, const AudioFragment *prev, const int window, const int delta_max, const int drift, FFTSample *correlation, RDFTContext *complex_to_real) { int best_offset = -drift; FFTSample best_metric = -FLT_MAX; FFTSample *xcorr; int i0; int i1; int i; yae_xcorr_via_rdft(correlation, complex_to_real, (const FFTComplex *)prev->xdat, (const FFTComplex *)frag->xdat, window); // identify search window boundaries: i0 = FFMAX(window / 2 - delta_max - drift, 0); i0 = FFMIN(i0, window); i1 = FFMIN(window / 2 + delta_max - drift, window - window / 16); i1 = FFMAX(i1, 0); // identify cross-correlation peaks within search window: xcorr = correlation + i0; for (i = i0; i < i1; i++, xcorr++) { FFTSample metric = *xcorr; // normalize: FFTSample drifti = (FFTSample)(drift + i); metric *= drifti * (FFTSample)(i - i0) * (FFTSample)(i1 - i); if (metric > best_metric) { best_metric = metric; best_offset = i - window / 2; } } return best_offset; } /** * Adjust current fragment position for better alignment * with previous fragment. * * @return alignment correction. */ static int yae_adjust_position(ATempoContext *atempo) { const AudioFragment *prev = yae_prev_frag(atempo); AudioFragment *frag = yae_curr_frag(atempo); const double prev_output_position = (double)(prev->position[1] - atempo->origin[1] + atempo->window / 2) * atempo->tempo; const double ideal_output_position = (double)(prev->position[0] - atempo->origin[0] + atempo->window / 2); const int drift = (int)(prev_output_position - ideal_output_position); const int delta_max = atempo->window / 2; const int correction = yae_align(frag, prev, atempo->window, delta_max, drift, atempo->correlation, atempo->complex_to_real); if (correction) { // adjust fragment position: frag->position[0] -= correction; // clear so that the fragment can be reloaded: frag->nsamples = 0; } return correction; } /** * A helper macro for blending the overlap region of previous * and current audio fragment. */ #define yae_blend(scalar_type) \ do { \ const scalar_type *aaa = (const scalar_type *)a; \ const scalar_type *bbb = (const scalar_type *)b; \ \ scalar_type *out = (scalar_type *)dst; \ scalar_type *out_end = (scalar_type *)dst_end; \ int64_t i; \ \ for (i = 0; i < overlap && out < out_end; \ i++, atempo->position[1]++, wa++, wb++) { \ float w0 = *wa; \ float w1 = *wb; \ int j; \ \ for (j = 0; j < atempo->channels; \ j++, aaa++, bbb++, out++) { \ float t0 = (float)*aaa; \ float t1 = (float)*bbb; \ \ *out = \ frag->position[0] + i < 0 ? \ *aaa : \ (scalar_type)(t0 * w0 + t1 * w1); \ } \ } \ dst = (uint8_t *)out; \ } while (0) /** * Blend the overlap region of previous and current audio fragment * and output the results to the given destination buffer. * * @return * 0 if the overlap region was completely stored in the dst buffer, * AVERROR(EAGAIN) if more destination buffer space is required. */ static int yae_overlap_add(ATempoContext *atempo, uint8_t **dst_ref, uint8_t *dst_end) { // shortcuts: const AudioFragment *prev = yae_prev_frag(atempo); const AudioFragment *frag = yae_curr_frag(atempo); const int64_t start_here = FFMAX(atempo->position[1], frag->position[1]); const int64_t stop_here = FFMIN(prev->position[1] + prev->nsamples, frag->position[1] + frag->nsamples); const int64_t overlap = stop_here - start_here; const int64_t ia = start_here - prev->position[1]; const int64_t ib = start_here - frag->position[1]; const float *wa = atempo->hann + ia; const float *wb = atempo->hann + ib; const uint8_t *a = prev->data + ia * atempo->stride; const uint8_t *b = frag->data + ib * atempo->stride; uint8_t *dst = *dst_ref; av_assert0(start_here <= stop_here && frag->position[1] <= start_here && overlap <= frag->nsamples); if (atempo->format == AV_SAMPLE_FMT_U8) { yae_blend(uint8_t); } else if (atempo->format == AV_SAMPLE_FMT_S16) { yae_blend(int16_t); } else if (atempo->format == AV_SAMPLE_FMT_S32) { yae_blend(int); } else if (atempo->format == AV_SAMPLE_FMT_FLT) { yae_blend(float); } else if (atempo->format == AV_SAMPLE_FMT_DBL) { yae_blend(double); } // pass-back the updated destination buffer pointer: *dst_ref = dst; return atempo->position[1] == stop_here ? 0 : AVERROR(EAGAIN); } /** * Feed as much data to the filter as it is able to consume * and receive as much processed data in the destination buffer * as it is able to produce or store. */ static void yae_apply(ATempoContext *atempo, const uint8_t **src_ref, const uint8_t *src_end, uint8_t **dst_ref, uint8_t *dst_end) { while (1) { if (atempo->state == YAE_LOAD_FRAGMENT) { // load additional data for the current fragment: if (yae_load_frag(atempo, src_ref, src_end) != 0) { break; } // down-mix to mono: yae_downmix(atempo, yae_curr_frag(atempo)); // apply rDFT: av_rdft_calc(atempo->real_to_complex, yae_curr_frag(atempo)->xdat); // must load the second fragment before alignment can start: if (!atempo->nfrag) { yae_advance_to_next_frag(atempo); continue; } atempo->state = YAE_ADJUST_POSITION; } if (atempo->state == YAE_ADJUST_POSITION) { // adjust position for better alignment: if (yae_adjust_position(atempo)) { // reload the fragment at the corrected position, so that the // Hann window blending would not require normalization: atempo->state = YAE_RELOAD_FRAGMENT; } else { atempo->state = YAE_OUTPUT_OVERLAP_ADD; } } if (atempo->state == YAE_RELOAD_FRAGMENT) { // load additional data if necessary due to position adjustment: if (yae_load_frag(atempo, src_ref, src_end) != 0) { break; } // down-mix to mono: yae_downmix(atempo, yae_curr_frag(atempo)); // apply rDFT: av_rdft_calc(atempo->real_to_complex, yae_curr_frag(atempo)->xdat); atempo->state = YAE_OUTPUT_OVERLAP_ADD; } if (atempo->state == YAE_OUTPUT_OVERLAP_ADD) { // overlap-add and output the result: if (yae_overlap_add(atempo, dst_ref, dst_end) != 0) { break; } // advance to the next fragment, repeat: yae_advance_to_next_frag(atempo); atempo->state = YAE_LOAD_FRAGMENT; } } } /** * Flush any buffered data from the filter. * * @return * 0 if all data was completely stored in the dst buffer, * AVERROR(EAGAIN) if more destination buffer space is required. */ static int yae_flush(ATempoContext *atempo, uint8_t **dst_ref, uint8_t *dst_end) { AudioFragment *frag = yae_curr_frag(atempo); int64_t overlap_end; int64_t start_here; int64_t stop_here; int64_t offset; const uint8_t *src; uint8_t *dst; int src_size; int dst_size; int nbytes; atempo->state = YAE_FLUSH_OUTPUT; if (!atempo->nfrag) { // there is nothing to flush: return 0; } if (atempo->position[0] == frag->position[0] + frag->nsamples && atempo->position[1] == frag->position[1] + frag->nsamples) { // the current fragment is already flushed: return 0; } if (frag->position[0] + frag->nsamples < atempo->position[0]) { // finish loading the current (possibly partial) fragment: yae_load_frag(atempo, NULL, NULL); if (atempo->nfrag) { // down-mix to mono: yae_downmix(atempo, frag); // apply rDFT: av_rdft_calc(atempo->real_to_complex, frag->xdat); // align current fragment to previous fragment: if (yae_adjust_position(atempo)) { // reload the current fragment due to adjusted position: yae_load_frag(atempo, NULL, NULL); } } } // flush the overlap region: overlap_end = frag->position[1] + FFMIN(atempo->window / 2, frag->nsamples); while (atempo->position[1] < overlap_end) { if (yae_overlap_add(atempo, dst_ref, dst_end) != 0) { return AVERROR(EAGAIN); } } // check whether all of the input samples have been consumed: if (frag->position[0] + frag->nsamples < atempo->position[0]) { yae_advance_to_next_frag(atempo); return AVERROR(EAGAIN); } // flush the remainder of the current fragment: start_here = FFMAX(atempo->position[1], overlap_end); stop_here = frag->position[1] + frag->nsamples; offset = start_here - frag->position[1]; av_assert0(start_here <= stop_here && frag->position[1] <= start_here); src = frag->data + offset * atempo->stride; dst = (uint8_t *)*dst_ref; src_size = (int)(stop_here - start_here) * atempo->stride; dst_size = dst_end - dst; nbytes = FFMIN(src_size, dst_size); memcpy(dst, src, nbytes); dst += nbytes; atempo->position[1] += (nbytes / atempo->stride); // pass-back the updated destination buffer pointer: *dst_ref = (uint8_t *)dst; return atempo->position[1] == stop_here ? 0 : AVERROR(EAGAIN); } static av_cold int init(AVFilterContext *ctx) { ATempoContext *atempo = ctx->priv; atempo->format = AV_SAMPLE_FMT_NONE; atempo->state = YAE_LOAD_FRAGMENT; return 0; } static av_cold void uninit(AVFilterContext *ctx) { ATempoContext *atempo = ctx->priv; yae_release_buffers(atempo); } static int query_formats(AVFilterContext *ctx) { AVFilterChannelLayouts *layouts = NULL; AVFilterFormats *formats = NULL; // WSOLA necessitates an internal sliding window ring buffer // for incoming audio stream. // // Planar sample formats are too cumbersome to store in a ring buffer, // therefore planar sample formats are not supported. // static const enum AVSampleFormat sample_fmts[] = { AV_SAMPLE_FMT_U8, AV_SAMPLE_FMT_S16, AV_SAMPLE_FMT_S32, AV_SAMPLE_FMT_FLT, AV_SAMPLE_FMT_DBL, AV_SAMPLE_FMT_NONE }; int ret; layouts = ff_all_channel_counts(); if (!layouts) { return AVERROR(ENOMEM); } ret = ff_set_common_channel_layouts(ctx, layouts); if (ret < 0) return ret; formats = ff_make_format_list(sample_fmts); if (!formats) { return AVERROR(ENOMEM); } ret = ff_set_common_formats(ctx, formats); if (ret < 0) return ret; formats = ff_all_samplerates(); if (!formats) { return AVERROR(ENOMEM); } return ff_set_common_samplerates(ctx, formats); } static int config_props(AVFilterLink *inlink) { AVFilterContext *ctx = inlink->dst; ATempoContext *atempo = ctx->priv; enum AVSampleFormat format = inlink->format; int sample_rate = (int)inlink->sample_rate; return yae_reset(atempo, format, sample_rate, inlink->channels); } static int push_samples(ATempoContext *atempo, AVFilterLink *outlink, int n_out) { int ret; atempo->dst_buffer->sample_rate = outlink->sample_rate; atempo->dst_buffer->nb_samples = n_out; // adjust the PTS: atempo->dst_buffer->pts = atempo->start_pts + av_rescale_q(atempo->nsamples_out, (AVRational){ 1, outlink->sample_rate }, outlink->time_base); ret = ff_filter_frame(outlink, atempo->dst_buffer); atempo->dst_buffer = NULL; atempo->dst = NULL; atempo->dst_end = NULL; if (ret < 0) return ret; atempo->nsamples_out += n_out; return 0; } static int filter_frame(AVFilterLink *inlink, AVFrame *src_buffer) { AVFilterContext *ctx = inlink->dst; ATempoContext *atempo = ctx->priv; AVFilterLink *outlink = ctx->outputs[0]; int ret = 0; int n_in = src_buffer->nb_samples; int n_out = (int)(0.5 + ((double)n_in) / atempo->tempo); const uint8_t *src = src_buffer->data[0]; const uint8_t *src_end = src + n_in * atempo->stride; if (atempo->start_pts == AV_NOPTS_VALUE) atempo->start_pts = av_rescale_q(src_buffer->pts, inlink->time_base, outlink->time_base); while (src < src_end) { if (!atempo->dst_buffer) { atempo->dst_buffer = ff_get_audio_buffer(outlink, n_out); if (!atempo->dst_buffer) { av_frame_free(&src_buffer); return AVERROR(ENOMEM); } av_frame_copy_props(atempo->dst_buffer, src_buffer); atempo->dst = atempo->dst_buffer->data[0]; atempo->dst_end = atempo->dst + n_out * atempo->stride; } yae_apply(atempo, &src, src_end, &atempo->dst, atempo->dst_end); if (atempo->dst == atempo->dst_end) { int n_samples = ((atempo->dst - atempo->dst_buffer->data[0]) / atempo->stride); ret = push_samples(atempo, outlink, n_samples); if (ret < 0) goto end; } } atempo->nsamples_in += n_in; end: av_frame_free(&src_buffer); return ret; } static int request_frame(AVFilterLink *outlink) { AVFilterContext *ctx = outlink->src; ATempoContext *atempo = ctx->priv; int ret; ret = ff_request_frame(ctx->inputs[0]); if (ret == AVERROR_EOF) { // flush the filter: int n_max = atempo->ring; int n_out; int err = AVERROR(EAGAIN); while (err == AVERROR(EAGAIN)) { if (!atempo->dst_buffer) { atempo->dst_buffer = ff_get_audio_buffer(outlink, n_max); if (!atempo->dst_buffer) return AVERROR(ENOMEM); atempo->dst = atempo->dst_buffer->data[0]; atempo->dst_end = atempo->dst + n_max * atempo->stride; } err = yae_flush(atempo, &atempo->dst, atempo->dst_end); n_out = ((atempo->dst - atempo->dst_buffer->data[0]) / atempo->stride); if (n_out) { ret = push_samples(atempo, outlink, n_out); if (ret < 0) return ret; } } av_frame_free(&atempo->dst_buffer); atempo->dst = NULL; atempo->dst_end = NULL; return AVERROR_EOF; } return ret; } static int process_command(AVFilterContext *ctx, const char *cmd, const char *arg, char *res, int res_len, int flags) { int ret = ff_filter_process_command(ctx, cmd, arg, res, res_len, flags); if (ret < 0) return ret; return yae_update(ctx); } static const AVFilterPad atempo_inputs[] = { { .name = "default", .type = AVMEDIA_TYPE_AUDIO, .filter_frame = filter_frame, .config_props = config_props, }, { NULL } }; static const AVFilterPad atempo_outputs[] = { { .name = "default", .request_frame = request_frame, .type = AVMEDIA_TYPE_AUDIO, }, { NULL } }; AVFilter ff_af_atempo = { .name = "atempo", .description = NULL_IF_CONFIG_SMALL("Adjust audio tempo."), .init = init, .uninit = uninit, .query_formats = query_formats, .process_command = process_command, .priv_size = sizeof(ATempoContext), .priv_class = &atempo_class, .inputs = atempo_inputs, .outputs = atempo_outputs, };