diff options
Diffstat (limited to 'libavdevice/decklink_enc.cpp')
-rw-r--r-- | libavdevice/decklink_enc.cpp | 426 |
1 files changed, 426 insertions, 0 deletions
diff --git a/libavdevice/decklink_enc.cpp b/libavdevice/decklink_enc.cpp new file mode 100644 index 0000000..6c5450f --- /dev/null +++ b/libavdevice/decklink_enc.cpp @@ -0,0 +1,426 @@ +/* + * Blackmagic DeckLink output + * Copyright (c) 2013-2014 Ramiro Polla + * + * This file is part of FFmpeg. + * + * FFmpeg is free software; you can redistribute it and/or + * modify it under the terms of the GNU Lesser General Public + * License as published by the Free Software Foundation; either + * version 2.1 of the License, or (at your option) any later version. + * + * FFmpeg is distributed in the hope that it will be useful, + * but WITHOUT ANY WARRANTY; without even the implied warranty of + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU + * Lesser General Public License for more details. + * + * You should have received a copy of the GNU Lesser General Public + * License along with FFmpeg; if not, write to the Free Software + * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA + */ + +#include <DeckLinkAPI.h> + +#include <pthread.h> +#include <semaphore.h> + +extern "C" { +#include "libavformat/avformat.h" +#include "libavformat/internal.h" +#include "libavutil/imgutils.h" +} + +#include "decklink_common.h" +#include "decklink_enc.h" + + +/* DeckLink callback class declaration */ +class decklink_frame : public IDeckLinkVideoFrame +{ +public: + decklink_frame(struct decklink_ctx *ctx, AVFrame *avframe, long width, + long height, void *buffer) : + _ctx(ctx), _avframe(avframe), _width(width), + _height(height), _buffer(buffer), _refs(0) { } + + virtual long STDMETHODCALLTYPE GetWidth (void) { return _width; } + virtual long STDMETHODCALLTYPE GetHeight (void) { return _height; } + virtual long STDMETHODCALLTYPE GetRowBytes (void) { return _width<<1; } + virtual BMDPixelFormat STDMETHODCALLTYPE GetPixelFormat(void) { return bmdFormat8BitYUV; } + virtual BMDFrameFlags STDMETHODCALLTYPE GetFlags (void) { return bmdVideoOutputFlagDefault; } + virtual HRESULT STDMETHODCALLTYPE GetBytes (void **buffer) { *buffer = _buffer; return S_OK; } + + virtual HRESULT STDMETHODCALLTYPE GetTimecode (BMDTimecodeFormat format, IDeckLinkTimecode **timecode) { return S_FALSE; } + virtual HRESULT STDMETHODCALLTYPE GetAncillaryData(IDeckLinkVideoFrameAncillary **ancillary) { return S_FALSE; } + + virtual HRESULT STDMETHODCALLTYPE QueryInterface(REFIID iid, LPVOID *ppv) { return E_NOINTERFACE; } + virtual ULONG STDMETHODCALLTYPE AddRef(void) { return ++_refs; } + virtual ULONG STDMETHODCALLTYPE Release(void) { if (!--_refs) delete this; return _refs; } + + struct decklink_ctx *_ctx; + AVFrame *_avframe; + +private: + long _width; + long _height; + void *_buffer; + int _refs; +}; + +class decklink_output_callback : public IDeckLinkVideoOutputCallback +{ +public: + virtual HRESULT STDMETHODCALLTYPE ScheduledFrameCompleted(IDeckLinkVideoFrame *_frame, BMDOutputFrameCompletionResult result) + { + decklink_frame *frame = static_cast<decklink_frame *>(_frame); + struct decklink_ctx *ctx = frame->_ctx; + AVFrame *avframe = frame->_avframe; + + av_frame_free(&avframe); + + sem_post(&ctx->semaphore); + + return S_OK; + } + virtual HRESULT STDMETHODCALLTYPE ScheduledPlaybackHasStopped(void) { return S_OK; } + virtual HRESULT STDMETHODCALLTYPE QueryInterface(REFIID iid, LPVOID *ppv) { return E_NOINTERFACE; } + virtual ULONG STDMETHODCALLTYPE AddRef(void) { return 1; } + virtual ULONG STDMETHODCALLTYPE Release(void) { return 1; } +}; + +static int decklink_setup_video(AVFormatContext *avctx, AVStream *st) +{ + struct decklink_cctx *cctx = (struct decklink_cctx *) avctx->priv_data; + struct decklink_ctx *ctx = (struct decklink_ctx *) cctx->ctx; + AVCodecContext *c = st->codec; + + if (ctx->video) { + av_log(avctx, AV_LOG_ERROR, "Only one video stream is supported!\n"); + return -1; + } + + if (c->pix_fmt != AV_PIX_FMT_UYVY422) { + av_log(avctx, AV_LOG_ERROR, "Unsupported pixel format!" + " Only AV_PIX_FMT_UYVY422 is supported.\n"); + return -1; + } + if (ff_decklink_set_format(avctx, c->width, c->height, + c->time_base.num, c->time_base.den)) { + av_log(avctx, AV_LOG_ERROR, "Unsupported video size or framerate!" + " Check available formats with -list_formats 1.\n"); + return -1; + } + if (ctx->dlo->EnableVideoOutput(ctx->bmd_mode, + bmdVideoOutputFlagDefault) != S_OK) { + av_log(avctx, AV_LOG_ERROR, "Could not enable video output!\n"); + return -1; + } + + /* Set callback. */ + ctx->output_callback = new decklink_output_callback(); + ctx->dlo->SetScheduledFrameCompletionCallback(ctx->output_callback); + + /* Start video semaphore. */ + ctx->frames_preroll = c->time_base.den * ctx->preroll; + if (c->time_base.den > 1000) + ctx->frames_preroll /= 1000; + + /* Buffer twice as many frames as the preroll. */ + ctx->frames_buffer = ctx->frames_preroll * 2; + ctx->frames_buffer = FFMIN(ctx->frames_buffer, 60); + sem_init(&ctx->semaphore, 0, ctx->frames_buffer); + + /* The device expects the framerate to be fixed. */ + avpriv_set_pts_info(st, 64, c->time_base.num, c->time_base.den); + + ctx->video = 1; + + return 0; +} + +static int decklink_setup_audio(AVFormatContext *avctx, AVStream *st) +{ + struct decklink_cctx *cctx = (struct decklink_cctx *) avctx->priv_data; + struct decklink_ctx *ctx = (struct decklink_ctx *) cctx->ctx; + AVCodecContext *c = st->codec; + + if (ctx->audio) { + av_log(avctx, AV_LOG_ERROR, "Only one audio stream is supported!\n"); + return -1; + } + if (c->sample_rate != 48000) { + av_log(avctx, AV_LOG_ERROR, "Unsupported sample rate!" + " Only 48kHz is supported.\n"); + return -1; + } + if (c->channels != 2 && c->channels != 8) { + av_log(avctx, AV_LOG_ERROR, "Unsupported number of channels!" + " Only stereo and 7.1 are supported.\n"); + return -1; + } + if (ctx->dlo->EnableAudioOutput(bmdAudioSampleRate48kHz, + bmdAudioSampleType16bitInteger, + c->channels, + bmdAudioOutputStreamTimestamped) != S_OK) { + av_log(avctx, AV_LOG_ERROR, "Could not enable audio output!\n"); + return -1; + } + if (ctx->dlo->BeginAudioPreroll() != S_OK) { + av_log(avctx, AV_LOG_ERROR, "Could not begin audio preroll!\n"); + return -1; + } + + /* The device expects the sample rate to be fixed. */ + avpriv_set_pts_info(st, 64, 1, c->sample_rate); + ctx->channels = c->channels; + + ctx->audio = 1; + + return 0; +} + +av_cold int ff_decklink_write_trailer(AVFormatContext *avctx) +{ + struct decklink_cctx *cctx = (struct decklink_cctx *) avctx->priv_data; + struct decklink_ctx *ctx = (struct decklink_ctx *) cctx->ctx; + + if (ctx->playback_started) { + BMDTimeValue actual; + ctx->dlo->StopScheduledPlayback(ctx->last_pts * ctx->bmd_tb_num, + &actual, ctx->bmd_tb_den); + ctx->dlo->DisableVideoOutput(); + if (ctx->audio) + ctx->dlo->DisableAudioOutput(); + } + + if (ctx->dlo) + ctx->dlo->Release(); + if (ctx->dl) + ctx->dl->Release(); + + if (ctx->output_callback) + delete ctx->output_callback; + + sem_destroy(&ctx->semaphore); + + av_freep(&cctx->ctx); + + return 0; +} + +static int decklink_write_video_packet(AVFormatContext *avctx, AVPacket *pkt) +{ + struct decklink_cctx *cctx = (struct decklink_cctx *) avctx->priv_data; + struct decklink_ctx *ctx = (struct decklink_ctx *) cctx->ctx; + AVPicture *avpicture = (AVPicture *) pkt->data; + AVFrame *avframe, *tmp; + decklink_frame *frame; + buffercount_type buffered; + HRESULT hr; + + /* HACK while av_uncoded_frame() isn't implemented */ + int ret; + + tmp = av_frame_alloc(); + if (!tmp) + return AVERROR(ENOMEM); + tmp->format = AV_PIX_FMT_UYVY422; + tmp->width = ctx->bmd_width; + tmp->height = ctx->bmd_height; + ret = av_frame_get_buffer(tmp, 32); + if (ret < 0) { + av_frame_free(&tmp); + return ret; + } + av_image_copy(tmp->data, tmp->linesize, (const uint8_t **) avpicture->data, + avpicture->linesize, (AVPixelFormat) tmp->format, tmp->width, + tmp->height); + avframe = av_frame_clone(tmp); + av_frame_free(&tmp); + if (!avframe) { + av_log(avctx, AV_LOG_ERROR, "Could not clone video frame.\n"); + return AVERROR(EIO); + } + /* end HACK */ + + frame = new decklink_frame(ctx, avframe, ctx->bmd_width, ctx->bmd_height, + (void *) avframe->data[0]); + if (!frame) { + av_log(avctx, AV_LOG_ERROR, "Could not create new frame.\n"); + return AVERROR(EIO); + } + + /* Always keep at most one second of frames buffered. */ + sem_wait(&ctx->semaphore); + + /* Schedule frame for playback. */ + hr = ctx->dlo->ScheduleVideoFrame((struct IDeckLinkVideoFrame *) frame, + pkt->pts * ctx->bmd_tb_num, + ctx->bmd_tb_num, ctx->bmd_tb_den); + if (hr != S_OK) { + av_log(avctx, AV_LOG_ERROR, "Could not schedule video frame." + " error %08x.\n", (uint32_t) hr); + frame->Release(); + return AVERROR(EIO); + } + + ctx->dlo->GetBufferedVideoFrameCount(&buffered); + av_log(avctx, AV_LOG_DEBUG, "Buffered video frames: %d.\n", (int) buffered); + if (pkt->pts > 2 && buffered <= 2) + av_log(avctx, AV_LOG_WARNING, "There are not enough buffered video frames." + " Video may misbehave!\n"); + + /* Preroll video frames. */ + if (!ctx->playback_started && pkt->pts > ctx->frames_preroll) { + av_log(avctx, AV_LOG_DEBUG, "Ending audio preroll.\n"); + if (ctx->audio && ctx->dlo->EndAudioPreroll() != S_OK) { + av_log(avctx, AV_LOG_ERROR, "Could not end audio preroll!\n"); + return AVERROR(EIO); + } + av_log(avctx, AV_LOG_DEBUG, "Starting scheduled playback.\n"); + if (ctx->dlo->StartScheduledPlayback(0, ctx->bmd_tb_den, 1.0) != S_OK) { + av_log(avctx, AV_LOG_ERROR, "Could not start scheduled playback!\n"); + return AVERROR(EIO); + } + ctx->playback_started = 1; + } + + return 0; +} + +static int decklink_write_audio_packet(AVFormatContext *avctx, AVPacket *pkt) +{ + struct decklink_cctx *cctx = (struct decklink_cctx *) avctx->priv_data; + struct decklink_ctx *ctx = (struct decklink_ctx *) cctx->ctx; + int sample_count = pkt->size / (ctx->channels << 1); + buffercount_type buffered; + + ctx->dlo->GetBufferedAudioSampleFrameCount(&buffered); + if (pkt->pts > 1 && !buffered) + av_log(avctx, AV_LOG_WARNING, "There's no buffered audio." + " Audio will misbehave!\n"); + + if (ctx->dlo->ScheduleAudioSamples(pkt->data, sample_count, pkt->pts, + bmdAudioSampleRate48kHz, NULL) != S_OK) { + av_log(avctx, AV_LOG_ERROR, "Could not schedule audio samples.\n"); + return AVERROR(EIO); + } + + return 0; +} + +extern "C" { + +av_cold int ff_decklink_write_header(AVFormatContext *avctx) +{ + struct decklink_cctx *cctx = (struct decklink_cctx *) avctx->priv_data; + struct decklink_ctx *ctx; + IDeckLinkDisplayModeIterator *itermode; + IDeckLinkIterator *iter; + IDeckLink *dl = NULL; + unsigned int n; + + ctx = (struct decklink_ctx *) av_mallocz(sizeof(struct decklink_ctx)); + if (!ctx) + return AVERROR(ENOMEM); + ctx->list_devices = cctx->list_devices; + ctx->list_formats = cctx->list_formats; + ctx->preroll = cctx->preroll; + cctx->ctx = ctx; + + iter = CreateDeckLinkIteratorInstance(); + if (!iter) { + av_log(avctx, AV_LOG_ERROR, "Could not create DeckLink iterator\n"); + return AVERROR(EIO); + } + + /* List available devices. */ + if (ctx->list_devices) { + ff_decklink_list_devices(avctx); + return AVERROR_EXIT; + } + + /* Open device. */ + while (iter->Next(&dl) == S_OK) { + const char *displayName; + ff_decklink_get_display_name(dl, &displayName); + if (!strcmp(avctx->filename, displayName)) { + av_free((void *) displayName); + ctx->dl = dl; + break; + } + av_free((void *) displayName); + dl->Release(); + } + iter->Release(); + if (!ctx->dl) { + av_log(avctx, AV_LOG_ERROR, "Could not open '%s'\n", avctx->filename); + return AVERROR(EIO); + } + + /* Get output device. */ + if (ctx->dl->QueryInterface(IID_IDeckLinkOutput, (void **) &ctx->dlo) != S_OK) { + av_log(avctx, AV_LOG_ERROR, "Could not open output device from '%s'\n", + avctx->filename); + ctx->dl->Release(); + return AVERROR(EIO); + } + + /* List supported formats. */ + if (ctx->list_formats) { + ff_decklink_list_formats(avctx); + ctx->dlo->Release(); + ctx->dl->Release(); + return AVERROR_EXIT; + } + + if (ctx->dlo->GetDisplayModeIterator(&itermode) != S_OK) { + av_log(avctx, AV_LOG_ERROR, "Could not get Display Mode Iterator\n"); + ctx->dl->Release(); + return AVERROR(EIO); + } + + /* Setup streams. */ + for (n = 0; n < avctx->nb_streams; n++) { + AVStream *st = avctx->streams[n]; + AVCodecContext *c = st->codec; + if (c->codec_type == AVMEDIA_TYPE_AUDIO) { + if (decklink_setup_audio(avctx, st)) + goto error; + } else if (c->codec_type == AVMEDIA_TYPE_VIDEO) { + if (decklink_setup_video(avctx, st)) + goto error; + } else { + av_log(avctx, AV_LOG_ERROR, "Unsupported stream type.\n"); + goto error; + } + } + itermode->Release(); + + return 0; + +error: + + ctx->dlo->Release(); + ctx->dl->Release(); + + return AVERROR(EIO); +} + +int ff_decklink_write_packet(AVFormatContext *avctx, AVPacket *pkt) +{ + struct decklink_cctx *cctx = (struct decklink_cctx *) avctx->priv_data; + struct decklink_ctx *ctx = (struct decklink_ctx *) cctx->ctx; + AVStream *st = avctx->streams[pkt->stream_index]; + + ctx->last_pts = FFMAX(ctx->last_pts, pkt->pts); + + if (st->codec->codec_type == AVMEDIA_TYPE_VIDEO) + return decklink_write_video_packet(avctx, pkt); + else if (st->codec->codec_type == AVMEDIA_TYPE_AUDIO) + return decklink_write_audio_packet(avctx, pkt); + + return AVERROR(EIO); +} + +} /* extern "C" */ |