/* * Blackmagic DeckLink output * Copyright (c) 2013-2014 Ramiro Polla * * This file is part of FFmpeg. * * FFmpeg is free software; you can redistribute it and/or * modify it under the terms of the GNU Lesser General Public * License as published by the Free Software Foundation; either * version 2.1 of the License, or (at your option) any later version. * * FFmpeg is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * Lesser General Public License for more details. * * You should have received a copy of the GNU Lesser General Public * License along with FFmpeg; if not, write to the Free Software * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA */ #include using std::atomic; /* Include internal.h first to avoid conflict between winsock.h (used by * DeckLink headers) and winsock2.h (used by libavformat) in MSVC++ builds */ extern "C" { #include "libavformat/internal.h" } #include extern "C" { #include "libavformat/avformat.h" #include "libavcodec/bytestream.h" #include "libavutil/internal.h" #include "libavutil/imgutils.h" #include "avdevice.h" } #include "decklink_common.h" #include "decklink_enc.h" #if CONFIG_LIBKLVANC #include "libklvanc/vanc.h" #include "libklvanc/vanc-lines.h" #include "libklvanc/pixels.h" #endif /* DeckLink callback class declaration */ class decklink_frame : public IDeckLinkVideoFrame { public: decklink_frame(struct decklink_ctx *ctx, AVFrame *avframe, AVCodecID codec_id, int height, int width) : _ctx(ctx), _avframe(avframe), _avpacket(NULL), _codec_id(codec_id), _ancillary(NULL), _height(height), _width(width), _refs(1) { } decklink_frame(struct decklink_ctx *ctx, AVPacket *avpacket, AVCodecID codec_id, int height, int width) : _ctx(ctx), _avframe(NULL), _avpacket(avpacket), _codec_id(codec_id), _ancillary(NULL), _height(height), _width(width), _refs(1) { } virtual long STDMETHODCALLTYPE GetWidth (void) { return _width; } virtual long STDMETHODCALLTYPE GetHeight (void) { return _height; } virtual long STDMETHODCALLTYPE GetRowBytes (void) { if (_codec_id == AV_CODEC_ID_WRAPPED_AVFRAME) return _avframe->linesize[0] < 0 ? -_avframe->linesize[0] : _avframe->linesize[0]; else return ((GetWidth() + 47) / 48) * 128; } virtual BMDPixelFormat STDMETHODCALLTYPE GetPixelFormat(void) { if (_codec_id == AV_CODEC_ID_WRAPPED_AVFRAME) return bmdFormat8BitYUV; else return bmdFormat10BitYUV; } virtual BMDFrameFlags STDMETHODCALLTYPE GetFlags (void) { if (_codec_id == AV_CODEC_ID_WRAPPED_AVFRAME) return _avframe->linesize[0] < 0 ? bmdFrameFlagFlipVertical : bmdFrameFlagDefault; else return bmdFrameFlagDefault; } virtual HRESULT STDMETHODCALLTYPE GetBytes (void **buffer) { if (_codec_id == AV_CODEC_ID_WRAPPED_AVFRAME) { if (_avframe->linesize[0] < 0) *buffer = (void *)(_avframe->data[0] + _avframe->linesize[0] * (_avframe->height - 1)); else *buffer = (void *)(_avframe->data[0]); } else { *buffer = (void *)(_avpacket->data); } return S_OK; } virtual HRESULT STDMETHODCALLTYPE GetTimecode (BMDTimecodeFormat format, IDeckLinkTimecode **timecode) { return S_FALSE; } virtual HRESULT STDMETHODCALLTYPE GetAncillaryData(IDeckLinkVideoFrameAncillary **ancillary) { *ancillary = _ancillary; if (_ancillary) { _ancillary->AddRef(); return S_OK; } else { return S_FALSE; } } virtual HRESULT STDMETHODCALLTYPE SetAncillaryData(IDeckLinkVideoFrameAncillary *ancillary) { if (_ancillary) _ancillary->Release(); _ancillary = ancillary; _ancillary->AddRef(); return S_OK; } virtual HRESULT STDMETHODCALLTYPE QueryInterface(REFIID iid, LPVOID *ppv) { return E_NOINTERFACE; } virtual ULONG STDMETHODCALLTYPE AddRef(void) { return ++_refs; } virtual ULONG STDMETHODCALLTYPE Release(void) { int ret = --_refs; if (!ret) { av_frame_free(&_avframe); av_packet_free(&_avpacket); if (_ancillary) _ancillary->Release(); delete this; } return ret; } struct decklink_ctx *_ctx; AVFrame *_avframe; AVPacket *_avpacket; AVCodecID _codec_id; IDeckLinkVideoFrameAncillary *_ancillary; int _height; int _width; private: std::atomic _refs; }; class decklink_output_callback : public IDeckLinkVideoOutputCallback { public: virtual HRESULT STDMETHODCALLTYPE ScheduledFrameCompleted(IDeckLinkVideoFrame *_frame, BMDOutputFrameCompletionResult result) { decklink_frame *frame = static_cast(_frame); struct decklink_ctx *ctx = frame->_ctx; if (frame->_avframe) av_frame_unref(frame->_avframe); if (frame->_avpacket) av_packet_unref(frame->_avpacket); pthread_mutex_lock(&ctx->mutex); ctx->frames_buffer_available_spots++; pthread_cond_broadcast(&ctx->cond); pthread_mutex_unlock(&ctx->mutex); return S_OK; } virtual HRESULT STDMETHODCALLTYPE ScheduledPlaybackHasStopped(void) { return S_OK; } virtual HRESULT STDMETHODCALLTYPE QueryInterface(REFIID iid, LPVOID *ppv) { return E_NOINTERFACE; } virtual ULONG STDMETHODCALLTYPE AddRef(void) { return 1; } virtual ULONG STDMETHODCALLTYPE Release(void) { return 1; } }; static int decklink_setup_video(AVFormatContext *avctx, AVStream *st) { struct decklink_cctx *cctx = (struct decklink_cctx *)avctx->priv_data; struct decklink_ctx *ctx = (struct decklink_ctx *)cctx->ctx; AVCodecParameters *c = st->codecpar; if (ctx->video) { av_log(avctx, AV_LOG_ERROR, "Only one video stream is supported!\n"); return -1; } if (c->codec_id == AV_CODEC_ID_WRAPPED_AVFRAME) { if (c->format != AV_PIX_FMT_UYVY422) { av_log(avctx, AV_LOG_ERROR, "Unsupported pixel format!" " Only AV_PIX_FMT_UYVY422 is supported.\n"); return -1; } ctx->raw_format = bmdFormat8BitYUV; } else if (c->codec_id != AV_CODEC_ID_V210) { av_log(avctx, AV_LOG_ERROR, "Unsupported codec type!" " Only V210 and wrapped frame with AV_PIX_FMT_UYVY422 are supported.\n"); return -1; } else { ctx->raw_format = bmdFormat10BitYUV; } if (ff_decklink_set_configs(avctx, DIRECTION_OUT) < 0) { av_log(avctx, AV_LOG_ERROR, "Could not set output configuration\n"); return -1; } if (ff_decklink_set_format(avctx, c->width, c->height, st->time_base.num, st->time_base.den, c->field_order)) { av_log(avctx, AV_LOG_ERROR, "Unsupported video size, framerate or field order!" " Check available formats with -list_formats 1.\n"); return -1; } if (ctx->supports_vanc && ctx->dlo->EnableVideoOutput(ctx->bmd_mode, bmdVideoOutputVANC) != S_OK) { av_log(avctx, AV_LOG_WARNING, "Could not enable video output with VANC! Trying without...\n"); ctx->supports_vanc = 0; } if (!ctx->supports_vanc && ctx->dlo->EnableVideoOutput(ctx->bmd_mode, bmdVideoOutputFlagDefault) != S_OK) { av_log(avctx, AV_LOG_ERROR, "Could not enable video output!\n"); return -1; } /* Set callback. */ ctx->output_callback = new decklink_output_callback(); ctx->dlo->SetScheduledFrameCompletionCallback(ctx->output_callback); ctx->frames_preroll = st->time_base.den * ctx->preroll; if (st->time_base.den > 1000) ctx->frames_preroll /= 1000; /* Buffer twice as many frames as the preroll. */ ctx->frames_buffer = ctx->frames_preroll * 2; ctx->frames_buffer = FFMIN(ctx->frames_buffer, 60); pthread_mutex_init(&ctx->mutex, NULL); pthread_cond_init(&ctx->cond, NULL); ctx->frames_buffer_available_spots = ctx->frames_buffer; av_log(avctx, AV_LOG_DEBUG, "output: %s, preroll: %d, frames buffer size: %d\n", avctx->url, ctx->frames_preroll, ctx->frames_buffer); /* The device expects the framerate to be fixed. */ avpriv_set_pts_info(st, 64, st->time_base.num, st->time_base.den); ctx->video = 1; return 0; } static int decklink_setup_audio(AVFormatContext *avctx, AVStream *st) { struct decklink_cctx *cctx = (struct decklink_cctx *)avctx->priv_data; struct decklink_ctx *ctx = (struct decklink_ctx *)cctx->ctx; AVCodecParameters *c = st->codecpar; if (ctx->audio) { av_log(avctx, AV_LOG_ERROR, "Only one audio stream is supported!\n"); return -1; } if (c->codec_id == AV_CODEC_ID_AC3) { /* Regardless of the number of channels in the codec, we're only using 2 SDI audio channels at 48000Hz */ ctx->channels = 2; } else if (c->codec_id == AV_CODEC_ID_PCM_S16LE) { if (c->sample_rate != 48000) { av_log(avctx, AV_LOG_ERROR, "Unsupported sample rate!" " Only 48kHz is supported.\n"); return -1; } if (c->ch_layout.nb_channels != 2 && c->ch_layout.nb_channels != 8 && c->ch_layout.nb_channels != 16) { av_log(avctx, AV_LOG_ERROR, "Unsupported number of channels!" " Only 2, 8 or 16 channels are supported.\n"); return -1; } ctx->channels = c->ch_layout.nb_channels; } else { av_log(avctx, AV_LOG_ERROR, "Unsupported codec specified!" " Only PCM_S16LE and AC-3 are supported.\n"); return -1; } if (ctx->dlo->EnableAudioOutput(bmdAudioSampleRate48kHz, bmdAudioSampleType16bitInteger, ctx->channels, bmdAudioOutputStreamTimestamped) != S_OK) { av_log(avctx, AV_LOG_ERROR, "Could not enable audio output!\n"); return -1; } if (ctx->dlo->BeginAudioPreroll() != S_OK) { av_log(avctx, AV_LOG_ERROR, "Could not begin audio preroll!\n"); return -1; } /* The device expects the sample rate to be fixed. */ avpriv_set_pts_info(st, 64, 1, 48000); ctx->audio = 1; return 0; } /* Wrap the AC-3 packet into an S337 payload that is in S16LE format which can be easily injected into the PCM stream. Note: despite the function name, only AC-3 is implemented */ static int create_s337_payload(AVPacket *pkt, uint8_t **outbuf, int *outsize) { /* Note: if the packet size is not divisible by four, we need to make the actual payload larger to ensure it ends on an two channel S16LE boundary */ int payload_size = FFALIGN(pkt->size, 4) + 8; uint16_t bitcount = pkt->size * 8; uint8_t *s337_payload; PutByteContext pb; /* Sanity check: According to SMPTE ST 340:2015 Sec 4.1, the AC-3 sync frame will exactly match the 1536 samples of baseband (PCM) audio that it represents. */ if (pkt->size > 1536) return AVERROR(EINVAL); /* Encapsulate AC3 syncframe into SMPTE 337 packet */ s337_payload = (uint8_t *) av_malloc(payload_size); if (s337_payload == NULL) return AVERROR(ENOMEM); bytestream2_init_writer(&pb, s337_payload, payload_size); bytestream2_put_le16u(&pb, 0xf872); /* Sync word 1 */ bytestream2_put_le16u(&pb, 0x4e1f); /* Sync word 1 */ bytestream2_put_le16u(&pb, 0x0001); /* Burst Info, including data type (1=ac3) */ bytestream2_put_le16u(&pb, bitcount); /* Length code */ for (int i = 0; i < (pkt->size - 1); i += 2) bytestream2_put_le16u(&pb, (pkt->data[i] << 8) | pkt->data[i+1]); /* Ensure final payload is aligned on 4-byte boundary */ if (pkt->size & 1) bytestream2_put_le16u(&pb, pkt->data[pkt->size - 1] << 8); if ((pkt->size & 3) == 1 || (pkt->size & 3) == 2) bytestream2_put_le16u(&pb, 0); *outsize = payload_size; *outbuf = s337_payload; return 0; } static int decklink_setup_subtitle(AVFormatContext *avctx, AVStream *st) { int ret = -1; switch(st->codecpar->codec_id) { #if CONFIG_LIBKLVANC case AV_CODEC_ID_EIA_608: /* No special setup required */ ret = 0; break; #endif default: av_log(avctx, AV_LOG_ERROR, "Unsupported subtitle codec specified\n"); break; } return ret; } av_cold int ff_decklink_write_trailer(AVFormatContext *avctx) { struct decklink_cctx *cctx = (struct decklink_cctx *)avctx->priv_data; struct decklink_ctx *ctx = (struct decklink_ctx *)cctx->ctx; if (ctx->playback_started) { BMDTimeValue actual; ctx->dlo->StopScheduledPlayback(ctx->last_pts * ctx->bmd_tb_num, &actual, ctx->bmd_tb_den); ctx->dlo->DisableVideoOutput(); if (ctx->audio) ctx->dlo->DisableAudioOutput(); } ff_decklink_cleanup(avctx); if (ctx->output_callback) delete ctx->output_callback; pthread_mutex_destroy(&ctx->mutex); pthread_cond_destroy(&ctx->cond); #if CONFIG_LIBKLVANC klvanc_context_destroy(ctx->vanc_ctx); #endif ff_ccfifo_uninit(&ctx->cc_fifo); av_freep(&cctx->ctx); return 0; } #if CONFIG_LIBKLVANC static void construct_cc(AVFormatContext *avctx, struct decklink_ctx *ctx, AVPacket *pkt, struct klvanc_line_set_s *vanc_lines) { struct klvanc_packet_eia_708b_s *cdp; uint16_t *cdp_words; uint16_t len; uint8_t cc_count; size_t size; int ret, i; const uint8_t *data = av_packet_get_side_data(pkt, AV_PKT_DATA_A53_CC, &size); if (!data) return; cc_count = size / 3; ret = klvanc_create_eia708_cdp(&cdp); if (ret) return; ret = klvanc_set_framerate_EIA_708B(cdp, ctx->bmd_tb_num, ctx->bmd_tb_den); if (ret) { av_log(avctx, AV_LOG_ERROR, "Invalid framerate specified: %lld/%lld\n", ctx->bmd_tb_num, ctx->bmd_tb_den); klvanc_destroy_eia708_cdp(cdp); return; } if (cc_count > KLVANC_MAX_CC_COUNT) { av_log(avctx, AV_LOG_ERROR, "Illegal cc_count received: %d\n", cc_count); cc_count = KLVANC_MAX_CC_COUNT; } /* CC data */ cdp->header.ccdata_present = 1; cdp->header.caption_service_active = 1; cdp->ccdata.cc_count = cc_count; for (i = 0; i < cc_count; i++) { if (data [3*i] & 0x04) cdp->ccdata.cc[i].cc_valid = 1; cdp->ccdata.cc[i].cc_type = data[3*i] & 0x03; cdp->ccdata.cc[i].cc_data[0] = data[3*i+1]; cdp->ccdata.cc[i].cc_data[1] = data[3*i+2]; } klvanc_finalize_EIA_708B(cdp, ctx->cdp_sequence_num++); ret = klvanc_convert_EIA_708B_to_words(cdp, &cdp_words, &len); klvanc_destroy_eia708_cdp(cdp); if (ret != 0) { av_log(avctx, AV_LOG_ERROR, "Failed converting 708 packet to words\n"); return; } ret = klvanc_line_insert(ctx->vanc_ctx, vanc_lines, cdp_words, len, 11, 0); free(cdp_words); if (ret != 0) { av_log(avctx, AV_LOG_ERROR, "VANC line insertion failed\n"); return; } } /* See SMPTE ST 2016-3:2009 */ static void construct_afd(AVFormatContext *avctx, struct decklink_ctx *ctx, AVPacket *pkt, struct klvanc_line_set_s *vanc_lines, AVStream *st) { struct klvanc_packet_afd_s *afd = NULL; uint16_t *afd_words = NULL; uint16_t len; size_t size; int f1_line = 12, f2_line = 0, ret; const uint8_t *data = av_packet_get_side_data(pkt, AV_PKT_DATA_AFD, &size); if (!data || size == 0) return; ret = klvanc_create_AFD(&afd); if (ret) return; ret = klvanc_set_AFD_val(afd, data[0]); if (ret) { av_log(avctx, AV_LOG_ERROR, "Invalid AFD value specified: %d\n", data[0]); klvanc_destroy_AFD(afd); return; } /* Compute the AR flag based on the DAR (see ST 2016-1:2009 Sec 9.1). Note, we treat anything below 1.4 as 4:3 (as opposed to the standard 1.33), because there are lots of streams in the field that aren't *exactly* 4:3 but a tiny bit larger after doing the math... */ if (av_cmp_q((AVRational) {st->codecpar->width * st->codecpar->sample_aspect_ratio.num, st->codecpar->height * st->codecpar->sample_aspect_ratio.den}, (AVRational) {14, 10}) == 1) afd->aspectRatio = ASPECT_16x9; else afd->aspectRatio = ASPECT_4x3; ret = klvanc_convert_AFD_to_words(afd, &afd_words, &len); if (ret) { av_log(avctx, AV_LOG_ERROR, "Failed converting AFD packet to words\n"); goto out; } ret = klvanc_line_insert(ctx->vanc_ctx, vanc_lines, afd_words, len, f1_line, 0); if (ret) { av_log(avctx, AV_LOG_ERROR, "VANC line insertion failed\n"); goto out; } /* For interlaced video, insert into both fields. Switching lines for field 2 derived from SMPTE RP 168:2009, Sec 6, Table 2. */ switch (ctx->bmd_mode) { case bmdModeNTSC: case bmdModeNTSC2398: f2_line = 273 - 10 + f1_line; break; case bmdModePAL: f2_line = 319 - 6 + f1_line; break; case bmdModeHD1080i50: case bmdModeHD1080i5994: case bmdModeHD1080i6000: f2_line = 569 - 7 + f1_line; break; default: f2_line = 0; break; } if (f2_line > 0) { ret = klvanc_line_insert(ctx->vanc_ctx, vanc_lines, afd_words, len, f2_line, 0); if (ret) { av_log(avctx, AV_LOG_ERROR, "VANC line insertion failed\n"); goto out; } } out: if (afd) klvanc_destroy_AFD(afd); if (afd_words) free(afd_words); } /* Parse any EIA-608 subtitles sitting on the queue, and write packet side data that will later be handled by construct_cc... */ static void parse_608subs(AVFormatContext *avctx, struct decklink_ctx *ctx, AVPacket *pkt) { size_t cc_size = ff_ccfifo_getoutputsize(&ctx->cc_fifo); uint8_t *cc_data; if (!ff_ccfifo_ccdetected(&ctx->cc_fifo)) return; cc_data = av_packet_new_side_data(pkt, AV_PKT_DATA_A53_CC, cc_size); if (cc_data) ff_ccfifo_injectbytes(&ctx->cc_fifo, cc_data, cc_size); } static int decklink_construct_vanc(AVFormatContext *avctx, struct decklink_ctx *ctx, AVPacket *pkt, decklink_frame *frame, AVStream *st) { struct klvanc_line_set_s vanc_lines = { 0 }; int ret = 0, i; if (!ctx->supports_vanc) return 0; parse_608subs(avctx, ctx, pkt); construct_cc(avctx, ctx, pkt, &vanc_lines); construct_afd(avctx, ctx, pkt, &vanc_lines, st); IDeckLinkVideoFrameAncillary *vanc; int result = ctx->dlo->CreateAncillaryData(bmdFormat10BitYUV, &vanc); if (result != S_OK) { av_log(avctx, AV_LOG_ERROR, "Failed to create vanc\n"); ret = AVERROR(EIO); goto done; } /* Now that we've got all the VANC lines in a nice orderly manner, generate the final VANC sections for the Decklink output */ for (i = 0; i < vanc_lines.num_lines; i++) { struct klvanc_line_s *line = vanc_lines.lines[i]; int real_line; void *buf; if (!line) break; /* FIXME: include hack for certain Decklink cards which mis-represent line numbers for pSF frames */ real_line = line->line_number; result = vanc->GetBufferForVerticalBlankingLine(real_line, &buf); if (result != S_OK) { av_log(avctx, AV_LOG_ERROR, "Failed to get VANC line %d: %d", real_line, result); continue; } /* Generate the full line taking into account all VANC packets on that line */ result = klvanc_generate_vanc_line_v210(ctx->vanc_ctx, line, (uint8_t *) buf, ctx->bmd_width); if (result) { av_log(avctx, AV_LOG_ERROR, "Failed to generate VANC line\n"); continue; } } result = frame->SetAncillaryData(vanc); vanc->Release(); if (result != S_OK) { av_log(avctx, AV_LOG_ERROR, "Failed to set vanc: %d", result); ret = AVERROR(EIO); } done: for (i = 0; i < vanc_lines.num_lines; i++) klvanc_line_free(vanc_lines.lines[i]); return ret; } #endif static int decklink_write_video_packet(AVFormatContext *avctx, AVPacket *pkt) { struct decklink_cctx *cctx = (struct decklink_cctx *)avctx->priv_data; struct decklink_ctx *ctx = (struct decklink_ctx *)cctx->ctx; AVStream *st = avctx->streams[pkt->stream_index]; AVFrame *avframe = NULL, *tmp = (AVFrame *)pkt->data; AVPacket *avpacket = NULL; decklink_frame *frame; uint32_t buffered; HRESULT hr; ctx->last_pts = FFMAX(ctx->last_pts, pkt->pts); if (st->codecpar->codec_id == AV_CODEC_ID_WRAPPED_AVFRAME) { if (tmp->format != AV_PIX_FMT_UYVY422 || tmp->width != ctx->bmd_width || tmp->height != ctx->bmd_height) { av_log(avctx, AV_LOG_ERROR, "Got a frame with invalid pixel format or dimension.\n"); return AVERROR(EINVAL); } avframe = av_frame_clone(tmp); if (!avframe) { av_log(avctx, AV_LOG_ERROR, "Could not clone video frame.\n"); return AVERROR(EIO); } frame = new decklink_frame(ctx, avframe, st->codecpar->codec_id, avframe->height, avframe->width); } else { avpacket = av_packet_clone(pkt); if (!avpacket) { av_log(avctx, AV_LOG_ERROR, "Could not clone video frame.\n"); return AVERROR(EIO); } frame = new decklink_frame(ctx, avpacket, st->codecpar->codec_id, ctx->bmd_height, ctx->bmd_width); #if CONFIG_LIBKLVANC if (decklink_construct_vanc(avctx, ctx, pkt, frame, st)) av_log(avctx, AV_LOG_ERROR, "Failed to construct VANC\n"); #endif } if (!frame) { av_log(avctx, AV_LOG_ERROR, "Could not create new frame.\n"); av_frame_free(&avframe); av_packet_free(&avpacket); return AVERROR(EIO); } /* Always keep at most one second of frames buffered. */ pthread_mutex_lock(&ctx->mutex); while (ctx->frames_buffer_available_spots == 0) { pthread_cond_wait(&ctx->cond, &ctx->mutex); } ctx->frames_buffer_available_spots--; pthread_mutex_unlock(&ctx->mutex); if (ctx->first_pts == AV_NOPTS_VALUE) ctx->first_pts = pkt->pts; /* Schedule frame for playback. */ hr = ctx->dlo->ScheduleVideoFrame((class IDeckLinkVideoFrame *) frame, pkt->pts * ctx->bmd_tb_num, ctx->bmd_tb_num, ctx->bmd_tb_den); /* Pass ownership to DeckLink, or release on failure */ frame->Release(); if (hr != S_OK) { av_log(avctx, AV_LOG_ERROR, "Could not schedule video frame." " error %08x.\n", (uint32_t) hr); return AVERROR(EIO); } ctx->dlo->GetBufferedVideoFrameCount(&buffered); av_log(avctx, AV_LOG_DEBUG, "Buffered video frames: %d.\n", (int) buffered); if (pkt->pts > 2 && buffered <= 2) av_log(avctx, AV_LOG_WARNING, "There are not enough buffered video frames." " Video may misbehave!\n"); /* Preroll video frames. */ if (!ctx->playback_started && pkt->pts > (ctx->first_pts + ctx->frames_preroll)) { av_log(avctx, AV_LOG_DEBUG, "Ending audio preroll.\n"); if (ctx->audio && ctx->dlo->EndAudioPreroll() != S_OK) { av_log(avctx, AV_LOG_ERROR, "Could not end audio preroll!\n"); return AVERROR(EIO); } av_log(avctx, AV_LOG_DEBUG, "Starting scheduled playback.\n"); if (ctx->dlo->StartScheduledPlayback(ctx->first_pts * ctx->bmd_tb_num, ctx->bmd_tb_den, 1.0) != S_OK) { av_log(avctx, AV_LOG_ERROR, "Could not start scheduled playback!\n"); return AVERROR(EIO); } ctx->playback_started = 1; } return 0; } static int decklink_write_audio_packet(AVFormatContext *avctx, AVPacket *pkt) { struct decklink_cctx *cctx = (struct decklink_cctx *)avctx->priv_data; struct decklink_ctx *ctx = (struct decklink_ctx *)cctx->ctx; AVStream *st = avctx->streams[pkt->stream_index]; int sample_count; uint32_t buffered; uint8_t *outbuf = NULL; int ret = 0; ctx->dlo->GetBufferedAudioSampleFrameCount(&buffered); if (pkt->pts > 1 && !buffered) av_log(avctx, AV_LOG_WARNING, "There's no buffered audio." " Audio will misbehave!\n"); if (st->codecpar->codec_id == AV_CODEC_ID_AC3) { /* Encapsulate AC3 syncframe into SMPTE 337 packet */ int outbuf_size; ret = create_s337_payload(pkt, &outbuf, &outbuf_size); if (ret < 0) return ret; sample_count = outbuf_size / 4; } else { sample_count = pkt->size / (ctx->channels << 1); outbuf = pkt->data; } if (ctx->dlo->ScheduleAudioSamples(outbuf, sample_count, pkt->pts, bmdAudioSampleRate48kHz, NULL) != S_OK) { av_log(avctx, AV_LOG_ERROR, "Could not schedule audio samples.\n"); ret = AVERROR(EIO); } if (st->codecpar->codec_id == AV_CODEC_ID_AC3) av_freep(&outbuf); return ret; } static int decklink_write_subtitle_packet(AVFormatContext *avctx, AVPacket *pkt) { struct decklink_cctx *cctx = (struct decklink_cctx *)avctx->priv_data; struct decklink_ctx *ctx = (struct decklink_ctx *)cctx->ctx; ff_ccfifo_extractbytes(&ctx->cc_fifo, pkt->data, pkt->size); return 0; } extern "C" { av_cold int ff_decklink_write_header(AVFormatContext *avctx) { struct decklink_cctx *cctx = (struct decklink_cctx *)avctx->priv_data; struct decklink_ctx *ctx; unsigned int n; int ret; ctx = (struct decklink_ctx *) av_mallocz(sizeof(struct decklink_ctx)); if (!ctx) return AVERROR(ENOMEM); ctx->list_devices = cctx->list_devices; ctx->list_formats = cctx->list_formats; ctx->preroll = cctx->preroll; ctx->duplex_mode = cctx->duplex_mode; ctx->first_pts = AV_NOPTS_VALUE; if (cctx->link > 0 && (unsigned int)cctx->link < FF_ARRAY_ELEMS(decklink_link_conf_map)) ctx->link = decklink_link_conf_map[cctx->link]; cctx->ctx = ctx; #if CONFIG_LIBKLVANC if (klvanc_context_create(&ctx->vanc_ctx) < 0) { av_log(avctx, AV_LOG_ERROR, "Cannot create VANC library context\n"); return AVERROR(ENOMEM); } ctx->supports_vanc = 1; #endif /* List available devices and exit. */ if (ctx->list_devices) { ff_decklink_list_devices_legacy(avctx, 0, 1); return AVERROR_EXIT; } ret = ff_decklink_init_device(avctx, avctx->url); if (ret < 0) return ret; /* Get output device. */ if (ctx->dl->QueryInterface(IID_IDeckLinkOutput, (void **) &ctx->dlo) != S_OK) { av_log(avctx, AV_LOG_ERROR, "Could not open output device from '%s'\n", avctx->url); ret = AVERROR(EIO); goto error; } /* List supported formats. */ if (ctx->list_formats) { ff_decklink_list_formats(avctx); ret = AVERROR_EXIT; goto error; } /* Setup streams. */ ret = AVERROR(EIO); for (n = 0; n < avctx->nb_streams; n++) { AVStream *st = avctx->streams[n]; AVCodecParameters *c = st->codecpar; if (c->codec_type == AVMEDIA_TYPE_AUDIO) { if (decklink_setup_audio(avctx, st)) goto error; } else if (c->codec_type == AVMEDIA_TYPE_VIDEO) { if (decklink_setup_video(avctx, st)) goto error; } else if (c->codec_type == AVMEDIA_TYPE_SUBTITLE) { if (decklink_setup_subtitle(avctx, st)) goto error; } else { av_log(avctx, AV_LOG_ERROR, "Unsupported stream type.\n"); goto error; } } for (n = 0; n < avctx->nb_streams; n++) { AVStream *st = avctx->streams[n]; AVCodecParameters *c = st->codecpar; if(c->codec_type == AVMEDIA_TYPE_SUBTITLE) avpriv_set_pts_info(st, 64, ctx->bmd_tb_num, ctx->bmd_tb_den); } ret = ff_ccfifo_init(&ctx->cc_fifo, av_make_q(ctx->bmd_tb_den, ctx->bmd_tb_num), avctx); if (ret < 0) { av_log(ctx, AV_LOG_ERROR, "Failure to setup CC FIFO queue\n"); goto error; } return 0; error: ff_decklink_cleanup(avctx); return ret; } int ff_decklink_write_packet(AVFormatContext *avctx, AVPacket *pkt) { AVStream *st = avctx->streams[pkt->stream_index]; if (st->codecpar->codec_type == AVMEDIA_TYPE_VIDEO) return decklink_write_video_packet(avctx, pkt); else if (st->codecpar->codec_type == AVMEDIA_TYPE_AUDIO) return decklink_write_audio_packet(avctx, pkt); else if (st->codecpar->codec_type == AVMEDIA_TYPE_SUBTITLE) return decklink_write_subtitle_packet(avctx, pkt); return AVERROR(EIO); } int ff_decklink_list_output_devices(AVFormatContext *avctx, struct AVDeviceInfoList *device_list) { return ff_decklink_list_devices(avctx, device_list, 0, 1); } } /* extern "C" */