mirror of
https://github.com/Ponce/slackbuilds
synced 2024-11-16 19:50:19 +01:00
7b49360918
Signed-off-by: Willy Sudiarto Raharjo <willysr@slackbuilds.org>
505 lines
20 KiB
Diff
505 lines
20 KiB
Diff
From c5f8d4b5b7d43c99ddcd2b9bcbe4ec4b7beddad9 Mon Sep 17 00:00:00 2001
|
|
From: letsfindaway <me@letsfindaway.de>
|
|
Date: Wed, 18 May 2022 12:13:46 +0200
|
|
Subject: [PATCH 1/2] fix: make UBFFmpegVideoEncoder compatible with ffmpeg-5
|
|
|
|
- keep separate reference to AVCodecContext
|
|
- copy parameters to AVCodecParameters
|
|
- use new API in writeFrame
|
|
- use AVChannelLayout
|
|
---
|
|
src/podcast/ffmpeg/UBFFmpegVideoEncoder.cpp | 182 ++++++++++++++------
|
|
src/podcast/ffmpeg/UBFFmpegVideoEncoder.h | 2 +
|
|
2 files changed, 134 insertions(+), 50 deletions(-)
|
|
|
|
diff --git a/src/podcast/ffmpeg/UBFFmpegVideoEncoder.cpp b/src/podcast/ffmpeg/UBFFmpegVideoEncoder.cpp
|
|
index 0e4829c49..4fd84d740 100644
|
|
--- a/src/podcast/ffmpeg/UBFFmpegVideoEncoder.cpp
|
|
+++ b/src/podcast/ffmpeg/UBFFmpegVideoEncoder.cpp
|
|
@@ -45,20 +45,20 @@
|
|
AVFormatContext *s = avformat_alloc_context();
|
|
int ret = 0;
|
|
|
|
- *avctx = NULL;
|
|
+ *avctx = nullptr;
|
|
if (!s)
|
|
goto nomem;
|
|
|
|
if (!oformat) {
|
|
if (format) {
|
|
- oformat = av_guess_format(format, NULL, NULL);
|
|
+ oformat = av_guess_format(format, nullptr, nullptr);
|
|
if (!oformat) {
|
|
av_log(s, AV_LOG_ERROR, "Requested output format '%s' is not a suitable output format\n", format);
|
|
ret = AVERROR(EINVAL);
|
|
goto error;
|
|
}
|
|
} else {
|
|
- oformat = av_guess_format(NULL, filename, NULL);
|
|
+ oformat = av_guess_format(nullptr, filename, nullptr);
|
|
if (!oformat) {
|
|
ret = AVERROR(EINVAL);
|
|
av_log(s, AV_LOG_ERROR, "Unable to find a suitable output format for '%s'\n",
|
|
@@ -78,7 +78,7 @@
|
|
av_opt_set_defaults(s->priv_data);
|
|
}
|
|
} else
|
|
- s->priv_data = NULL;
|
|
+ s->priv_data = nullptr;
|
|
|
|
if (filename)
|
|
av_strlcpy(s->filename, filename, sizeof(s->filename));
|
|
@@ -168,15 +168,17 @@ QString avErrorToQString(int errnum)
|
|
/**
|
|
* @brief Write a given frame to the audio stream or, if a null frame is passed, flush the stream.
|
|
*
|
|
- * @param frame An AVFrame to be written to the stream, or NULL to flush the stream
|
|
+ * @param frame An AVFrame to be written to the stream, or nullptr to flush the stream
|
|
* @param packet A (reusable) packet, used to temporarily store frame data
|
|
* @param stream The stream to write to
|
|
* @param outputFormatContext The output format context
|
|
*/
|
|
-void writeFrame(AVFrame *frame, AVPacket *packet, AVStream *stream, AVFormatContext *outputFormatContext)
|
|
+void writeFrame(AVFrame *frame, AVPacket *packet, AVStream *stream, AVCodecContext* c, AVFormatContext *outputFormatContext)
|
|
{
|
|
- int gotOutput, ret;
|
|
+ int ret;
|
|
|
|
+#if LIBAVFORMAT_VERSION_MAJOR < 58
|
|
+ int gotOutput;
|
|
av_init_packet(packet);
|
|
|
|
do {
|
|
@@ -200,11 +202,34 @@ void writeFrame(AVFrame *frame, AVPacket *packet, AVStream *stream, AVFormatCont
|
|
}
|
|
|
|
} while (gotOutput && !frame);
|
|
+#else
|
|
+ // send the frame to the encoder
|
|
+ ret = avcodec_send_frame(c, frame);
|
|
+
|
|
+ while (ret >= 0) {
|
|
+ ret = avcodec_receive_packet(c, packet);
|
|
+ if (ret == AVERROR(EAGAIN) || ret == AVERROR_EOF)
|
|
+ break;
|
|
+ else if (ret < 0) {
|
|
+ qWarning() << "Couldn't encode audio frame: " << avErrorToQString(ret);
|
|
+ }
|
|
+
|
|
+ /* rescale output packet timestamp values from codec to stream timebase */
|
|
+ av_packet_rescale_ts(packet, c->time_base, stream->time_base);
|
|
+ packet->stream_index = stream->index;
|
|
+
|
|
+ /* Write the compressed frame to the media file. */
|
|
+ ret = av_interleaved_write_frame(outputFormatContext, packet);
|
|
+ /* pkt is now blank (av_interleaved_write_frame() takes ownership of
|
|
+ * its contents and resets pkt), so that no unreferencing is necessary.
|
|
+ * This would be different if one used av_write_frame(). */
|
|
+ }
|
|
+#endif
|
|
}
|
|
|
|
-void flushStream(AVPacket *packet, AVStream *stream, AVFormatContext *outputFormatContext)
|
|
+void flushStream(AVPacket *packet, AVStream *stream, AVCodecContext* c, AVFormatContext *outputFormatContext)
|
|
{
|
|
- writeFrame(NULL, packet, stream, outputFormatContext);
|
|
+ writeFrame(nullptr, packet, stream, c, outputFormatContext);
|
|
}
|
|
|
|
//-------------------------------------------------------------------------
|
|
@@ -213,12 +238,12 @@ void flushStream(AVPacket *packet, AVStream *stream, AVFormatContext *outputForm
|
|
|
|
UBFFmpegVideoEncoder::UBFFmpegVideoEncoder(QObject* parent)
|
|
: UBAbstractVideoEncoder(parent)
|
|
- , mOutputFormatContext(NULL)
|
|
- , mSwsContext(NULL)
|
|
+ , mOutputFormatContext(nullptr)
|
|
+ , mSwsContext(nullptr)
|
|
, mShouldRecordAudio(true)
|
|
- , mAudioInput(NULL)
|
|
- , mSwrContext(NULL)
|
|
- , mAudioOutBuffer(NULL)
|
|
+ , mAudioInput(nullptr)
|
|
+ , mSwrContext(nullptr)
|
|
+ , mAudioOutBuffer(nullptr)
|
|
, mAudioSampleRate(44100)
|
|
, mAudioFrameCount(0)
|
|
{
|
|
@@ -288,16 +313,18 @@ bool UBFFmpegVideoEncoder::stop()
|
|
|
|
bool UBFFmpegVideoEncoder::init()
|
|
{
|
|
+#if LIBAVFORMAT_VERSION_MAJOR < 58
|
|
av_register_all();
|
|
avcodec_register_all();
|
|
+#endif
|
|
|
|
- AVDictionary * options = NULL;
|
|
+ AVDictionary * options = nullptr;
|
|
int ret;
|
|
|
|
// Output format and context
|
|
// --------------------------------------
|
|
- if (avformat_alloc_output_context2(&mOutputFormatContext, NULL,
|
|
- "mp4", NULL) < 0)
|
|
+ if (avformat_alloc_output_context2(&mOutputFormatContext, nullptr,
|
|
+ "mp4", nullptr) < 0)
|
|
{
|
|
setLastErrorMessage("Couldn't allocate video format context");
|
|
return false;
|
|
@@ -308,16 +335,25 @@ bool UBFFmpegVideoEncoder::init()
|
|
|
|
// Video codec and context
|
|
// -------------------------------------
|
|
- mVideoStream = avformat_new_stream(mOutputFormatContext, 0);
|
|
+ mVideoStream = avformat_new_stream(mOutputFormatContext, nullptr);
|
|
+ if (!mVideoStream) {
|
|
+ setLastErrorMessage("Could not allocate stream");
|
|
+ return false;
|
|
+ }
|
|
|
|
- AVCodec * videoCodec = avcodec_find_encoder(mOutputFormatContext->oformat->video_codec);
|
|
+ auto videoCodec = avcodec_find_encoder(mOutputFormatContext->oformat->video_codec);
|
|
if (!videoCodec) {
|
|
setLastErrorMessage("Video codec not found");
|
|
return false;
|
|
}
|
|
|
|
AVCodecContext* c = avcodec_alloc_context3(videoCodec);
|
|
+ if (!c) {
|
|
+ setLastErrorMessage("Could not allocate encoding context");
|
|
+ return false;
|
|
+ }
|
|
|
|
+ c->codec_id = mOutputFormatContext->oformat->video_codec;
|
|
c->bit_rate = videoBitsPerSecond();
|
|
c->width = videoSize().width();
|
|
c->height = videoSize().height();
|
|
@@ -329,6 +365,8 @@ bool UBFFmpegVideoEncoder::init()
|
|
if (mOutputFormatContext->oformat->flags & AVFMT_GLOBALHEADER)
|
|
c->flags |= AV_CODEC_FLAG_GLOBAL_HEADER;
|
|
|
|
+ mVideoStream->time_base = c->time_base;
|
|
+
|
|
/*
|
|
* Supported pixel formats for h264 are:
|
|
* AV_PIX_FMT_YUV420P
|
|
@@ -347,13 +385,20 @@ bool UBFFmpegVideoEncoder::init()
|
|
return false;
|
|
}
|
|
|
|
- mVideoStream->codec = c;
|
|
+ mVideoCodecContext = c;
|
|
+
|
|
+ // copy the stream parameters to the muxer
|
|
+ ret = avcodec_parameters_from_context(mVideoStream->codecpar, c);
|
|
+ if (ret < 0) {
|
|
+ setLastErrorMessage("Could not copy the stream parameters");
|
|
+ return false;
|
|
+ }
|
|
|
|
// Source images are RGB32, and should be converted to YUV for h264 video
|
|
mSwsContext = sws_getCachedContext(mSwsContext,
|
|
c->width, c->height, AV_PIX_FMT_RGB32,
|
|
c->width, c->height, c->pix_fmt,
|
|
- SWS_BICUBIC, 0, 0, 0);
|
|
+ SWS_BICUBIC, nullptr, nullptr, nullptr);
|
|
|
|
// Audio codec and context
|
|
// -------------------------------------
|
|
@@ -381,7 +426,7 @@ bool UBFFmpegVideoEncoder::init()
|
|
|
|
// Codec
|
|
|
|
- AVCodec * audioCodec = avcodec_find_encoder(mOutputFormatContext->oformat->audio_codec);
|
|
+ auto audioCodec = avcodec_find_encoder(mOutputFormatContext->oformat->audio_codec);
|
|
|
|
if (!audioCodec) {
|
|
setLastErrorMessage("Audio codec not found");
|
|
@@ -389,15 +434,29 @@ bool UBFFmpegVideoEncoder::init()
|
|
}
|
|
|
|
mAudioStream = avformat_new_stream(mOutputFormatContext, audioCodec);
|
|
+ if (!mAudioStream) {
|
|
+ setLastErrorMessage("Could not allocate stream");
|
|
+ return false;
|
|
+ }
|
|
+
|
|
mAudioStream->id = mOutputFormatContext->nb_streams-1;
|
|
|
|
- c = mAudioStream->codec;
|
|
+ c = avcodec_alloc_context3(audioCodec);
|
|
+ if (!c) {
|
|
+ setLastErrorMessage("Could not allocate encoding context");
|
|
+ return false;
|
|
+ }
|
|
|
|
c->bit_rate = 96000;
|
|
c->sample_fmt = audioCodec->sample_fmts ? audioCodec->sample_fmts[0] : AV_SAMPLE_FMT_FLTP;// FLTP by default for AAC
|
|
c->sample_rate = mAudioSampleRate;
|
|
+
|
|
+#if LIBAVUTIL_VERSION_INT < AV_VERSION_INT(57, 25, 100)
|
|
c->channel_layout = AV_CH_LAYOUT_STEREO;
|
|
c->channels = av_get_channel_layout_nb_channels(c->channel_layout);
|
|
+#else
|
|
+ av_channel_layout_copy(&c->ch_layout, &(AVChannelLayout)AV_CHANNEL_LAYOUT_STEREO);
|
|
+#endif
|
|
|
|
//deprecated on ffmpeg 4
|
|
c->strict_std_compliance = -2;// Enable use of experimental codec
|
|
@@ -412,13 +471,22 @@ bool UBFFmpegVideoEncoder::init()
|
|
if (mOutputFormatContext->oformat->flags & AVFMT_GLOBALHEADER)
|
|
c->flags |= AV_CODEC_FLAG_GLOBAL_HEADER;
|
|
|
|
- ret = avcodec_open2(c, audioCodec, NULL);
|
|
+ ret = avcodec_open2(c, audioCodec, nullptr);
|
|
|
|
if (ret < 0) {
|
|
setLastErrorMessage(QString("Couldn't open audio codec: ") + avErrorToQString(ret));
|
|
return false;
|
|
}
|
|
|
|
+ mAudioCodecContext = c;
|
|
+
|
|
+ // copy the stream parameters to the muxer
|
|
+ ret = avcodec_parameters_from_context(mAudioStream->codecpar, c);
|
|
+ if (ret < 0) {
|
|
+ setLastErrorMessage("Could not copy the stream parameters");
|
|
+ return false;
|
|
+ }
|
|
+
|
|
// The input (raw sound from the microphone) may not match the codec's sampling rate,
|
|
// sample format or number of channels; we use libswresample to convert and resample it
|
|
mSwrContext = swr_alloc();
|
|
@@ -427,15 +495,23 @@ bool UBFFmpegVideoEncoder::init()
|
|
return false;
|
|
}
|
|
|
|
- av_opt_set_int(mSwrContext, "in_channel_count", inChannelCount, 0);
|
|
- av_opt_set_int(mSwrContext, "in_channel_layout", av_get_default_channel_layout(inChannelCount), 0);
|
|
av_opt_set_int(mSwrContext, "in_sample_rate", inSampleRate, 0);
|
|
av_opt_set_sample_fmt(mSwrContext, "in_sample_fmt", (AVSampleFormat)mAudioInput->sampleFormat(), 0);
|
|
- av_opt_set_int(mSwrContext, "out_channel_count", c->channels, 0);
|
|
- av_opt_set_int(mSwrContext, "out_channel_layout", c->channel_layout, 0);
|
|
av_opt_set_int(mSwrContext, "out_sample_rate", c->sample_rate, 0);
|
|
av_opt_set_sample_fmt(mSwrContext, "out_sample_fmt", c->sample_fmt, 0);
|
|
|
|
+#if LIBAVUTIL_VERSION_INT < AV_VERSION_INT(57, 25, 100)
|
|
+ av_opt_set_int(mSwrContext, "in_channel_count", inChannelCount, 0);
|
|
+ av_opt_set_int(mSwrContext, "in_channel_layout", av_get_default_channel_layout(inChannelCount), 0);
|
|
+ av_opt_set_int(mSwrContext, "out_channel_count", c->channels, 0);
|
|
+ av_opt_set_int(mSwrContext, "out_channel_layout", c->channel_layout, 0);
|
|
+#else
|
|
+ AVChannelLayout inChannelLayout;
|
|
+ av_channel_layout_default(&inChannelLayout, inChannelCount);
|
|
+ av_opt_set_chlayout (mSwrContext, "in_chlayout", &inChannelLayout, 0);
|
|
+ av_opt_set_chlayout (mSwrContext, "out_chlayout", &c->ch_layout, 0);
|
|
+#endif
|
|
+
|
|
ret = swr_init(mSwrContext);
|
|
if (ret < 0) {
|
|
setLastErrorMessage(QString("Couldn't initialize the resampling context: ") + avErrorToQString(ret));
|
|
@@ -456,7 +532,7 @@ bool UBFFmpegVideoEncoder::init()
|
|
}
|
|
|
|
// Write stream header
|
|
- ret = avformat_write_header(mOutputFormatContext, NULL);
|
|
+ ret = avformat_write_header(mOutputFormatContext, nullptr);
|
|
|
|
if (ret < 0) {
|
|
setLastErrorMessage(QString("Couldn't write header to file: ") + avErrorToQString(ret));
|
|
@@ -506,9 +582,9 @@ AVFrame* UBFFmpegVideoEncoder::convertImageFrame(ImageFrame frame)
|
|
{
|
|
AVFrame* avFrame = av_frame_alloc();
|
|
|
|
- avFrame->format = mVideoStream->codec->pix_fmt;
|
|
- avFrame->width = mVideoStream->codec->width;
|
|
- avFrame->height = mVideoStream->codec->height;
|
|
+ avFrame->format = mVideoCodecContext->pix_fmt;
|
|
+ avFrame->width = mVideoCodecContext->width;
|
|
+ avFrame->height = mVideoCodecContext->height;
|
|
avFrame->pts = mVideoTimebase * frame.timestamp / 1000;
|
|
|
|
const uchar * rgbImage = frame.image.bits();
|
|
@@ -516,18 +592,18 @@ AVFrame* UBFFmpegVideoEncoder::convertImageFrame(ImageFrame frame)
|
|
const int in_linesize[1] = { frame.image.bytesPerLine() };
|
|
|
|
// Allocate the output image
|
|
- if (av_image_alloc(avFrame->data, avFrame->linesize, mVideoStream->codec->width,
|
|
- mVideoStream->codec->height, mVideoStream->codec->pix_fmt, 32) < 0)
|
|
+ if (av_image_alloc(avFrame->data, avFrame->linesize, mVideoCodecContext->width,
|
|
+ mVideoCodecContext->height, mVideoCodecContext->pix_fmt, 32) < 0)
|
|
{
|
|
qWarning() << "Couldn't allocate image";
|
|
- return NULL;
|
|
+ return nullptr;
|
|
}
|
|
|
|
sws_scale(mSwsContext,
|
|
(const uint8_t* const*)&rgbImage,
|
|
in_linesize,
|
|
0,
|
|
- mVideoStream->codec->height,
|
|
+ mVideoCodecContext->height,
|
|
avFrame->data,
|
|
avFrame->linesize);
|
|
|
|
@@ -548,7 +624,7 @@ void UBFFmpegVideoEncoder::onAudioAvailable(QByteArray data)
|
|
void UBFFmpegVideoEncoder::processAudio(QByteArray &data)
|
|
{
|
|
int ret;
|
|
- AVCodecContext* codecContext = mAudioStream->codec;
|
|
+ AVCodecContext* codecContext = mAudioCodecContext;
|
|
|
|
const char * inSamples = data.constData();
|
|
|
|
@@ -559,7 +635,7 @@ void UBFFmpegVideoEncoder::processAudio(QByteArray &data)
|
|
int outSamplesCount = swr_get_out_samples(mSwrContext, inSamplesCount);
|
|
|
|
// Allocate output samples
|
|
- uint8_t ** outSamples = NULL;
|
|
+ uint8_t ** outSamples = nullptr;
|
|
int outSamplesLineSize;
|
|
|
|
ret = av_samples_alloc_array_and_samples(&outSamples, &outSamplesLineSize,
|
|
@@ -597,13 +673,19 @@ void UBFFmpegVideoEncoder::processAudio(QByteArray &data)
|
|
|
|
AVFrame * avFrame = av_frame_alloc();
|
|
avFrame->nb_samples = codecContext->frame_size;
|
|
+
|
|
+#if LIBAVUTIL_VERSION_INT < AV_VERSION_INT(57, 25, 100)
|
|
avFrame->channel_layout = codecContext->channel_layout;
|
|
+#else
|
|
+ av_channel_layout_copy(&avFrame->ch_layout, &codecContext->ch_layout);
|
|
+#endif
|
|
+
|
|
avFrame->format = codecContext->sample_fmt;
|
|
avFrame->sample_rate = codecContext->sample_rate;
|
|
avFrame->pts = mAudioFrameCount;
|
|
|
|
#if LIBAVFORMAT_VERSION_MICRO < 100
|
|
- int buffer_size = av_samples_get_buffer_size(NULL, codecContext->channels, codecContext->frame_size, codecContext->sample_fmt, 0);
|
|
+ int buffer_size = av_samples_get_buffer_size(nullptr, codecContext->channels, codecContext->frame_size, codecContext->sample_fmt, 0);
|
|
audio_samples_buffer = (uint8_t*)av_malloc(buffer_size);
|
|
if (!audio_samples_buffer) {
|
|
qWarning() << "Couldn't allocate samples for audio frame: " << avErrorToQString(ret);
|
|
@@ -645,19 +727,19 @@ void UBFFmpegVideoEncoder::finishEncoding()
|
|
{
|
|
qDebug() << "VideoEncoder::finishEncoding called";
|
|
|
|
- flushStream(mVideoWorker->mVideoPacket, mVideoStream, mOutputFormatContext);
|
|
+ flushStream(mVideoWorker->mVideoPacket, mVideoStream, mVideoCodecContext, mOutputFormatContext);
|
|
|
|
if (mShouldRecordAudio)
|
|
- flushStream(mVideoWorker->mAudioPacket, mAudioStream, mOutputFormatContext);
|
|
+ flushStream(mVideoWorker->mAudioPacket, mAudioStream, mAudioCodecContext, mOutputFormatContext);
|
|
|
|
av_write_trailer(mOutputFormatContext);
|
|
avio_close(mOutputFormatContext->pb);
|
|
|
|
- avcodec_close(mVideoStream->codec);
|
|
+ avcodec_close(mVideoCodecContext);
|
|
sws_freeContext(mSwsContext);
|
|
|
|
if (mShouldRecordAudio) {
|
|
- avcodec_close(mAudioStream->codec);
|
|
+ avcodec_close(mAudioCodecContext);
|
|
swr_free(&mSwrContext);
|
|
}
|
|
|
|
@@ -676,17 +758,17 @@ UBFFmpegVideoEncoderWorker::UBFFmpegVideoEncoderWorker(UBFFmpegVideoEncoder* con
|
|
{
|
|
mStopRequested = false;
|
|
mIsRunning = false;
|
|
- mVideoPacket = new AVPacket();
|
|
- mAudioPacket = new AVPacket();
|
|
+ mVideoPacket = av_packet_alloc();
|
|
+ mAudioPacket = av_packet_alloc();
|
|
}
|
|
|
|
UBFFmpegVideoEncoderWorker::~UBFFmpegVideoEncoderWorker()
|
|
{
|
|
if (mVideoPacket)
|
|
- delete mVideoPacket;
|
|
+ av_packet_free(&mVideoPacket);
|
|
|
|
if (mAudioPacket)
|
|
- delete mAudioPacket;
|
|
+ av_packet_free(&mAudioPacket);
|
|
}
|
|
|
|
void UBFFmpegVideoEncoderWorker::stopEncoding()
|
|
@@ -743,7 +825,7 @@ void UBFFmpegVideoEncoderWorker::runEncoding()
|
|
void UBFFmpegVideoEncoderWorker::writeLatestVideoFrame()
|
|
{
|
|
AVFrame* frame = mImageQueue.dequeue();
|
|
- writeFrame(frame, mVideoPacket, mController->mVideoStream, mController->mOutputFormatContext);
|
|
+ writeFrame(frame, mVideoPacket, mController->mVideoStream, mController->mVideoCodecContext, mController->mOutputFormatContext);
|
|
av_freep(&frame->data[0]);
|
|
av_frame_free(&frame);
|
|
}
|
|
@@ -751,14 +833,14 @@ void UBFFmpegVideoEncoderWorker::writeLatestVideoFrame()
|
|
void UBFFmpegVideoEncoderWorker::writeLatestAudioFrame()
|
|
{
|
|
AVFrame *frame = mAudioQueue.dequeue();
|
|
- writeFrame(frame, mAudioPacket, mController->mAudioStream, mController->mOutputFormatContext);
|
|
+ writeFrame(frame, mAudioPacket, mController->mAudioStream, mController->mAudioCodecContext, mController->mOutputFormatContext);
|
|
av_frame_free(&frame);
|
|
|
|
#if LIBAVFORMAT_VERSION_MICRO < 100
|
|
if (audio_samples_buffer) {
|
|
av_free(audio_samples_buffer);
|
|
av_freep(&frame->data[0]);
|
|
- audio_samples_buffer = NULL;
|
|
+ audio_samples_buffer = nullptr;
|
|
}
|
|
#endif
|
|
}
|
|
diff --git a/src/podcast/ffmpeg/UBFFmpegVideoEncoder.h b/src/podcast/ffmpeg/UBFFmpegVideoEncoder.h
|
|
index 1e8a12cfa..14b735064 100644
|
|
--- a/src/podcast/ffmpeg/UBFFmpegVideoEncoder.h
|
|
+++ b/src/podcast/ffmpeg/UBFFmpegVideoEncoder.h
|
|
@@ -123,6 +123,7 @@ private slots:
|
|
|
|
// Video
|
|
// ------------------------------------------
|
|
+ AVCodecContext* mVideoCodecContext;
|
|
QQueue<ImageFrame> mPendingFrames;
|
|
struct SwsContext * mSwsContext;
|
|
|
|
@@ -133,6 +134,7 @@ private slots:
|
|
bool mShouldRecordAudio;
|
|
|
|
UBMicrophoneInput * mAudioInput;
|
|
+ AVCodecContext* mAudioCodecContext;
|
|
struct SwrContext * mSwrContext;
|
|
/// Queue for audio that has been rescaled/converted but not encoded yet
|
|
AVAudioFifo *mAudioOutBuffer;
|
|
|
|
From 347af2a56804100c589766ff20768914d8a04903 Mon Sep 17 00:00:00 2001
|
|
From: letsfindaway <me@letsfindaway.de>
|
|
Date: Thu, 15 Sep 2022 11:55:17 +0200
|
|
Subject: [PATCH 2/2] fix: avoid taking address of initializer list
|
|
|
|
- FFmpeg example code was used for UBFFmpegVideoEncoder.cpp
|
|
- channel layout was initialized by taking address of an initializer
|
|
- at least for some compilers this fails, as it is an rvalue
|
|
- using an explicit variable avoids this
|
|
---
|
|
src/podcast/ffmpeg/UBFFmpegVideoEncoder.cpp | 3 ++-
|
|
1 file changed, 2 insertions(+), 1 deletion(-)
|
|
|
|
diff --git a/src/podcast/ffmpeg/UBFFmpegVideoEncoder.cpp b/src/podcast/ffmpeg/UBFFmpegVideoEncoder.cpp
|
|
index 4fd84d740..805ff77e1 100644
|
|
--- a/src/podcast/ffmpeg/UBFFmpegVideoEncoder.cpp
|
|
+++ b/src/podcast/ffmpeg/UBFFmpegVideoEncoder.cpp
|
|
@@ -455,7 +455,8 @@ bool UBFFmpegVideoEncoder::init()
|
|
c->channel_layout = AV_CH_LAYOUT_STEREO;
|
|
c->channels = av_get_channel_layout_nb_channels(c->channel_layout);
|
|
#else
|
|
- av_channel_layout_copy(&c->ch_layout, &(AVChannelLayout)AV_CHANNEL_LAYOUT_STEREO);
|
|
+ AVChannelLayout layout = AV_CHANNEL_LAYOUT_STEREO;
|
|
+ av_channel_layout_copy(&c->ch_layout, &layout);
|
|
#endif
|
|
|
|
//deprecated on ffmpeg 4
|