213 lines
9.7 KiB
C++
213 lines
9.7 KiB
C++
#include "videosplitthread.h"
|
|
#include <QImage>
|
|
#include <QPainter>
|
|
#include <QDebug>
|
|
extern "C"{
|
|
#include <libavformat/avformat.h>
|
|
#include <libavcodec/avcodec.h>
|
|
#include <libswscale/swscale.h>
|
|
}
|
|
|
|
static void imgCleanupHandler(void *info) {
|
|
delete [] (uchar*)info;
|
|
}
|
|
|
|
VideoSplitThread::VideoSplitThread(int elew, int eleh, int dpw, int sph, QList<int> &widths, QPointF pos, QByteArray file) : mEleW(elew), mEleH(eleh), mDPW(dpw), mSPH(sph), mWidths(widths), pos(pos), file(file) {
|
|
connect(this, &QThread::finished, this, &QThread::deleteLater);
|
|
}
|
|
|
|
void VideoSplitThread::run() {
|
|
AVFormatContext *in_fmt = avformat_alloc_context(), *out_fmt = 0;
|
|
AVCodecContext *de_ctx = 0, *en_ctx = 0;
|
|
QString err;
|
|
char buf[AV_ERROR_MAX_STRING_SIZE];
|
|
int ret;
|
|
{
|
|
if((ret = avformat_open_input(&in_fmt, file.constData(), 0, 0)) < 0) {
|
|
err = QString("Couldn't open input stream. ")+av_make_error_string(buf, AV_ERROR_MAX_STRING_SIZE, ret);
|
|
goto free;
|
|
}
|
|
if((ret = avformat_find_stream_info(in_fmt, nullptr)) < 0) {
|
|
err = QString("Couldn't find stream information. ")+av_make_error_string(buf, AV_ERROR_MAX_STRING_SIZE, ret);
|
|
goto free;
|
|
}
|
|
auto outfile = file+"-square.mp4";
|
|
if((ret = avformat_alloc_output_context2(&out_fmt, 0, "mp4", outfile.constData())) < 0) {
|
|
err = QString("avformat_alloc_output_context2 fail. ")+av_make_error_string(buf, AV_ERROR_MAX_STRING_SIZE, ret);
|
|
goto free;
|
|
}
|
|
int vi_idx = -1;
|
|
AVStream *out_vi_stream;
|
|
for(uint ss=0; ss<in_fmt->nb_streams; ss++) {
|
|
AVStream *stream = in_fmt->streams[ss];
|
|
qDebug()<<"codec_type"<<av_get_media_type_string(stream->codecpar->codec_type);
|
|
if(stream->codecpar->codec_type == AVMEDIA_TYPE_DATA) continue;
|
|
AVStream *out_stream = avformat_new_stream(out_fmt, 0);
|
|
if(vi_idx == -1 && stream->codecpar->codec_type == AVMEDIA_TYPE_VIDEO) {
|
|
vi_idx = ss;
|
|
out_vi_stream = out_stream;
|
|
} else {
|
|
if((ret = avcodec_parameters_copy(out_stream->codecpar, stream->codecpar)) < 0) {
|
|
err = QString("avcodec_parameters_copy fail. ") + av_make_error_string(buf, AV_ERROR_MAX_STRING_SIZE, ret);
|
|
goto free;
|
|
}
|
|
}
|
|
out_stream->time_base = stream->time_base;
|
|
out_stream->start_time = stream->start_time;
|
|
out_stream->duration = stream->duration;
|
|
}
|
|
if(vi_idx == -1) {
|
|
err = "Didn't find a Video Stream";
|
|
goto free;
|
|
}
|
|
|
|
auto codecpar = in_fmt->streams[vi_idx]->codecpar;
|
|
|
|
auto decoder = avcodec_find_decoder(codecpar->codec_id);
|
|
if(decoder==0) {
|
|
err = "Could not found Video Decoder";
|
|
goto free;
|
|
}
|
|
de_ctx = avcodec_alloc_context3(decoder);
|
|
de_ctx->thread_count = 4;
|
|
avcodec_parameters_to_context(de_ctx, codecpar);
|
|
if(avcodec_open2(de_ctx, decoder, 0) < 0) {
|
|
err = "Could not open Video decode Ctx";
|
|
goto free;
|
|
}
|
|
|
|
auto outPar = out_vi_stream->codecpar;
|
|
outPar->codec_type = AVMEDIA_TYPE_VIDEO;
|
|
outPar->codec_id = AV_CODEC_ID_H264;
|
|
outPar->format = AV_PIX_FMT_YUV420P;
|
|
outPar->width = mDPW;
|
|
outPar->height = mSPH*mWidths.size();
|
|
|
|
auto encoder = avcodec_find_encoder(outPar->codec_id);
|
|
if(encoder==0) {
|
|
fprintf(stderr, "Codec not found\n");
|
|
goto free;
|
|
}
|
|
en_ctx = avcodec_alloc_context3(encoder);
|
|
en_ctx->thread_count = 4;
|
|
avcodec_parameters_to_context(en_ctx, outPar);
|
|
en_ctx->bit_rate = outPar->width * outPar->height * 6;
|
|
en_ctx->gop_size = de_ctx->gop_size;
|
|
en_ctx->max_b_frames = 3;
|
|
en_ctx->time_base = out_vi_stream->time_base;
|
|
|
|
if((ret = avcodec_open2(en_ctx, encoder, 0)) < 0) {
|
|
err = QString("Open video encode ctx failed. ") + av_make_error_string(buf, AV_ERROR_MAX_STRING_SIZE, ret);
|
|
goto free;
|
|
}
|
|
if(out_fmt->flags & AVFMT_NOFILE) qDebug()<<"AVFMT_NOFILE";
|
|
else if((ret = avio_open(&out_fmt->pb, outfile.constData(), AVIO_FLAG_WRITE)) < 0) {
|
|
err = QString("avio_open fail. ")+av_make_error_string(buf, AV_ERROR_MAX_STRING_SIZE, ret);
|
|
goto free;
|
|
}
|
|
if((ret = avformat_write_header(out_fmt, 0)) < 0) {
|
|
err = QString("avformat_write_header fail. ")+av_make_error_string(buf, AV_ERROR_MAX_STRING_SIZE, ret);
|
|
goto free;
|
|
}
|
|
auto sws_ctx = sws_getContext(de_ctx->width, de_ctx->height, de_ctx->pix_fmt, mEleW, mEleH, AV_PIX_FMT_RGB32, SWS_FAST_BILINEAR, 0, 0, 0);
|
|
auto out_sws_ctx = sws_getContext(outPar->width, outPar->height, AV_PIX_FMT_RGB32, outPar->width, outPar->height, AV_PIX_FMT_YUV420P, SWS_FAST_BILINEAR, 0, 0, 0);
|
|
auto packet = av_packet_alloc();
|
|
auto frm = av_frame_alloc();
|
|
|
|
int img_linesize[4]{(mEleW*4+63)/64*64};
|
|
uint8_t *img_data[4]{new uchar[img_linesize[0] * mEleH]};
|
|
QImage img(img_data[0], mEleW, mEleH, img_linesize[0], QImage::Format_ARGB32, imgCleanupHandler, img_data[0]);
|
|
|
|
int out_img_linesize[4]{(outPar->width*4+63)/64*64};
|
|
uint8_t *out_img_data[4]{new uchar[out_img_linesize[0] * outPar->height]};
|
|
QImage out_img(out_img_data[0], outPar->width, outPar->height, out_img_linesize[0], QImage::Format_ARGB32, imgCleanupHandler, out_img_data[0]);
|
|
QPainter painter(&out_img);
|
|
auto end = mWidths.size() - 1;
|
|
while(1) {
|
|
if((ret = av_read_frame(in_fmt, packet)) < 0) {
|
|
if(ret!=AVERROR_EOF) {
|
|
err = QString("Read packet fail: ")+av_make_error_string(buf, AV_ERROR_MAX_STRING_SIZE, ret);
|
|
break;
|
|
}
|
|
ret = avcodec_send_packet(de_ctx, 0);
|
|
} else {
|
|
if(packet->stream_index != vi_idx) {
|
|
av_interleaved_write_frame(out_fmt, packet);
|
|
continue;
|
|
}
|
|
ret = avcodec_send_packet(de_ctx, packet);
|
|
}
|
|
if(ret < 0) {
|
|
err = QString("avcodec_send_packet fail. ")+av_make_error_string(buf, AV_ERROR_MAX_STRING_SIZE, ret);
|
|
break;
|
|
}
|
|
while((ret = avcodec_receive_frame(de_ctx, frm)) != AVERROR(EAGAIN)) {
|
|
if(ret < 0) {
|
|
if(ret!=AVERROR_EOF) {
|
|
err = QString("Receive frame fail: ")+av_make_error_string(buf, AV_ERROR_MAX_STRING_SIZE, ret);
|
|
goto free2;
|
|
}
|
|
ret = avcodec_send_frame(en_ctx, 0);
|
|
} else {
|
|
sws_scale(sws_ctx, frm->data, frm->linesize, 0, de_ctx->height, img_data, img_linesize);
|
|
auto apos = pos;
|
|
painter.drawImage(apos, img, QRectF(0, 0, mWidths[0]-apos.rx(), img.height()));
|
|
for(int i=0; i<end; i++) {
|
|
apos.rx() -= mWidths[i];
|
|
apos.ry() += mSPH;
|
|
painter.drawImage(apos, img, QRectF(0, 0, mWidths[i+1]-apos.rx(), img.height()));
|
|
}
|
|
auto pts = frm->pts;
|
|
auto dur = frm->pkt_duration;
|
|
av_frame_unref(frm);
|
|
frm->pts = pts;
|
|
frm->pkt_duration = dur;
|
|
frm->format = AV_PIX_FMT_YUV420P;
|
|
frm->width = outPar->width;
|
|
frm->height = outPar->height;
|
|
if((ret = av_frame_get_buffer(frm, 0)) < 0) {
|
|
err = QString("av_frame_get_buffer fail. ")+av_make_error_string(buf, AV_ERROR_MAX_STRING_SIZE, ret);
|
|
goto free2;
|
|
}
|
|
sws_scale(out_sws_ctx, out_img_data, out_img_linesize, 0, outPar->height, frm->data, frm->linesize);
|
|
ret = avcodec_send_frame(en_ctx, frm);
|
|
int pro = frm->pts*100/out_vi_stream->duration;
|
|
if(pro > lastPro) {
|
|
lastPro = pro;
|
|
emit emProgress(pro);
|
|
}
|
|
}
|
|
if(ret < 0) {
|
|
err = QString("avcodec_send_frame fail. ")+av_make_error_string(buf, AV_ERROR_MAX_STRING_SIZE, ret);
|
|
goto free2;
|
|
}
|
|
while((ret = avcodec_receive_packet(en_ctx, packet)) != AVERROR(EAGAIN)) {
|
|
if(ret < 0) {
|
|
if(ret!=AVERROR_EOF) err = QString("Receive frame fail: ")+av_make_error_string(buf, AV_ERROR_MAX_STRING_SIZE, ret);
|
|
else {
|
|
av_interleaved_write_frame(out_fmt, 0);
|
|
av_write_trailer(out_fmt);
|
|
emit emProgress(100);
|
|
}
|
|
goto free2;
|
|
} else {
|
|
av_interleaved_write_frame(out_fmt, packet);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
free2:
|
|
av_frame_free(&frm);
|
|
av_packet_free(&packet);
|
|
sws_freeContext(sws_ctx);
|
|
sws_freeContext(out_sws_ctx);
|
|
}
|
|
free:
|
|
avcodec_free_context(&de_ctx);
|
|
avcodec_free_context(&en_ctx);
|
|
avformat_close_input(&in_fmt);
|
|
avio_closep(&out_fmt->pb);
|
|
if(out_fmt) avformat_free_context(out_fmt);
|
|
emit emErr(err);
|
|
}
|