1 | // Copyright (C) 2024 The Qt Company Ltd. |
2 | // SPDX-License-Identifier: LicenseRef-Qt-Commercial OR LGPL-3.0-only OR GPL-2.0-only OR GPL-3.0-only |
3 | |
4 | #include "qffmpegconverter_p.h" |
5 | #include "qffmpeg_p.h" |
6 | #include <QtMultimedia/qvideoframeformat.h> |
7 | #include <QtMultimedia/qvideoframe.h> |
8 | #include <QtCore/qloggingcategory.h> |
9 | #include <private/qvideotexturehelper_p.h> |
10 | |
11 | extern "C" { |
12 | #include <libswscale/swscale.h> |
13 | } |
14 | |
15 | QT_BEGIN_NAMESPACE |
16 | |
17 | namespace { |
18 | |
19 | Q_LOGGING_CATEGORY(lc, "qt.multimedia.ffmpeg.converter" ); |
20 | |
21 | |
22 | // Converts to FFmpeg pixel format. This function differs from |
23 | // QFFmpegVideoBuffer::toAVPixelFormat which only covers the subset |
24 | // of pixel formats required for encoding. Here we need to cover more |
25 | // pixel formats to be able to generate test images for decoding/display |
26 | AVPixelFormat toAVPixelFormat(QVideoFrameFormat::PixelFormat pixelFormat) |
27 | { |
28 | switch (pixelFormat) { |
29 | default: |
30 | case QVideoFrameFormat::Format_Invalid: |
31 | return AV_PIX_FMT_NONE; |
32 | case QVideoFrameFormat::Format_AYUV: |
33 | case QVideoFrameFormat::Format_AYUV_Premultiplied: |
34 | return AV_PIX_FMT_NONE; // TODO: Fixme (No corresponding FFmpeg format available) |
35 | case QVideoFrameFormat::Format_YV12: |
36 | case QVideoFrameFormat::Format_IMC1: |
37 | case QVideoFrameFormat::Format_IMC3: |
38 | case QVideoFrameFormat::Format_IMC2: |
39 | case QVideoFrameFormat::Format_IMC4: |
40 | return AV_PIX_FMT_YUV420P; |
41 | case QVideoFrameFormat::Format_Jpeg: |
42 | return AV_PIX_FMT_BGRA; |
43 | case QVideoFrameFormat::Format_ARGB8888: |
44 | return AV_PIX_FMT_ARGB; |
45 | case QVideoFrameFormat::Format_ARGB8888_Premultiplied: |
46 | case QVideoFrameFormat::Format_XRGB8888: |
47 | return AV_PIX_FMT_0RGB; |
48 | case QVideoFrameFormat::Format_BGRA8888: |
49 | return AV_PIX_FMT_BGRA; |
50 | case QVideoFrameFormat::Format_BGRA8888_Premultiplied: |
51 | case QVideoFrameFormat::Format_BGRX8888: |
52 | return AV_PIX_FMT_BGR0; |
53 | case QVideoFrameFormat::Format_ABGR8888: |
54 | return AV_PIX_FMT_ABGR; |
55 | case QVideoFrameFormat::Format_XBGR8888: |
56 | return AV_PIX_FMT_0BGR; |
57 | case QVideoFrameFormat::Format_RGBA8888: |
58 | return AV_PIX_FMT_RGBA; |
59 | case QVideoFrameFormat::Format_RGBX8888: |
60 | return AV_PIX_FMT_RGB0; |
61 | case QVideoFrameFormat::Format_YUV422P: |
62 | return AV_PIX_FMT_YUV422P; |
63 | case QVideoFrameFormat::Format_YUV420P: |
64 | return AV_PIX_FMT_YUV420P; |
65 | case QVideoFrameFormat::Format_YUV420P10: |
66 | return AV_PIX_FMT_YUV420P10; |
67 | case QVideoFrameFormat::Format_UYVY: |
68 | return AV_PIX_FMT_UYVY422; |
69 | case QVideoFrameFormat::Format_YUYV: |
70 | return AV_PIX_FMT_YUYV422; |
71 | case QVideoFrameFormat::Format_NV12: |
72 | return AV_PIX_FMT_NV12; |
73 | case QVideoFrameFormat::Format_NV21: |
74 | return AV_PIX_FMT_NV21; |
75 | case QVideoFrameFormat::Format_Y8: |
76 | return AV_PIX_FMT_GRAY8; |
77 | case QVideoFrameFormat::Format_Y16: |
78 | return AV_PIX_FMT_GRAY16; |
79 | case QVideoFrameFormat::Format_P010: |
80 | return AV_PIX_FMT_P010; |
81 | case QVideoFrameFormat::Format_P016: |
82 | return AV_PIX_FMT_P016; |
83 | case QVideoFrameFormat::Format_SamplerExternalOES: |
84 | return AV_PIX_FMT_MEDIACODEC; |
85 | } |
86 | } |
87 | |
88 | struct SwsFrameData |
89 | { |
90 | static constexpr int arraySize = 4; // Array size required by sws_scale |
91 | std::array<uchar *, arraySize> bits; |
92 | std::array<int, arraySize> stride; |
93 | }; |
94 | |
95 | SwsFrameData getSwsData(QVideoFrame &dst) |
96 | { |
97 | switch (dst.pixelFormat()) { |
98 | case QVideoFrameFormat::Format_YV12: |
99 | case QVideoFrameFormat::Format_IMC1: |
100 | return { .bits: { dst.bits(plane: 0), dst.bits(plane: 2), dst.bits(plane: 1), nullptr }, |
101 | .stride: { dst.bytesPerLine(plane: 0), dst.bytesPerLine(plane: 2), dst.bytesPerLine(plane: 1), 0 } }; |
102 | |
103 | case QVideoFrameFormat::Format_IMC2: |
104 | return { .bits: { dst.bits(plane: 0), dst.bits(plane: 1) + dst.bytesPerLine(plane: 1) / 2, dst.bits(plane: 1), nullptr }, |
105 | .stride: { dst.bytesPerLine(plane: 0), dst.bytesPerLine(plane: 1), dst.bytesPerLine(plane: 1), 0 } }; |
106 | |
107 | case QVideoFrameFormat::Format_IMC4: |
108 | return { .bits: { dst.bits(plane: 0), dst.bits(plane: 1), dst.bits(plane: 1) + dst.bytesPerLine(plane: 1) / 2, nullptr }, |
109 | .stride: { dst.bytesPerLine(plane: 0), dst.bytesPerLine(plane: 1), dst.bytesPerLine(plane: 1), 0 } }; |
110 | default: |
111 | return { .bits: { dst.bits(plane: 0), dst.bits(plane: 1), dst.bits(plane: 2), nullptr }, |
112 | .stride: { dst.bytesPerLine(plane: 0), dst.bytesPerLine(plane: 1), dst.bytesPerLine(plane: 2), 0 } }; |
113 | } |
114 | } |
115 | |
116 | struct SwsColorSpace |
117 | { |
118 | int colorSpace; |
119 | int colorRange; // 0 - mpeg/video, 1 - jpeg/full |
120 | }; |
121 | |
122 | // Qt heuristics for determining color space requires checking |
123 | // both frame color space and range. This function mimics logic |
124 | // used elsewhere in Qt Multimedia. |
125 | SwsColorSpace toSwsColorSpace(QVideoFrameFormat::ColorRange colorRange, |
126 | QVideoFrameFormat::ColorSpace colorSpace) |
127 | { |
128 | const int avRange = colorRange == QVideoFrameFormat::ColorRange_Video ? 0 : 1; |
129 | |
130 | switch (colorSpace) { |
131 | case QVideoFrameFormat::ColorSpace_BT601: |
132 | if (colorRange == QVideoFrameFormat::ColorRange_Full) |
133 | return { SWS_CS_ITU709, .colorRange: 1 }; // TODO: FIXME - Not exact match |
134 | return { SWS_CS_ITU601, .colorRange: 0 }; |
135 | case QVideoFrameFormat::ColorSpace_BT709: |
136 | return { SWS_CS_ITU709, .colorRange: avRange }; |
137 | case QVideoFrameFormat::ColorSpace_AdobeRgb: |
138 | return { SWS_CS_ITU601, .colorRange: 1 }; // TODO: Why do ITU601 and Adobe RGB match well? |
139 | case QVideoFrameFormat::ColorSpace_BT2020: |
140 | return { SWS_CS_BT2020, .colorRange: avRange }; |
141 | case QVideoFrameFormat::ColorSpace_Undefined: |
142 | default: |
143 | return { SWS_CS_DEFAULT, .colorRange: avRange }; |
144 | } |
145 | } |
146 | |
147 | using PixelFormat = QVideoFrameFormat::PixelFormat; |
148 | |
149 | // clang-format off |
150 | |
151 | QFFmpeg::SwsContextUPtr createConverter(const QSize &srcSize, PixelFormat srcPixFmt, |
152 | const QSize &dstSize, PixelFormat dstPixFmt) |
153 | { |
154 | return QFFmpeg::createSwsContext(srcSize, srcPixFmt: toAVPixelFormat(pixelFormat: srcPixFmt), dstSize, dstPixFmt: toAVPixelFormat(pixelFormat: dstPixFmt), SWS_BILINEAR); |
155 | } |
156 | |
157 | bool setColorSpaceDetails(SwsContext *context, |
158 | const QVideoFrameFormat &srcFormat, |
159 | const QVideoFrameFormat &dstFormat) |
160 | { |
161 | const SwsColorSpace src = toSwsColorSpace(colorRange: srcFormat.colorRange(), colorSpace: srcFormat.colorSpace()); |
162 | const SwsColorSpace dst = toSwsColorSpace(colorRange: dstFormat.colorRange(), colorSpace: dstFormat.colorSpace()); |
163 | |
164 | constexpr int brightness = 0; |
165 | constexpr int contrast = 0; |
166 | constexpr int saturation = 0; |
167 | const int status = sws_setColorspaceDetails(c: context, |
168 | inv_table: sws_getCoefficients(colorspace: src.colorSpace), srcRange: src.colorRange, |
169 | table: sws_getCoefficients(colorspace: dst.colorSpace), dstRange: dst.colorRange, |
170 | brightness, contrast, saturation); |
171 | |
172 | return status == 0; |
173 | } |
174 | |
175 | bool convert(SwsContext *context, QVideoFrame &src, int srcHeight, QVideoFrame &dst) |
176 | { |
177 | if (!src.map(mode: QVideoFrame::ReadOnly)) |
178 | return false; |
179 | |
180 | QScopeGuard unmapSrc{[&] { |
181 | src.unmap(); |
182 | }}; |
183 | |
184 | if (!dst.map(mode: QVideoFrame::WriteOnly)) |
185 | return false; |
186 | |
187 | QScopeGuard unmapDst{[&] { |
188 | dst.unmap(); |
189 | }}; |
190 | |
191 | const SwsFrameData srcData = getSwsData(dst&: src); |
192 | const SwsFrameData dstData = getSwsData(dst); |
193 | |
194 | constexpr int firstSrcSliceRow = 0; |
195 | const int scaledHeight = sws_scale(c: context, |
196 | srcSlice: srcData.bits.data(), srcStride: srcData.stride.data(), |
197 | srcSliceY: firstSrcSliceRow, srcSliceH: srcHeight, |
198 | dst: dstData.bits.data(), dstStride: dstData.stride.data()); |
199 | |
200 | if (scaledHeight != srcHeight) |
201 | return false; |
202 | |
203 | return true; |
204 | } |
205 | |
206 | // Ensure even size if using planar format with chroma subsampling |
207 | QSize adjustSize(const QSize& size, PixelFormat srcFmt, PixelFormat dstFmt) |
208 | { |
209 | const auto* srcDesc = QVideoTextureHelper::textureDescription(format: srcFmt); |
210 | const auto* dstDesc = QVideoTextureHelper::textureDescription(format: dstFmt); |
211 | |
212 | QSize output = size; |
213 | for (const auto desc : { srcDesc, dstDesc }) { |
214 | for (int i = 0; i < desc->nplanes; ++i) { |
215 | // TODO: Assumes that max subsampling is 2 |
216 | if (desc->sizeScale[i].x != 1) |
217 | output.setWidth(output.width() & ~1); // Make even |
218 | |
219 | if (desc->sizeScale[i].y != 1) |
220 | output.setHeight(output.height() & ~1); // Make even |
221 | } |
222 | } |
223 | |
224 | return output; |
225 | } |
226 | |
227 | } // namespace |
228 | |
229 | // Converts a video frame to the dstFormat video frame format. |
230 | QVideoFrame convertFrame(QVideoFrame &src, const QVideoFrameFormat &dstFormat) |
231 | { |
232 | if (src.size() != dstFormat.frameSize()) { |
233 | qCCritical(lc) << "Resizing is not supported" ; |
234 | return {}; |
235 | } |
236 | |
237 | // Adjust size to even width/height if we have chroma subsampling |
238 | const QSize size = adjustSize(size: src.size(), srcFmt: src.pixelFormat(), dstFmt: dstFormat.pixelFormat()); |
239 | if (size != src.size()) |
240 | qCWarning(lc) << "Input truncated to even width/height" ; |
241 | |
242 | const QFFmpeg::SwsContextUPtr conv = createConverter( |
243 | srcSize: size, srcPixFmt: src.pixelFormat(), dstSize: size, dstPixFmt: dstFormat.pixelFormat()); |
244 | |
245 | if (!conv) { |
246 | qCCritical(lc) << "Failed to create SW converter" ; |
247 | return {}; |
248 | } |
249 | |
250 | if (!setColorSpaceDetails(context: conv.get(), srcFormat: src.surfaceFormat(), dstFormat)) { |
251 | qCCritical(lc) << "Failed to set color space details" ; |
252 | return {}; |
253 | } |
254 | |
255 | QVideoFrame dst{ dstFormat }; |
256 | |
257 | if (!convert(context: conv.get(), src, srcHeight: size.height(), dst)) { |
258 | qCCritical(lc) << "Frame conversion failed" ; |
259 | return {}; |
260 | } |
261 | |
262 | return dst; |
263 | } |
264 | |
265 | // clang-format on |
266 | |
267 | QT_END_NAMESPACE |
268 | |