1// Copyright (C) 2022 The Qt Company Ltd.
2// SPDX-License-Identifier: LicenseRef-Qt-Commercial OR LGPL-3.0-only OR GPL-2.0-only OR GPL-3.0-only
3
4#include "qvideoframeconverter_p.h"
5#include "qvideoframeconversionhelper_p.h"
6#include "qvideoframeformat.h"
7#include "qvideoframe_p.h"
8#include "qmultimediautils_p.h"
9#include "qabstractvideobuffer.h"
10
11#include <QtCore/qcoreapplication.h>
12#include <QtCore/qsize.h>
13#include <QtCore/qhash.h>
14#include <QtCore/qfile.h>
15#include <QtCore/qthreadstorage.h>
16#include <QtGui/qimage.h>
17#include <QtGui/qoffscreensurface.h>
18#include <qpa/qplatformintegration.h>
19#include <private/qvideotexturehelper_p.h>
20#include <private/qguiapplication_p.h>
21#include <rhi/qrhi.h>
22
23#ifdef Q_OS_DARWIN
24#include <QtCore/private/qcore_mac_p.h>
25#endif
26
27QT_BEGIN_NAMESPACE
28
29static Q_LOGGING_CATEGORY(qLcVideoFrameConverter, "qt.multimedia.video.frameconverter")
30
31namespace {
32
33struct State
34{
35 QRhi *rhi = nullptr;
36#if QT_CONFIG(opengl)
37 QOffscreenSurface *fallbackSurface = nullptr;
38#endif
39 bool cpuOnly = false;
40#if defined(Q_OS_ANDROID)
41 QMetaObject::Connection appStateChangedConnection;
42#endif
43 ~State() {
44 resetRhi();
45 }
46
47 void resetRhi() {
48 delete rhi;
49 rhi = nullptr;
50#if QT_CONFIG(opengl)
51 delete fallbackSurface;
52 fallbackSurface = nullptr;
53#endif
54 cpuOnly = false;
55 }
56};
57
58}
59
60static QThreadStorage<State> g_state;
61static QHash<QString, QShader> g_shaderCache;
62
63static const float g_quad[] = {
64 // Rotation 0 CW
65 1.f, -1.f, 1.f, 1.f,
66 1.f, 1.f, 1.f, 0.f,
67 -1.f, -1.f, 0.f, 1.f,
68 -1.f, 1.f, 0.f, 0.f,
69 // Rotation 90 CW
70 1.f, -1.f, 1.f, 0.f,
71 1.f, 1.f, 0.f, 0.f,
72 -1.f, -1.f, 1.f, 1.f,
73 -1.f, 1.f, 0.f, 1.f,
74 // Rotation 180 CW
75 1.f, -1.f, 0.f, 0.f,
76 1.f, 1.f, 0.f, 1.f,
77 -1.f, -1.f, 1.f, 0.f,
78 -1.f, 1.f, 1.f, 1.f,
79 // Rotation 270 CW
80 1.f, -1.f, 0.f, 1.f,
81 1.f, 1.f, 1.f, 1.f,
82 -1.f, -1.f, 0.f, 0.f,
83 -1.f, 1.f, 1.f, 0.f,
84};
85
86static bool pixelFormatHasAlpha(QVideoFrameFormat::PixelFormat format)
87{
88 switch (format) {
89 case QVideoFrameFormat::Format_ARGB8888:
90 case QVideoFrameFormat::Format_ARGB8888_Premultiplied:
91 case QVideoFrameFormat::Format_BGRA8888:
92 case QVideoFrameFormat::Format_BGRA8888_Premultiplied:
93 case QVideoFrameFormat::Format_ABGR8888:
94 case QVideoFrameFormat::Format_RGBA8888:
95 case QVideoFrameFormat::Format_AYUV:
96 case QVideoFrameFormat::Format_AYUV_Premultiplied:
97 return true;
98 default:
99 return false;
100 }
101};
102
103static QShader vfcGetShader(const QString &name)
104{
105 QShader shader = g_shaderCache.value(key: name);
106 if (shader.isValid())
107 return shader;
108
109 QFile f(name);
110 if (f.open(flags: QIODevice::ReadOnly))
111 shader = QShader::fromSerialized(data: f.readAll());
112
113 if (shader.isValid())
114 g_shaderCache[name] = shader;
115
116 return shader;
117}
118
119static void rasterTransform(QImage &image, VideoTransformation transformation)
120{
121 QTransform t;
122 if (transformation.rotation != QtVideo::Rotation::None)
123 t.rotate(a: qreal(transformation.rotation));
124 if (transformation.mirrorredHorizontallyAfterRotation)
125 t.scale(sx: -1., sy: 1);
126 if (!t.isIdentity())
127 image = image.transformed(matrix: t);
128}
129
130static void imageCleanupHandler(void *info)
131{
132 QByteArray *imageData = reinterpret_cast<QByteArray *>(info);
133 delete imageData;
134}
135
136static QRhi *initializeRHI(QRhi *videoFrameRhi)
137{
138 if (g_state.localData().rhi || g_state.localData().cpuOnly)
139 return g_state.localData().rhi;
140
141 QRhi::Implementation backend = videoFrameRhi ? videoFrameRhi->backend() : QRhi::Null;
142 const QPlatformIntegration *qpa = QGuiApplicationPrivate::platformIntegration();
143
144 if (qpa && qpa->hasCapability(cap: QPlatformIntegration::RhiBasedRendering)) {
145
146#if defined(Q_OS_MACOS) || defined(Q_OS_IOS)
147 if (backend == QRhi::Metal || backend == QRhi::Null) {
148 QRhiMetalInitParams params;
149 g_state.localData().rhi = QRhi::create(QRhi::Metal, &params);
150 }
151#endif
152
153#if defined(Q_OS_WIN)
154 if (backend == QRhi::D3D11 || backend == QRhi::Null) {
155 QRhiD3D11InitParams params;
156 g_state.localData().rhi = QRhi::create(QRhi::D3D11, &params);
157 }
158#endif
159
160#if QT_CONFIG(opengl)
161 if (!g_state.localData().rhi && (backend == QRhi::OpenGLES2 || backend == QRhi::Null)) {
162 if (qpa->hasCapability(cap: QPlatformIntegration::OpenGL)
163 && qpa->hasCapability(cap: QPlatformIntegration::RasterGLSurface)
164 && !QCoreApplication::testAttribute(attribute: Qt::AA_ForceRasterWidgets)) {
165
166 g_state.localData().fallbackSurface = QRhiGles2InitParams::newFallbackSurface();
167 QRhiGles2InitParams params;
168 params.fallbackSurface = g_state.localData().fallbackSurface;
169 if (backend == QRhi::OpenGLES2)
170 params.shareContext = static_cast<const QRhiGles2NativeHandles*>(videoFrameRhi->nativeHandles())->context;
171 g_state.localData().rhi = QRhi::create(impl: QRhi::OpenGLES2, params: &params);
172
173#if defined(Q_OS_ANDROID)
174 // reset RHI state on application suspension, as this will be invalid after resuming
175 if (!g_state.localData().appStateChangedConnection) {
176 g_state.localData().appStateChangedConnection = QObject::connect(qApp, &QGuiApplication::applicationStateChanged, qApp, [](auto state) {
177 if (state == Qt::ApplicationSuspended)
178 g_state.localData().resetRhi();
179 });
180 }
181#endif
182 }
183 }
184#endif
185 }
186
187 if (!g_state.localData().rhi) {
188 g_state.localData().cpuOnly = true;
189 qWarning() << Q_FUNC_INFO << ": No RHI backend. Using CPU conversion.";
190 }
191
192 return g_state.localData().rhi;
193}
194
195static bool updateTextures(QRhi *rhi,
196 std::unique_ptr<QRhiBuffer> &uniformBuffer,
197 std::unique_ptr<QRhiSampler> &textureSampler,
198 std::unique_ptr<QRhiShaderResourceBindings> &shaderResourceBindings,
199 std::unique_ptr<QRhiGraphicsPipeline> &graphicsPipeline,
200 std::unique_ptr<QRhiRenderPassDescriptor> &renderPass,
201 QVideoFrame &frame,
202 const std::unique_ptr<QVideoFrameTextures> &videoFrameTextures)
203{
204 auto format = frame.surfaceFormat();
205 auto pixelFormat = format.pixelFormat();
206
207 auto textureDesc = QVideoTextureHelper::textureDescription(format: pixelFormat);
208
209 QRhiShaderResourceBinding bindings[4];
210 auto *b = bindings;
211 *b++ = QRhiShaderResourceBinding::uniformBuffer(binding: 0, stage: QRhiShaderResourceBinding::VertexStage | QRhiShaderResourceBinding::FragmentStage,
212 buf: uniformBuffer.get());
213 for (int i = 0; i < textureDesc->nplanes; ++i)
214 *b++ = QRhiShaderResourceBinding::sampledTexture(binding: i + 1, stage: QRhiShaderResourceBinding::FragmentStage,
215 tex: videoFrameTextures->texture(plane: i), sampler: textureSampler.get());
216 shaderResourceBindings->setBindings(first: bindings, last: b);
217 shaderResourceBindings->create();
218
219 graphicsPipeline.reset(p: rhi->newGraphicsPipeline());
220 graphicsPipeline->setTopology(QRhiGraphicsPipeline::TriangleStrip);
221
222 QShader vs = vfcGetShader(name: QVideoTextureHelper::vertexShaderFileName(format));
223 if (!vs.isValid())
224 return false;
225
226 QShader fs = vfcGetShader(name: QVideoTextureHelper::fragmentShaderFileName(format));
227 if (!fs.isValid())
228 return false;
229
230 graphicsPipeline->setShaderStages({
231 { QRhiShaderStage::Vertex, vs },
232 { QRhiShaderStage::Fragment, fs }
233 });
234
235 QRhiVertexInputLayout inputLayout;
236 inputLayout.setBindings({
237 { 4 * sizeof(float) }
238 });
239 inputLayout.setAttributes({
240 { 0, 0, QRhiVertexInputAttribute::Float2, 0 },
241 { 0, 1, QRhiVertexInputAttribute::Float2, 2 * sizeof(float) }
242 });
243
244 graphicsPipeline->setVertexInputLayout(inputLayout);
245 graphicsPipeline->setShaderResourceBindings(shaderResourceBindings.get());
246 graphicsPipeline->setRenderPassDescriptor(renderPass.get());
247 graphicsPipeline->create();
248
249 return true;
250}
251
252static QImage convertJPEG(const QVideoFrame &frame, const VideoTransformation &transform)
253{
254 QVideoFrame varFrame = frame;
255 if (!varFrame.map(mode: QVideoFrame::ReadOnly)) {
256 qCDebug(qLcVideoFrameConverter) << Q_FUNC_INFO << ": frame mapping failed";
257 return {};
258 }
259 QImage image;
260 image.loadFromData(buf: varFrame.bits(plane: 0), len: varFrame.mappedBytes(plane: 0), format: "JPG");
261 varFrame.unmap();
262 rasterTransform(image, transformation: transform);
263 return image;
264}
265
266static QImage convertCPU(const QVideoFrame &frame, const VideoTransformation &transform)
267{
268 VideoFrameConvertFunc convert = qConverterForFormat(format: frame.pixelFormat());
269 if (!convert) {
270 qCDebug(qLcVideoFrameConverter) << Q_FUNC_INFO << ": unsupported pixel format" << frame.pixelFormat();
271 return {};
272 } else {
273 QVideoFrame varFrame = frame;
274 if (!varFrame.map(mode: QVideoFrame::ReadOnly)) {
275 qCDebug(qLcVideoFrameConverter) << Q_FUNC_INFO << ": frame mapping failed";
276 return {};
277 }
278 auto format = pixelFormatHasAlpha(format: varFrame.pixelFormat()) ? QImage::Format_ARGB32_Premultiplied : QImage::Format_RGB32;
279 QImage image = QImage(varFrame.width(), varFrame.height(), format);
280 convert(varFrame, image.bits());
281 varFrame.unmap();
282 rasterTransform(image, transformation: transform);
283 return image;
284 }
285}
286
287QImage qImageFromVideoFrame(const QVideoFrame &frame, bool forceCpu)
288{
289 // by default, surface transformation is applied, as full transformation is used for presentation only
290 return qImageFromVideoFrame(frame, transformation: qNormalizedSurfaceTransformation(format: frame.surfaceFormat()),
291 forceCpu);
292}
293
294QImage qImageFromVideoFrame(const QVideoFrame &frame, const VideoTransformation &transformation,
295 bool forceCpu)
296{
297#ifdef Q_OS_DARWIN
298 QMacAutoReleasePool releasePool;
299#endif
300
301 if (!g_state.hasLocalData())
302 g_state.setLocalData({});
303
304 std::unique_ptr<QRhiRenderPassDescriptor> renderPass;
305 std::unique_ptr<QRhiBuffer> vertexBuffer;
306 std::unique_ptr<QRhiBuffer> uniformBuffer;
307 std::unique_ptr<QRhiTexture> targetTexture;
308 std::unique_ptr<QRhiTextureRenderTarget> renderTarget;
309 std::unique_ptr<QRhiSampler> textureSampler;
310 std::unique_ptr<QRhiShaderResourceBindings> shaderResourceBindings;
311 std::unique_ptr<QRhiGraphicsPipeline> graphicsPipeline;
312
313 if (frame.size().isEmpty() || frame.pixelFormat() == QVideoFrameFormat::Format_Invalid)
314 return {};
315
316 if (frame.pixelFormat() == QVideoFrameFormat::Format_Jpeg)
317 return convertJPEG(frame, transform: transformation);
318
319 if (forceCpu) // For test purposes
320 return convertCPU(frame, transform: transformation);
321
322 QRhi *rhi = nullptr;
323
324 if (QHwVideoBuffer *buffer = QVideoFramePrivate::hwBuffer(frame))
325 rhi = buffer->rhi();
326
327 if (!rhi || rhi->thread() != QThread::currentThread())
328 rhi = initializeRHI(videoFrameRhi: rhi);
329
330 if (!rhi || rhi->isRecordingFrame())
331 return convertCPU(frame, transform: transformation);
332
333 // Do conversion using shaders
334
335 const QSize frameSize = qRotatedFrameSize(size: frame.size(), rotation: frame.surfaceFormat().rotation());
336
337 vertexBuffer.reset(p: rhi->newBuffer(type: QRhiBuffer::Immutable, usage: QRhiBuffer::VertexBuffer, size: sizeof(g_quad)));
338 vertexBuffer->create();
339
340 uniformBuffer.reset(p: rhi->newBuffer(type: QRhiBuffer::Dynamic, usage: QRhiBuffer::UniformBuffer, size: 64 + 64 + 4 + 4 + 4 + 4));
341 uniformBuffer->create();
342
343 textureSampler.reset(p: rhi->newSampler(magFilter: QRhiSampler::Linear, minFilter: QRhiSampler::Linear, mipmapMode: QRhiSampler::None,
344 addressU: QRhiSampler::ClampToEdge, addressV: QRhiSampler::ClampToEdge));
345 textureSampler->create();
346
347 shaderResourceBindings.reset(p: rhi->newShaderResourceBindings());
348
349 targetTexture.reset(p: rhi->newTexture(format: QRhiTexture::RGBA8, pixelSize: frameSize, sampleCount: 1, flags: QRhiTexture::RenderTarget));
350 if (!targetTexture->create()) {
351 qCDebug(qLcVideoFrameConverter) << "Failed to create target texture. Using CPU conversion.";
352 return convertCPU(frame, transform: transformation);
353 }
354
355 renderTarget.reset(p: rhi->newTextureRenderTarget(desc: { { targetTexture.get() } }));
356 renderPass.reset(p: renderTarget->newCompatibleRenderPassDescriptor());
357 renderTarget->setRenderPassDescriptor(renderPass.get());
358 renderTarget->create();
359
360 QRhiCommandBuffer *cb = nullptr;
361 QRhi::FrameOpResult r = rhi->beginOffscreenFrame(cb: &cb);
362 if (r != QRhi::FrameOpSuccess) {
363 qCDebug(qLcVideoFrameConverter) << "Failed to set up offscreen frame. Using CPU conversion.";
364 return convertCPU(frame, transform: transformation);
365 }
366
367 QRhiResourceUpdateBatch *rub = rhi->nextResourceUpdateBatch();
368
369 rub->uploadStaticBuffer(buf: vertexBuffer.get(), data: g_quad);
370
371 QVideoFrame frameTmp = frame;
372 auto videoFrameTextures = QVideoTextureHelper::createTextures(frame&: frameTmp, rhi, rub, oldTextures: {});
373 if (!videoFrameTextures) {
374 qCDebug(qLcVideoFrameConverter) << "Failed obtain textures. Using CPU conversion.";
375 return convertCPU(frame, transform: transformation);
376 }
377
378 if (!updateTextures(rhi, uniformBuffer, textureSampler, shaderResourceBindings,
379 graphicsPipeline, renderPass, frame&: frameTmp, videoFrameTextures)) {
380 qCDebug(qLcVideoFrameConverter) << "Failed to update textures. Using CPU conversion.";
381 return convertCPU(frame, transform: transformation);
382 }
383
384 float xScale = transformation.mirrorredHorizontallyAfterRotation ? -1.0 : 1.0;
385 float yScale = 1.f;
386
387 if (rhi->isYUpInFramebuffer())
388 yScale = -yScale;
389
390 QMatrix4x4 transform;
391 transform.scale(x: xScale, y: yScale);
392
393 QByteArray uniformData(64 + 64 + 4 + 4, Qt::Uninitialized);
394 QVideoTextureHelper::updateUniformData(dst: &uniformData, format: frame.surfaceFormat(), frame, transform, opacity: 1.f);
395 rub->updateDynamicBuffer(buf: uniformBuffer.get(), offset: 0, size: uniformData.size(), data: uniformData.constData());
396
397 cb->beginPass(rt: renderTarget.get(), colorClearValue: Qt::black, depthStencilClearValue: { 1.0f, 0 }, resourceUpdates: rub);
398 cb->setGraphicsPipeline(graphicsPipeline.get());
399
400 cb->setViewport({ 0, 0, float(frameSize.width()), float(frameSize.height()) });
401 cb->setShaderResources(srb: shaderResourceBindings.get());
402
403 const quint32 vertexOffset = quint32(sizeof(float)) * 16 * transformation.rotationIndex();
404 const QRhiCommandBuffer::VertexInput vbufBinding(vertexBuffer.get(), vertexOffset);
405 cb->setVertexInput(startBinding: 0, bindingCount: 1, bindings: &vbufBinding);
406 cb->draw(vertexCount: 4);
407
408 QRhiReadbackDescription readDesc(targetTexture.get());
409 QRhiReadbackResult readResult;
410 bool readCompleted = false;
411
412 readResult.completed = [&readCompleted] { readCompleted = true; };
413
414 rub = rhi->nextResourceUpdateBatch();
415 rub->readBackTexture(rb: readDesc, result: &readResult);
416
417 cb->endPass(resourceUpdates: rub);
418
419 rhi->endOffscreenFrame();
420
421 if (!readCompleted) {
422 qCDebug(qLcVideoFrameConverter) << "Failed to read back texture. Using CPU conversion.";
423 return convertCPU(frame, transform: transformation);
424 }
425
426 QByteArray *imageData = new QByteArray(readResult.data);
427
428 return QImage(reinterpret_cast<const uchar *>(imageData->constData()),
429 readResult.pixelSize.width(), readResult.pixelSize.height(),
430 QImage::Format_RGBA8888_Premultiplied, imageCleanupHandler, imageData);
431}
432
433QImage videoFramePlaneAsImage(QVideoFrame &frame, int plane, QImage::Format targetFormat,
434 QSize targetSize)
435{
436 if (plane >= frame.planeCount())
437 return {};
438
439 if (!frame.map(mode: QVideoFrame::ReadOnly)) {
440 qWarning() << "Cannot map a video frame in ReadOnly mode!";
441 return {};
442 }
443
444 auto frameHandle = QVideoFramePrivate::handle(frame);
445
446 // With incrementing the reference counter, we share the mapped QVideoFrame
447 // with the target QImage. The function imageCleanupFunction is going to adopt
448 // the frameHandle by QVideoFrame and dereference it upon the destruction.
449 frameHandle->ref.ref();
450
451 auto imageCleanupFunction = [](void *data) {
452 QVideoFrame frame = reinterpret_cast<QVideoFramePrivate *>(data)->adoptThisByVideoFrame();
453 Q_ASSERT(frame.isMapped());
454 frame.unmap();
455 };
456
457 const auto bytesPerLine = frame.bytesPerLine(plane);
458 const auto height =
459 bytesPerLine ? qMin(a: targetSize.height(), b: frame.mappedBytes(plane) / bytesPerLine) : 0;
460
461 return QImage(reinterpret_cast<const uchar *>(frame.bits(plane)), targetSize.width(), height,
462 bytesPerLine, targetFormat, imageCleanupFunction, frameHandle);
463}
464
465QT_END_NAMESPACE
466
467

Provided by KDAB

Privacy Policy
Learn to use CMake with our Intro Training
Find out more

source code of qtmultimedia/src/multimedia/video/qvideoframeconverter.cpp