1 | // Copyright (C) 2022 The Qt Company Ltd. |
2 | // SPDX-License-Identifier: LicenseRef-Qt-Commercial OR LGPL-3.0-only OR GPL-2.0-only OR GPL-3.0-only |
3 | |
4 | #include "qvideoframeconverter_p.h" |
5 | #include "qvideoframeconversionhelper_p.h" |
6 | #include "qvideoframeformat.h" |
7 | #include "qvideoframe_p.h" |
8 | #include "qmultimediautils_p.h" |
9 | #include "qabstractvideobuffer.h" |
10 | |
11 | #include <QtCore/qcoreapplication.h> |
12 | #include <QtCore/qsize.h> |
13 | #include <QtCore/qhash.h> |
14 | #include <QtCore/qfile.h> |
15 | #include <QtCore/qthreadstorage.h> |
16 | #include <QtGui/qimage.h> |
17 | #include <QtGui/qoffscreensurface.h> |
18 | #include <qpa/qplatformintegration.h> |
19 | #include <private/qvideotexturehelper_p.h> |
20 | #include <private/qguiapplication_p.h> |
21 | #include <rhi/qrhi.h> |
22 | |
23 | #ifdef Q_OS_DARWIN |
24 | #include <QtCore/private/qcore_mac_p.h> |
25 | #endif |
26 | |
27 | QT_BEGIN_NAMESPACE |
28 | |
29 | static Q_LOGGING_CATEGORY(qLcVideoFrameConverter, "qt.multimedia.video.frameconverter" ) |
30 | |
31 | namespace { |
32 | |
33 | struct State |
34 | { |
35 | QRhi *rhi = nullptr; |
36 | #if QT_CONFIG(opengl) |
37 | QOffscreenSurface *fallbackSurface = nullptr; |
38 | #endif |
39 | bool cpuOnly = false; |
40 | #if defined(Q_OS_ANDROID) |
41 | QMetaObject::Connection appStateChangedConnection; |
42 | #endif |
43 | ~State() { |
44 | resetRhi(); |
45 | } |
46 | |
47 | void resetRhi() { |
48 | delete rhi; |
49 | rhi = nullptr; |
50 | #if QT_CONFIG(opengl) |
51 | delete fallbackSurface; |
52 | fallbackSurface = nullptr; |
53 | #endif |
54 | cpuOnly = false; |
55 | } |
56 | }; |
57 | |
58 | } |
59 | |
60 | static QThreadStorage<State> g_state; |
61 | static QHash<QString, QShader> g_shaderCache; |
62 | |
63 | static const float g_quad[] = { |
64 | // Rotation 0 CW |
65 | 1.f, -1.f, 1.f, 1.f, |
66 | 1.f, 1.f, 1.f, 0.f, |
67 | -1.f, -1.f, 0.f, 1.f, |
68 | -1.f, 1.f, 0.f, 0.f, |
69 | // Rotation 90 CW |
70 | 1.f, -1.f, 1.f, 0.f, |
71 | 1.f, 1.f, 0.f, 0.f, |
72 | -1.f, -1.f, 1.f, 1.f, |
73 | -1.f, 1.f, 0.f, 1.f, |
74 | // Rotation 180 CW |
75 | 1.f, -1.f, 0.f, 0.f, |
76 | 1.f, 1.f, 0.f, 1.f, |
77 | -1.f, -1.f, 1.f, 0.f, |
78 | -1.f, 1.f, 1.f, 1.f, |
79 | // Rotation 270 CW |
80 | 1.f, -1.f, 0.f, 1.f, |
81 | 1.f, 1.f, 1.f, 1.f, |
82 | -1.f, -1.f, 0.f, 0.f, |
83 | -1.f, 1.f, 1.f, 0.f, |
84 | }; |
85 | |
86 | static bool pixelFormatHasAlpha(QVideoFrameFormat::PixelFormat format) |
87 | { |
88 | switch (format) { |
89 | case QVideoFrameFormat::Format_ARGB8888: |
90 | case QVideoFrameFormat::Format_ARGB8888_Premultiplied: |
91 | case QVideoFrameFormat::Format_BGRA8888: |
92 | case QVideoFrameFormat::Format_BGRA8888_Premultiplied: |
93 | case QVideoFrameFormat::Format_ABGR8888: |
94 | case QVideoFrameFormat::Format_RGBA8888: |
95 | case QVideoFrameFormat::Format_AYUV: |
96 | case QVideoFrameFormat::Format_AYUV_Premultiplied: |
97 | return true; |
98 | default: |
99 | return false; |
100 | } |
101 | }; |
102 | |
103 | static QShader vfcGetShader(const QString &name) |
104 | { |
105 | QShader shader = g_shaderCache.value(key: name); |
106 | if (shader.isValid()) |
107 | return shader; |
108 | |
109 | QFile f(name); |
110 | if (f.open(flags: QIODevice::ReadOnly)) |
111 | shader = QShader::fromSerialized(data: f.readAll()); |
112 | |
113 | if (shader.isValid()) |
114 | g_shaderCache[name] = shader; |
115 | |
116 | return shader; |
117 | } |
118 | |
119 | static void rasterTransform(QImage &image, VideoTransformation transformation) |
120 | { |
121 | QTransform t; |
122 | if (transformation.rotation != QtVideo::Rotation::None) |
123 | t.rotate(a: qreal(transformation.rotation)); |
124 | if (transformation.mirrorredHorizontallyAfterRotation) |
125 | t.scale(sx: -1., sy: 1); |
126 | if (!t.isIdentity()) |
127 | image = image.transformed(matrix: t); |
128 | } |
129 | |
130 | static void imageCleanupHandler(void *info) |
131 | { |
132 | QByteArray *imageData = reinterpret_cast<QByteArray *>(info); |
133 | delete imageData; |
134 | } |
135 | |
136 | static QRhi *initializeRHI(QRhi *videoFrameRhi) |
137 | { |
138 | if (g_state.localData().rhi || g_state.localData().cpuOnly) |
139 | return g_state.localData().rhi; |
140 | |
141 | QRhi::Implementation backend = videoFrameRhi ? videoFrameRhi->backend() : QRhi::Null; |
142 | const QPlatformIntegration *qpa = QGuiApplicationPrivate::platformIntegration(); |
143 | |
144 | if (qpa && qpa->hasCapability(cap: QPlatformIntegration::RhiBasedRendering)) { |
145 | |
146 | #if defined(Q_OS_MACOS) || defined(Q_OS_IOS) |
147 | if (backend == QRhi::Metal || backend == QRhi::Null) { |
148 | QRhiMetalInitParams params; |
149 | g_state.localData().rhi = QRhi::create(QRhi::Metal, ¶ms); |
150 | } |
151 | #endif |
152 | |
153 | #if defined(Q_OS_WIN) |
154 | if (backend == QRhi::D3D11 || backend == QRhi::Null) { |
155 | QRhiD3D11InitParams params; |
156 | g_state.localData().rhi = QRhi::create(QRhi::D3D11, ¶ms); |
157 | } |
158 | #endif |
159 | |
160 | #if QT_CONFIG(opengl) |
161 | if (!g_state.localData().rhi && (backend == QRhi::OpenGLES2 || backend == QRhi::Null)) { |
162 | if (qpa->hasCapability(cap: QPlatformIntegration::OpenGL) |
163 | && qpa->hasCapability(cap: QPlatformIntegration::RasterGLSurface) |
164 | && !QCoreApplication::testAttribute(attribute: Qt::AA_ForceRasterWidgets)) { |
165 | |
166 | g_state.localData().fallbackSurface = QRhiGles2InitParams::newFallbackSurface(); |
167 | QRhiGles2InitParams params; |
168 | params.fallbackSurface = g_state.localData().fallbackSurface; |
169 | if (backend == QRhi::OpenGLES2) |
170 | params.shareContext = static_cast<const QRhiGles2NativeHandles*>(videoFrameRhi->nativeHandles())->context; |
171 | g_state.localData().rhi = QRhi::create(impl: QRhi::OpenGLES2, params: ¶ms); |
172 | |
173 | #if defined(Q_OS_ANDROID) |
174 | // reset RHI state on application suspension, as this will be invalid after resuming |
175 | if (!g_state.localData().appStateChangedConnection) { |
176 | g_state.localData().appStateChangedConnection = QObject::connect(qApp, &QGuiApplication::applicationStateChanged, qApp, [](auto state) { |
177 | if (state == Qt::ApplicationSuspended) |
178 | g_state.localData().resetRhi(); |
179 | }); |
180 | } |
181 | #endif |
182 | } |
183 | } |
184 | #endif |
185 | } |
186 | |
187 | if (!g_state.localData().rhi) { |
188 | g_state.localData().cpuOnly = true; |
189 | qWarning() << Q_FUNC_INFO << ": No RHI backend. Using CPU conversion." ; |
190 | } |
191 | |
192 | return g_state.localData().rhi; |
193 | } |
194 | |
195 | static bool updateTextures(QRhi *rhi, |
196 | std::unique_ptr<QRhiBuffer> &uniformBuffer, |
197 | std::unique_ptr<QRhiSampler> &textureSampler, |
198 | std::unique_ptr<QRhiShaderResourceBindings> &shaderResourceBindings, |
199 | std::unique_ptr<QRhiGraphicsPipeline> &graphicsPipeline, |
200 | std::unique_ptr<QRhiRenderPassDescriptor> &renderPass, |
201 | QVideoFrame &frame, |
202 | const std::unique_ptr<QVideoFrameTextures> &videoFrameTextures) |
203 | { |
204 | auto format = frame.surfaceFormat(); |
205 | auto pixelFormat = format.pixelFormat(); |
206 | |
207 | auto textureDesc = QVideoTextureHelper::textureDescription(format: pixelFormat); |
208 | |
209 | QRhiShaderResourceBinding bindings[4]; |
210 | auto *b = bindings; |
211 | *b++ = QRhiShaderResourceBinding::uniformBuffer(binding: 0, stage: QRhiShaderResourceBinding::VertexStage | QRhiShaderResourceBinding::FragmentStage, |
212 | buf: uniformBuffer.get()); |
213 | for (int i = 0; i < textureDesc->nplanes; ++i) |
214 | *b++ = QRhiShaderResourceBinding::sampledTexture(binding: i + 1, stage: QRhiShaderResourceBinding::FragmentStage, |
215 | tex: videoFrameTextures->texture(plane: i), sampler: textureSampler.get()); |
216 | shaderResourceBindings->setBindings(first: bindings, last: b); |
217 | shaderResourceBindings->create(); |
218 | |
219 | graphicsPipeline.reset(p: rhi->newGraphicsPipeline()); |
220 | graphicsPipeline->setTopology(QRhiGraphicsPipeline::TriangleStrip); |
221 | |
222 | QShader vs = vfcGetShader(name: QVideoTextureHelper::vertexShaderFileName(format)); |
223 | if (!vs.isValid()) |
224 | return false; |
225 | |
226 | QShader fs = vfcGetShader(name: QVideoTextureHelper::fragmentShaderFileName(format)); |
227 | if (!fs.isValid()) |
228 | return false; |
229 | |
230 | graphicsPipeline->setShaderStages({ |
231 | { QRhiShaderStage::Vertex, vs }, |
232 | { QRhiShaderStage::Fragment, fs } |
233 | }); |
234 | |
235 | QRhiVertexInputLayout inputLayout; |
236 | inputLayout.setBindings({ |
237 | { 4 * sizeof(float) } |
238 | }); |
239 | inputLayout.setAttributes({ |
240 | { 0, 0, QRhiVertexInputAttribute::Float2, 0 }, |
241 | { 0, 1, QRhiVertexInputAttribute::Float2, 2 * sizeof(float) } |
242 | }); |
243 | |
244 | graphicsPipeline->setVertexInputLayout(inputLayout); |
245 | graphicsPipeline->setShaderResourceBindings(shaderResourceBindings.get()); |
246 | graphicsPipeline->setRenderPassDescriptor(renderPass.get()); |
247 | graphicsPipeline->create(); |
248 | |
249 | return true; |
250 | } |
251 | |
252 | static QImage convertJPEG(const QVideoFrame &frame, const VideoTransformation &transform) |
253 | { |
254 | QVideoFrame varFrame = frame; |
255 | if (!varFrame.map(mode: QVideoFrame::ReadOnly)) { |
256 | qCDebug(qLcVideoFrameConverter) << Q_FUNC_INFO << ": frame mapping failed" ; |
257 | return {}; |
258 | } |
259 | QImage image; |
260 | image.loadFromData(buf: varFrame.bits(plane: 0), len: varFrame.mappedBytes(plane: 0), format: "JPG" ); |
261 | varFrame.unmap(); |
262 | rasterTransform(image, transformation: transform); |
263 | return image; |
264 | } |
265 | |
266 | static QImage convertCPU(const QVideoFrame &frame, const VideoTransformation &transform) |
267 | { |
268 | VideoFrameConvertFunc convert = qConverterForFormat(format: frame.pixelFormat()); |
269 | if (!convert) { |
270 | qCDebug(qLcVideoFrameConverter) << Q_FUNC_INFO << ": unsupported pixel format" << frame.pixelFormat(); |
271 | return {}; |
272 | } else { |
273 | QVideoFrame varFrame = frame; |
274 | if (!varFrame.map(mode: QVideoFrame::ReadOnly)) { |
275 | qCDebug(qLcVideoFrameConverter) << Q_FUNC_INFO << ": frame mapping failed" ; |
276 | return {}; |
277 | } |
278 | auto format = pixelFormatHasAlpha(format: varFrame.pixelFormat()) ? QImage::Format_ARGB32_Premultiplied : QImage::Format_RGB32; |
279 | QImage image = QImage(varFrame.width(), varFrame.height(), format); |
280 | convert(varFrame, image.bits()); |
281 | varFrame.unmap(); |
282 | rasterTransform(image, transformation: transform); |
283 | return image; |
284 | } |
285 | } |
286 | |
287 | QImage qImageFromVideoFrame(const QVideoFrame &frame, bool forceCpu) |
288 | { |
289 | // by default, surface transformation is applied, as full transformation is used for presentation only |
290 | return qImageFromVideoFrame(frame, transformation: qNormalizedSurfaceTransformation(format: frame.surfaceFormat()), |
291 | forceCpu); |
292 | } |
293 | |
294 | QImage qImageFromVideoFrame(const QVideoFrame &frame, const VideoTransformation &transformation, |
295 | bool forceCpu) |
296 | { |
297 | #ifdef Q_OS_DARWIN |
298 | QMacAutoReleasePool releasePool; |
299 | #endif |
300 | |
301 | if (!g_state.hasLocalData()) |
302 | g_state.setLocalData({}); |
303 | |
304 | std::unique_ptr<QRhiRenderPassDescriptor> renderPass; |
305 | std::unique_ptr<QRhiBuffer> vertexBuffer; |
306 | std::unique_ptr<QRhiBuffer> uniformBuffer; |
307 | std::unique_ptr<QRhiTexture> targetTexture; |
308 | std::unique_ptr<QRhiTextureRenderTarget> renderTarget; |
309 | std::unique_ptr<QRhiSampler> textureSampler; |
310 | std::unique_ptr<QRhiShaderResourceBindings> shaderResourceBindings; |
311 | std::unique_ptr<QRhiGraphicsPipeline> graphicsPipeline; |
312 | |
313 | if (frame.size().isEmpty() || frame.pixelFormat() == QVideoFrameFormat::Format_Invalid) |
314 | return {}; |
315 | |
316 | if (frame.pixelFormat() == QVideoFrameFormat::Format_Jpeg) |
317 | return convertJPEG(frame, transform: transformation); |
318 | |
319 | if (forceCpu) // For test purposes |
320 | return convertCPU(frame, transform: transformation); |
321 | |
322 | QRhi *rhi = nullptr; |
323 | |
324 | if (QHwVideoBuffer *buffer = QVideoFramePrivate::hwBuffer(frame)) |
325 | rhi = buffer->rhi(); |
326 | |
327 | if (!rhi || rhi->thread() != QThread::currentThread()) |
328 | rhi = initializeRHI(videoFrameRhi: rhi); |
329 | |
330 | if (!rhi || rhi->isRecordingFrame()) |
331 | return convertCPU(frame, transform: transformation); |
332 | |
333 | // Do conversion using shaders |
334 | |
335 | const QSize frameSize = qRotatedFrameSize(size: frame.size(), rotation: frame.surfaceFormat().rotation()); |
336 | |
337 | vertexBuffer.reset(p: rhi->newBuffer(type: QRhiBuffer::Immutable, usage: QRhiBuffer::VertexBuffer, size: sizeof(g_quad))); |
338 | vertexBuffer->create(); |
339 | |
340 | uniformBuffer.reset(p: rhi->newBuffer(type: QRhiBuffer::Dynamic, usage: QRhiBuffer::UniformBuffer, size: 64 + 64 + 4 + 4 + 4 + 4)); |
341 | uniformBuffer->create(); |
342 | |
343 | textureSampler.reset(p: rhi->newSampler(magFilter: QRhiSampler::Linear, minFilter: QRhiSampler::Linear, mipmapMode: QRhiSampler::None, |
344 | addressU: QRhiSampler::ClampToEdge, addressV: QRhiSampler::ClampToEdge)); |
345 | textureSampler->create(); |
346 | |
347 | shaderResourceBindings.reset(p: rhi->newShaderResourceBindings()); |
348 | |
349 | targetTexture.reset(p: rhi->newTexture(format: QRhiTexture::RGBA8, pixelSize: frameSize, sampleCount: 1, flags: QRhiTexture::RenderTarget)); |
350 | if (!targetTexture->create()) { |
351 | qCDebug(qLcVideoFrameConverter) << "Failed to create target texture. Using CPU conversion." ; |
352 | return convertCPU(frame, transform: transformation); |
353 | } |
354 | |
355 | renderTarget.reset(p: rhi->newTextureRenderTarget(desc: { { targetTexture.get() } })); |
356 | renderPass.reset(p: renderTarget->newCompatibleRenderPassDescriptor()); |
357 | renderTarget->setRenderPassDescriptor(renderPass.get()); |
358 | renderTarget->create(); |
359 | |
360 | QRhiCommandBuffer *cb = nullptr; |
361 | QRhi::FrameOpResult r = rhi->beginOffscreenFrame(cb: &cb); |
362 | if (r != QRhi::FrameOpSuccess) { |
363 | qCDebug(qLcVideoFrameConverter) << "Failed to set up offscreen frame. Using CPU conversion." ; |
364 | return convertCPU(frame, transform: transformation); |
365 | } |
366 | |
367 | QRhiResourceUpdateBatch *rub = rhi->nextResourceUpdateBatch(); |
368 | |
369 | rub->uploadStaticBuffer(buf: vertexBuffer.get(), data: g_quad); |
370 | |
371 | QVideoFrame frameTmp = frame; |
372 | auto videoFrameTextures = QVideoTextureHelper::createTextures(frame&: frameTmp, rhi, rub, oldTextures: {}); |
373 | if (!videoFrameTextures) { |
374 | qCDebug(qLcVideoFrameConverter) << "Failed obtain textures. Using CPU conversion." ; |
375 | return convertCPU(frame, transform: transformation); |
376 | } |
377 | |
378 | if (!updateTextures(rhi, uniformBuffer, textureSampler, shaderResourceBindings, |
379 | graphicsPipeline, renderPass, frame&: frameTmp, videoFrameTextures)) { |
380 | qCDebug(qLcVideoFrameConverter) << "Failed to update textures. Using CPU conversion." ; |
381 | return convertCPU(frame, transform: transformation); |
382 | } |
383 | |
384 | float xScale = transformation.mirrorredHorizontallyAfterRotation ? -1.0 : 1.0; |
385 | float yScale = 1.f; |
386 | |
387 | if (rhi->isYUpInFramebuffer()) |
388 | yScale = -yScale; |
389 | |
390 | QMatrix4x4 transform; |
391 | transform.scale(x: xScale, y: yScale); |
392 | |
393 | QByteArray uniformData(64 + 64 + 4 + 4, Qt::Uninitialized); |
394 | QVideoTextureHelper::updateUniformData(dst: &uniformData, format: frame.surfaceFormat(), frame, transform, opacity: 1.f); |
395 | rub->updateDynamicBuffer(buf: uniformBuffer.get(), offset: 0, size: uniformData.size(), data: uniformData.constData()); |
396 | |
397 | cb->beginPass(rt: renderTarget.get(), colorClearValue: Qt::black, depthStencilClearValue: { 1.0f, 0 }, resourceUpdates: rub); |
398 | cb->setGraphicsPipeline(graphicsPipeline.get()); |
399 | |
400 | cb->setViewport({ 0, 0, float(frameSize.width()), float(frameSize.height()) }); |
401 | cb->setShaderResources(srb: shaderResourceBindings.get()); |
402 | |
403 | const quint32 vertexOffset = quint32(sizeof(float)) * 16 * transformation.rotationIndex(); |
404 | const QRhiCommandBuffer::VertexInput vbufBinding(vertexBuffer.get(), vertexOffset); |
405 | cb->setVertexInput(startBinding: 0, bindingCount: 1, bindings: &vbufBinding); |
406 | cb->draw(vertexCount: 4); |
407 | |
408 | QRhiReadbackDescription readDesc(targetTexture.get()); |
409 | QRhiReadbackResult readResult; |
410 | bool readCompleted = false; |
411 | |
412 | readResult.completed = [&readCompleted] { readCompleted = true; }; |
413 | |
414 | rub = rhi->nextResourceUpdateBatch(); |
415 | rub->readBackTexture(rb: readDesc, result: &readResult); |
416 | |
417 | cb->endPass(resourceUpdates: rub); |
418 | |
419 | rhi->endOffscreenFrame(); |
420 | |
421 | if (!readCompleted) { |
422 | qCDebug(qLcVideoFrameConverter) << "Failed to read back texture. Using CPU conversion." ; |
423 | return convertCPU(frame, transform: transformation); |
424 | } |
425 | |
426 | QByteArray *imageData = new QByteArray(readResult.data); |
427 | |
428 | return QImage(reinterpret_cast<const uchar *>(imageData->constData()), |
429 | readResult.pixelSize.width(), readResult.pixelSize.height(), |
430 | QImage::Format_RGBA8888_Premultiplied, imageCleanupHandler, imageData); |
431 | } |
432 | |
433 | QImage videoFramePlaneAsImage(QVideoFrame &frame, int plane, QImage::Format targetFormat, |
434 | QSize targetSize) |
435 | { |
436 | if (plane >= frame.planeCount()) |
437 | return {}; |
438 | |
439 | if (!frame.map(mode: QVideoFrame::ReadOnly)) { |
440 | qWarning() << "Cannot map a video frame in ReadOnly mode!" ; |
441 | return {}; |
442 | } |
443 | |
444 | auto frameHandle = QVideoFramePrivate::handle(frame); |
445 | |
446 | // With incrementing the reference counter, we share the mapped QVideoFrame |
447 | // with the target QImage. The function imageCleanupFunction is going to adopt |
448 | // the frameHandle by QVideoFrame and dereference it upon the destruction. |
449 | frameHandle->ref.ref(); |
450 | |
451 | auto imageCleanupFunction = [](void *data) { |
452 | QVideoFrame frame = reinterpret_cast<QVideoFramePrivate *>(data)->adoptThisByVideoFrame(); |
453 | Q_ASSERT(frame.isMapped()); |
454 | frame.unmap(); |
455 | }; |
456 | |
457 | const auto bytesPerLine = frame.bytesPerLine(plane); |
458 | const auto height = |
459 | bytesPerLine ? qMin(a: targetSize.height(), b: frame.mappedBytes(plane) / bytesPerLine) : 0; |
460 | |
461 | return QImage(reinterpret_cast<const uchar *>(frame.bits(plane)), targetSize.width(), height, |
462 | bytesPerLine, targetFormat, imageCleanupFunction, frameHandle); |
463 | } |
464 | |
465 | QT_END_NAMESPACE |
466 | |
467 | |