blob: ab9688a43f13df05df26f4bd8f88fb9446b2c7cc [file] [log] [blame]
Magnus Jedvert577bc192016-10-19 15:29:02 +02001/*
2 * Copyright 2015 The WebRTC project authors. All Rights Reserved.
3 *
4 * Use of this source code is governed by a BSD-style license
5 * that can be found in the LICENSE file in the root of the source
6 * tree. An additional intellectual property rights grant can be found
7 * in the file PATENTS. All contributing project authors may
8 * be found in the AUTHORS file in the root of the source tree.
9 */
10
11package org.webrtc;
12
Magnus Jedvert65c61dc2018-06-15 09:33:20 +020013import android.graphics.Matrix;
Magnus Jedvert577bc192016-10-19 15:29:02 +020014import android.opengl.GLES20;
Fabian Bergmark1bb36d22021-06-17 11:53:24 +020015import android.opengl.GLException;
Byoungchan Lee02334e02021-08-14 11:41:59 +090016import androidx.annotation.Nullable;
Magnus Jedvert577bc192016-10-19 15:29:02 +020017import java.nio.ByteBuffer;
Sami Kalliomäkicb98b112017-10-16 11:20:26 +020018import org.webrtc.VideoFrame.I420Buffer;
19import org.webrtc.VideoFrame.TextureBuffer;
Magnus Jedvert577bc192016-10-19 15:29:02 +020020
21/**
Magnus Jedvert1d270f82018-04-16 16:28:29 +020022 * Class for converting OES textures to a YUV ByteBuffer. It can be constructed on any thread, but
23 * should only be operated from a single thread with an active EGL context.
Magnus Jedvert577bc192016-10-19 15:29:02 +020024 */
Fabian Bergmark1bb36d22021-06-17 11:53:24 +020025public final class YuvConverter {
26 private static final String TAG = "YuvConverter";
27
Magnus Jedvert65c61dc2018-06-15 09:33:20 +020028 private static final String FRAGMENT_SHADER =
Magnus Jedvert577bc192016-10-19 15:29:02 +020029 // Difference in texture coordinate corresponding to one
30 // sub-pixel in the x direction.
Magnus Jedvert65c61dc2018-06-15 09:33:20 +020031 "uniform vec2 xUnit;\n"
Magnus Jedvert577bc192016-10-19 15:29:02 +020032 // Color conversion coefficients, including constant term
33 + "uniform vec4 coeffs;\n"
34 + "\n"
35 + "void main() {\n"
36 // Since the alpha read from the texture is always 1, this could
37 // be written as a mat4 x vec4 multiply. However, that seems to
38 // give a worse framerate, possibly because the additional
39 // multiplies by 1.0 consume resources. TODO(nisse): Could also
40 // try to do it as a vec3 x mat3x4, followed by an add in of a
41 // constant vector.
42 + " gl_FragColor.r = coeffs.a + dot(coeffs.rgb,\n"
Magnus Jedvert65c61dc2018-06-15 09:33:20 +020043 + " sample(tc - 1.5 * xUnit).rgb);\n"
Magnus Jedvert577bc192016-10-19 15:29:02 +020044 + " gl_FragColor.g = coeffs.a + dot(coeffs.rgb,\n"
Magnus Jedvert65c61dc2018-06-15 09:33:20 +020045 + " sample(tc - 0.5 * xUnit).rgb);\n"
Magnus Jedvert577bc192016-10-19 15:29:02 +020046 + " gl_FragColor.b = coeffs.a + dot(coeffs.rgb,\n"
Magnus Jedvert65c61dc2018-06-15 09:33:20 +020047 + " sample(tc + 0.5 * xUnit).rgb);\n"
Magnus Jedvert577bc192016-10-19 15:29:02 +020048 + " gl_FragColor.a = coeffs.a + dot(coeffs.rgb,\n"
Magnus Jedvert65c61dc2018-06-15 09:33:20 +020049 + " sample(tc + 1.5 * xUnit).rgb);\n"
Sami Kalliomäkicb98b112017-10-16 11:20:26 +020050 + "}\n";
51
Magnus Jedvert65c61dc2018-06-15 09:33:20 +020052 private static class ShaderCallbacks implements GlGenericDrawer.ShaderCallbacks {
Sami Kalliomäki8ccddff2018-09-05 11:43:38 +020053 // Y'UV444 to RGB888, see https://en.wikipedia.org/wiki/YUV#Y%E2%80%B2UV444_to_RGB888_conversion
54 // We use the ITU-R BT.601 coefficients for Y, U and V.
55 // The values in Wikipedia are inaccurate, the accurate values derived from the spec are:
56 // Y = 0.299 * R + 0.587 * G + 0.114 * B
57 // U = -0.168736 * R - 0.331264 * G + 0.5 * B + 0.5
58 // V = 0.5 * R - 0.418688 * G - 0.0813124 * B + 0.5
59 // To map the Y-values to range [16-235] and U- and V-values to range [16-240], the matrix has
60 // been multiplied with matrix:
61 // {{219 / 255, 0, 0, 16 / 255},
62 // {0, 224 / 255, 0, 16 / 255},
63 // {0, 0, 224 / 255, 16 / 255},
64 // {0, 0, 0, 1}}
65 private static final float[] yCoeffs =
66 new float[] {0.256788f, 0.504129f, 0.0979059f, 0.0627451f};
Magnus Jedvert65c61dc2018-06-15 09:33:20 +020067 private static final float[] uCoeffs =
Sami Kalliomäki8ccddff2018-09-05 11:43:38 +020068 new float[] {-0.148223f, -0.290993f, 0.439216f, 0.501961f};
69 private static final float[] vCoeffs =
70 new float[] {0.439216f, -0.367788f, -0.0714274f, 0.501961f};
Magnus Jedvert65c61dc2018-06-15 09:33:20 +020071
72 private int xUnitLoc;
73 private int coeffsLoc;
74
75 private float[] coeffs;
76 private float stepSize;
77
78 public void setPlaneY() {
79 coeffs = yCoeffs;
80 stepSize = 1.0f;
81 }
82
83 public void setPlaneU() {
84 coeffs = uCoeffs;
85 stepSize = 2.0f;
86 }
87
88 public void setPlaneV() {
89 coeffs = vCoeffs;
90 stepSize = 2.0f;
91 }
92
93 @Override
94 public void onNewShader(GlShader shader) {
95 xUnitLoc = shader.getUniformLocation("xUnit");
96 coeffsLoc = shader.getUniformLocation("coeffs");
97 }
98
99 @Override
100 public void onPrepareShader(GlShader shader, float[] texMatrix, int frameWidth, int frameHeight,
101 int viewportWidth, int viewportHeight) {
102 GLES20.glUniform4fv(coeffsLoc, /* count= */ 1, coeffs, /* offset= */ 0);
103 // Matrix * (1;0;0;0) / (width / stepSize). Note that OpenGL uses column major order.
104 GLES20.glUniform2f(
105 xUnitLoc, stepSize * texMatrix[0] / frameWidth, stepSize * texMatrix[1] / frameWidth);
106 }
107 }
Magnus Jedvert577bc192016-10-19 15:29:02 +0200108
magjed1cb48232016-10-20 03:19:16 -0700109 private final ThreadUtils.ThreadChecker threadChecker = new ThreadUtils.ThreadChecker();
Magnus Jedvert65c61dc2018-06-15 09:33:20 +0200110 private final GlTextureFrameBuffer i420TextureFrameBuffer =
111 new GlTextureFrameBuffer(GLES20.GL_RGBA);
Magnus Jedvert65c61dc2018-06-15 09:33:20 +0200112 private final ShaderCallbacks shaderCallbacks = new ShaderCallbacks();
113 private final GlGenericDrawer drawer = new GlGenericDrawer(FRAGMENT_SHADER, shaderCallbacks);
Åsa Perssonf2889bb2019-02-25 16:20:01 +0100114 private final VideoFrameDrawer videoFrameDrawer;
Magnus Jedvert577bc192016-10-19 15:29:02 +0200115
magjed1cb48232016-10-20 03:19:16 -0700116 /**
117 * This class should be constructed on a thread that has an active EGL context.
118 */
119 public YuvConverter() {
Åsa Perssonf2889bb2019-02-25 16:20:01 +0100120 this(new VideoFrameDrawer());
121 }
122
123 public YuvConverter(VideoFrameDrawer videoFrameDrawer) {
124 this.videoFrameDrawer = videoFrameDrawer;
Magnus Jedvert1d270f82018-04-16 16:28:29 +0200125 threadChecker.detachThread();
Sami Kalliomäkicb98b112017-10-16 11:20:26 +0200126 }
127
128 /** Converts the texture buffer to I420. */
Fabian Bergmark1bb36d22021-06-17 11:53:24 +0200129 @Nullable
Magnus Jedvert65c61dc2018-06-15 09:33:20 +0200130 public I420Buffer convert(TextureBuffer inputTextureBuffer) {
Fabian Bergmark1bb36d22021-06-17 11:53:24 +0200131 try {
132 return convertInternal(inputTextureBuffer);
133 } catch (GLException e) {
134 Logging.w(TAG, "Failed to convert TextureBuffer", e);
135 }
136 return null;
137 }
Åsa Perssonf2889bb2019-02-25 16:20:01 +0100138
Fabian Bergmark1bb36d22021-06-17 11:53:24 +0200139 private I420Buffer convertInternal(TextureBuffer inputTextureBuffer) {
Åsa Perssonf2889bb2019-02-25 16:20:01 +0100140 TextureBuffer preparedBuffer = (TextureBuffer) videoFrameDrawer.prepareBufferForViewportSize(
141 inputTextureBuffer, inputTextureBuffer.getWidth(), inputTextureBuffer.getHeight());
142
Magnus Jedvert577bc192016-10-19 15:29:02 +0200143 // We draw into a buffer laid out like
144 //
145 // +---------+
146 // | |
147 // | Y |
148 // | |
149 // | |
150 // +----+----+
151 // | U | V |
152 // | | |
153 // +----+----+
154 //
155 // In memory, we use the same stride for all of Y, U and V. The
Artem Titovd7ac5812021-07-27 12:23:39 +0200156 // U data starts at offset `height` * `stride` from the Y data,
Magnus Jedvert577bc192016-10-19 15:29:02 +0200157 // and the V data starts at at offset |stride/2| from the U
158 // data, with rows of U and V data alternating.
159 //
160 // Now, it would have made sense to allocate a pixel buffer with
161 // a single byte per pixel (EGL10.EGL_COLOR_BUFFER_TYPE,
162 // EGL10.EGL_LUMINANCE_BUFFER,), but that seems to be
163 // unsupported by devices. So do the following hack: Allocate an
Artem Titovd7ac5812021-07-27 12:23:39 +0200164 // RGBA buffer, of width `stride`/4. To render each of these
Magnus Jedvert577bc192016-10-19 15:29:02 +0200165 // large pixels, sample the texture at 4 different x coordinates
166 // and store the results in the four components.
167 //
168 // Since the V data needs to start on a boundary of such a
Artem Titovd7ac5812021-07-27 12:23:39 +0200169 // larger pixel, it is not sufficient that `stride` is even, it
Magnus Jedvert577bc192016-10-19 15:29:02 +0200170 // has to be a multiple of 8 pixels.
Åsa Perssonf2889bb2019-02-25 16:20:01 +0100171 final int frameWidth = preparedBuffer.getWidth();
172 final int frameHeight = preparedBuffer.getHeight();
Magnus Jedvert65c61dc2018-06-15 09:33:20 +0200173 final int stride = ((frameWidth + 7) / 8) * 8;
174 final int uvHeight = (frameHeight + 1) / 2;
175 // Total height of the combined memory layout.
176 final int totalHeight = frameHeight + uvHeight;
177 final ByteBuffer i420ByteBuffer = JniCommon.nativeAllocateByteBuffer(stride * totalHeight);
178 // Viewport width is divided by four since we are squeezing in four color bytes in each RGBA
179 // pixel.
180 final int viewportWidth = stride / 4;
Magnus Jedvert577bc192016-10-19 15:29:02 +0200181
Magnus Jedvert65c61dc2018-06-15 09:33:20 +0200182 // Produce a frame buffer starting at top-left corner, not bottom-left.
183 final Matrix renderMatrix = new Matrix();
184 renderMatrix.preTranslate(0.5f, 0.5f);
185 renderMatrix.preScale(1f, -1f);
186 renderMatrix.preTranslate(-0.5f, -0.5f);
Magnus Jedvert577bc192016-10-19 15:29:02 +0200187
Magnus Jedvert65c61dc2018-06-15 09:33:20 +0200188 i420TextureFrameBuffer.setSize(viewportWidth, totalHeight);
sakal2fcd2dd2017-01-18 03:21:10 -0800189
magjed1cb48232016-10-20 03:19:16 -0700190 // Bind our framebuffer.
Magnus Jedvert65c61dc2018-06-15 09:33:20 +0200191 GLES20.glBindFramebuffer(GLES20.GL_FRAMEBUFFER, i420TextureFrameBuffer.getFrameBufferId());
magjed1cb48232016-10-20 03:19:16 -0700192 GlUtil.checkNoGLES2Error("glBindFramebuffer");
193
Magnus Jedvert65c61dc2018-06-15 09:33:20 +0200194 // Draw Y.
195 shaderCallbacks.setPlaneY();
Åsa Perssonf2889bb2019-02-25 16:20:01 +0100196 VideoFrameDrawer.drawTexture(drawer, preparedBuffer, renderMatrix, frameWidth, frameHeight,
Magnus Jedvert65c61dc2018-06-15 09:33:20 +0200197 /* viewportX= */ 0, /* viewportY= */ 0, viewportWidth,
198 /* viewportHeight= */ frameHeight);
Magnus Jedvert577bc192016-10-19 15:29:02 +0200199
Magnus Jedvert65c61dc2018-06-15 09:33:20 +0200200 // Draw U.
201 shaderCallbacks.setPlaneU();
Åsa Perssonf2889bb2019-02-25 16:20:01 +0100202 VideoFrameDrawer.drawTexture(drawer, preparedBuffer, renderMatrix, frameWidth, frameHeight,
Magnus Jedvert65c61dc2018-06-15 09:33:20 +0200203 /* viewportX= */ 0, /* viewportY= */ frameHeight, viewportWidth / 2,
204 /* viewportHeight= */ uvHeight);
Magnus Jedvert577bc192016-10-19 15:29:02 +0200205
Magnus Jedvert65c61dc2018-06-15 09:33:20 +0200206 // Draw V.
207 shaderCallbacks.setPlaneV();
Åsa Perssonf2889bb2019-02-25 16:20:01 +0100208 VideoFrameDrawer.drawTexture(drawer, preparedBuffer, renderMatrix, frameWidth, frameHeight,
Magnus Jedvert65c61dc2018-06-15 09:33:20 +0200209 /* viewportX= */ viewportWidth / 2, /* viewportY= */ frameHeight, viewportWidth / 2,
210 /* viewportHeight= */ uvHeight);
Magnus Jedvert577bc192016-10-19 15:29:02 +0200211
Magnus Jedvert65c61dc2018-06-15 09:33:20 +0200212 GLES20.glReadPixels(0, 0, i420TextureFrameBuffer.getWidth(), i420TextureFrameBuffer.getHeight(),
213 GLES20.GL_RGBA, GLES20.GL_UNSIGNED_BYTE, i420ByteBuffer);
Magnus Jedvert577bc192016-10-19 15:29:02 +0200214
215 GlUtil.checkNoGLES2Error("YuvConverter.convert");
216
magjed1cb48232016-10-20 03:19:16 -0700217 // Restore normal framebuffer.
218 GLES20.glBindFramebuffer(GLES20.GL_FRAMEBUFFER, 0);
magjed1cb48232016-10-20 03:19:16 -0700219
Magnus Jedvert65c61dc2018-06-15 09:33:20 +0200220 // Prepare Y, U, and V ByteBuffer slices.
221 final int yPos = 0;
222 final int uPos = yPos + stride * frameHeight;
223 // Rows of U and V alternate in the buffer, so V data starts after the first row of U.
224 final int vPos = uPos + stride / 2;
225
226 i420ByteBuffer.position(yPos);
227 i420ByteBuffer.limit(yPos + stride * frameHeight);
228 final ByteBuffer dataY = i420ByteBuffer.slice();
229
230 i420ByteBuffer.position(uPos);
231 // The last row does not have padding.
232 final int uvSize = stride * (uvHeight - 1) + stride / 2;
233 i420ByteBuffer.limit(uPos + uvSize);
234 final ByteBuffer dataU = i420ByteBuffer.slice();
235
236 i420ByteBuffer.position(vPos);
237 i420ByteBuffer.limit(vPos + uvSize);
238 final ByteBuffer dataV = i420ByteBuffer.slice();
239
Åsa Perssonf2889bb2019-02-25 16:20:01 +0100240 preparedBuffer.release();
241
Magnus Jedvert65c61dc2018-06-15 09:33:20 +0200242 return JavaI420Buffer.wrap(frameWidth, frameHeight, dataY, stride, dataU, stride, dataV, stride,
243 () -> { JniCommon.nativeFreeByteBuffer(i420ByteBuffer); });
Magnus Jedvert577bc192016-10-19 15:29:02 +0200244 }
245
magjed1cb48232016-10-20 03:19:16 -0700246 public void release() {
247 threadChecker.checkIsOnValidThread();
Magnus Jedvert65c61dc2018-06-15 09:33:20 +0200248 drawer.release();
249 i420TextureFrameBuffer.release();
Åsa Perssonf2889bb2019-02-25 16:20:01 +0100250 videoFrameDrawer.release();
Magnus Jedvert7b875302018-08-09 13:51:42 +0200251 // Allow this class to be reused.
252 threadChecker.detachThread();
Magnus Jedvert577bc192016-10-19 15:29:02 +0200253 }
254}