ObjC: Split out NV12 texture uploading into separate class
This CL splits out the code in RTCNV12Shader.mm for uploading byte buffers to textures into its own class RTCNV12TextureCache. The purpose is to prepare for allowing clients to inject their own shaders. RTCNV12TextureCache will be used in the generic code, while the actual shaders in RTCNV12Shader will be customizable by the client. BUG=webrtc:7473 Review-Url: https://codereview.webrtc.org/2835203003 Cr-Commit-Position: refs/heads/master@{#17902}
This commit is contained in:
@ -209,6 +209,8 @@ if (is_ios || is_mac) {
|
||||
if (is_ios) {
|
||||
sources += [
|
||||
"objc/Framework/Classes/RTCEAGLVideoView.m",
|
||||
"objc/Framework/Classes/RTCNV12TextureCache.h",
|
||||
"objc/Framework/Classes/RTCNV12TextureCache.m",
|
||||
"objc/Framework/Classes/RTCNativeNV12Shader.mm",
|
||||
"objc/Framework/Headers/WebRTC/RTCEAGLVideoView.h",
|
||||
]
|
||||
|
||||
30
webrtc/sdk/objc/Framework/Classes/RTCNV12TextureCache.h
Normal file
30
webrtc/sdk/objc/Framework/Classes/RTCNV12TextureCache.h
Normal file
@ -0,0 +1,30 @@
|
||||
/*
|
||||
* Copyright 2017 The WebRTC project authors. All Rights Reserved.
|
||||
*
|
||||
* Use of this source code is governed by a BSD-style license
|
||||
* that can be found in the LICENSE file in the root of the source
|
||||
* tree. An additional intellectual property rights grant can be found
|
||||
* in the file PATENTS. All contributing project authors may
|
||||
* be found in the AUTHORS file in the root of the source tree.
|
||||
*/
|
||||
|
||||
#import <GLKit/GLKit.h>
|
||||
|
||||
@class RTCVideoFrame;
|
||||
|
||||
NS_ASSUME_NONNULL_BEGIN
|
||||
|
||||
@interface RTCNV12TextureCache : NSObject
|
||||
|
||||
@property(nonatomic, readonly) GLuint yTexture;
|
||||
@property(nonatomic, readonly) GLuint uvTexture;
|
||||
|
||||
- (nullable instancetype)initWithContext:(EAGLContext *)context;
|
||||
|
||||
- (BOOL)uploadFrameToTextures:(RTCVideoFrame *)frame;
|
||||
|
||||
- (void)releaseTextures;
|
||||
|
||||
@end
|
||||
|
||||
NS_ASSUME_NONNULL_END
|
||||
108
webrtc/sdk/objc/Framework/Classes/RTCNV12TextureCache.m
Normal file
108
webrtc/sdk/objc/Framework/Classes/RTCNV12TextureCache.m
Normal file
@ -0,0 +1,108 @@
|
||||
/*
|
||||
* Copyright 2017 The WebRTC project authors. All Rights Reserved.
|
||||
*
|
||||
* Use of this source code is governed by a BSD-style license
|
||||
* that can be found in the LICENSE file in the root of the source
|
||||
* tree. An additional intellectual property rights grant can be found
|
||||
* in the file PATENTS. All contributing project authors may
|
||||
* be found in the AUTHORS file in the root of the source tree.
|
||||
*/
|
||||
|
||||
#import "RTCNV12TextureCache.h"
|
||||
|
||||
#import "WebRTC/RTCVideoFrame.h"
|
||||
|
||||
@implementation RTCNV12TextureCache {
|
||||
CVOpenGLESTextureCacheRef _textureCache;
|
||||
CVOpenGLESTextureRef _yTextureRef;
|
||||
CVOpenGLESTextureRef _uvTextureRef;
|
||||
}
|
||||
|
||||
- (GLuint)yTexture {
|
||||
return CVOpenGLESTextureGetName(_yTextureRef);
|
||||
}
|
||||
|
||||
- (GLuint)uvTexture {
|
||||
return CVOpenGLESTextureGetName(_uvTextureRef);
|
||||
}
|
||||
|
||||
- (instancetype)initWithContext:(EAGLContext *)context {
|
||||
if (self = [super init]) {
|
||||
CVReturn ret = CVOpenGLESTextureCacheCreate(
|
||||
kCFAllocatorDefault, NULL,
|
||||
#if COREVIDEO_USE_EAGLCONTEXT_CLASS_IN_API
|
||||
context,
|
||||
#else
|
||||
(__bridge void *)context,
|
||||
#endif
|
||||
NULL, &_textureCache);
|
||||
if (ret != kCVReturnSuccess) {
|
||||
self = nil;
|
||||
}
|
||||
}
|
||||
return self;
|
||||
}
|
||||
|
||||
- (BOOL)loadTexture:(CVOpenGLESTextureRef *)textureOut
|
||||
pixelBuffer:(CVPixelBufferRef)pixelBuffer
|
||||
planeIndex:(int)planeIndex
|
||||
pixelFormat:(GLenum)pixelFormat {
|
||||
const int width = CVPixelBufferGetWidthOfPlane(pixelBuffer, planeIndex);
|
||||
const int height = CVPixelBufferGetHeightOfPlane(pixelBuffer, planeIndex);
|
||||
|
||||
if (*textureOut) {
|
||||
CFRelease(*textureOut);
|
||||
*textureOut = nil;
|
||||
}
|
||||
CVReturn ret = CVOpenGLESTextureCacheCreateTextureFromImage(
|
||||
kCFAllocatorDefault, _textureCache, pixelBuffer, NULL, GL_TEXTURE_2D, pixelFormat, width,
|
||||
height, pixelFormat, GL_UNSIGNED_BYTE, planeIndex, textureOut);
|
||||
if (ret != kCVReturnSuccess) {
|
||||
CFRelease(*textureOut);
|
||||
*textureOut = nil;
|
||||
return NO;
|
||||
}
|
||||
NSAssert(CVOpenGLESTextureGetTarget(*textureOut) == GL_TEXTURE_2D,
|
||||
@"Unexpected GLES texture target");
|
||||
glBindTexture(GL_TEXTURE_2D, CVOpenGLESTextureGetName(*textureOut));
|
||||
glTexParameteri(GL_TEXTURE_2D, GL_TEXTURE_MIN_FILTER, GL_LINEAR);
|
||||
glTexParameteri(GL_TEXTURE_2D, GL_TEXTURE_MAG_FILTER, GL_LINEAR);
|
||||
glTexParameteri(GL_TEXTURE_2D, GL_TEXTURE_WRAP_S, GL_CLAMP_TO_EDGE);
|
||||
glTexParameteri(GL_TEXTURE_2D, GL_TEXTURE_WRAP_T, GL_CLAMP_TO_EDGE);
|
||||
return YES;
|
||||
}
|
||||
|
||||
- (BOOL)uploadFrameToTextures:(RTCVideoFrame *)frame {
|
||||
CVPixelBufferRef pixelBuffer = frame.nativeHandle;
|
||||
NSParameterAssert(pixelBuffer);
|
||||
return [self loadTexture:&_yTextureRef
|
||||
pixelBuffer:pixelBuffer
|
||||
planeIndex:0
|
||||
pixelFormat:GL_LUMINANCE] &&
|
||||
[self loadTexture:&_uvTextureRef
|
||||
pixelBuffer:pixelBuffer
|
||||
planeIndex:1
|
||||
pixelFormat:GL_LUMINANCE_ALPHA];
|
||||
}
|
||||
|
||||
- (void)releaseTextures {
|
||||
if (_uvTextureRef) {
|
||||
CFRelease(_uvTextureRef);
|
||||
_uvTextureRef = nil;
|
||||
}
|
||||
if (_yTextureRef) {
|
||||
CFRelease(_yTextureRef);
|
||||
_yTextureRef = nil;
|
||||
}
|
||||
}
|
||||
|
||||
- (void)dealloc {
|
||||
[self releaseTextures];
|
||||
if (_textureCache) {
|
||||
CFRelease(_textureCache);
|
||||
_textureCache = nil;
|
||||
}
|
||||
}
|
||||
|
||||
@end
|
||||
|
||||
@ -10,12 +10,7 @@
|
||||
|
||||
#import "RTCShader.h"
|
||||
|
||||
// Native CVPixelBufferRef rendering is only supported on iPhone because it
|
||||
// depends on CVOpenGLESTextureCacheCreate.
|
||||
#if TARGET_OS_IPHONE
|
||||
|
||||
#import <CoreVideo/CVOpenGLESTextureCache.h>
|
||||
|
||||
#import "RTCNV12TextureCache.h"
|
||||
#import "RTCShader+Private.h"
|
||||
#import "WebRTC/RTCLogging.h"
|
||||
#import "WebRTC/RTCVideoFrame.h"
|
||||
@ -47,7 +42,7 @@ static const char kNV12FragmentShaderSource[] =
|
||||
GLuint _nv12Program;
|
||||
GLint _ySampler;
|
||||
GLint _uvSampler;
|
||||
CVOpenGLESTextureCacheRef _textureCache;
|
||||
RTCNV12TextureCache *_textureCache;
|
||||
// Store current rotation and only upload new vertex data when rotation
|
||||
// changes.
|
||||
rtc::Optional<RTCVideoRotation> _currentRotation;
|
||||
@ -55,7 +50,8 @@ static const char kNV12FragmentShaderSource[] =
|
||||
|
||||
- (instancetype)initWithContext:(GlContextType *)context {
|
||||
if (self = [super init]) {
|
||||
if (![self setupNV12Program] || ![self setupTextureCacheWithContext:context] ||
|
||||
_textureCache = [[RTCNV12TextureCache alloc] initWithContext:context];
|
||||
if (!_textureCache || ![self setupNV12Program] ||
|
||||
!RTCSetupVerticesForProgram(_nv12Program, &_vertexBuffer, nullptr)) {
|
||||
RTCLog(@"Failed to initialize RTCNativeNV12Shader.");
|
||||
self = nil;
|
||||
@ -67,10 +63,6 @@ static const char kNV12FragmentShaderSource[] =
|
||||
- (void)dealloc {
|
||||
glDeleteProgram(_nv12Program);
|
||||
glDeleteBuffers(1, &_vertexBuffer);
|
||||
if (_textureCache) {
|
||||
CFRelease(_textureCache);
|
||||
_textureCache = nullptr;
|
||||
}
|
||||
}
|
||||
|
||||
- (BOOL)setupNV12Program {
|
||||
@ -84,75 +76,21 @@ static const char kNV12FragmentShaderSource[] =
|
||||
return (_ySampler >= 0 && _uvSampler >= 0);
|
||||
}
|
||||
|
||||
- (BOOL)setupTextureCacheWithContext:(GlContextType *)context {
|
||||
CVReturn ret = CVOpenGLESTextureCacheCreate(
|
||||
kCFAllocatorDefault, NULL,
|
||||
#if COREVIDEO_USE_EAGLCONTEXT_CLASS_IN_API
|
||||
context,
|
||||
#else
|
||||
(__bridge void *)context,
|
||||
#endif
|
||||
NULL, &_textureCache);
|
||||
return ret == kCVReturnSuccess;
|
||||
}
|
||||
|
||||
- (BOOL)drawFrame:(RTCVideoFrame *)frame {
|
||||
CVPixelBufferRef pixelBuffer = frame.nativeHandle;
|
||||
RTC_CHECK(pixelBuffer);
|
||||
glUseProgram(_nv12Program);
|
||||
const OSType pixelFormat = CVPixelBufferGetPixelFormatType(pixelBuffer);
|
||||
RTC_CHECK(pixelFormat == kCVPixelFormatType_420YpCbCr8BiPlanarFullRange ||
|
||||
pixelFormat == kCVPixelFormatType_420YpCbCr8BiPlanarVideoRange)
|
||||
<< "Unsupported native pixel format: " << pixelFormat;
|
||||
if (![_textureCache uploadFrameToTextures:frame]) {
|
||||
return NO;
|
||||
}
|
||||
|
||||
// Y-plane.
|
||||
const int lumaWidth = CVPixelBufferGetWidthOfPlane(pixelBuffer, 0);
|
||||
const int lumaHeight = CVPixelBufferGetHeightOfPlane(pixelBuffer, 0);
|
||||
|
||||
CVOpenGLESTextureRef lumaTexture = nullptr;
|
||||
glActiveTexture(GL_TEXTURE0);
|
||||
glUniform1i(_ySampler, 0);
|
||||
CVReturn ret = CVOpenGLESTextureCacheCreateTextureFromImage(
|
||||
kCFAllocatorDefault, _textureCache, pixelBuffer, NULL, GL_TEXTURE_2D,
|
||||
RTC_PIXEL_FORMAT, lumaWidth, lumaHeight, RTC_PIXEL_FORMAT,
|
||||
GL_UNSIGNED_BYTE, 0, &lumaTexture);
|
||||
if (ret != kCVReturnSuccess) {
|
||||
CFRelease(lumaTexture);
|
||||
return NO;
|
||||
}
|
||||
|
||||
RTC_CHECK_EQ(static_cast<GLenum>(GL_TEXTURE_2D),
|
||||
CVOpenGLESTextureGetTarget(lumaTexture));
|
||||
glBindTexture(GL_TEXTURE_2D, CVOpenGLESTextureGetName(lumaTexture));
|
||||
glTexParameteri(GL_TEXTURE_2D, GL_TEXTURE_MIN_FILTER, GL_LINEAR);
|
||||
glTexParameteri(GL_TEXTURE_2D, GL_TEXTURE_MAG_FILTER, GL_LINEAR);
|
||||
glTexParameteri(GL_TEXTURE_2D, GL_TEXTURE_WRAP_S, GL_CLAMP_TO_EDGE);
|
||||
glTexParameteri(GL_TEXTURE_2D, GL_TEXTURE_WRAP_T, GL_CLAMP_TO_EDGE);
|
||||
glBindTexture(GL_TEXTURE_2D, _textureCache.yTexture);
|
||||
|
||||
// UV-plane.
|
||||
const int chromaWidth = CVPixelBufferGetWidthOfPlane(pixelBuffer, 1);
|
||||
const int chromeHeight = CVPixelBufferGetHeightOfPlane(pixelBuffer, 1);
|
||||
|
||||
CVOpenGLESTextureRef chromaTexture = nullptr;
|
||||
glActiveTexture(GL_TEXTURE1);
|
||||
glUniform1i(_uvSampler, 1);
|
||||
ret = CVOpenGLESTextureCacheCreateTextureFromImage(
|
||||
kCFAllocatorDefault, _textureCache, pixelBuffer, NULL, GL_TEXTURE_2D,
|
||||
GL_LUMINANCE_ALPHA, chromaWidth, chromeHeight, GL_LUMINANCE_ALPHA,
|
||||
GL_UNSIGNED_BYTE, 1, &chromaTexture);
|
||||
if (ret != kCVReturnSuccess) {
|
||||
CFRelease(chromaTexture);
|
||||
CFRelease(lumaTexture);
|
||||
return NO;
|
||||
}
|
||||
|
||||
RTC_CHECK_EQ(static_cast<GLenum>(GL_TEXTURE_2D),
|
||||
CVOpenGLESTextureGetTarget(chromaTexture));
|
||||
glBindTexture(GL_TEXTURE_2D, CVOpenGLESTextureGetName(chromaTexture));
|
||||
glTexParameteri(GL_TEXTURE_2D, GL_TEXTURE_MIN_FILTER, GL_LINEAR);
|
||||
glTexParameteri(GL_TEXTURE_2D, GL_TEXTURE_MAG_FILTER, GL_LINEAR);
|
||||
glTexParameteri(GL_TEXTURE_2D, GL_TEXTURE_WRAP_S, GL_CLAMP_TO_EDGE);
|
||||
glTexParameteri(GL_TEXTURE_2D, GL_TEXTURE_WRAP_T, GL_CLAMP_TO_EDGE);
|
||||
glBindTexture(GL_TEXTURE_2D, _textureCache.uvTexture);
|
||||
|
||||
glBindBuffer(GL_ARRAY_BUFFER, _vertexBuffer);
|
||||
if (!_currentRotation || frame.rotation != *_currentRotation) {
|
||||
@ -161,11 +99,9 @@ static const char kNV12FragmentShaderSource[] =
|
||||
}
|
||||
glDrawArrays(GL_TRIANGLE_FAN, 0, 4);
|
||||
|
||||
CFRelease(chromaTexture);
|
||||
CFRelease(lumaTexture);
|
||||
[_textureCache releaseTextures];
|
||||
|
||||
return YES;
|
||||
}
|
||||
|
||||
@end
|
||||
#endif // TARGET_OS_IPHONE
|
||||
|
||||
Reference in New Issue
Block a user