Separate AndroidVideoTrackSource::OnFrameCaptured from adaptation

AndroidVideoTrackSource::OnFrameCaptured currently does adaptation
before passing frames on. We want to add video processing between
adaptation and delivering the frame to the rest WebRTC C++. This
CL prepares for that by splitting OnFrameCaptured() into a separate
adaptation step and delivery step.

Bug: webrtc:10247
Change-Id: Iab759bac7f3072d4552ece80d0b81fc3e634c64c
Reviewed-on: https://webrtc-review.googlesource.com/c/119952
Commit-Queue: Magnus Jedvert <magjed@webrtc.org>
Reviewed-by: Sami Kalliomäki <sakal@webrtc.org>
Cr-Commit-Position: refs/heads/master@{#26571}
This commit is contained in:
Magnus Jedvert
2019-02-06 14:48:57 +01:00
committed by Commit Bot
parent bb87f8a4a4
commit 9025bd5142
5 changed files with 130 additions and 32 deletions

View File

@ -43,7 +43,20 @@ public class VideoSource extends MediaSource {
@Override
public void onFrameCaptured(VideoFrame frame) {
nativeAndroidVideoTrackSource.onFrameCaptured(frame);
final NativeAndroidVideoTrackSource.FrameAdaptationParameters parameters =
nativeAndroidVideoTrackSource.adaptFrame(frame);
if (parameters == null) {
// Drop frame.
return;
}
final VideoFrame.Buffer adaptedBuffer =
frame.getBuffer().cropAndScale(parameters.cropX, parameters.cropY, parameters.cropWidth,
parameters.cropHeight, parameters.scaleWidth, parameters.scaleHeight);
// TODO(magjed): Add video processing hook here.
nativeAndroidVideoTrackSource.onFrameCaptured(
new VideoFrame(adaptedBuffer, frame.getRotation(), parameters.timestampNs));
adaptedBuffer.release();
}
};

View File

@ -17,11 +17,36 @@ import org.webrtc.VideoFrame;
* This class is meant to be a simple layer that only handles the JNI wrapping of a C++
* AndroidVideoTrackSource, that can easily be mocked out in Java unit tests. Refrain from adding
* any unnecessary logic to this class.
* This class is thred safe and methods can be called from any thread, but if frames A, B, ..., are
* sent to adaptFrame(), the adapted frames adaptedA, adaptedB, ..., needs to be passed in the same
* order to onFrameCaptured().
*/
class NativeAndroidVideoTrackSource {
// Pointer to webrtc::jni::AndroidVideoTrackSource.
private final long nativeAndroidVideoTrackSource;
public static class FrameAdaptationParameters {
public final int cropX;
public final int cropY;
public final int cropWidth;
public final int cropHeight;
public final int scaleWidth;
public final int scaleHeight;
public final long timestampNs;
@CalledByNative("FrameAdaptationParameters")
FrameAdaptationParameters(int cropX, int cropY, int cropWidth, int cropHeight, int scaleWidth,
int scaleHeight, long timestampNs) {
this.cropX = cropX;
this.cropY = cropY;
this.cropWidth = cropWidth;
this.cropHeight = cropHeight;
this.scaleWidth = scaleWidth;
this.scaleHeight = scaleHeight;
this.timestampNs = timestampNs;
}
}
public NativeAndroidVideoTrackSource(long nativeAndroidVideoTrackSource) {
this.nativeAndroidVideoTrackSource = nativeAndroidVideoTrackSource;
}
@ -34,11 +59,25 @@ class NativeAndroidVideoTrackSource {
nativeSetState(nativeAndroidVideoTrackSource, isLive);
}
/** Pass a frame to the native AndroidVideoTrackSource. */
/**
* This function should be called before delivering any frame to determine if the frame should be
* dropped or what the cropping and scaling parameters should be. If the return value is null, the
* frame should be dropped, otherwise the frame should be adapted in accordance to the frame
* adaptation parameters before calling onFrameCaptured().
*/
@Nullable
public FrameAdaptationParameters adaptFrame(VideoFrame frame) {
return nativeAdaptFrame(nativeAndroidVideoTrackSource, frame.getBuffer().getWidth(),
frame.getBuffer().getHeight(), frame.getRotation(), frame.getTimestampNs());
}
/**
* Pass an adapted frame to the native AndroidVideoTrackSource. Note that adaptFrame() is
* expected to be called first and that the passed frame conforms to those parameters.
*/
public void onFrameCaptured(VideoFrame frame) {
nativeOnFrameCaptured(nativeAndroidVideoTrackSource, frame.getBuffer().getWidth(),
frame.getBuffer().getHeight(), frame.getRotation(), frame.getTimestampNs(),
frame.getBuffer());
nativeOnFrameCaptured(nativeAndroidVideoTrackSource, frame.getRotation(),
frame.getTimestampNs(), frame.getBuffer());
}
/**
@ -59,6 +98,9 @@ class NativeAndroidVideoTrackSource {
int landscapeWidth, int landscapeHeight, @Nullable Integer maxLandscapePixelCount,
int portraitWidth, int portraitHeight, @Nullable Integer maxPortraitPixelCount,
@Nullable Integer maxFps);
private static native void nativeOnFrameCaptured(long nativeAndroidVideoTrackSource, int width,
int height, int rotation, long timestampNs, VideoFrame.Buffer buffer);
@Nullable
private static native FrameAdaptationParameters nativeAdaptFrame(
long nativeAndroidVideoTrackSource, int width, int height, int rotation, long timestampNs);
private static native void nativeOnFrameCaptured(
long nativeAndroidVideoTrackSource, int rotation, long timestampNs, VideoFrame.Buffer buffer);
}

View File

@ -36,6 +36,18 @@ class NativeCapturerObserver implements CapturerObserver {
@Override
public void onFrameCaptured(VideoFrame frame) {
nativeAndroidVideoTrackSource.onFrameCaptured(frame);
final NativeAndroidVideoTrackSource.FrameAdaptationParameters parameters =
nativeAndroidVideoTrackSource.adaptFrame(frame);
if (parameters == null) {
// Drop frame.
return;
}
final VideoFrame.Buffer adaptedBuffer =
frame.getBuffer().cropAndScale(parameters.cropX, parameters.cropY, parameters.cropWidth,
parameters.cropHeight, parameters.scaleWidth, parameters.scaleHeight);
nativeAndroidVideoTrackSource.onFrameCaptured(
new VideoFrame(adaptedBuffer, frame.getRotation(), parameters.timestampNs));
adaptedBuffer.release();
}
}

View File

@ -86,21 +86,21 @@ bool AndroidVideoTrackSource::remote() const {
return false;
}
void AndroidVideoTrackSource::OnFrameCaptured(
ScopedJavaLocalRef<jobject> AndroidVideoTrackSource::AdaptFrame(
JNIEnv* env,
const JavaRef<jobject>& j_caller,
jint j_width,
jint j_height,
jint j_rotation,
jlong j_timestamp_ns,
const JavaRef<jobject>& j_video_frame_buffer) {
jlong j_timestamp_ns) {
const VideoRotation rotation = jintToVideoRotation(j_rotation);
int64_t camera_time_us = j_timestamp_ns / rtc::kNumNanosecsPerMicrosec;
int64_t translated_camera_time_us =
align_timestamps_ ? timestamp_aligner_.TranslateTimestamp(
camera_time_us, rtc::TimeMicros())
: camera_time_us;
const int64_t camera_time_us = j_timestamp_ns / rtc::kNumNanosecsPerMicrosec;
const int64_t aligned_timestamp_ns =
align_timestamps_ ? rtc::kNumNanosecsPerMicrosec *
timestamp_aligner_.TranslateTimestamp(
camera_time_us, rtc::TimeMicros())
: j_timestamp_ns;
int adapted_width;
int adapted_height;
@ -109,35 +109,46 @@ void AndroidVideoTrackSource::OnFrameCaptured(
int crop_x;
int crop_y;
// TODO(magjed): Move this logic to users of NativeAndroidVideoTrackSource
// instead, in order to keep this native wrapping layer as thin as possible.
if (rotation % 180 == 0) {
if (!AdaptFrame(j_width, j_height, camera_time_us, &adapted_width,
&adapted_height, &crop_width, &crop_height, &crop_x,
&crop_y)) {
return;
if (!rtc::AdaptedVideoTrackSource::AdaptFrame(
j_width, j_height, camera_time_us, &adapted_width, &adapted_height,
&crop_width, &crop_height, &crop_x, &crop_y)) {
return nullptr;
}
} else {
// Swap all width/height and x/y.
if (!AdaptFrame(j_height, j_width, camera_time_us, &adapted_height,
&adapted_width, &crop_height, &crop_width, &crop_y,
&crop_x)) {
return;
if (!rtc::AdaptedVideoTrackSource::AdaptFrame(
j_height, j_width, camera_time_us, &adapted_height, &adapted_width,
&crop_height, &crop_width, &crop_y, &crop_x)) {
return nullptr;
}
}
return Java_FrameAdaptationParameters_Constructor(
env, crop_x, crop_y, crop_width, crop_height, adapted_width,
adapted_height, aligned_timestamp_ns);
}
void AndroidVideoTrackSource::OnFrameCaptured(
JNIEnv* env,
const JavaRef<jobject>& j_caller,
jint j_rotation,
jlong j_timestamp_ns,
const JavaRef<jobject>& j_video_frame_buffer) {
rtc::scoped_refptr<VideoFrameBuffer> buffer =
AndroidVideoBuffer::Create(env, j_video_frame_buffer)
->CropAndScale(env, crop_x, crop_y, crop_width, crop_height,
adapted_width, adapted_height);
AndroidVideoBuffer::Create(env, j_video_frame_buffer);
const VideoRotation rotation = jintToVideoRotation(j_rotation);
// AdaptedVideoTrackSource handles applying rotation for I420 frames.
if (apply_rotation() && rotation != kVideoRotation_0) {
if (apply_rotation() && rotation != kVideoRotation_0)
buffer = buffer->ToI420();
}
OnFrame(VideoFrame::Builder()
.set_video_frame_buffer(buffer)
.set_rotation(rotation)
.set_timestamp_us(translated_camera_time_us)
.set_timestamp_us(j_timestamp_ns / rtc::kNumNanosecsPerMicrosec)
.build());
}

View File

@ -24,6 +24,11 @@
namespace webrtc {
namespace jni {
// This class needs to be used in conjunction with the Java corresponding class
// NativeAndroidVideoTrackSource. This class is thred safe and methods can be
// called from any thread, but if frames A, B, ..., are sent to adaptFrame(),
// the adapted frames adaptedA, adaptedB, ..., needs to be passed in the same
// order to onFrameCaptured().
class AndroidVideoTrackSource : public rtc::AdaptedVideoTrackSource {
public:
AndroidVideoTrackSource(rtc::Thread* signaling_thread,
@ -45,10 +50,25 @@ class AndroidVideoTrackSource : public rtc::AdaptedVideoTrackSource {
bool remote() const override;
// This function should be called before delivering any frame to determine if
// the frame should be dropped or what the cropping and scaling parameters
// should be. This function is thread safe and can be called from any thread.
// This function returns
// NativeAndroidVideoTrackSource.FrameAdaptationParameters, or null if the
// frame should be dropped.
ScopedJavaLocalRef<jobject> AdaptFrame(JNIEnv* env,
const JavaRef<jobject>& j_caller,
jint j_width,
jint j_height,
jint j_rotation,
jlong j_timestamp_ns);
// This function converts and passes the frame on to the rest of the C++
// WebRTC layer. Note that GetFrameAdaptationParameters() is expected to be
// called first and that the delivered frame conforms to those parameters.
// This function is thread safe and can be called from any thread.
void OnFrameCaptured(JNIEnv* env,
const JavaRef<jobject>& j_caller,
jint j_width,
jint j_height,
jint j_rotation,
jlong j_timestamp_ns,
const JavaRef<jobject>& j_video_frame_buffer);