The FFmpeg video decoder requires up to 8 additional bytes to be allocated for its encoded image buffer input, due to optimized byte readers over-reading on some platforms. We plan to use FFmpeg for a soon-to-land H.264 enc/dec. This CL adds support for padding encoded image buffers based on codec type, and makes sure calls to VCMEncodedFrame::VerifyAndAllocate use the padding. All padding constants are 0 but making H.264 pad with 8 bytes will be a one-line change. Also, added -framework CoreFoundation to webrtc_h264_video_toolbox which was missing. BUG=chromium:468365 BUG=https://bugs.chromium.org/p/webrtc/issues/detail?id=5424 NOTRY=True Review URL: https://codereview.webrtc.org/1602523004 Cr-Commit-Position: refs/heads/master@{#11337}
268 lines
8.6 KiB
C++
268 lines
8.6 KiB
C++
/*
|
|
* Copyright (c) 2012 The WebRTC project authors. All Rights Reserved.
|
|
*
|
|
* Use of this source code is governed by a BSD-style license
|
|
* that can be found in the LICENSE file in the root of the source
|
|
* tree. An additional intellectual property rights grant can be found
|
|
* in the file PATENTS. All contributing project authors may
|
|
* be found in the AUTHORS file in the root of the source tree.
|
|
*/
|
|
|
|
#include "webrtc/video_frame.h"
|
|
|
|
#include <string.h>
|
|
|
|
#include <algorithm> // swap
|
|
|
|
#include "webrtc/base/bind.h"
|
|
#include "webrtc/base/checks.h"
|
|
|
|
namespace webrtc {
|
|
|
|
// TODO(hbos): The FFmpeg video decoder will require up to 8 bytes, update this
|
|
// when the FFmpeg decoder is added.
|
|
const size_t EncodedImage::kBufferPaddingBytesH264 = 0;
|
|
|
|
bool EqualPlane(const uint8_t* data1,
|
|
const uint8_t* data2,
|
|
int stride,
|
|
int width,
|
|
int height) {
|
|
for (int y = 0; y < height; ++y) {
|
|
if (memcmp(data1, data2, width) != 0)
|
|
return false;
|
|
data1 += stride;
|
|
data2 += stride;
|
|
}
|
|
return true;
|
|
}
|
|
|
|
int ExpectedSize(int plane_stride, int image_height, PlaneType type) {
|
|
if (type == kYPlane)
|
|
return plane_stride * image_height;
|
|
return plane_stride * ((image_height + 1) / 2);
|
|
}
|
|
|
|
VideoFrame::VideoFrame() {
|
|
// Intentionally using Reset instead of initializer list so that any missed
|
|
// fields in Reset will be caught by memory checkers.
|
|
Reset();
|
|
}
|
|
|
|
VideoFrame::VideoFrame(const rtc::scoped_refptr<VideoFrameBuffer>& buffer,
|
|
uint32_t timestamp,
|
|
int64_t render_time_ms,
|
|
VideoRotation rotation)
|
|
: video_frame_buffer_(buffer),
|
|
timestamp_(timestamp),
|
|
ntp_time_ms_(0),
|
|
render_time_ms_(render_time_ms),
|
|
rotation_(rotation) {
|
|
}
|
|
|
|
int VideoFrame::CreateEmptyFrame(int width,
|
|
int height,
|
|
int stride_y,
|
|
int stride_u,
|
|
int stride_v) {
|
|
const int half_width = (width + 1) / 2;
|
|
RTC_DCHECK_GT(width, 0);
|
|
RTC_DCHECK_GT(height, 0);
|
|
RTC_DCHECK_GE(stride_y, width);
|
|
RTC_DCHECK_GE(stride_u, half_width);
|
|
RTC_DCHECK_GE(stride_v, half_width);
|
|
|
|
// Creating empty frame - reset all values.
|
|
timestamp_ = 0;
|
|
ntp_time_ms_ = 0;
|
|
render_time_ms_ = 0;
|
|
rotation_ = kVideoRotation_0;
|
|
|
|
// Check if it's safe to reuse allocation.
|
|
if (video_frame_buffer_ && video_frame_buffer_->HasOneRef() &&
|
|
!video_frame_buffer_->native_handle() &&
|
|
width == video_frame_buffer_->width() &&
|
|
height == video_frame_buffer_->height() && stride_y == stride(kYPlane) &&
|
|
stride_u == stride(kUPlane) && stride_v == stride(kVPlane)) {
|
|
return 0;
|
|
}
|
|
|
|
// Need to allocate new buffer.
|
|
video_frame_buffer_ = new rtc::RefCountedObject<I420Buffer>(
|
|
width, height, stride_y, stride_u, stride_v);
|
|
return 0;
|
|
}
|
|
|
|
int VideoFrame::CreateFrame(const uint8_t* buffer_y,
|
|
const uint8_t* buffer_u,
|
|
const uint8_t* buffer_v,
|
|
int width,
|
|
int height,
|
|
int stride_y,
|
|
int stride_u,
|
|
int stride_v) {
|
|
return CreateFrame(buffer_y, buffer_u, buffer_v, width, height, stride_y,
|
|
stride_u, stride_v, kVideoRotation_0);
|
|
}
|
|
|
|
int VideoFrame::CreateFrame(const uint8_t* buffer_y,
|
|
const uint8_t* buffer_u,
|
|
const uint8_t* buffer_v,
|
|
int width,
|
|
int height,
|
|
int stride_y,
|
|
int stride_u,
|
|
int stride_v,
|
|
VideoRotation rotation) {
|
|
const int half_height = (height + 1) / 2;
|
|
const int expected_size_y = height * stride_y;
|
|
const int expected_size_u = half_height * stride_u;
|
|
const int expected_size_v = half_height * stride_v;
|
|
CreateEmptyFrame(width, height, stride_y, stride_u, stride_v);
|
|
memcpy(buffer(kYPlane), buffer_y, expected_size_y);
|
|
memcpy(buffer(kUPlane), buffer_u, expected_size_u);
|
|
memcpy(buffer(kVPlane), buffer_v, expected_size_v);
|
|
rotation_ = rotation;
|
|
return 0;
|
|
}
|
|
|
|
int VideoFrame::CreateFrame(const uint8_t* buffer,
|
|
int width,
|
|
int height,
|
|
VideoRotation rotation) {
|
|
const int stride_y = width;
|
|
const int stride_uv = (width + 1) / 2;
|
|
|
|
const uint8_t* buffer_y = buffer;
|
|
const uint8_t* buffer_u = buffer_y + stride_y * height;
|
|
const uint8_t* buffer_v = buffer_u + stride_uv * ((height + 1) / 2);
|
|
return CreateFrame(buffer_y, buffer_u, buffer_v, width, height, stride_y,
|
|
stride_uv, stride_uv, rotation);
|
|
}
|
|
|
|
int VideoFrame::CopyFrame(const VideoFrame& videoFrame) {
|
|
if (videoFrame.IsZeroSize()) {
|
|
video_frame_buffer_ = nullptr;
|
|
} else if (videoFrame.native_handle()) {
|
|
video_frame_buffer_ = videoFrame.video_frame_buffer();
|
|
} else {
|
|
CreateFrame(videoFrame.buffer(kYPlane), videoFrame.buffer(kUPlane),
|
|
videoFrame.buffer(kVPlane), videoFrame.width(),
|
|
videoFrame.height(), videoFrame.stride(kYPlane),
|
|
videoFrame.stride(kUPlane), videoFrame.stride(kVPlane));
|
|
}
|
|
|
|
timestamp_ = videoFrame.timestamp_;
|
|
ntp_time_ms_ = videoFrame.ntp_time_ms_;
|
|
render_time_ms_ = videoFrame.render_time_ms_;
|
|
rotation_ = videoFrame.rotation_;
|
|
return 0;
|
|
}
|
|
|
|
void VideoFrame::ShallowCopy(const VideoFrame& videoFrame) {
|
|
video_frame_buffer_ = videoFrame.video_frame_buffer();
|
|
timestamp_ = videoFrame.timestamp_;
|
|
ntp_time_ms_ = videoFrame.ntp_time_ms_;
|
|
render_time_ms_ = videoFrame.render_time_ms_;
|
|
rotation_ = videoFrame.rotation_;
|
|
}
|
|
|
|
void VideoFrame::Reset() {
|
|
video_frame_buffer_ = nullptr;
|
|
timestamp_ = 0;
|
|
ntp_time_ms_ = 0;
|
|
render_time_ms_ = 0;
|
|
rotation_ = kVideoRotation_0;
|
|
}
|
|
|
|
uint8_t* VideoFrame::buffer(PlaneType type) {
|
|
return video_frame_buffer_ ? video_frame_buffer_->MutableData(type)
|
|
: nullptr;
|
|
}
|
|
|
|
const uint8_t* VideoFrame::buffer(PlaneType type) const {
|
|
return video_frame_buffer_ ? video_frame_buffer_->data(type) : nullptr;
|
|
}
|
|
|
|
int VideoFrame::allocated_size(PlaneType type) const {
|
|
const int plane_height = (type == kYPlane) ? height() : (height() + 1) / 2;
|
|
return plane_height * stride(type);
|
|
}
|
|
|
|
int VideoFrame::stride(PlaneType type) const {
|
|
return video_frame_buffer_ ? video_frame_buffer_->stride(type) : 0;
|
|
}
|
|
|
|
int VideoFrame::width() const {
|
|
return video_frame_buffer_ ? video_frame_buffer_->width() : 0;
|
|
}
|
|
|
|
int VideoFrame::height() const {
|
|
return video_frame_buffer_ ? video_frame_buffer_->height() : 0;
|
|
}
|
|
|
|
bool VideoFrame::IsZeroSize() const {
|
|
return !video_frame_buffer_;
|
|
}
|
|
|
|
void* VideoFrame::native_handle() const {
|
|
return video_frame_buffer_ ? video_frame_buffer_->native_handle() : nullptr;
|
|
}
|
|
|
|
rtc::scoped_refptr<VideoFrameBuffer> VideoFrame::video_frame_buffer() const {
|
|
return video_frame_buffer_;
|
|
}
|
|
|
|
void VideoFrame::set_video_frame_buffer(
|
|
const rtc::scoped_refptr<webrtc::VideoFrameBuffer>& buffer) {
|
|
video_frame_buffer_ = buffer;
|
|
}
|
|
|
|
VideoFrame VideoFrame::ConvertNativeToI420Frame() const {
|
|
RTC_DCHECK(native_handle());
|
|
VideoFrame frame;
|
|
frame.ShallowCopy(*this);
|
|
frame.set_video_frame_buffer(video_frame_buffer_->NativeToI420Buffer());
|
|
return frame;
|
|
}
|
|
|
|
bool VideoFrame::EqualsFrame(const VideoFrame& frame) const {
|
|
if (width() != frame.width() || height() != frame.height() ||
|
|
stride(kYPlane) != frame.stride(kYPlane) ||
|
|
stride(kUPlane) != frame.stride(kUPlane) ||
|
|
stride(kVPlane) != frame.stride(kVPlane) ||
|
|
timestamp() != frame.timestamp() ||
|
|
ntp_time_ms() != frame.ntp_time_ms() ||
|
|
render_time_ms() != frame.render_time_ms()) {
|
|
return false;
|
|
}
|
|
const int half_width = (width() + 1) / 2;
|
|
const int half_height = (height() + 1) / 2;
|
|
return EqualPlane(buffer(kYPlane), frame.buffer(kYPlane),
|
|
stride(kYPlane), width(), height()) &&
|
|
EqualPlane(buffer(kUPlane), frame.buffer(kUPlane),
|
|
stride(kUPlane), half_width, half_height) &&
|
|
EqualPlane(buffer(kVPlane), frame.buffer(kVPlane),
|
|
stride(kVPlane), half_width, half_height);
|
|
}
|
|
|
|
size_t EncodedImage::GetBufferPaddingBytes(VideoCodecType codec_type) {
|
|
switch (codec_type) {
|
|
case kVideoCodecVP8:
|
|
case kVideoCodecVP9:
|
|
return 0;
|
|
case kVideoCodecH264:
|
|
return kBufferPaddingBytesH264;
|
|
case kVideoCodecI420:
|
|
case kVideoCodecRED:
|
|
case kVideoCodecULPFEC:
|
|
case kVideoCodecGeneric:
|
|
case kVideoCodecUnknown:
|
|
return 0;
|
|
}
|
|
RTC_NOTREACHED();
|
|
return 0;
|
|
}
|
|
|
|
} // namespace webrtc
|