Add async simulcast support to VideoProcessor.
For the buffering of |input_frames_|, we assume that frames are ordered per simulcast layer but we make no assumptions between layers. For SVC, we still assume ordering of encode callbacks for the spatial layers. If we ever add async codecs that support SVC, they should still obey this assumption. Bug: webrtc:8448 Change-Id: I4ebb0c1e1d0eef41d850ed5b92aacc79d0a11137 Reviewed-on: https://webrtc-review.googlesource.com/60801 Commit-Queue: Rasmus Brandt <brandtr@webrtc.org> Reviewed-by: Sergey Silkin <ssilkin@webrtc.org> Cr-Commit-Position: refs/heads/master@{#22380}
This commit is contained in:

committed by
Commit Bot

parent
0fdb317b67
commit
0f1c0bd326
@ -119,10 +119,6 @@ void TestConfig::SetCodecSettings(VideoCodecType codec_type,
|
||||
// Spatial scalability is only available with VP9.
|
||||
RTC_CHECK(num_spatial_layers < 2 || codec_type == kVideoCodecVP9);
|
||||
|
||||
// Simulcast/SVC is only supposed to work with software codecs.
|
||||
RTC_CHECK((!hw_encoder && !hw_decoder) ||
|
||||
(num_simulcast_streams == 1 && num_spatial_layers == 1));
|
||||
|
||||
// Some base code requires numberOfSimulcastStreams to be set to zero
|
||||
// when simulcast is not used.
|
||||
codec_settings.numberOfSimulcastStreams =
|
||||
|
@ -31,7 +31,9 @@ namespace webrtc {
|
||||
namespace test {
|
||||
|
||||
namespace {
|
||||
|
||||
const int kMsToRtpTimestamp = kVideoPayloadTypeFrequency / 1000;
|
||||
const int kMaxBufferedInputFrames = 10;
|
||||
|
||||
std::unique_ptr<VideoBitrateAllocator> CreateBitrateAllocator(
|
||||
TestConfig* config) {
|
||||
@ -88,10 +90,10 @@ int GetElapsedTimeMicroseconds(int64_t start_ns, int64_t stop_ns) {
|
||||
return static_cast<int>(diff_us);
|
||||
}
|
||||
|
||||
void ExtractBufferWithSize(const VideoFrame& image,
|
||||
int width,
|
||||
int height,
|
||||
rtc::Buffer* buffer) {
|
||||
void ExtractI420BufferWithSize(const VideoFrame& image,
|
||||
int width,
|
||||
int height,
|
||||
rtc::Buffer* buffer) {
|
||||
if (image.width() != width || image.height() != height) {
|
||||
EXPECT_DOUBLE_EQ(static_cast<double>(width) / height,
|
||||
static_cast<double>(image.width()) / image.height());
|
||||
@ -166,14 +168,15 @@ VideoProcessor::VideoProcessor(webrtc::VideoEncoder* encoder,
|
||||
encode_callback_(this),
|
||||
decode_callback_(this),
|
||||
input_frame_reader_(input_frame_reader),
|
||||
merged_encoded_frames_(num_simulcast_or_spatial_layers_),
|
||||
encoded_frame_writers_(encoded_frame_writers),
|
||||
decoded_frame_writers_(decoded_frame_writers),
|
||||
last_inputed_frame_num_(0),
|
||||
last_inputed_timestamp_(0),
|
||||
first_encoded_frame(true),
|
||||
last_encoded_frame_num_(0),
|
||||
last_encoded_simulcast_svc_idx_(0),
|
||||
last_decoded_frame_num_(0) {
|
||||
first_encoded_frame_(num_simulcast_or_spatial_layers_, true),
|
||||
last_encoded_frame_num_(num_simulcast_or_spatial_layers_),
|
||||
first_decoded_frame_(num_simulcast_or_spatial_layers_, true),
|
||||
last_decoded_frame_num_(num_simulcast_or_spatial_layers_) {
|
||||
// Sanity checks.
|
||||
RTC_CHECK(rtc::TaskQueue::Current())
|
||||
<< "VideoProcessor must be run on a task queue.";
|
||||
@ -196,7 +199,6 @@ VideoProcessor::VideoProcessor(webrtc::VideoEncoder* encoder,
|
||||
static_cast<int>(config_.NumberOfCores()),
|
||||
config_.max_payload_size_bytes),
|
||||
WEBRTC_VIDEO_CODEC_OK);
|
||||
|
||||
for (auto& decoder : *decoders_) {
|
||||
RTC_CHECK_EQ(decoder->InitDecode(&config_.codec_settings,
|
||||
static_cast<int>(config_.NumberOfCores())),
|
||||
@ -218,7 +220,14 @@ VideoProcessor::~VideoProcessor() {
|
||||
decoder->RegisterDecodeCompleteCallback(nullptr);
|
||||
}
|
||||
|
||||
RTC_CHECK(last_encoded_frames_.empty());
|
||||
for (size_t simulcast_svc_idx = 0;
|
||||
simulcast_svc_idx < num_simulcast_or_spatial_layers_;
|
||||
++simulcast_svc_idx) {
|
||||
uint8_t* buffer = merged_encoded_frames_.at(simulcast_svc_idx)._buffer;
|
||||
if (buffer) {
|
||||
delete[] buffer;
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
void VideoProcessor::ProcessFrame() {
|
||||
@ -260,7 +269,6 @@ void VideoProcessor::ProcessFrame() {
|
||||
config_.FrameTypeForFrame(frame_number);
|
||||
const int encode_return_code =
|
||||
encoder_->Encode(input_frame, nullptr, &frame_types);
|
||||
|
||||
for (size_t simulcast_svc_idx = 0;
|
||||
simulcast_svc_idx < num_simulcast_or_spatial_layers_;
|
||||
++simulcast_svc_idx) {
|
||||
@ -268,41 +276,6 @@ void VideoProcessor::ProcessFrame() {
|
||||
stats_->GetFrame(frame_number, simulcast_svc_idx);
|
||||
frame_stat->encode_return_code = encode_return_code;
|
||||
}
|
||||
|
||||
// For async codecs frame decoding is done in frame encode callback.
|
||||
if (!config_.IsAsyncCodec()) {
|
||||
for (size_t simulcast_svc_idx = 0;
|
||||
simulcast_svc_idx < num_simulcast_or_spatial_layers_;
|
||||
++simulcast_svc_idx) {
|
||||
if (last_encoded_frames_.find(simulcast_svc_idx) !=
|
||||
last_encoded_frames_.end()) {
|
||||
EncodedImage& encoded_image = last_encoded_frames_[simulcast_svc_idx];
|
||||
|
||||
FrameStatistics* frame_stat =
|
||||
stats_->GetFrame(frame_number, simulcast_svc_idx);
|
||||
|
||||
if (encoded_frame_writers_) {
|
||||
RTC_CHECK(encoded_frame_writers_->at(simulcast_svc_idx)
|
||||
->WriteFrame(encoded_image,
|
||||
config_.codec_settings.codecType));
|
||||
}
|
||||
|
||||
// For the highest measurement accuracy of the decode time, the
|
||||
// start/stop time recordings should wrap the Decode call as tightly as
|
||||
// possible.
|
||||
frame_stat->decode_start_ns = rtc::TimeNanos();
|
||||
frame_stat->decode_return_code =
|
||||
decoders_->at(simulcast_svc_idx)
|
||||
->Decode(encoded_image, false, nullptr);
|
||||
|
||||
RTC_CHECK(encoded_image._buffer);
|
||||
delete[] encoded_image._buffer;
|
||||
encoded_image._buffer = nullptr;
|
||||
|
||||
last_encoded_frames_.erase(simulcast_svc_idx);
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
void VideoProcessor::SetRates(size_t bitrate_kbps, size_t framerate_fps) {
|
||||
@ -342,18 +315,24 @@ void VideoProcessor::FrameEncoded(
|
||||
encoded_image._timeStamp, simulcast_svc_idx);
|
||||
const size_t frame_number = frame_stat->frame_number;
|
||||
|
||||
// Reordering is unexpected. Frames of different layers have the same value
|
||||
// of frame_number. VP8 multi-res delivers frames starting from hires layer.
|
||||
RTC_CHECK_GE(frame_number, last_encoded_frame_num_);
|
||||
// Ensure that the encode order is monotonically increasing, within this
|
||||
// simulcast/spatial layer.
|
||||
RTC_CHECK(first_encoded_frame_[simulcast_svc_idx] ||
|
||||
last_encoded_frame_num_[simulcast_svc_idx] < frame_number);
|
||||
|
||||
// Ensure SVC spatial layers are delivered in ascending order.
|
||||
if (config_.NumberOfSpatialLayers() > 1) {
|
||||
RTC_CHECK(first_encoded_frame || frame_number >= last_encoded_frame_num_ ||
|
||||
simulcast_svc_idx > last_encoded_simulcast_svc_idx_);
|
||||
if (!first_encoded_frame_[simulcast_svc_idx] &&
|
||||
config_.NumberOfSpatialLayers() > 1) {
|
||||
for (size_t i = 0; i < simulcast_svc_idx; ++i) {
|
||||
RTC_CHECK_EQ(last_encoded_frame_num_[i], frame_number);
|
||||
}
|
||||
for (size_t i = simulcast_svc_idx + 1; i < num_simulcast_or_spatial_layers_;
|
||||
++i) {
|
||||
RTC_CHECK_GT(frame_number, last_encoded_frame_num_[i]);
|
||||
}
|
||||
}
|
||||
first_encoded_frame = false;
|
||||
last_encoded_frame_num_ = frame_number;
|
||||
last_encoded_simulcast_svc_idx_ = simulcast_svc_idx;
|
||||
first_encoded_frame_[simulcast_svc_idx] = false;
|
||||
last_encoded_frame_num_[simulcast_svc_idx] = frame_number;
|
||||
|
||||
// Update frame statistics.
|
||||
frame_stat->encoding_successful = true;
|
||||
@ -373,7 +352,6 @@ void VideoProcessor::FrameEncoded(
|
||||
500) /
|
||||
1000;
|
||||
}
|
||||
|
||||
frame_stat->length_bytes = encoded_image._length;
|
||||
frame_stat->frame_type = encoded_image._frameType;
|
||||
frame_stat->temporal_layer_idx = temporal_idx;
|
||||
@ -381,18 +359,21 @@ void VideoProcessor::FrameEncoded(
|
||||
frame_stat->max_nalu_size_bytes = GetMaxNaluSizeBytes(encoded_image, config_);
|
||||
frame_stat->qp = encoded_image.qp_;
|
||||
|
||||
if (!config_.IsAsyncCodec()) {
|
||||
// Store encoded frame. It will be decoded after all layers are encoded.
|
||||
CopyEncodedImage(encoded_image, codec_type, frame_number,
|
||||
simulcast_svc_idx);
|
||||
} else {
|
||||
const size_t simulcast_idx =
|
||||
codec_type == kVideoCodecVP8
|
||||
? codec_specific.codecSpecific.VP8.simulcastIdx
|
||||
: 0;
|
||||
frame_stat->decode_start_ns = rtc::TimeNanos();
|
||||
frame_stat->decode_return_code =
|
||||
decoders_->at(simulcast_idx)->Decode(encoded_image, false, nullptr);
|
||||
// Decode.
|
||||
const webrtc::EncodedImage* encoded_image_for_decode = &encoded_image;
|
||||
if (config_.NumberOfSpatialLayers() > 1) {
|
||||
encoded_image_for_decode = MergeAndStoreEncodedImageForSvcDecoding(
|
||||
encoded_image, codec_type, frame_number, simulcast_svc_idx);
|
||||
}
|
||||
frame_stat->decode_start_ns = rtc::TimeNanos();
|
||||
frame_stat->decode_return_code =
|
||||
decoders_->at(simulcast_svc_idx)
|
||||
->Decode(*encoded_image_for_decode, false, nullptr);
|
||||
|
||||
if (encoded_frame_writers_) {
|
||||
RTC_CHECK(
|
||||
encoded_frame_writers_->at(simulcast_svc_idx)
|
||||
->WriteFrame(encoded_image, config_.codec_settings.codecType));
|
||||
}
|
||||
}
|
||||
|
||||
@ -403,8 +384,6 @@ void VideoProcessor::FrameDecoded(const VideoFrame& decoded_frame) {
|
||||
// time recordings should wrap the Decode call as tightly as possible.
|
||||
const int64_t decode_stop_ns = rtc::TimeNanos();
|
||||
|
||||
RTC_CHECK(frame_wxh_to_simulcast_svc_idx_.find(decoded_frame.size()) !=
|
||||
frame_wxh_to_simulcast_svc_idx_.end());
|
||||
// Layer metadata.
|
||||
const size_t simulcast_svc_idx =
|
||||
frame_wxh_to_simulcast_svc_idx_.at(decoded_frame.size());
|
||||
@ -412,10 +391,12 @@ void VideoProcessor::FrameDecoded(const VideoFrame& decoded_frame) {
|
||||
decoded_frame.timestamp(), simulcast_svc_idx);
|
||||
const size_t frame_number = frame_stat->frame_number;
|
||||
|
||||
// Reordering is unexpected. Frames of different layers have the same value
|
||||
// of frame_number.
|
||||
RTC_CHECK_GE(frame_number, last_decoded_frame_num_);
|
||||
last_decoded_frame_num_ = frame_number;
|
||||
// Ensure that the decode order is monotonically increasing, within this
|
||||
// simulcast/spatial layer.
|
||||
RTC_CHECK(first_decoded_frame_[simulcast_svc_idx] ||
|
||||
last_decoded_frame_num_[simulcast_svc_idx] < frame_number);
|
||||
first_decoded_frame_[simulcast_svc_idx] = false;
|
||||
last_decoded_frame_num_[simulcast_svc_idx] = frame_number;
|
||||
|
||||
// Update frame statistics.
|
||||
frame_stat->decoding_successful = true;
|
||||
@ -426,21 +407,29 @@ void VideoProcessor::FrameDecoded(const VideoFrame& decoded_frame) {
|
||||
|
||||
// Skip quality metrics calculation to not affect CPU usage.
|
||||
if (!config_.measure_cpu) {
|
||||
CalculateFrameQuality(input_frames_.at(frame_number), decoded_frame,
|
||||
frame_stat);
|
||||
const auto reference_frame = input_frames_.find(frame_number);
|
||||
RTC_CHECK(reference_frame != input_frames_.cend())
|
||||
<< "The codecs are either buffering too much, dropping too much, or "
|
||||
"being too slow relative the input frame rate.";
|
||||
CalculateFrameQuality(reference_frame->second, decoded_frame, frame_stat);
|
||||
}
|
||||
|
||||
// Delay erasing of input frames by one frame. The current frame might
|
||||
// still be needed for other simulcast stream or spatial layer.
|
||||
if (frame_number > 0) {
|
||||
auto input_frame_erase_to = input_frames_.lower_bound(frame_number - 1);
|
||||
input_frames_.erase(input_frames_.begin(), input_frame_erase_to);
|
||||
}
|
||||
// Erase all buffered input frames that we have moved past for all
|
||||
// simulcast/spatial layers. Never buffer more than |kMaxBufferedInputFrames|
|
||||
// frames, to protect against long runs of consecutive frame drops for a
|
||||
// particular layer.
|
||||
const auto min_last_decoded_frame_num = std::min_element(
|
||||
last_decoded_frame_num_.cbegin(), last_decoded_frame_num_.cend());
|
||||
const size_t min_buffered_frame_num =
|
||||
std::max(0, static_cast<int>(frame_number) - kMaxBufferedInputFrames + 1);
|
||||
RTC_CHECK(min_last_decoded_frame_num != last_decoded_frame_num_.cend());
|
||||
const auto input_frames_erase_before = input_frames_.lower_bound(
|
||||
std::max(*min_last_decoded_frame_num, min_buffered_frame_num));
|
||||
input_frames_.erase(input_frames_.cbegin(), input_frames_erase_before);
|
||||
|
||||
if (decoded_frame_writers_) {
|
||||
ExtractBufferWithSize(decoded_frame, config_.codec_settings.width,
|
||||
config_.codec_settings.height, &tmp_i420_buffer_);
|
||||
RTC_CHECK(simulcast_svc_idx < decoded_frame_writers_->size());
|
||||
ExtractI420BufferWithSize(decoded_frame, config_.codec_settings.width,
|
||||
config_.codec_settings.height, &tmp_i420_buffer_);
|
||||
RTC_CHECK_EQ(tmp_i420_buffer_.size(),
|
||||
decoded_frame_writers_->at(simulcast_svc_idx)->FrameLength());
|
||||
RTC_CHECK(decoded_frame_writers_->at(simulcast_svc_idx)
|
||||
@ -448,23 +437,24 @@ void VideoProcessor::FrameDecoded(const VideoFrame& decoded_frame) {
|
||||
}
|
||||
}
|
||||
|
||||
void VideoProcessor::CopyEncodedImage(const EncodedImage& encoded_image,
|
||||
const VideoCodecType codec,
|
||||
size_t frame_number,
|
||||
size_t simulcast_svc_idx) {
|
||||
RTC_DCHECK_CALLED_SEQUENTIALLY(&sequence_checker_);
|
||||
const webrtc::EncodedImage*
|
||||
VideoProcessor::MergeAndStoreEncodedImageForSvcDecoding(
|
||||
const EncodedImage& encoded_image,
|
||||
const VideoCodecType codec,
|
||||
size_t frame_number,
|
||||
size_t simulcast_svc_idx) {
|
||||
// Should only be called for SVC.
|
||||
RTC_CHECK_GT(config_.NumberOfSpatialLayers(), 1);
|
||||
|
||||
EncodedImage base_image;
|
||||
RTC_CHECK_EQ(base_image._length, 0);
|
||||
|
||||
// Each SVC layer is decoded with dedicated decoder. Add data of base layers
|
||||
// to current coded frame buffer.
|
||||
if (config_.NumberOfSpatialLayers() > 1 && simulcast_svc_idx > 0) {
|
||||
RTC_CHECK(last_encoded_frames_.find(simulcast_svc_idx - 1) !=
|
||||
last_encoded_frames_.end());
|
||||
base_image = last_encoded_frames_[simulcast_svc_idx - 1];
|
||||
if (simulcast_svc_idx > 0) {
|
||||
base_image = merged_encoded_frames_.at(simulcast_svc_idx - 1);
|
||||
RTC_CHECK_EQ(base_image._timeStamp, encoded_image._timeStamp);
|
||||
}
|
||||
|
||||
const size_t payload_size_bytes = base_image._length + encoded_image._length;
|
||||
const size_t buffer_size_bytes =
|
||||
payload_size_bytes + EncodedImage::GetBufferPaddingBytes(codec);
|
||||
@ -473,9 +463,9 @@ void VideoProcessor::CopyEncodedImage(const EncodedImage& encoded_image,
|
||||
RTC_CHECK(copied_buffer);
|
||||
|
||||
if (base_image._length) {
|
||||
RTC_CHECK(base_image._buffer);
|
||||
memcpy(copied_buffer, base_image._buffer, base_image._length);
|
||||
}
|
||||
|
||||
memcpy(copied_buffer + base_image._length, encoded_image._buffer,
|
||||
encoded_image._length);
|
||||
|
||||
@ -485,7 +475,14 @@ void VideoProcessor::CopyEncodedImage(const EncodedImage& encoded_image,
|
||||
copied_image._length = payload_size_bytes;
|
||||
copied_image._size = buffer_size_bytes;
|
||||
|
||||
last_encoded_frames_[simulcast_svc_idx] = copied_image;
|
||||
// Replace previous EncodedImage for this spatial layer.
|
||||
uint8_t* old_buffer = merged_encoded_frames_.at(simulcast_svc_idx)._buffer;
|
||||
if (old_buffer) {
|
||||
delete[] old_buffer;
|
||||
}
|
||||
merged_encoded_frames_.at(simulcast_svc_idx) = copied_image;
|
||||
|
||||
return &merged_encoded_frames_.at(simulcast_svc_idx);
|
||||
}
|
||||
|
||||
} // namespace test
|
||||
|
@ -164,10 +164,13 @@ class VideoProcessor {
|
||||
// Invoked by the callback adapter when a frame has completed decoding.
|
||||
void FrameDecoded(const webrtc::VideoFrame& image);
|
||||
|
||||
void CopyEncodedImage(const EncodedImage& encoded_image,
|
||||
const VideoCodecType codec,
|
||||
size_t frame_number,
|
||||
size_t simulcast_svc_idx);
|
||||
// In order to supply the SVC decoders with super frames containing all
|
||||
// lower layer frames, we merge and store the layer frames in this method.
|
||||
const webrtc::EncodedImage* MergeAndStoreEncodedImageForSvcDecoding(
|
||||
const EncodedImage& encoded_image,
|
||||
const VideoCodecType codec,
|
||||
size_t frame_number,
|
||||
size_t simulcast_svc_idx) RTC_RUN_ON(sequence_checker_);
|
||||
|
||||
// Test input/output.
|
||||
TestConfig config_ RTC_GUARDED_BY(sequence_checker_);
|
||||
@ -199,7 +202,8 @@ class VideoProcessor {
|
||||
// then, after all layers are encoded, decode them. Such separation of
|
||||
// frame processing on superframe level simplifies encoding/decoding time
|
||||
// measurement.
|
||||
std::map<size_t, EncodedImage> last_encoded_frames_
|
||||
// simulcast_svc_idx -> merged SVC encoded frame.
|
||||
std::vector<EncodedImage> merged_encoded_frames_
|
||||
RTC_GUARDED_BY(sequence_checker_);
|
||||
|
||||
// These (optional) file writers are used to persistently store the encoded
|
||||
@ -208,16 +212,21 @@ class VideoProcessor {
|
||||
FrameWriterList* const decoded_frame_writers_;
|
||||
rtc::Buffer tmp_i420_buffer_; // Temp storage for format conversion.
|
||||
|
||||
// Metadata of inputed/encoded/decoded frames. Used for frame drop detection
|
||||
// and other purposes.
|
||||
// Metadata for inputed/encoded/decoded frames. Used for frame identification,
|
||||
// frame drop detection, etc. We assume that encoded/decoded frames are
|
||||
// ordered within each simulcast/spatial layer, but we do not make any
|
||||
// assumptions of frame ordering between layers.
|
||||
size_t last_inputed_frame_num_ RTC_GUARDED_BY(sequence_checker_);
|
||||
size_t last_inputed_timestamp_ RTC_GUARDED_BY(sequence_checker_);
|
||||
bool first_encoded_frame RTC_GUARDED_BY(sequence_checker_);
|
||||
size_t last_encoded_frame_num_ RTC_GUARDED_BY(sequence_checker_);
|
||||
size_t last_encoded_simulcast_svc_idx_ RTC_GUARDED_BY(sequence_checker_);
|
||||
size_t last_decoded_frame_num_ RTC_GUARDED_BY(sequence_checker_);
|
||||
|
||||
// Map of frame size (in pixels) to simulcast/spatial layer index.
|
||||
// simulcast_svc_idx -> encode status.
|
||||
std::vector<bool> first_encoded_frame_ RTC_GUARDED_BY(sequence_checker_);
|
||||
// simulcast_svc_idx -> frame_number.
|
||||
std::vector<size_t> last_encoded_frame_num_ RTC_GUARDED_BY(sequence_checker_);
|
||||
// simulcast_svc_idx -> decode status.
|
||||
std::vector<bool> first_decoded_frame_ RTC_GUARDED_BY(sequence_checker_);
|
||||
// simulcast_svc_idx -> frame_number.
|
||||
std::vector<size_t> last_decoded_frame_num_ RTC_GUARDED_BY(sequence_checker_);
|
||||
// frame size (pixels) -> simulcast_svc_idx.
|
||||
std::map<size_t, size_t> frame_wxh_to_simulcast_svc_idx_
|
||||
RTC_GUARDED_BY(sequence_checker_);
|
||||
|
||||
|
@ -196,7 +196,7 @@ TEST_F(VideoProcessorIntegrationTestLibvpx, HighBitrateVP8) {
|
||||
{5, 1, 0, 0.1, 0.2, 0.1, 0, 1}};
|
||||
|
||||
// std::vector<QualityThresholds> quality_thresholds = {{37, 35, 0.93, 0.91}};
|
||||
// TODO(webrtc:8757): AMR VP8 encoder's quality is significantly worse
|
||||
// TODO(webrtc:8757): ARM VP8 encoder's quality is significantly worse
|
||||
// than quality of x86 version. Use lower thresholds for now.
|
||||
std::vector<QualityThresholds> quality_thresholds = {{35, 33, 0.91, 0.89}};
|
||||
|
||||
@ -237,7 +237,7 @@ TEST_F(VideoProcessorIntegrationTestLibvpx, MAYBE_ChangeBitrateVP8) {
|
||||
|
||||
// std::vector<QualityThresholds> quality_thresholds = {
|
||||
// {33, 32, 0.89, 0.88}, {38, 36, 0.94, 0.93}, {35, 34, 0.92, 0.91}};
|
||||
// TODO(webrtc:8757): AMR VP8 encoder's quality is significantly worse
|
||||
// TODO(webrtc:8757): ARM VP8 encoder's quality is significantly worse
|
||||
// than quality of x86 version. Use lower thresholds for now.
|
||||
std::vector<QualityThresholds> quality_thresholds = {
|
||||
{31.8, 31, 0.86, 0.85}, {36, 34.8, 0.92, 0.90}, {33.5, 32, 0.90, 0.88}};
|
||||
@ -265,7 +265,7 @@ TEST_F(VideoProcessorIntegrationTestLibvpx, MAYBE_ChangeFramerateVP8) {
|
||||
// {10, 2, 20, 0.4, 0.3, 0.1, 0, 1},
|
||||
// {5, 2, 5, 0.3, 0.3, 0.1, 0, 0},
|
||||
// {4, 2, 1, 0.2, 0.3, 0.2, 0, 0}};
|
||||
// TODO(webrtc:8757): AMR VP8 drops more frames than x86 version. Use lower
|
||||
// TODO(webrtc:8757): ARM VP8 drops more frames than x86 version. Use lower
|
||||
// thresholds for now.
|
||||
std::vector<RateControlThresholds> rc_thresholds = {
|
||||
{10, 2, 60, 0.5, 0.3, 0.3, 0, 1},
|
||||
@ -274,7 +274,7 @@ TEST_F(VideoProcessorIntegrationTestLibvpx, MAYBE_ChangeFramerateVP8) {
|
||||
|
||||
// std::vector<QualityThresholds> quality_thresholds = {
|
||||
// {31, 30, 0.87, 0.86}, {32, 31, 0.89, 0.86}, {32, 30, 0.87, 0.82}};
|
||||
// TODO(webrtc:8757): AMR VP8 encoder's quality is significantly worse
|
||||
// TODO(webrtc:8757): ARM VP8 encoder's quality is significantly worse
|
||||
// than quality of x86 version. Use lower thresholds for now.
|
||||
std::vector<QualityThresholds> quality_thresholds = {
|
||||
{31, 30, 0.85, 0.84}, {31.5, 30.5, 0.86, 0.84}, {30.5, 29, 0.83, 0.78}};
|
||||
@ -298,7 +298,7 @@ TEST_F(VideoProcessorIntegrationTestLibvpx, MAYBE_TemporalLayersVP8) {
|
||||
|
||||
// std::vector<RateControlThresholds> rc_thresholds = {
|
||||
// {5, 1, 0, 0.1, 0.2, 0.1, 0, 1}, {10, 2, 0, 0.1, 0.2, 0.1, 0, 1}};
|
||||
// TODO(webrtc:8757): AMR VP8 drops more frames than x86 version. Use lower
|
||||
// TODO(webrtc:8757): ARM VP8 drops more frames than x86 version. Use lower
|
||||
// thresholds for now.
|
||||
std::vector<RateControlThresholds> rc_thresholds = {
|
||||
{10, 1, 2, 0.3, 0.2, 0.1, 0, 1}, {12, 2, 3, 0.1, 0.2, 0.1, 0, 1}};
|
||||
@ -306,7 +306,7 @@ TEST_F(VideoProcessorIntegrationTestLibvpx, MAYBE_TemporalLayersVP8) {
|
||||
// Min SSIM drops because of high motion scene with complex backgound (trees).
|
||||
// std::vector<QualityThresholds> quality_thresholds = {{32, 30, 0.88, 0.85},
|
||||
// {33, 30, 0.89, 0.83}};
|
||||
// TODO(webrtc:8757): AMR VP8 encoder's quality is significantly worse
|
||||
// TODO(webrtc:8757): ARM VP8 encoder's quality is significantly worse
|
||||
// than quality of x86 version. Use lower thresholds for now.
|
||||
std::vector<QualityThresholds> quality_thresholds = {{31, 30, 0.85, 0.84},
|
||||
{31, 28, 0.85, 0.75}};
|
||||
@ -315,6 +315,29 @@ TEST_F(VideoProcessorIntegrationTestLibvpx, MAYBE_TemporalLayersVP8) {
|
||||
&quality_thresholds, nullptr, nullptr);
|
||||
}
|
||||
|
||||
// Might be too slow on mobile platforms.
|
||||
#if defined(WEBRTC_ANDROID) || defined(WEBRTC_IOS)
|
||||
#define MAYBE_MultiresVP8 DISABLED_MultiresVP8
|
||||
#else
|
||||
#define MAYBE_MultiresVP8 MultiresVP8
|
||||
#endif
|
||||
TEST_F(VideoProcessorIntegrationTestLibvpx, MAYBE_MultiresVP8) {
|
||||
config_.filename = "ConferenceMotion_1280_720_50";
|
||||
config_.filepath = ResourcePath(config_.filename, "yuv");
|
||||
config_.num_frames = 100;
|
||||
config_.SetCodecSettings(kVideoCodecVP8, 3, 1, 3, true, true, false,
|
||||
kResilienceOn, 1280, 720);
|
||||
|
||||
std::vector<RateProfile> rate_profiles = {{1500, 30, config_.num_frames}};
|
||||
|
||||
std::vector<RateControlThresholds> rc_thresholds = {
|
||||
{5, 1, 5, 0.2, 0.3, 0.1, 0, 1}};
|
||||
std::vector<QualityThresholds> quality_thresholds = {{34, 32, 0.90, 0.88}};
|
||||
|
||||
ProcessFramesAndMaybeVerify(rate_profiles, &rc_thresholds,
|
||||
&quality_thresholds, nullptr, nullptr);
|
||||
}
|
||||
|
||||
// Might be too slow on mobile platforms.
|
||||
#if defined(WEBRTC_ANDROID) || defined(WEBRTC_IOS)
|
||||
#define MAYBE_SimulcastVP8 DISABLED_SimulcastVP8
|
||||
@ -325,13 +348,14 @@ TEST_F(VideoProcessorIntegrationTestLibvpx, MAYBE_SimulcastVP8) {
|
||||
config_.filename = "ConferenceMotion_1280_720_50";
|
||||
config_.filepath = ResourcePath(config_.filename, "yuv");
|
||||
config_.num_frames = 100;
|
||||
config_.simulcast_adapted_encoder = true;
|
||||
config_.SetCodecSettings(kVideoCodecVP8, 3, 1, 3, true, true, false,
|
||||
kResilienceOn, 1280, 720);
|
||||
|
||||
std::vector<RateProfile> rate_profiles = {{1500, 30, config_.num_frames}};
|
||||
|
||||
std::vector<RateControlThresholds> rc_thresholds = {
|
||||
{5, 1, 5, 0.2, 0.3, 0.1, 0, 1}};
|
||||
{20, 5, 90, 0.75, 0.5, 0.3, 0, 1}};
|
||||
std::vector<QualityThresholds> quality_thresholds = {{34, 32, 0.90, 0.88}};
|
||||
|
||||
ProcessFramesAndMaybeVerify(rate_profiles, &rc_thresholds,
|
||||
@ -361,7 +385,7 @@ TEST_F(VideoProcessorIntegrationTestLibvpx, MAYBE_SvcVP9) {
|
||||
&quality_thresholds, nullptr, nullptr);
|
||||
}
|
||||
|
||||
TEST_F(VideoProcessorIntegrationTestLibvpx, DISABLED_SimulcastVP8RdPerf) {
|
||||
TEST_F(VideoProcessorIntegrationTestLibvpx, DISABLED_MultiresVP8RdPerf) {
|
||||
config_.filename = "FourPeople_1280x720_30";
|
||||
config_.filepath = ResourcePath(config_.filename, "yuv");
|
||||
config_.num_frames = 300;
|
||||
|
Reference in New Issue
Block a user