FrameBuffer for the new jitter buffer.
BUG=webrtc:5514 R=danilchap@webrtc.org, mflodman@webrtc.org Review URL: https://codereview.webrtc.org/1969403007 . Cr-Commit-Position: refs/heads/master@{#12798}
This commit is contained in:
@ -362,6 +362,7 @@
|
||||
'video_coding/codecs/vp8/simulcast_encoder_adapter_unittest.cc',
|
||||
'video_coding/codecs/vp8/simulcast_unittest.cc',
|
||||
'video_coding/codecs/vp8/simulcast_unittest.h',
|
||||
'video_coding/frame_buffer2_unittest.cc',
|
||||
'video_coding/include/mock/mock_vcm_callbacks.h',
|
||||
'video_coding/decoding_state_unittest.cc',
|
||||
'video_coding/histogram_unittest.cc',
|
||||
|
@ -21,6 +21,8 @@ source_set("video_coding") {
|
||||
"fec_tables_xor.h",
|
||||
"frame_buffer.cc",
|
||||
"frame_buffer.h",
|
||||
"frame_buffer2.cc",
|
||||
"frame_buffer2.h",
|
||||
"frame_object.cc",
|
||||
"frame_object.h",
|
||||
"generic_decoder.cc",
|
||||
|
154
webrtc/modules/video_coding/frame_buffer2.cc
Normal file
154
webrtc/modules/video_coding/frame_buffer2.cc
Normal file
@ -0,0 +1,154 @@
|
||||
/*
|
||||
* Copyright (c) 2016 The WebRTC project authors. All Rights Reserved.
|
||||
*
|
||||
* Use of this source code is governed by a BSD-style license
|
||||
* that can be found in the LICENSE file in the root of the source
|
||||
* tree. An additional intellectual property rights grant can be found
|
||||
* in the file PATENTS. All contributing project authors may
|
||||
* be found in the AUTHORS file in the root of the source tree.
|
||||
*/
|
||||
|
||||
#include "webrtc/modules/video_coding/frame_buffer2.h"
|
||||
|
||||
#include <algorithm>
|
||||
|
||||
#include "webrtc/base/checks.h"
|
||||
#include "webrtc/modules/video_coding/frame_object.h"
|
||||
#include "webrtc/modules/video_coding/jitter_estimator.h"
|
||||
#include "webrtc/modules/video_coding/sequence_number_util.h"
|
||||
#include "webrtc/modules/video_coding/timing.h"
|
||||
#include "webrtc/system_wrappers/include/clock.h"
|
||||
|
||||
namespace webrtc {
|
||||
namespace video_coding {
|
||||
|
||||
namespace {
|
||||
// The maximum age of decoded frames tracked by frame buffer, compared to
|
||||
// |newest_picture_id_|.
|
||||
constexpr int kMaxFrameAge = 4096;
|
||||
|
||||
// The maximum number of decoded frames being tracked by the frame buffer.
|
||||
constexpr int kMaxNumHistoryFrames = 256;
|
||||
} // namespace
|
||||
|
||||
bool FrameBuffer::FrameComp::operator()(const FrameKey& f1,
|
||||
const FrameKey& f2) const {
|
||||
// first = picture id
|
||||
// second = spatial layer
|
||||
if (f1.first == f2.first)
|
||||
return f1.second < f2.second;
|
||||
return AheadOf(f2.first, f1.first);
|
||||
}
|
||||
|
||||
FrameBuffer::FrameBuffer(Clock* clock,
|
||||
VCMJitterEstimator* jitter_estimator,
|
||||
const VCMTiming* timing)
|
||||
: clock_(clock),
|
||||
frame_inserted_event_(false, false),
|
||||
jitter_estimator_(jitter_estimator),
|
||||
timing_(timing),
|
||||
newest_picture_id_(-1) {}
|
||||
|
||||
std::unique_ptr<FrameObject> FrameBuffer::NextFrame(int64_t max_wait_time_ms) {
|
||||
int64_t latest_return_time = clock_->TimeInMilliseconds() + max_wait_time_ms;
|
||||
while (true) {
|
||||
int64_t now = clock_->TimeInMilliseconds();
|
||||
int64_t wait_ms = max_wait_time_ms;
|
||||
|
||||
crit_.Enter();
|
||||
frame_inserted_event_.Reset();
|
||||
auto next_frame = frames_.end();
|
||||
for (auto frame_it = frames_.begin(); frame_it != frames_.end();
|
||||
++frame_it) {
|
||||
const FrameObject& frame = *frame_it->second;
|
||||
if (IsContinuous(frame)) {
|
||||
next_frame = frame_it;
|
||||
int64_t render_time = timing_->RenderTimeMs(frame.timestamp, now);
|
||||
wait_ms = timing_->MaxWaitingTime(render_time, now);
|
||||
|
||||
// This will cause the frame buffer to prefer high framerate rather
|
||||
// than high resolution in the case of the decoder not decoding fast
|
||||
// enough and the stream has multiple spatial and temporal layers.
|
||||
if (wait_ms == 0)
|
||||
continue;
|
||||
|
||||
break;
|
||||
}
|
||||
}
|
||||
crit_.Leave();
|
||||
|
||||
// If the timout occures, return. Otherwise a new frame has been inserted
|
||||
// and the best frame to decode next will be selected again.
|
||||
wait_ms = std::min<int64_t>(wait_ms, latest_return_time - now);
|
||||
wait_ms = std::max<int64_t>(wait_ms, 0);
|
||||
if (!frame_inserted_event_.Wait(wait_ms)) {
|
||||
crit_.Enter();
|
||||
if (next_frame != frames_.end()) {
|
||||
// TODO(philipel): update jitter estimator with correct values.
|
||||
jitter_estimator_->UpdateEstimate(100, 100);
|
||||
|
||||
decoded_frames_.insert(next_frame->first);
|
||||
std::unique_ptr<FrameObject> frame = std::move(next_frame->second);
|
||||
frames_.erase(frames_.begin(), ++next_frame);
|
||||
crit_.Leave();
|
||||
return frame;
|
||||
} else {
|
||||
crit_.Leave();
|
||||
return std::unique_ptr<FrameObject>();
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
void FrameBuffer::InsertFrame(std::unique_ptr<FrameObject> frame) {
|
||||
rtc::CritScope lock(&crit_);
|
||||
if (newest_picture_id_ == -1)
|
||||
newest_picture_id_ = frame->picture_id;
|
||||
|
||||
if (AheadOf<uint16_t>(frame->picture_id, newest_picture_id_))
|
||||
newest_picture_id_ = frame->picture_id;
|
||||
|
||||
// Remove frames as long as we have too many, |kMaxNumHistoryFrames|.
|
||||
while (decoded_frames_.size() > kMaxNumHistoryFrames)
|
||||
decoded_frames_.erase(decoded_frames_.begin());
|
||||
|
||||
// Remove frames that are too old, |kMaxNumHistoryFrames|.
|
||||
uint16_t old_picture_id = Subtract<1 << 16>(newest_picture_id_, kMaxFrameAge);
|
||||
auto old_decoded_it =
|
||||
decoded_frames_.lower_bound(FrameKey(old_picture_id, 0));
|
||||
decoded_frames_.erase(decoded_frames_.begin(), old_decoded_it);
|
||||
|
||||
FrameKey key(frame->picture_id, frame->spatial_layer);
|
||||
frames_[key] = std::move(frame);
|
||||
frame_inserted_event_.Set();
|
||||
}
|
||||
|
||||
bool FrameBuffer::IsContinuous(const FrameObject& frame) const {
|
||||
// If a frame with an earlier picture id was inserted compared to the last
|
||||
// decoded frames picture id then that frame arrived too late.
|
||||
if (!decoded_frames_.empty() &&
|
||||
AheadOf(decoded_frames_.rbegin()->first, frame.picture_id)) {
|
||||
return false;
|
||||
}
|
||||
|
||||
// Have we decoded all frames that this frame depend on?
|
||||
for (size_t r = 0; r < frame.num_references; ++r) {
|
||||
FrameKey ref_key(frame.references[r], frame.spatial_layer);
|
||||
if (decoded_frames_.find(ref_key) == decoded_frames_.end())
|
||||
return false;
|
||||
}
|
||||
|
||||
// If this is a layer frame, have we decoded the lower layer of this
|
||||
// super frame.
|
||||
if (frame.inter_layer_predicted) {
|
||||
RTC_DCHECK_GT(frame.spatial_layer, 0);
|
||||
FrameKey ref_key(frame.picture_id, frame.spatial_layer - 1);
|
||||
if (decoded_frames_.find(ref_key) == decoded_frames_.end())
|
||||
return false;
|
||||
}
|
||||
|
||||
return true;
|
||||
}
|
||||
|
||||
} // namespace video_coding
|
||||
} // namespace webrtc
|
83
webrtc/modules/video_coding/frame_buffer2.h
Normal file
83
webrtc/modules/video_coding/frame_buffer2.h
Normal file
@ -0,0 +1,83 @@
|
||||
/*
|
||||
* Copyright (c) 2016 The WebRTC project authors. All Rights Reserved.
|
||||
*
|
||||
* Use of this source code is governed by a BSD-style license
|
||||
* that can be found in the LICENSE file in the root of the source
|
||||
* tree. An additional intellectual property rights grant can be found
|
||||
* in the file PATENTS. All contributing project authors may
|
||||
* be found in the AUTHORS file in the root of the source tree.
|
||||
*/
|
||||
|
||||
#ifndef WEBRTC_MODULES_VIDEO_CODING_FRAME_BUFFER2_H_
|
||||
#define WEBRTC_MODULES_VIDEO_CODING_FRAME_BUFFER2_H_
|
||||
|
||||
#include <array>
|
||||
#include <map>
|
||||
#include <memory>
|
||||
#include <set>
|
||||
#include <utility>
|
||||
|
||||
#include "webrtc/base/constructormagic.h"
|
||||
#include "webrtc/base/criticalsection.h"
|
||||
#include "webrtc/base/event.h"
|
||||
#include "webrtc/base/thread_annotations.h"
|
||||
|
||||
namespace webrtc {
|
||||
|
||||
class Clock;
|
||||
class VCMJitterEstimator;
|
||||
class VCMTiming;
|
||||
|
||||
namespace video_coding {
|
||||
|
||||
class FrameObject;
|
||||
|
||||
class FrameBuffer {
|
||||
public:
|
||||
FrameBuffer(Clock* clock,
|
||||
VCMJitterEstimator* jitter_estimator,
|
||||
const VCMTiming* timing);
|
||||
|
||||
// Insert a frame into the frame buffer.
|
||||
void InsertFrame(std::unique_ptr<FrameObject> frame);
|
||||
|
||||
// Get the next frame for decoding. Will return at latest after
|
||||
// |max_wait_time_ms|, with either a managed FrameObject or an empty
|
||||
// unique ptr if there is no available frame for decoding.
|
||||
std::unique_ptr<FrameObject> NextFrame(int64_t max_wait_time_ms);
|
||||
|
||||
private:
|
||||
// FrameKey is a pair of (picture id, spatial layer).
|
||||
using FrameKey = std::pair<uint16_t, uint8_t>;
|
||||
|
||||
// Comparator used to sort frames, first on their picture id, and second
|
||||
// on their spatial layer.
|
||||
struct FrameComp {
|
||||
bool operator()(const FrameKey& f1, const FrameKey& f2) const;
|
||||
};
|
||||
|
||||
// Determines whether a frame is continuous.
|
||||
bool IsContinuous(const FrameObject& frame) const
|
||||
EXCLUSIVE_LOCKS_REQUIRED(crit_);
|
||||
|
||||
// Keep track of decoded frames.
|
||||
std::set<FrameKey, FrameComp> decoded_frames_ GUARDED_BY(crit_);
|
||||
|
||||
// The actual buffer that holds the FrameObjects.
|
||||
std::map<FrameKey, std::unique_ptr<FrameObject>, FrameComp> frames_
|
||||
GUARDED_BY(crit_);
|
||||
|
||||
rtc::CriticalSection crit_;
|
||||
Clock* const clock_;
|
||||
rtc::Event frame_inserted_event_;
|
||||
VCMJitterEstimator* const jitter_estimator_;
|
||||
const VCMTiming* const timing_;
|
||||
int newest_picture_id_ GUARDED_BY(crit_);
|
||||
|
||||
RTC_DISALLOW_IMPLICIT_CONSTRUCTORS(FrameBuffer);
|
||||
};
|
||||
|
||||
} // namespace video_coding
|
||||
} // namespace webrtc
|
||||
|
||||
#endif // WEBRTC_MODULES_VIDEO_CODING_FRAME_BUFFER2_H_
|
329
webrtc/modules/video_coding/frame_buffer2_unittest.cc
Normal file
329
webrtc/modules/video_coding/frame_buffer2_unittest.cc
Normal file
@ -0,0 +1,329 @@
|
||||
/*
|
||||
* Copyright (c) 2016 The WebRTC project authors. All Rights Reserved.
|
||||
*
|
||||
* Use of this source code is governed by a BSD-style license
|
||||
* that can be found in the LICENSE file in the root of the source
|
||||
* tree. An additional intellectual property rights grant can be found
|
||||
* in the file PATENTS. All contributing project authors may
|
||||
* be found in the AUTHORS file in the root of the source tree.
|
||||
*/
|
||||
|
||||
#include "webrtc/modules/video_coding/frame_buffer2.h"
|
||||
|
||||
#include <algorithm>
|
||||
#include <cstring>
|
||||
#include <limits>
|
||||
#include <vector>
|
||||
|
||||
#include "testing/gmock/include/gmock/gmock.h"
|
||||
#include "testing/gtest/include/gtest/gtest.h"
|
||||
#include "webrtc/base/platform_thread.h"
|
||||
#include "webrtc/base/random.h"
|
||||
#include "webrtc/modules/video_coding/frame_object.h"
|
||||
#include "webrtc/modules/video_coding/jitter_estimator.h"
|
||||
#include "webrtc/modules/video_coding/sequence_number_util.h"
|
||||
#include "webrtc/modules/video_coding/timing.h"
|
||||
#include "webrtc/system_wrappers/include/clock.h"
|
||||
|
||||
namespace webrtc {
|
||||
namespace video_coding {
|
||||
|
||||
class VCMTimingFake : public VCMTiming {
|
||||
public:
|
||||
explicit VCMTimingFake(Clock* clock) : VCMTiming(clock) {}
|
||||
|
||||
int64_t RenderTimeMs(uint32_t frame_timestamp,
|
||||
int64_t now_ms) const override {
|
||||
if (last_ms_ == -1) {
|
||||
last_ms_ = now_ms + kDelayMs;
|
||||
last_timestamp_ = frame_timestamp;
|
||||
}
|
||||
|
||||
uint32_t diff = MinDiff(frame_timestamp, last_timestamp_);
|
||||
if (AheadOf(frame_timestamp, last_timestamp_))
|
||||
last_ms_ += diff / 90;
|
||||
else
|
||||
last_ms_ -= diff / 90;
|
||||
|
||||
last_timestamp_ = frame_timestamp;
|
||||
return last_ms_;
|
||||
}
|
||||
|
||||
uint32_t MaxWaitingTime(int64_t render_time_ms,
|
||||
int64_t now_ms) const override {
|
||||
return std::max<int>(0, render_time_ms - now_ms - kDecodeTime);
|
||||
}
|
||||
|
||||
private:
|
||||
static constexpr int kDelayMs = 50;
|
||||
static constexpr int kDecodeTime = kDelayMs / 2;
|
||||
mutable uint32_t last_timestamp_ = 0;
|
||||
mutable int64_t last_ms_ = -1;
|
||||
};
|
||||
|
||||
class VCMJitterEstimatorMock : public VCMJitterEstimator {
|
||||
public:
|
||||
explicit VCMJitterEstimatorMock(Clock* clock) : VCMJitterEstimator(clock) {}
|
||||
|
||||
MOCK_METHOD1(UpdateRtt, void(int64_t rttMs));
|
||||
MOCK_METHOD3(UpdateEstimate,
|
||||
void(int64_t frameDelayMs,
|
||||
uint32_t frameSizeBytes,
|
||||
bool incompleteFrame));
|
||||
};
|
||||
|
||||
class FrameObjectMock : public FrameObject {
|
||||
public:
|
||||
MOCK_CONST_METHOD1(GetBitstream, bool(uint8_t* destination));
|
||||
};
|
||||
|
||||
class TestFrameBuffer2 : public ::testing::Test {
|
||||
protected:
|
||||
static constexpr int kMaxReferences = 5;
|
||||
static constexpr int kFps1 = 1000;
|
||||
static constexpr int kFps10 = kFps1 / 10;
|
||||
static constexpr int kFps20 = kFps1 / 20;
|
||||
|
||||
TestFrameBuffer2()
|
||||
: clock_(0),
|
||||
timing_(&clock_),
|
||||
jitter_estimator_(&clock_),
|
||||
buffer_(&clock_, &jitter_estimator_, &timing_),
|
||||
rand_(0x34678213),
|
||||
tear_down_(false),
|
||||
extract_thread_(&ExtractLoop, this, "Extract Thread"),
|
||||
trigger_extract_event_(false, false),
|
||||
crit_acquired_event_(false, false) {}
|
||||
|
||||
void SetUp() override { extract_thread_.Start(); }
|
||||
|
||||
void TearDown() override {
|
||||
tear_down_ = true;
|
||||
trigger_extract_event_.Set();
|
||||
extract_thread_.Stop();
|
||||
}
|
||||
|
||||
template <typename... T>
|
||||
void InsertFrame(uint16_t picture_id,
|
||||
uint8_t spatial_layer,
|
||||
int64_t ts_ms,
|
||||
bool inter_layer_predicted,
|
||||
T... refs) {
|
||||
static_assert(sizeof...(refs) <= kMaxReferences,
|
||||
"To many references specified for FrameObject.");
|
||||
std::array<uint16_t, sizeof...(refs)> references = {{refs...}};
|
||||
|
||||
std::unique_ptr<FrameObjectMock> frame(new FrameObjectMock());
|
||||
frame->picture_id = picture_id;
|
||||
frame->spatial_layer = spatial_layer;
|
||||
frame->timestamp = ts_ms * 90;
|
||||
frame->num_references = references.size();
|
||||
frame->inter_layer_predicted = inter_layer_predicted;
|
||||
for (size_t r = 0; r < references.size(); ++r)
|
||||
frame->references[r] = references[r];
|
||||
|
||||
buffer_.InsertFrame(std::move(frame));
|
||||
}
|
||||
|
||||
void ExtractFrame(int64_t max_wait_time = 0) {
|
||||
crit_.Enter();
|
||||
if (max_wait_time == 0) {
|
||||
frames_.emplace_back(buffer_.NextFrame(0));
|
||||
crit_.Leave();
|
||||
} else {
|
||||
max_wait_time_ = max_wait_time;
|
||||
trigger_extract_event_.Set();
|
||||
crit_.Leave();
|
||||
// Make sure |crit_| is aquired by |extract_thread_| before returning.
|
||||
crit_acquired_event_.Wait(rtc::Event::kForever);
|
||||
}
|
||||
}
|
||||
|
||||
void CheckFrame(size_t index, int picture_id, int spatial_layer) {
|
||||
rtc::CritScope lock(&crit_);
|
||||
ASSERT_LT(index, frames_.size());
|
||||
ASSERT_TRUE(frames_[index]);
|
||||
ASSERT_EQ(picture_id, frames_[index]->picture_id);
|
||||
ASSERT_EQ(spatial_layer, frames_[index]->spatial_layer);
|
||||
}
|
||||
|
||||
void CheckNoFrame(size_t index) {
|
||||
rtc::CritScope lock(&crit_);
|
||||
ASSERT_LT(index, frames_.size());
|
||||
ASSERT_FALSE(frames_[index]);
|
||||
}
|
||||
|
||||
static bool ExtractLoop(void* obj) {
|
||||
TestFrameBuffer2* tfb = static_cast<TestFrameBuffer2*>(obj);
|
||||
while (true) {
|
||||
tfb->trigger_extract_event_.Wait(rtc::Event::kForever);
|
||||
{
|
||||
rtc::CritScope lock(&tfb->crit_);
|
||||
tfb->crit_acquired_event_.Set();
|
||||
if (tfb->tear_down_)
|
||||
return false;
|
||||
|
||||
tfb->frames_.emplace_back(tfb->buffer_.NextFrame(tfb->max_wait_time_));
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
uint32_t Rand() { return rand_.Rand<uint32_t>(); }
|
||||
|
||||
SimulatedClock clock_;
|
||||
VCMTimingFake timing_;
|
||||
VCMJitterEstimatorMock jitter_estimator_;
|
||||
FrameBuffer buffer_;
|
||||
std::vector<std::unique_ptr<FrameObject>> frames_;
|
||||
Random rand_;
|
||||
|
||||
int64_t max_wait_time_;
|
||||
bool tear_down_;
|
||||
rtc::PlatformThread extract_thread_;
|
||||
rtc::Event trigger_extract_event_;
|
||||
rtc::Event crit_acquired_event_;
|
||||
rtc::CriticalSection crit_;
|
||||
};
|
||||
|
||||
TEST_F(TestFrameBuffer2, ExtractFromEmptyBuffer) {
|
||||
ExtractFrame();
|
||||
CheckNoFrame(0);
|
||||
}
|
||||
|
||||
TEST_F(TestFrameBuffer2, WaitForFrame) {
|
||||
uint16_t pid = Rand();
|
||||
uint32_t ts = Rand();
|
||||
|
||||
ExtractFrame(20);
|
||||
InsertFrame(pid, 0, ts, false);
|
||||
CheckFrame(0, pid, 0);
|
||||
}
|
||||
|
||||
TEST_F(TestFrameBuffer2, OneSuperFrame) {
|
||||
uint16_t pid = Rand();
|
||||
uint32_t ts = Rand();
|
||||
|
||||
ExtractFrame(20);
|
||||
InsertFrame(pid, 1, ts, true);
|
||||
InsertFrame(pid, 0, ts, false);
|
||||
ExtractFrame();
|
||||
|
||||
CheckFrame(0, pid, 0);
|
||||
CheckFrame(1, pid, 1);
|
||||
}
|
||||
|
||||
TEST_F(TestFrameBuffer2, OneLayerStream) {
|
||||
uint16_t pid = Rand();
|
||||
uint32_t ts = Rand();
|
||||
|
||||
InsertFrame(pid, 0, ts, false);
|
||||
ExtractFrame();
|
||||
CheckFrame(0, pid, 0);
|
||||
for (int i = 1; i < 10; ++i) {
|
||||
InsertFrame(pid + i, 0, ts + i * kFps10, false, pid + i - 1);
|
||||
ExtractFrame();
|
||||
clock_.AdvanceTimeMilliseconds(kFps10);
|
||||
CheckFrame(i, pid + i, 0);
|
||||
}
|
||||
}
|
||||
|
||||
TEST_F(TestFrameBuffer2, OneLayerStreamReordered) {
|
||||
uint16_t pid = Rand();
|
||||
uint32_t ts = Rand();
|
||||
|
||||
InsertFrame(pid, 0, ts, false);
|
||||
ExtractFrame();
|
||||
CheckFrame(0, pid, 0);
|
||||
for (int i = 1; i < 10; i += 2) {
|
||||
ExtractFrame(15);
|
||||
InsertFrame(pid + i + 1, 0, ts + (i + 1) * kFps10, false, pid + i);
|
||||
clock_.AdvanceTimeMilliseconds(kFps10);
|
||||
InsertFrame(pid + i, 0, ts + i * kFps10, false, pid + i - 1);
|
||||
clock_.AdvanceTimeMilliseconds(kFps10);
|
||||
ExtractFrame();
|
||||
CheckFrame(i, pid + i, 0);
|
||||
CheckFrame(i + 1, pid + i + 1, 0);
|
||||
}
|
||||
}
|
||||
|
||||
TEST_F(TestFrameBuffer2, DropTemporalLayerSlowDecoder) {
|
||||
uint16_t pid = Rand();
|
||||
uint32_t ts = Rand();
|
||||
|
||||
InsertFrame(pid, 0, ts, false);
|
||||
InsertFrame(pid + 1, 0, ts + kFps20, false);
|
||||
for (int i = 2; i < 10; i += 2) {
|
||||
uint32_t ts_tl0 = ts + i / 2 * kFps10;
|
||||
InsertFrame(pid + i, 0, ts_tl0, false, pid + i - 2);
|
||||
InsertFrame(pid + i + 1, 0, ts_tl0 + kFps20, false, pid + i, pid + i - 1);
|
||||
}
|
||||
|
||||
for (int i = 0; i < 10; ++i) {
|
||||
ExtractFrame();
|
||||
clock_.AdvanceTimeMilliseconds(60);
|
||||
}
|
||||
|
||||
CheckFrame(0, pid, 0);
|
||||
CheckFrame(1, pid + 1, 0);
|
||||
CheckFrame(2, pid + 2, 0);
|
||||
CheckFrame(3, pid + 4, 0);
|
||||
CheckFrame(4, pid + 6, 0);
|
||||
CheckFrame(5, pid + 8, 0);
|
||||
CheckNoFrame(6);
|
||||
CheckNoFrame(7);
|
||||
CheckNoFrame(8);
|
||||
CheckNoFrame(9);
|
||||
}
|
||||
|
||||
TEST_F(TestFrameBuffer2, DropSpatialLayerSlowDecoder) {
|
||||
uint16_t pid = Rand();
|
||||
uint32_t ts = Rand();
|
||||
|
||||
InsertFrame(pid, 0, ts, false);
|
||||
InsertFrame(pid, 1, ts, false);
|
||||
for (int i = 1; i < 6; ++i) {
|
||||
uint32_t ts_tl0 = ts + i * kFps10;
|
||||
InsertFrame(pid + i, 0, ts_tl0, false, pid + i - 1);
|
||||
InsertFrame(pid + i, 1, ts_tl0, false, pid + i - 1);
|
||||
}
|
||||
|
||||
ExtractFrame();
|
||||
ExtractFrame();
|
||||
clock_.AdvanceTimeMilliseconds(55);
|
||||
for (int i = 2; i < 12; ++i) {
|
||||
ExtractFrame();
|
||||
clock_.AdvanceTimeMilliseconds(55);
|
||||
}
|
||||
|
||||
CheckFrame(0, pid, 0);
|
||||
CheckFrame(1, pid, 1);
|
||||
CheckFrame(2, pid + 1, 0);
|
||||
CheckFrame(3, pid + 1, 1);
|
||||
CheckFrame(4, pid + 2, 0);
|
||||
CheckFrame(5, pid + 2, 1);
|
||||
CheckFrame(6, pid + 3, 0);
|
||||
CheckFrame(7, pid + 4, 0);
|
||||
CheckFrame(8, pid + 5, 0);
|
||||
CheckNoFrame(9);
|
||||
CheckNoFrame(10);
|
||||
CheckNoFrame(11);
|
||||
}
|
||||
|
||||
TEST_F(TestFrameBuffer2, InsertLateFrame) {
|
||||
uint16_t pid = Rand();
|
||||
uint32_t ts = Rand();
|
||||
|
||||
InsertFrame(pid, 0, ts, false);
|
||||
ExtractFrame();
|
||||
InsertFrame(pid + 2, 0, ts, false);
|
||||
ExtractFrame();
|
||||
InsertFrame(pid + 1, 0, ts, false, pid);
|
||||
ExtractFrame();
|
||||
|
||||
CheckFrame(0, pid, 0);
|
||||
CheckFrame(1, pid + 2, 0);
|
||||
CheckNoFrame(2);
|
||||
}
|
||||
|
||||
} // namespace video_coding
|
||||
} // namespace webrtc
|
@ -18,6 +18,7 @@ namespace video_coding {
|
||||
FrameObject::FrameObject()
|
||||
: picture_id(0),
|
||||
spatial_layer(0),
|
||||
timestamp(0),
|
||||
num_references(0),
|
||||
inter_layer_predicted(false) {}
|
||||
|
||||
|
@ -31,6 +31,7 @@ class FrameObject {
|
||||
// have to be constructed from the header data relevant to that codec.
|
||||
uint16_t picture_id;
|
||||
uint8_t spatial_layer;
|
||||
uint32_t timestamp;
|
||||
|
||||
size_t num_references;
|
||||
uint16_t references[kMaxFrameReferences];
|
||||
|
@ -28,7 +28,7 @@ class VCMTiming {
|
||||
// The primary timing component should be passed
|
||||
// if this is the dual timing component.
|
||||
explicit VCMTiming(Clock* clock, VCMTiming* master_timing = NULL);
|
||||
~VCMTiming();
|
||||
virtual ~VCMTiming();
|
||||
|
||||
// Resets the timing to the initial state.
|
||||
void Reset();
|
||||
@ -69,11 +69,11 @@ class VCMTiming {
|
||||
// Returns the receiver system time when the frame with timestamp
|
||||
// frame_timestamp should be rendered, assuming that the system time currently
|
||||
// is now_ms.
|
||||
int64_t RenderTimeMs(uint32_t frame_timestamp, int64_t now_ms) const;
|
||||
virtual int64_t RenderTimeMs(uint32_t frame_timestamp, int64_t now_ms) const;
|
||||
|
||||
// Returns the maximum time in ms that we can wait for a frame to become
|
||||
// complete before we must pass it to the decoder.
|
||||
uint32_t MaxWaitingTime(int64_t render_time_ms, int64_t now_ms) const;
|
||||
virtual uint32_t MaxWaitingTime(int64_t render_time_ms, int64_t now_ms) const;
|
||||
|
||||
// Returns the current target delay which is required delay + decode time +
|
||||
// render delay.
|
||||
|
@ -32,6 +32,7 @@
|
||||
'encoded_frame.h',
|
||||
'fec_tables_xor.h',
|
||||
'frame_buffer.h',
|
||||
'frame_buffer2.h',
|
||||
'frame_object.h',
|
||||
'rtp_frame_reference_finder.h',
|
||||
'generic_decoder.h',
|
||||
@ -62,6 +63,7 @@
|
||||
'decoding_state.cc',
|
||||
'encoded_frame.cc',
|
||||
'frame_buffer.cc',
|
||||
'frame_buffer2.cc',
|
||||
'frame_object.cc',
|
||||
'rtp_frame_reference_finder.cc',
|
||||
'generic_decoder.cc',
|
||||
|
Reference in New Issue
Block a user