SSRC knowledge is contained withing VideoSendStream. That also means that debug recording is moved to VideoSendStream. I think that make sence since that allows debug recording with external encoder implementations one day. BUG=webrtc:5687 Review-Url: https://codereview.webrtc.org/1936503002 Cr-Commit-Position: refs/heads/master@{#12632}
409 lines
13 KiB
C++
409 lines
13 KiB
C++
/*
|
|
* Copyright (c) 2012 The WebRTC project authors. All Rights Reserved.
|
|
*
|
|
* Use of this source code is governed by a BSD-style license
|
|
* that can be found in the LICENSE file in the root of the source
|
|
* tree. An additional intellectual property rights grant can be found
|
|
* in the file PATENTS. All contributing project authors may
|
|
* be found in the AUTHORS file in the root of the source tree.
|
|
*/
|
|
|
|
#include "webrtc/video/vie_encoder.h"
|
|
|
|
#include <assert.h>
|
|
|
|
#include <algorithm>
|
|
|
|
#include "webrtc/base/checks.h"
|
|
#include "webrtc/base/logging.h"
|
|
#include "webrtc/base/trace_event.h"
|
|
#include "webrtc/modules/pacing/paced_sender.h"
|
|
#include "webrtc/modules/video_coding/include/video_coding.h"
|
|
#include "webrtc/modules/video_coding/include/video_coding_defines.h"
|
|
#include "webrtc/system_wrappers/include/metrics.h"
|
|
#include "webrtc/system_wrappers/include/tick_util.h"
|
|
#include "webrtc/video/overuse_frame_detector.h"
|
|
#include "webrtc/video/send_statistics_proxy.h"
|
|
#include "webrtc/video_frame.h"
|
|
|
|
namespace webrtc {
|
|
|
|
static const float kStopPaddingThresholdMs = 2000;
|
|
|
|
class QMVideoSettingsCallback : public VCMQMSettingsCallback {
|
|
public:
|
|
explicit QMVideoSettingsCallback(VideoProcessing* vpm);
|
|
|
|
~QMVideoSettingsCallback();
|
|
|
|
// Update VPM with QM (quality modes: frame size & frame rate) settings.
|
|
int32_t SetVideoQMSettings(const uint32_t frame_rate,
|
|
const uint32_t width,
|
|
const uint32_t height);
|
|
|
|
private:
|
|
VideoProcessing* vp_;
|
|
};
|
|
|
|
ViEEncoder::ViEEncoder(uint32_t number_of_cores,
|
|
ProcessThread* module_process_thread,
|
|
SendStatisticsProxy* stats_proxy,
|
|
OveruseFrameDetector* overuse_detector)
|
|
: number_of_cores_(number_of_cores),
|
|
vp_(VideoProcessing::Create()),
|
|
qm_callback_(new QMVideoSettingsCallback(vp_.get())),
|
|
video_sender_(Clock::GetRealTimeClock(),
|
|
this,
|
|
this,
|
|
qm_callback_.get(),
|
|
this),
|
|
stats_proxy_(stats_proxy),
|
|
overuse_detector_(overuse_detector),
|
|
time_of_last_frame_activity_ms_(0),
|
|
encoder_config_(),
|
|
min_transmit_bitrate_bps_(0),
|
|
last_observed_bitrate_bps_(0),
|
|
network_is_transmitting_(true),
|
|
encoder_paused_(true),
|
|
encoder_paused_and_dropped_frame_(false),
|
|
module_process_thread_(module_process_thread),
|
|
has_received_sli_(false),
|
|
picture_id_sli_(0),
|
|
has_received_rpsi_(false),
|
|
picture_id_rpsi_(0),
|
|
video_suspended_(false) {
|
|
module_process_thread_->RegisterModule(&video_sender_);
|
|
vp_->EnableTemporalDecimation(true);
|
|
|
|
// Enable/disable content analysis: off by default for now.
|
|
vp_->EnableContentAnalysis(false);
|
|
}
|
|
|
|
vcm::VideoSender* ViEEncoder::video_sender() {
|
|
return &video_sender_;
|
|
}
|
|
|
|
ViEEncoder::~ViEEncoder() {
|
|
module_process_thread_->DeRegisterModule(&video_sender_);
|
|
}
|
|
|
|
void ViEEncoder::SetNetworkTransmissionState(bool is_transmitting) {
|
|
{
|
|
rtc::CritScope lock(&data_cs_);
|
|
network_is_transmitting_ = is_transmitting;
|
|
}
|
|
}
|
|
|
|
void ViEEncoder::Pause() {
|
|
rtc::CritScope lock(&data_cs_);
|
|
encoder_paused_ = true;
|
|
}
|
|
|
|
void ViEEncoder::Start() {
|
|
rtc::CritScope lock(&data_cs_);
|
|
encoder_paused_ = false;
|
|
}
|
|
|
|
int32_t ViEEncoder::RegisterExternalEncoder(webrtc::VideoEncoder* encoder,
|
|
uint8_t pl_type,
|
|
bool internal_source) {
|
|
video_sender_.RegisterExternalEncoder(encoder, pl_type, internal_source);
|
|
return 0;
|
|
}
|
|
|
|
int32_t ViEEncoder::DeRegisterExternalEncoder(uint8_t pl_type) {
|
|
video_sender_.RegisterExternalEncoder(nullptr, pl_type, false);
|
|
return 0;
|
|
}
|
|
void ViEEncoder::SetEncoder(const webrtc::VideoCodec& video_codec,
|
|
int min_transmit_bitrate_bps,
|
|
size_t max_data_payload_length,
|
|
EncodedImageCallback* sink) {
|
|
// Setting target width and height for VPM.
|
|
RTC_CHECK_EQ(VPM_OK,
|
|
vp_->SetTargetResolution(video_codec.width, video_codec.height,
|
|
video_codec.maxFramerate));
|
|
|
|
// Cache codec before calling AddBitrateObserver (which calls OnBitrateUpdated
|
|
// that makes use of the number of simulcast streams configured).
|
|
{
|
|
rtc::CritScope lock(&data_cs_);
|
|
encoder_config_ = video_codec;
|
|
min_transmit_bitrate_bps_ = min_transmit_bitrate_bps;
|
|
}
|
|
{
|
|
rtc::CritScope lock(&sink_cs_);
|
|
sink_ = sink;
|
|
}
|
|
|
|
bool success = video_sender_.RegisterSendCodec(
|
|
&video_codec, number_of_cores_,
|
|
static_cast<uint32_t>(max_data_payload_length)) == VCM_OK;
|
|
if (!success) {
|
|
LOG(LS_ERROR) << "Failed to configure encoder.";
|
|
RTC_DCHECK(success);
|
|
}
|
|
|
|
if (stats_proxy_) {
|
|
VideoEncoderConfig::ContentType content_type =
|
|
VideoEncoderConfig::ContentType::kRealtimeVideo;
|
|
switch (video_codec.mode) {
|
|
case kRealtimeVideo:
|
|
content_type = VideoEncoderConfig::ContentType::kRealtimeVideo;
|
|
break;
|
|
case kScreensharing:
|
|
content_type = VideoEncoderConfig::ContentType::kScreen;
|
|
break;
|
|
default:
|
|
RTC_NOTREACHED();
|
|
break;
|
|
}
|
|
stats_proxy_->SetContentType(content_type);
|
|
}
|
|
}
|
|
|
|
int ViEEncoder::GetPaddingNeededBps() const {
|
|
int64_t time_of_last_frame_activity_ms;
|
|
int min_transmit_bitrate_bps;
|
|
int bitrate_bps;
|
|
VideoCodec send_codec;
|
|
{
|
|
rtc::CritScope lock(&data_cs_);
|
|
bool send_padding = encoder_config_.numberOfSimulcastStreams > 1 ||
|
|
video_suspended_ || min_transmit_bitrate_bps_ > 0;
|
|
if (!send_padding)
|
|
return 0;
|
|
time_of_last_frame_activity_ms = time_of_last_frame_activity_ms_;
|
|
min_transmit_bitrate_bps = min_transmit_bitrate_bps_;
|
|
bitrate_bps = last_observed_bitrate_bps_;
|
|
send_codec = encoder_config_;
|
|
}
|
|
|
|
bool video_is_suspended = video_sender_.VideoSuspended();
|
|
|
|
// Find the max amount of padding we can allow ourselves to send at this
|
|
// point, based on which streams are currently active and what our current
|
|
// available bandwidth is.
|
|
int pad_up_to_bitrate_bps = 0;
|
|
if (send_codec.numberOfSimulcastStreams == 0) {
|
|
pad_up_to_bitrate_bps = send_codec.minBitrate * 1000;
|
|
} else {
|
|
SimulcastStream* stream_configs = send_codec.simulcastStream;
|
|
pad_up_to_bitrate_bps =
|
|
stream_configs[send_codec.numberOfSimulcastStreams - 1].minBitrate *
|
|
1000;
|
|
for (int i = 0; i < send_codec.numberOfSimulcastStreams - 1; ++i) {
|
|
pad_up_to_bitrate_bps += stream_configs[i].targetBitrate * 1000;
|
|
}
|
|
}
|
|
|
|
// Disable padding if only sending one stream and video isn't suspended and
|
|
// min-transmit bitrate isn't used (applied later).
|
|
if (!video_is_suspended && send_codec.numberOfSimulcastStreams <= 1)
|
|
pad_up_to_bitrate_bps = 0;
|
|
|
|
// The amount of padding should decay to zero if no frames are being
|
|
// captured/encoded unless a min-transmit bitrate is used.
|
|
int64_t now_ms = TickTime::MillisecondTimestamp();
|
|
if (now_ms - time_of_last_frame_activity_ms > kStopPaddingThresholdMs)
|
|
pad_up_to_bitrate_bps = 0;
|
|
|
|
// Pad up to min bitrate.
|
|
if (pad_up_to_bitrate_bps < min_transmit_bitrate_bps)
|
|
pad_up_to_bitrate_bps = min_transmit_bitrate_bps;
|
|
|
|
// Padding may never exceed bitrate estimate.
|
|
if (pad_up_to_bitrate_bps > bitrate_bps)
|
|
pad_up_to_bitrate_bps = bitrate_bps;
|
|
|
|
return pad_up_to_bitrate_bps;
|
|
}
|
|
|
|
bool ViEEncoder::EncoderPaused() const {
|
|
// Pause video if paused by caller or as long as the network is down or the
|
|
// pacer queue has grown too large in buffered mode.
|
|
if (encoder_paused_) {
|
|
return true;
|
|
}
|
|
if (video_suspended_ || last_observed_bitrate_bps_ == 0) {
|
|
return true;
|
|
}
|
|
return !network_is_transmitting_;
|
|
}
|
|
|
|
void ViEEncoder::TraceFrameDropStart() {
|
|
// Start trace event only on the first frame after encoder is paused.
|
|
if (!encoder_paused_and_dropped_frame_) {
|
|
TRACE_EVENT_ASYNC_BEGIN0("webrtc", "EncoderPaused", this);
|
|
}
|
|
encoder_paused_and_dropped_frame_ = true;
|
|
return;
|
|
}
|
|
|
|
void ViEEncoder::TraceFrameDropEnd() {
|
|
// End trace event on first frame after encoder resumes, if frame was dropped.
|
|
if (encoder_paused_and_dropped_frame_) {
|
|
TRACE_EVENT_ASYNC_END0("webrtc", "EncoderPaused", this);
|
|
}
|
|
encoder_paused_and_dropped_frame_ = false;
|
|
}
|
|
|
|
void ViEEncoder::EncodeVideoFrame(const VideoFrame& video_frame) {
|
|
VideoCodecType codec_type;
|
|
{
|
|
rtc::CritScope lock(&data_cs_);
|
|
time_of_last_frame_activity_ms_ = TickTime::MillisecondTimestamp();
|
|
if (EncoderPaused()) {
|
|
TraceFrameDropStart();
|
|
return;
|
|
}
|
|
TraceFrameDropEnd();
|
|
codec_type = encoder_config_.codecType;
|
|
}
|
|
|
|
TRACE_EVENT_ASYNC_STEP0("webrtc", "Video", video_frame.render_time_ms(),
|
|
"Encode");
|
|
const VideoFrame* frame_to_send = &video_frame;
|
|
// TODO(wuchengli): support texture frames.
|
|
if (!video_frame.video_frame_buffer()->native_handle()) {
|
|
// Pass frame via preprocessor.
|
|
frame_to_send = vp_->PreprocessFrame(video_frame);
|
|
if (!frame_to_send) {
|
|
// Drop this frame, or there was an error processing it.
|
|
return;
|
|
}
|
|
}
|
|
|
|
if (codec_type == webrtc::kVideoCodecVP8) {
|
|
webrtc::CodecSpecificInfo codec_specific_info;
|
|
codec_specific_info.codecType = webrtc::kVideoCodecVP8;
|
|
{
|
|
rtc::CritScope lock(&data_cs_);
|
|
codec_specific_info.codecSpecific.VP8.hasReceivedRPSI =
|
|
has_received_rpsi_;
|
|
codec_specific_info.codecSpecific.VP8.hasReceivedSLI =
|
|
has_received_sli_;
|
|
codec_specific_info.codecSpecific.VP8.pictureIdRPSI =
|
|
picture_id_rpsi_;
|
|
codec_specific_info.codecSpecific.VP8.pictureIdSLI =
|
|
picture_id_sli_;
|
|
has_received_sli_ = false;
|
|
has_received_rpsi_ = false;
|
|
}
|
|
|
|
video_sender_.AddVideoFrame(*frame_to_send, vp_->GetContentMetrics(),
|
|
&codec_specific_info);
|
|
return;
|
|
}
|
|
video_sender_.AddVideoFrame(*frame_to_send, nullptr, nullptr);
|
|
}
|
|
|
|
void ViEEncoder::SendKeyFrame() {
|
|
video_sender_.IntraFrameRequest(0);
|
|
}
|
|
|
|
void ViEEncoder::SetProtectionMethod(bool nack, bool fec) {
|
|
// Set Video Protection for VCM.
|
|
VCMVideoProtection protection_mode;
|
|
if (fec) {
|
|
protection_mode =
|
|
nack ? webrtc::kProtectionNackFEC : kProtectionFEC;
|
|
} else {
|
|
protection_mode = nack ? kProtectionNack : kProtectionNone;
|
|
}
|
|
video_sender_.SetVideoProtection(protection_mode);
|
|
}
|
|
|
|
void ViEEncoder::OnSetRates(uint32_t bitrate_bps, int framerate) {
|
|
if (stats_proxy_)
|
|
stats_proxy_->OnSetRates(bitrate_bps, framerate);
|
|
}
|
|
|
|
int32_t ViEEncoder::Encoded(const EncodedImage& encoded_image,
|
|
const CodecSpecificInfo* codec_specific_info,
|
|
const RTPFragmentationHeader* fragmentation) {
|
|
{
|
|
rtc::CritScope lock(&data_cs_);
|
|
time_of_last_frame_activity_ms_ = TickTime::MillisecondTimestamp();
|
|
}
|
|
if (stats_proxy_) {
|
|
stats_proxy_->OnSendEncodedImage(encoded_image, codec_specific_info);
|
|
}
|
|
|
|
int success = 0;
|
|
{
|
|
rtc::CritScope lock(&sink_cs_);
|
|
success = sink_->Encoded(encoded_image, codec_specific_info, fragmentation);
|
|
}
|
|
|
|
overuse_detector_->FrameSent(encoded_image._timeStamp);
|
|
return success;
|
|
}
|
|
|
|
void ViEEncoder::SendStatistics(uint32_t bit_rate,
|
|
uint32_t frame_rate,
|
|
const std::string& encoder_name) {
|
|
if (stats_proxy_)
|
|
stats_proxy_->OnEncoderStatsUpdate(frame_rate, bit_rate, encoder_name);
|
|
}
|
|
|
|
void ViEEncoder::OnReceivedSLI(uint8_t picture_id) {
|
|
rtc::CritScope lock(&data_cs_);
|
|
picture_id_sli_ = picture_id;
|
|
has_received_sli_ = true;
|
|
}
|
|
|
|
void ViEEncoder::OnReceivedRPSI(uint64_t picture_id) {
|
|
rtc::CritScope lock(&data_cs_);
|
|
picture_id_rpsi_ = picture_id;
|
|
has_received_rpsi_ = true;
|
|
}
|
|
|
|
void ViEEncoder::OnReceivedIntraFrameRequest(size_t stream_index) {
|
|
// Key frame request from remote side, signal to VCM.
|
|
TRACE_EVENT0("webrtc", "OnKeyFrameRequest");
|
|
video_sender_.IntraFrameRequest(stream_index);
|
|
}
|
|
|
|
void ViEEncoder::OnBitrateUpdated(uint32_t bitrate_bps,
|
|
uint8_t fraction_lost,
|
|
int64_t round_trip_time_ms) {
|
|
LOG(LS_VERBOSE) << "OnBitrateUpdated, bitrate " << bitrate_bps
|
|
<< " packet loss " << static_cast<int>(fraction_lost)
|
|
<< " rtt " << round_trip_time_ms;
|
|
video_sender_.SetChannelParameters(bitrate_bps, fraction_lost,
|
|
round_trip_time_ms);
|
|
bool video_is_suspended = video_sender_.VideoSuspended();
|
|
bool video_suspension_changed;
|
|
{
|
|
rtc::CritScope lock(&data_cs_);
|
|
last_observed_bitrate_bps_ = bitrate_bps;
|
|
video_suspension_changed = video_suspended_ != video_is_suspended;
|
|
video_suspended_ = video_is_suspended;
|
|
}
|
|
|
|
if (!video_suspension_changed)
|
|
return;
|
|
// Video suspend-state changed, inform codec observer.
|
|
LOG(LS_INFO) << "Video suspend state changed " << video_is_suspended;
|
|
|
|
if (stats_proxy_)
|
|
stats_proxy_->OnSuspendChange(video_is_suspended);
|
|
}
|
|
|
|
QMVideoSettingsCallback::QMVideoSettingsCallback(VideoProcessing* vpm)
|
|
: vp_(vpm) {
|
|
}
|
|
|
|
QMVideoSettingsCallback::~QMVideoSettingsCallback() {
|
|
}
|
|
|
|
int32_t QMVideoSettingsCallback::SetVideoQMSettings(
|
|
const uint32_t frame_rate,
|
|
const uint32_t width,
|
|
const uint32_t height) {
|
|
return vp_->SetTargetResolution(width, height, frame_rate);
|
|
}
|
|
|
|
} // namespace webrtc
|