/* * Copyright (c) 2012 The WebRTC project authors. All Rights Reserved. * * Use of this source code is governed by a BSD-style license * that can be found in the LICENSE file in the root of the source * tree. An additional intellectual property rights grant can be found * in the file PATENTS. All contributing project authors may * be found in the AUTHORS file in the root of the source tree. */ #include "webrtc/modules/rtp_rtcp/source/rtp_receiver_video.h" #include #include // assert #include // memcpy() #include "webrtc/modules/rtp_rtcp/source/receiver_fec.h" #include "webrtc/modules/rtp_rtcp/source/rtp_format_video_generic.h" #include "webrtc/modules/rtp_rtcp/source/rtp_payload_registry.h" #include "webrtc/modules/rtp_rtcp/source/rtp_rtcp_impl.h" #include "webrtc/modules/rtp_rtcp/source/rtp_utility.h" #include "webrtc/system_wrappers/interface/critical_section_wrapper.h" #include "webrtc/system_wrappers/interface/trace.h" #include "webrtc/system_wrappers/interface/trace_event.h" namespace webrtc { uint32_t BitRateBPS(uint16_t x) { return (x & 0x3fff) * uint32_t(pow(10.0f, (2 + (x >> 14)))); } RTPReceiverVideo::RTPReceiverVideo( const int32_t id, const RTPPayloadRegistry* rtp_rtp_payload_registry, RtpData* data_callback) : RTPReceiverStrategy(data_callback), id_(id), rtp_rtp_payload_registry_(rtp_rtp_payload_registry), critical_section_receiver_video_( CriticalSectionWrapper::CreateCriticalSection()), current_fec_frame_decoded_(false), receive_fec_(NULL) { } RTPReceiverVideo::~RTPReceiverVideo() { delete critical_section_receiver_video_; delete receive_fec_; } bool RTPReceiverVideo::ShouldReportCsrcChanges( uint8_t payload_type) const { // Always do this for video packets. return true; } int32_t RTPReceiverVideo::OnNewPayloadTypeCreated( const char payload_name[RTP_PAYLOAD_NAME_SIZE], const int8_t payload_type, const uint32_t frequency) { if (ModuleRTPUtility::StringCompare(payload_name, "ULPFEC", 6)) { // Enable FEC if not enabled. if (receive_fec_ == NULL) { receive_fec_ = new ReceiverFEC(id_, this); } receive_fec_->SetPayloadTypeFEC(payload_type); } return 0; } int32_t RTPReceiverVideo::ParseRtpPacket( WebRtcRTPHeader* rtp_header, const ModuleRTPUtility::PayloadUnion& specific_payload, const bool is_red, const uint8_t* packet, const uint16_t packet_length, const int64_t timestamp_ms, const bool is_first_packet) { TRACE_EVENT2("webrtc_rtp", "Video::ParseRtp", "seqnum", rtp_header->header.sequenceNumber, "timestamp", rtp_header->header.timestamp); const uint8_t* payload_data = ModuleRTPUtility::GetPayloadData(rtp_header->header, packet); const uint16_t payload_data_length = ModuleRTPUtility::GetPayloadDataLength(rtp_header->header, packet_length); return ParseVideoCodecSpecific(rtp_header, payload_data, payload_data_length, specific_payload.Video.videoCodecType, is_red, packet, packet_length, timestamp_ms, is_first_packet); } int32_t RTPReceiverVideo::GetFrequencyHz() const { return kDefaultVideoFrequency; } RTPAliveType RTPReceiverVideo::ProcessDeadOrAlive( uint16_t last_payload_length) const { return kRtpDead; } int32_t RTPReceiverVideo::InvokeOnInitializeDecoder( RtpFeedback* callback, const int32_t id, const int8_t payload_type, const char payload_name[RTP_PAYLOAD_NAME_SIZE], const ModuleRTPUtility::PayloadUnion& specific_payload) const { // For video we just go with default values. if (-1 == callback->OnInitializeDecoder( id, payload_type, payload_name, kDefaultVideoFrequency, 1, 0)) { WEBRTC_TRACE(kTraceError, kTraceRtpRtcp, id, "Failed to create video decoder for payload type:%d", payload_type); return -1; } return 0; } // we have no critext when calling this // we are not allowed to have any critsects when calling // CallbackOfReceivedPayloadData int32_t RTPReceiverVideo::ParseVideoCodecSpecific( WebRtcRTPHeader* rtp_header, const uint8_t* payload_data, const uint16_t payload_data_length, const RtpVideoCodecTypes video_type, const bool is_red, const uint8_t* incoming_rtp_packet, const uint16_t incoming_rtp_packet_size, const int64_t now_ms, const bool is_first_packet) { int32_t ret_val = 0; critical_section_receiver_video_->Enter(); if (is_red) { if (receive_fec_ == NULL) { critical_section_receiver_video_->Leave(); return -1; } bool FECpacket = false; ret_val = receive_fec_->AddReceivedFECPacket( rtp_header, incoming_rtp_packet, payload_data_length, FECpacket); if (ret_val != -1) { ret_val = receive_fec_->ProcessReceivedFEC(); } critical_section_receiver_video_->Leave(); if (ret_val == 0 && FECpacket) { // Callback with the received FEC packet. // The normal packets are delivered after parsing. // This contains the original RTP packet header but with // empty payload and data length. rtp_header->frameType = kFrameEmpty; // We need this for the routing. int32_t ret_val = SetCodecType(video_type, rtp_header); if (ret_val != 0) { return ret_val; } // Pass the length of FEC packets so that they can be accounted for in // the bandwidth estimator. ret_val = data_callback_->OnReceivedPayloadData( NULL, payload_data_length, rtp_header); } } else { // will leave the critical_section_receiver_video_ critsect ret_val = ParseVideoCodecSpecificSwitch(rtp_header, payload_data, payload_data_length, video_type, is_first_packet); } return ret_val; } int32_t RTPReceiverVideo::BuildRTPheader( const WebRtcRTPHeader* rtp_header, uint8_t* data_buffer) const { data_buffer[0] = static_cast(0x80); // version 2 data_buffer[1] = static_cast(rtp_header->header.payloadType); if (rtp_header->header.markerBit) { data_buffer[1] |= kRtpMarkerBitMask; // MarkerBit is 1 } ModuleRTPUtility::AssignUWord16ToBuffer(data_buffer + 2, rtp_header->header.sequenceNumber); ModuleRTPUtility::AssignUWord32ToBuffer(data_buffer + 4, rtp_header->header.timestamp); ModuleRTPUtility::AssignUWord32ToBuffer(data_buffer + 8, rtp_header->header.ssrc); int32_t rtp_header_length = 12; // Add the CSRCs if any if (rtp_header->header.numCSRCs > 0) { if (rtp_header->header.numCSRCs > 16) { // error assert(false); } uint8_t* ptr = &data_buffer[rtp_header_length]; for (uint32_t i = 0; i < rtp_header->header.numCSRCs; ++i) { ModuleRTPUtility::AssignUWord32ToBuffer(ptr, rtp_header->header.arrOfCSRCs[i]); ptr += 4; } data_buffer[0] = (data_buffer[0] & 0xf0) | rtp_header->header.numCSRCs; // Update length of header rtp_header_length += sizeof(uint32_t) * rtp_header->header.numCSRCs; } return rtp_header_length; } int32_t RTPReceiverVideo::ReceiveRecoveredPacketCallback( WebRtcRTPHeader* rtp_header, const uint8_t* payload_data, const uint16_t payload_data_length) { // TODO(pwestin) Re-factor this to avoid the messy critsect handling. critical_section_receiver_video_->Enter(); current_fec_frame_decoded_ = true; ModuleRTPUtility::Payload* payload = NULL; if (rtp_rtp_payload_registry_->PayloadTypeToPayload( rtp_header->header.payloadType, payload) != 0) { critical_section_receiver_video_->Leave(); return -1; } // here we can re-create the original lost packet so that we can use it for // the relay we need to re-create the RED header too uint8_t recovered_packet[IP_PACKET_SIZE]; uint16_t rtp_header_length = (uint16_t) BuildRTPheader(rtp_header, recovered_packet); const uint8_t kREDForFECHeaderLength = 1; // replace pltype recovered_packet[1] &= 0x80; // Reset. recovered_packet[1] += rtp_rtp_payload_registry_->red_payload_type(); // add RED header recovered_packet[rtp_header_length] = rtp_header->header.payloadType; // f-bit always 0 memcpy(recovered_packet + rtp_header_length + kREDForFECHeaderLength, payload_data, payload_data_length); // A recovered packet can be the first packet, but we lack the ability to // detect it at the moment since we do not store the history of recently // received packets. Most codecs like VP8 deal with this in other ways. bool is_first_packet = false; return ParseVideoCodecSpecificSwitch( rtp_header, payload_data, payload_data_length, payload->typeSpecific.Video.videoCodecType, is_first_packet); } int32_t RTPReceiverVideo::SetCodecType( const RtpVideoCodecTypes video_type, WebRtcRTPHeader* rtp_header) const { switch (video_type) { case kRtpGenericVideo: rtp_header->type.Video.codec = kRTPVideoGeneric; break; case kRtpVp8Video: rtp_header->type.Video.codec = kRTPVideoVP8; break; case kRtpFecVideo: rtp_header->type.Video.codec = kRTPVideoFEC; break; } return 0; } int32_t RTPReceiverVideo::ParseVideoCodecSpecificSwitch( WebRtcRTPHeader* rtp_header, const uint8_t* payload_data, const uint16_t payload_data_length, const RtpVideoCodecTypes video_type, const bool is_first_packet) { int32_t ret_val = SetCodecType(video_type, rtp_header); if (ret_val != 0) { critical_section_receiver_video_->Leave(); return ret_val; } WEBRTC_TRACE(kTraceStream, kTraceRtpRtcp, id_, "%s(timestamp:%u)", __FUNCTION__, rtp_header->header.timestamp); // All receive functions release critical_section_receiver_video_ before // returning. switch (video_type) { case kRtpGenericVideo: rtp_header->type.Video.isFirstPacket = is_first_packet; return ReceiveGenericCodec(rtp_header, payload_data, payload_data_length); case kRtpVp8Video: return ReceiveVp8Codec(rtp_header, payload_data, payload_data_length); case kRtpFecVideo: break; } critical_section_receiver_video_->Leave(); return -1; } int32_t RTPReceiverVideo::ReceiveVp8Codec( WebRtcRTPHeader* rtp_header, const uint8_t* payload_data, const uint16_t payload_data_length) { bool success; ModuleRTPUtility::RTPPayload parsed_packet; if (payload_data_length == 0) { success = true; parsed_packet.info.VP8.dataLength = 0; } else { ModuleRTPUtility::RTPPayloadParser rtp_payload_parser( kRtpVp8Video, payload_data, payload_data_length, id_); success = rtp_payload_parser.Parse(parsed_packet); } // from here down we only work on local data critical_section_receiver_video_->Leave(); if (!success) { return -1; } if (parsed_packet.info.VP8.dataLength == 0) { // we have an "empty" VP8 packet, it's ok, could be one way video // Inform the jitter buffer about this packet. rtp_header->frameType = kFrameEmpty; if (data_callback_->OnReceivedPayloadData(NULL, 0, rtp_header) != 0) { return -1; } return 0; } rtp_header->frameType = (parsed_packet.frameType == ModuleRTPUtility::kIFrame) ? kVideoFrameKey : kVideoFrameDelta; RTPVideoHeaderVP8* to_header = &rtp_header->type.Video.codecHeader.VP8; ModuleRTPUtility::RTPPayloadVP8* from_header = &parsed_packet.info.VP8; rtp_header->type.Video.isFirstPacket = from_header->beginningOfPartition && (from_header->partitionID == 0); to_header->nonReference = from_header->nonReferenceFrame; to_header->pictureId = from_header->hasPictureID ? from_header->pictureID : kNoPictureId; to_header->tl0PicIdx = from_header->hasTl0PicIdx ? from_header->tl0PicIdx : kNoTl0PicIdx; if (from_header->hasTID) { to_header->temporalIdx = from_header->tID; to_header->layerSync = from_header->layerSync; } else { to_header->temporalIdx = kNoTemporalIdx; to_header->layerSync = false; } to_header->keyIdx = from_header->hasKeyIdx ? from_header->keyIdx : kNoKeyIdx; rtp_header->type.Video.width = from_header->frameWidth; rtp_header->type.Video.height = from_header->frameHeight; to_header->partitionId = from_header->partitionID; to_header->beginningOfPartition = from_header->beginningOfPartition; if (data_callback_->OnReceivedPayloadData(parsed_packet.info.VP8.data, parsed_packet.info.VP8.dataLength, rtp_header) != 0) { return -1; } return 0; } int32_t RTPReceiverVideo::ReceiveGenericCodec( WebRtcRTPHeader* rtp_header, const uint8_t* payload_data, uint16_t payload_data_length) { uint8_t generic_header = *payload_data++; --payload_data_length; rtp_header->frameType = ((generic_header & RtpFormatVideoGeneric::kKeyFrameBit) != 0) ? kVideoFrameKey : kVideoFrameDelta; rtp_header->type.Video.isFirstPacket = (generic_header & RtpFormatVideoGeneric::kFirstPacketBit) != 0; critical_section_receiver_video_->Leave(); if (data_callback_->OnReceivedPayloadData( payload_data, payload_data_length, rtp_header) != 0) { return -1; } return 0; } } // namespace webrtc