Tommi | a4e71b9 | 2018-02-21 09:37:11 +0100 | [diff] [blame] | 1 | /* |
| 2 | * Copyright (c) 2012 The WebRTC project authors. All Rights Reserved. |
| 3 | * |
| 4 | * Use of this source code is governed by a BSD-style license |
| 5 | * that can be found in the LICENSE file in the root of the source |
| 6 | * tree. An additional intellectual property rights grant can be found |
| 7 | * in the file PATENTS. All contributing project authors may |
| 8 | * be found in the AUTHORS file in the root of the source tree. |
| 9 | */ |
| 10 | |
| 11 | #include "modules/video_coding/generic_decoder_android.h" |
| 12 | |
| 13 | #include <algorithm> |
| 14 | |
| 15 | #include "modules/video_coding/include/video_coding.h" |
| 16 | #include "modules/video_coding/internal_defines.h" |
| 17 | #include "rtc_base/checks.h" |
| 18 | #include "rtc_base/logging.h" |
| 19 | #include "rtc_base/timeutils.h" |
| 20 | #include "rtc_base/trace_event.h" |
| 21 | #include "system_wrappers/include/clock.h" |
| 22 | |
| 23 | namespace webrtc { |
| 24 | |
| 25 | VCMDecodedFrameCallback::VCMDecodedFrameCallback(VCMTiming* timing, |
| 26 | Clock* clock) |
| 27 | : _clock(clock), |
| 28 | _timing(timing), |
| 29 | _timestampMap(kDecoderFrameMemoryLength), |
| 30 | _lastReceivedPictureID(0) { |
| 31 | ntp_offset_ = |
| 32 | _clock->CurrentNtpInMilliseconds() - _clock->TimeInMilliseconds(); |
| 33 | } |
| 34 | |
| 35 | VCMDecodedFrameCallback::~VCMDecodedFrameCallback() { |
| 36 | } |
| 37 | |
| 38 | void VCMDecodedFrameCallback::SetUserReceiveCallback( |
| 39 | VCMReceiveCallback* receiveCallback) { |
| 40 | RTC_DCHECK(construction_thread_.CalledOnValidThread()); |
| 41 | RTC_DCHECK((!_receiveCallback && receiveCallback) || |
| 42 | (_receiveCallback && !receiveCallback)); |
| 43 | _receiveCallback = receiveCallback; |
| 44 | } |
| 45 | |
| 46 | VCMReceiveCallback* VCMDecodedFrameCallback::UserReceiveCallback() { |
| 47 | // Called on the decode thread via VCMCodecDataBase::GetDecoder. |
| 48 | // The callback must always have been set before this happens. |
| 49 | RTC_DCHECK(_receiveCallback); |
| 50 | return _receiveCallback; |
| 51 | } |
| 52 | |
| 53 | int32_t VCMDecodedFrameCallback::Decoded(VideoFrame& decodedImage) { |
| 54 | return Decoded(decodedImage, -1); |
| 55 | } |
| 56 | |
| 57 | int32_t VCMDecodedFrameCallback::Decoded(VideoFrame& decodedImage, |
| 58 | int64_t decode_time_ms) { |
| 59 | Decoded(decodedImage, |
| 60 | decode_time_ms >= 0 ? rtc::Optional<int32_t>(decode_time_ms) |
| 61 | : rtc::nullopt, |
| 62 | rtc::nullopt); |
| 63 | return WEBRTC_VIDEO_CODEC_OK; |
| 64 | } |
| 65 | |
| 66 | void VCMDecodedFrameCallback::Decoded(VideoFrame& decodedImage, |
| 67 | rtc::Optional<int32_t> decode_time_ms, |
| 68 | rtc::Optional<uint8_t> qp) { |
| 69 | RTC_DCHECK(_receiveCallback) << "Callback must not be null at this point"; |
| 70 | TRACE_EVENT_INSTANT1("webrtc", "VCMDecodedFrameCallback::Decoded", |
| 71 | "timestamp", decodedImage.timestamp()); |
| 72 | // TODO(holmer): We should improve this so that we can handle multiple |
| 73 | // callbacks from one call to Decode(). |
| 74 | VCMFrameInformation* frameInfo; |
| 75 | { |
| 76 | rtc::CritScope cs(&lock_); |
| 77 | frameInfo = _timestampMap.Pop(decodedImage.timestamp()); |
| 78 | } |
| 79 | |
| 80 | if (frameInfo == NULL) { |
| 81 | RTC_LOG(LS_WARNING) << "Too many frames backed up in the decoder, dropping " |
| 82 | "this one."; |
| 83 | return; |
| 84 | } |
| 85 | |
| 86 | const int64_t now_ms = _clock->TimeInMilliseconds(); |
| 87 | if (!decode_time_ms) { |
| 88 | decode_time_ms = now_ms - frameInfo->decodeStartTimeMs; |
| 89 | } |
| 90 | _timing->StopDecodeTimer(decodedImage.timestamp(), *decode_time_ms, now_ms, |
| 91 | frameInfo->renderTimeMs); |
| 92 | |
| 93 | // Report timing information. |
| 94 | if (frameInfo->timing.flags != TimingFrameFlags::kInvalid) { |
| 95 | int64_t capture_time_ms = decodedImage.ntp_time_ms() - ntp_offset_; |
| 96 | // Convert remote timestamps to local time from ntp timestamps. |
| 97 | frameInfo->timing.encode_start_ms -= ntp_offset_; |
| 98 | frameInfo->timing.encode_finish_ms -= ntp_offset_; |
| 99 | frameInfo->timing.packetization_finish_ms -= ntp_offset_; |
| 100 | frameInfo->timing.pacer_exit_ms -= ntp_offset_; |
| 101 | frameInfo->timing.network_timestamp_ms -= ntp_offset_; |
| 102 | frameInfo->timing.network2_timestamp_ms -= ntp_offset_; |
| 103 | |
| 104 | int64_t sender_delta_ms = 0; |
| 105 | if (decodedImage.ntp_time_ms() < 0) { |
| 106 | // Sender clock is not estimated yet. Make sure that sender times are all |
| 107 | // negative to indicate that. Yet they still should be relatively correct. |
| 108 | sender_delta_ms = |
| 109 | std::max({capture_time_ms, frameInfo->timing.encode_start_ms, |
| 110 | frameInfo->timing.encode_finish_ms, |
| 111 | frameInfo->timing.packetization_finish_ms, |
| 112 | frameInfo->timing.pacer_exit_ms, |
| 113 | frameInfo->timing.network_timestamp_ms, |
| 114 | frameInfo->timing.network2_timestamp_ms}) + |
| 115 | 1; |
| 116 | } |
| 117 | |
| 118 | TimingFrameInfo timing_frame_info; |
| 119 | |
| 120 | timing_frame_info.capture_time_ms = capture_time_ms - sender_delta_ms; |
| 121 | timing_frame_info.encode_start_ms = |
| 122 | frameInfo->timing.encode_start_ms - sender_delta_ms; |
| 123 | timing_frame_info.encode_finish_ms = |
| 124 | frameInfo->timing.encode_finish_ms - sender_delta_ms; |
| 125 | timing_frame_info.packetization_finish_ms = |
| 126 | frameInfo->timing.packetization_finish_ms - sender_delta_ms; |
| 127 | timing_frame_info.pacer_exit_ms = |
| 128 | frameInfo->timing.pacer_exit_ms - sender_delta_ms; |
| 129 | timing_frame_info.network_timestamp_ms = |
| 130 | frameInfo->timing.network_timestamp_ms - sender_delta_ms; |
| 131 | timing_frame_info.network2_timestamp_ms = |
| 132 | frameInfo->timing.network2_timestamp_ms - sender_delta_ms; |
| 133 | timing_frame_info.receive_start_ms = frameInfo->timing.receive_start_ms; |
| 134 | timing_frame_info.receive_finish_ms = frameInfo->timing.receive_finish_ms; |
| 135 | timing_frame_info.decode_start_ms = frameInfo->decodeStartTimeMs; |
| 136 | timing_frame_info.decode_finish_ms = now_ms; |
| 137 | timing_frame_info.render_time_ms = frameInfo->renderTimeMs; |
| 138 | timing_frame_info.rtp_timestamp = decodedImage.timestamp(); |
| 139 | timing_frame_info.flags = frameInfo->timing.flags; |
| 140 | |
| 141 | _timing->SetTimingFrameInfo(timing_frame_info); |
| 142 | } |
| 143 | |
| 144 | decodedImage.set_timestamp_us( |
| 145 | frameInfo->renderTimeMs * rtc::kNumMicrosecsPerMillisec); |
| 146 | decodedImage.set_rotation(frameInfo->rotation); |
| 147 | _receiveCallback->FrameToRender(decodedImage, qp, frameInfo->content_type); |
| 148 | } |
| 149 | |
| 150 | int32_t VCMDecodedFrameCallback::ReceivedDecodedReferenceFrame( |
| 151 | const uint64_t pictureId) { |
| 152 | return _receiveCallback->ReceivedDecodedReferenceFrame(pictureId); |
| 153 | } |
| 154 | |
| 155 | int32_t VCMDecodedFrameCallback::ReceivedDecodedFrame( |
| 156 | const uint64_t pictureId) { |
| 157 | _lastReceivedPictureID = pictureId; |
| 158 | return 0; |
| 159 | } |
| 160 | |
| 161 | uint64_t VCMDecodedFrameCallback::LastReceivedPictureID() const { |
| 162 | return _lastReceivedPictureID; |
| 163 | } |
| 164 | |
| 165 | void VCMDecodedFrameCallback::OnDecoderImplementationName( |
| 166 | const char* implementation_name) { |
| 167 | _receiveCallback->OnDecoderImplementationName(implementation_name); |
| 168 | } |
| 169 | |
| 170 | void VCMDecodedFrameCallback::Map(uint32_t timestamp, |
| 171 | VCMFrameInformation* frameInfo) { |
| 172 | rtc::CritScope cs(&lock_); |
| 173 | _timestampMap.Add(timestamp, frameInfo); |
| 174 | } |
| 175 | |
| 176 | int32_t VCMDecodedFrameCallback::Pop(uint32_t timestamp) { |
| 177 | rtc::CritScope cs(&lock_); |
| 178 | if (_timestampMap.Pop(timestamp) == NULL) { |
| 179 | return VCM_GENERAL_ERROR; |
| 180 | } |
| 181 | return VCM_OK; |
| 182 | } |
| 183 | |
| 184 | VCMGenericDecoder::VCMGenericDecoder(std::unique_ptr<VideoDecoder> decoder) |
| 185 | : VCMGenericDecoder(decoder.release(), false /* isExternal */) {} |
| 186 | |
| 187 | VCMGenericDecoder::VCMGenericDecoder(VideoDecoder* decoder, bool isExternal) |
| 188 | : _callback(NULL), |
| 189 | _frameInfos(), |
| 190 | _nextFrameInfoIdx(0), |
| 191 | decoder_(decoder), |
| 192 | _codecType(kVideoCodecUnknown), |
| 193 | _isExternal(isExternal), |
| 194 | _last_keyframe_content_type(VideoContentType::UNSPECIFIED) { |
| 195 | RTC_DCHECK(decoder_); |
| 196 | } |
| 197 | |
| 198 | VCMGenericDecoder::~VCMGenericDecoder() { |
| 199 | decoder_->Release(); |
| 200 | if (_isExternal) |
| 201 | decoder_.release(); |
| 202 | RTC_DCHECK(_isExternal || decoder_); |
| 203 | } |
| 204 | |
| 205 | int32_t VCMGenericDecoder::InitDecode(const VideoCodec* settings, |
| 206 | int32_t numberOfCores) { |
| 207 | TRACE_EVENT0("webrtc", "VCMGenericDecoder::InitDecode"); |
| 208 | _codecType = settings->codecType; |
| 209 | |
| 210 | return decoder_->InitDecode(settings, numberOfCores); |
| 211 | } |
| 212 | |
| 213 | int32_t VCMGenericDecoder::Decode(const VCMEncodedFrame& frame, int64_t nowMs) { |
| 214 | TRACE_EVENT1("webrtc", "VCMGenericDecoder::Decode", "timestamp", |
| 215 | frame.EncodedImage()._timeStamp); |
| 216 | _frameInfos[_nextFrameInfoIdx].decodeStartTimeMs = nowMs; |
| 217 | _frameInfos[_nextFrameInfoIdx].renderTimeMs = frame.RenderTimeMs(); |
| 218 | _frameInfos[_nextFrameInfoIdx].rotation = frame.rotation(); |
| 219 | _frameInfos[_nextFrameInfoIdx].timing = frame.video_timing(); |
| 220 | // Set correctly only for key frames. Thus, use latest key frame |
| 221 | // content type. If the corresponding key frame was lost, decode will fail |
| 222 | // and content type will be ignored. |
| 223 | if (frame.FrameType() == kVideoFrameKey) { |
| 224 | _frameInfos[_nextFrameInfoIdx].content_type = frame.contentType(); |
| 225 | _last_keyframe_content_type = frame.contentType(); |
| 226 | } else { |
| 227 | _frameInfos[_nextFrameInfoIdx].content_type = _last_keyframe_content_type; |
| 228 | } |
| 229 | _callback->Map(frame.TimeStamp(), &_frameInfos[_nextFrameInfoIdx]); |
| 230 | |
| 231 | _nextFrameInfoIdx = (_nextFrameInfoIdx + 1) % kDecoderFrameMemoryLength; |
| 232 | const RTPFragmentationHeader dummy_header; |
| 233 | int32_t ret = decoder_->Decode(frame.EncodedImage(), frame.MissingFrame(), |
| 234 | &dummy_header, |
| 235 | frame.CodecSpecific(), frame.RenderTimeMs()); |
| 236 | |
| 237 | _callback->OnDecoderImplementationName(decoder_->ImplementationName()); |
| 238 | if (ret < WEBRTC_VIDEO_CODEC_OK) { |
| 239 | RTC_LOG(LS_WARNING) << "Failed to decode frame with timestamp " |
| 240 | << frame.TimeStamp() << ", error code: " << ret; |
| 241 | _callback->Pop(frame.TimeStamp()); |
| 242 | return ret; |
| 243 | } else if (ret == WEBRTC_VIDEO_CODEC_NO_OUTPUT || |
| 244 | ret == WEBRTC_VIDEO_CODEC_REQUEST_SLI) { |
| 245 | // No output |
| 246 | _callback->Pop(frame.TimeStamp()); |
| 247 | } |
| 248 | return ret; |
| 249 | } |
| 250 | |
| 251 | int32_t VCMGenericDecoder::RegisterDecodeCompleteCallback( |
| 252 | VCMDecodedFrameCallback* callback) { |
| 253 | _callback = callback; |
| 254 | return decoder_->RegisterDecodeCompleteCallback(callback); |
| 255 | } |
| 256 | |
| 257 | bool VCMGenericDecoder::PrefersLateDecoding() const { |
| 258 | return decoder_->PrefersLateDecoding(); |
| 259 | } |
| 260 | |
| 261 | } // namespace webrtc |