blob: bbbe76d4df6df71fa1ea1a822be78e104be9094f [file] [log] [blame]
philipelbe7a9e52016-05-19 12:19:35 +02001/*
2 * Copyright (c) 2016 The WebRTC project authors. All Rights Reserved.
3 *
4 * Use of this source code is governed by a BSD-style license
5 * that can be found in the LICENSE file in the root of the source
6 * tree. An additional intellectual property rights grant can be found
7 * in the file PATENTS. All contributing project authors may
8 * be found in the AUTHORS file in the root of the source tree.
9 */
10
Mirko Bonadei92ea95e2017-09-15 06:47:31 +020011#include "modules/video_coding/frame_buffer2.h"
philipelbe7a9e52016-05-19 12:19:35 +020012
13#include <algorithm>
Yves Gerey3e707812018-11-28 16:47:49 +010014#include <cstdlib>
15#include <iterator>
philipele0b2f152016-09-28 10:23:49 +020016#include <queue>
Yves Gerey3e707812018-11-28 16:47:49 +010017#include <utility>
philipel798b2822018-06-11 13:10:14 +020018#include <vector>
philipelbe7a9e52016-05-19 12:19:35 +020019
Sebastian Jansson11d0d7b2019-04-11 12:39:34 +020020#include "absl/memory/memory.h"
Yves Gerey3e707812018-11-28 16:47:49 +010021#include "api/video/encoded_image.h"
22#include "api/video/video_timing.h"
Mirko Bonadei92ea95e2017-09-15 06:47:31 +020023#include "modules/video_coding/include/video_coding_defines.h"
24#include "modules/video_coding/jitter_estimator.h"
25#include "modules/video_coding/timing.h"
26#include "rtc_base/checks.h"
Yves Gerey3e707812018-11-28 16:47:49 +010027#include "rtc_base/experiments/rtt_mult_experiment.h"
Mirko Bonadei92ea95e2017-09-15 06:47:31 +020028#include "rtc_base/logging.h"
Yves Gerey3e707812018-11-28 16:47:49 +010029#include "rtc_base/numerics/sequence_number_util.h"
Mirko Bonadei92ea95e2017-09-15 06:47:31 +020030#include "rtc_base/trace_event.h"
31#include "system_wrappers/include/clock.h"
philipel707f2782017-10-02 14:10:28 +020032#include "system_wrappers/include/field_trial.h"
philipelbe7a9e52016-05-19 12:19:35 +020033
34namespace webrtc {
35namespace video_coding {
36
37namespace {
philipele0b2f152016-09-28 10:23:49 +020038// Max number of frames the buffer will hold.
Ilya Nikolaevskiy13717842019-01-14 13:24:22 +010039constexpr size_t kMaxFramesBuffered = 800;
philipelbe7a9e52016-05-19 12:19:35 +020040
Johannes Kron23bfff32021-09-28 21:31:46 +020041// Default value for the maximum decode queue size that is used when the
42// low-latency renderer is used.
43constexpr size_t kZeroPlayoutDelayDefaultMaxDecodeQueueSize = 8;
44
philipele0b2f152016-09-28 10:23:49 +020045// Max number of decoded frame info that will be saved.
Ilya Nikolaevskiy13717842019-01-14 13:24:22 +010046constexpr int kMaxFramesHistory = 1 << 13;
philipel65e1f942017-07-24 08:26:53 -070047
Ilya Nikolaevskiy8c4fe162018-02-27 15:49:47 +010048// The time it's allowed for a frame to be late to its rendering prediction and
49// still be rendered.
Ilya Nikolaevskiy7eef0072018-02-28 09:59:26 +010050constexpr int kMaxAllowedFrameDelayMs = 5;
Ilya Nikolaevskiy8c4fe162018-02-27 15:49:47 +010051
philipel65e1f942017-07-24 08:26:53 -070052constexpr int64_t kLogNonDecodedIntervalMs = 5000;
philipelbe7a9e52016-05-19 12:19:35 +020053} // namespace
54
philipelbe7a9e52016-05-19 12:19:35 +020055FrameBuffer::FrameBuffer(Clock* clock,
philipela45102f2017-02-22 05:30:39 -080056 VCMTiming* timing,
57 VCMReceiveStatisticsCallback* stats_callback)
Ilya Nikolaevskiy13717842019-01-14 13:24:22 +010058 : decoded_frames_history_(kMaxFramesHistory),
59 clock_(clock),
Sebastian Jansson11d0d7b2019-04-11 12:39:34 +020060 callback_queue_(nullptr),
Niels Möllerd9c2d942019-04-30 09:16:36 +020061 jitter_estimator_(clock),
philipelbe7a9e52016-05-19 12:19:35 +020062 timing_(timing),
philipel4f6cd6a2016-08-03 10:59:32 +020063 inter_frame_delay_(clock_->TimeInMilliseconds()),
philipel29f730e2017-03-15 08:10:08 -070064 stopped_(false),
philipela45102f2017-02-22 05:30:39 -080065 protection_mode_(kProtectionNack),
philipel65e1f942017-07-24 08:26:53 -070066 stats_callback_(stats_callback),
Elad Alone4b50232019-01-14 18:56:14 +010067 last_log_non_decoded_ms_(-kLogNonDecodedIntervalMs),
Johannes Kron2ddc39e2021-08-10 16:56:12 +020068 rtt_mult_settings_(RttMultExperiment::GetRttMultValue()),
Johannes Kron23bfff32021-09-28 21:31:46 +020069 zero_playout_delay_max_decode_queue_size_(
70 "max_decode_queue_size",
71 kZeroPlayoutDelayDefaultMaxDecodeQueueSize) {
Johannes Kron2ddc39e2021-08-10 16:56:12 +020072 ParseFieldTrial({&zero_playout_delay_max_decode_queue_size_},
73 field_trial::FindFullName("WebRTC-ZeroPlayoutDelay"));
Tommi430951a2020-05-19 23:27:29 +020074 callback_checker_.Detach();
75}
philipel266f0a42016-11-28 08:49:07 -080076
Tommi430951a2020-05-19 23:27:29 +020077FrameBuffer::~FrameBuffer() {
78 RTC_DCHECK_RUN_ON(&construction_checker_);
79}
philipelbe7a9e52016-05-19 12:19:35 +020080
Sebastian Jansson11d0d7b2019-04-11 12:39:34 +020081void FrameBuffer::NextFrame(
82 int64_t max_wait_time_ms,
83 bool keyframe_required,
84 rtc::TaskQueue* callback_queue,
85 std::function<void(std::unique_ptr<EncodedFrame>, ReturnReason)> handler) {
Tommi430951a2020-05-19 23:27:29 +020086 RTC_DCHECK_RUN_ON(&callback_checker_);
87 RTC_DCHECK(callback_queue->IsCurrent());
Sebastian Jansson11d0d7b2019-04-11 12:39:34 +020088 TRACE_EVENT0("webrtc", "FrameBuffer::NextFrame");
89 int64_t latest_return_time_ms =
90 clock_->TimeInMilliseconds() + max_wait_time_ms;
Tommi430951a2020-05-19 23:27:29 +020091
Markus Handell6deec382020-07-07 12:17:12 +020092 MutexLock lock(&mutex_);
Sebastian Jansson11d0d7b2019-04-11 12:39:34 +020093 if (stopped_) {
94 return;
95 }
96 latest_return_time_ms_ = latest_return_time_ms;
97 keyframe_required_ = keyframe_required;
98 frame_handler_ = handler;
99 callback_queue_ = callback_queue;
100 StartWaitForNextFrameOnQueue();
101}
102
103void FrameBuffer::StartWaitForNextFrameOnQueue() {
104 RTC_DCHECK(callback_queue_);
105 RTC_DCHECK(!callback_task_.Running());
106 int64_t wait_ms = FindNextFrame(clock_->TimeInMilliseconds());
107 callback_task_ = RepeatingTaskHandle::DelayedStart(
Danil Chapovalov55284022020-02-07 14:53:52 +0100108 callback_queue_->Get(), TimeDelta::Millis(wait_ms), [this] {
Tommi430951a2020-05-19 23:27:29 +0200109 RTC_DCHECK_RUN_ON(&callback_checker_);
Sebastian Jansson11d0d7b2019-04-11 12:39:34 +0200110 // If this task has not been cancelled, we did not get any new frames
111 // while waiting. Continue with frame delivery.
Markus Handell334b1fd2020-12-14 00:08:11 +0100112 std::unique_ptr<EncodedFrame> frame;
113 std::function<void(std::unique_ptr<EncodedFrame>, ReturnReason)>
114 frame_handler;
115 {
116 MutexLock lock(&mutex_);
117 if (!frames_to_decode_.empty()) {
118 // We have frames, deliver!
119 frame = absl::WrapUnique(GetNextFrame());
Rezaul Barbhuiya82c22482021-08-05 17:54:11 -0700120 timing_->SetLastDecodeScheduledTimestamp(
121 clock_->TimeInMilliseconds());
Markus Handell334b1fd2020-12-14 00:08:11 +0100122 } else if (clock_->TimeInMilliseconds() < latest_return_time_ms_) {
123 // If there's no frames to decode and there is still time left, it
124 // means that the frame buffer was cleared between creation and
125 // execution of this task. Continue waiting for the remaining time.
126 int64_t wait_ms = FindNextFrame(clock_->TimeInMilliseconds());
127 return TimeDelta::Millis(wait_ms);
128 }
129 frame_handler = std::move(frame_handler_);
Sebastian Jansson11d0d7b2019-04-11 12:39:34 +0200130 CancelCallback();
Sebastian Jansson11d0d7b2019-04-11 12:39:34 +0200131 }
Markus Handell334b1fd2020-12-14 00:08:11 +0100132 // Deliver frame, if any. Otherwise signal timeout.
133 ReturnReason reason = frame ? kFrameFound : kTimeout;
134 frame_handler(std::move(frame), reason);
135 return TimeDelta::Zero(); // Ignored.
Sebastian Jansson11d0d7b2019-04-11 12:39:34 +0200136 });
137}
138
Sebastian Jansson1c747f52019-04-04 13:01:39 +0200139int64_t FrameBuffer::FindNextFrame(int64_t now_ms) {
140 int64_t wait_ms = latest_return_time_ms_ - now_ms;
141 frames_to_decode_.clear();
142
Artem Titovdcd7fc72021-08-09 13:02:57 +0200143 // `last_continuous_frame_` may be empty below, but nullopt is smaller
Sebastian Jansson1c747f52019-04-04 13:01:39 +0200144 // than everything else and loop will immediately terminate as expected.
145 for (auto frame_it = frames_.begin();
146 frame_it != frames_.end() && frame_it->first <= last_continuous_frame_;
147 ++frame_it) {
148 if (!frame_it->second.continuous ||
149 frame_it->second.num_missing_decodable > 0) {
150 continue;
151 }
152
153 EncodedFrame* frame = frame_it->second.frame.get();
154
155 if (keyframe_required_ && !frame->is_keyframe())
156 continue;
157
158 auto last_decoded_frame_timestamp =
159 decoded_frames_history_.GetLastDecodedFrameTimestamp();
160
161 // TODO(https://bugs.webrtc.org/9974): consider removing this check
162 // as it may make a stream undecodable after a very long delay between
163 // frames.
164 if (last_decoded_frame_timestamp &&
165 AheadOf(*last_decoded_frame_timestamp, frame->Timestamp())) {
166 continue;
167 }
168
Sebastian Jansson1c747f52019-04-04 13:01:39 +0200169 // Gather all remaining frames for the same superframe.
170 std::vector<FrameMap::iterator> current_superframe;
171 current_superframe.push_back(frame_it);
172 bool last_layer_completed = frame_it->second.frame->is_last_spatial_layer;
173 FrameMap::iterator next_frame_it = frame_it;
philipela65d7852020-11-20 17:49:24 +0100174 while (!last_layer_completed) {
Sebastian Jansson1c747f52019-04-04 13:01:39 +0200175 ++next_frame_it;
philipela65d7852020-11-20 17:49:24 +0100176
177 if (next_frame_it == frames_.end() || !next_frame_it->second.frame) {
178 break;
179 }
180
181 if (next_frame_it->second.frame->Timestamp() != frame->Timestamp() ||
Sebastian Jansson1c747f52019-04-04 13:01:39 +0200182 !next_frame_it->second.continuous) {
183 break;
184 }
philipela65d7852020-11-20 17:49:24 +0100185
186 if (next_frame_it->second.num_missing_decodable > 0) {
philipelcb327d92020-12-10 10:49:20 +0100187 bool has_inter_layer_dependency = false;
188 for (size_t i = 0; i < EncodedFrame::kMaxFrameReferences &&
philipela65d7852020-11-20 17:49:24 +0100189 i < next_frame_it->second.frame->num_references;
190 ++i) {
philipel9aa9b8d2021-02-15 13:31:29 +0100191 if (next_frame_it->second.frame->references[i] >= frame_it->first) {
philipela65d7852020-11-20 17:49:24 +0100192 has_inter_layer_dependency = true;
philipelcb327d92020-12-10 10:49:20 +0100193 break;
philipela65d7852020-11-20 17:49:24 +0100194 }
195 }
196
197 // If the frame has an undecoded dependency that is not within the same
philipelcb327d92020-12-10 10:49:20 +0100198 // temporal unit then this frame is not yet ready to be decoded. If it
philipela65d7852020-11-20 17:49:24 +0100199 // is within the same temporal unit then the not yet decoded dependency
200 // is just a lower spatial frame, which is ok.
201 if (!has_inter_layer_dependency ||
202 next_frame_it->second.num_missing_decodable > 1) {
203 break;
204 }
Sebastian Jansson1c747f52019-04-04 13:01:39 +0200205 }
philipela65d7852020-11-20 17:49:24 +0100206
Sebastian Jansson1c747f52019-04-04 13:01:39 +0200207 current_superframe.push_back(next_frame_it);
208 last_layer_completed = next_frame_it->second.frame->is_last_spatial_layer;
209 }
210 // Check if the current superframe is complete.
211 // TODO(bugs.webrtc.org/10064): consider returning all available to
212 // decode frames even if the superframe is not complete yet.
213 if (!last_layer_completed) {
214 continue;
215 }
216
217 frames_to_decode_ = std::move(current_superframe);
218
219 if (frame->RenderTime() == -1) {
220 frame->SetRenderTime(timing_->RenderTimeMs(frame->Timestamp(), now_ms));
221 }
Johannes Kron2ddc39e2021-08-10 16:56:12 +0200222 bool too_many_frames_queued =
223 frames_.size() > zero_playout_delay_max_decode_queue_size_ ? true
224 : false;
225 wait_ms = timing_->MaxWaitingTime(frame->RenderTime(), now_ms,
226 too_many_frames_queued);
Sebastian Jansson1c747f52019-04-04 13:01:39 +0200227
228 // This will cause the frame buffer to prefer high framerate rather
229 // than high resolution in the case of the decoder not decoding fast
230 // enough and the stream has multiple spatial and temporal layers.
231 // For multiple temporal layers it may cause non-base layer frames to be
232 // skipped if they are late.
233 if (wait_ms < -kMaxAllowedFrameDelayMs)
234 continue;
235
236 break;
237 }
238 wait_ms = std::min<int64_t>(wait_ms, latest_return_time_ms_ - now_ms);
239 wait_ms = std::max<int64_t>(wait_ms, 0);
240 return wait_ms;
241}
242
243EncodedFrame* FrameBuffer::GetNextFrame() {
Tommi430951a2020-05-19 23:27:29 +0200244 RTC_DCHECK_RUN_ON(&callback_checker_);
Sebastian Jansson1c747f52019-04-04 13:01:39 +0200245 int64_t now_ms = clock_->TimeInMilliseconds();
Artem Titovdcd7fc72021-08-09 13:02:57 +0200246 // TODO(ilnik): remove `frames_out` use frames_to_decode_ directly.
Sebastian Jansson1c747f52019-04-04 13:01:39 +0200247 std::vector<EncodedFrame*> frames_out;
248
249 RTC_DCHECK(!frames_to_decode_.empty());
250 bool superframe_delayed_by_retransmission = false;
251 size_t superframe_size = 0;
252 EncodedFrame* first_frame = frames_to_decode_[0]->second.frame.get();
253 int64_t render_time_ms = first_frame->RenderTime();
254 int64_t receive_time_ms = first_frame->ReceivedTime();
255 // Gracefully handle bad RTP timestamps and render time issues.
256 if (HasBadRenderTiming(*first_frame, now_ms)) {
Niels Möllerd9c2d942019-04-30 09:16:36 +0200257 jitter_estimator_.Reset();
Sebastian Jansson1c747f52019-04-04 13:01:39 +0200258 timing_->Reset();
259 render_time_ms = timing_->RenderTimeMs(first_frame->Timestamp(), now_ms);
260 }
261
262 for (FrameMap::iterator& frame_it : frames_to_decode_) {
263 RTC_DCHECK(frame_it != frames_.end());
264 EncodedFrame* frame = frame_it->second.frame.release();
265
266 frame->SetRenderTime(render_time_ms);
267
268 superframe_delayed_by_retransmission |= frame->delayed_by_retransmission();
269 receive_time_ms = std::max(receive_time_ms, frame->ReceivedTime());
270 superframe_size += frame->size();
271
272 PropagateDecodability(frame_it->second);
273 decoded_frames_history_.InsertDecoded(frame_it->first, frame->Timestamp());
274
275 // Remove decoded frame and all undecoded frames before it.
Johannes Kron0c141c52019-08-26 15:04:43 +0200276 if (stats_callback_) {
philipel9aa9b8d2021-02-15 13:31:29 +0100277 unsigned int dropped_frames =
278 std::count_if(frames_.begin(), frame_it,
279 [](const std::pair<const int64_t, FrameInfo>& frame) {
280 return frame.second.frame != nullptr;
281 });
Johannes Kron0c141c52019-08-26 15:04:43 +0200282 if (dropped_frames > 0) {
283 stats_callback_->OnDroppedFrames(dropped_frames);
284 }
285 }
286
Sebastian Jansson1c747f52019-04-04 13:01:39 +0200287 frames_.erase(frames_.begin(), ++frame_it);
288
289 frames_out.push_back(frame);
290 }
291
292 if (!superframe_delayed_by_retransmission) {
293 int64_t frame_delay;
294
295 if (inter_frame_delay_.CalculateDelay(first_frame->Timestamp(),
296 &frame_delay, receive_time_ms)) {
Niels Möllerd9c2d942019-04-30 09:16:36 +0200297 jitter_estimator_.UpdateEstimate(frame_delay, superframe_size);
Sebastian Jansson1c747f52019-04-04 13:01:39 +0200298 }
299
300 float rtt_mult = protection_mode_ == kProtectionNackFEC ? 0.0 : 1.0;
“Michaeld3a4ebe2019-06-07 03:55:01 -0500301 absl::optional<float> rtt_mult_add_cap_ms = absl::nullopt;
302 if (rtt_mult_settings_.has_value()) {
303 rtt_mult = rtt_mult_settings_->rtt_mult_setting;
304 rtt_mult_add_cap_ms = rtt_mult_settings_->rtt_mult_add_cap_ms;
Sebastian Jansson1c747f52019-04-04 13:01:39 +0200305 }
“Michaele0f37042019-06-04 10:04:12 -0500306 timing_->SetJitterDelay(
307 jitter_estimator_.GetJitterEstimate(rtt_mult, rtt_mult_add_cap_ms));
Sebastian Jansson1c747f52019-04-04 13:01:39 +0200308 timing_->UpdateCurrentDelay(render_time_ms, now_ms);
309 } else {
Evan Shrubsole9146d762021-10-13 11:19:48 +0200310 if (RttMultExperiment::RttMultEnabled())
Niels Möllerd9c2d942019-04-30 09:16:36 +0200311 jitter_estimator_.FrameNacked();
Sebastian Jansson1c747f52019-04-04 13:01:39 +0200312 }
313
314 UpdateJitterDelay();
315 UpdateTimingFrameInfo();
316
317 if (frames_out.size() == 1) {
318 return frames_out[0];
319 } else {
320 return CombineAndDeleteFrames(frames_out);
321 }
322}
323
philipele7c891f2018-02-22 14:35:06 +0100324bool FrameBuffer::HasBadRenderTiming(const EncodedFrame& frame,
325 int64_t now_ms) {
stefan95e97542017-05-23 09:52:18 -0700326 // Assume that render timing errors are due to changes in the video stream.
327 int64_t render_time_ms = frame.RenderTimeMs();
Stefan Holmer812ceaf2018-05-15 13:00:10 +0200328 // Zero render time means render immediately.
329 if (render_time_ms == 0) {
330 return false;
331 }
stefan95e97542017-05-23 09:52:18 -0700332 if (render_time_ms < 0) {
333 return true;
334 }
Stefan Holmer812ceaf2018-05-15 13:00:10 +0200335 const int64_t kMaxVideoDelayMs = 10000;
stefan95e97542017-05-23 09:52:18 -0700336 if (std::abs(render_time_ms - now_ms) > kMaxVideoDelayMs) {
337 int frame_delay = static_cast<int>(std::abs(render_time_ms - now_ms));
Mirko Bonadei675513b2017-11-09 11:09:25 +0100338 RTC_LOG(LS_WARNING)
339 << "A frame about to be decoded is out of the configured "
Jonas Olssonb2b20312020-01-14 12:11:31 +0100340 "delay bounds ("
341 << frame_delay << " > " << kMaxVideoDelayMs
Mirko Bonadei675513b2017-11-09 11:09:25 +0100342 << "). Resetting the video jitter buffer.";
stefan95e97542017-05-23 09:52:18 -0700343 return true;
344 }
345 if (static_cast<int>(timing_->TargetVideoDelay()) > kMaxVideoDelayMs) {
Mirko Bonadei675513b2017-11-09 11:09:25 +0100346 RTC_LOG(LS_WARNING) << "The video target delay has grown larger than "
347 << kMaxVideoDelayMs << " ms.";
stefan95e97542017-05-23 09:52:18 -0700348 return true;
349 }
350 return false;
351}
352
philipel4f6cd6a2016-08-03 10:59:32 +0200353void FrameBuffer::SetProtectionMode(VCMVideoProtection mode) {
tommidb23ea62017-03-03 07:21:18 -0800354 TRACE_EVENT0("webrtc", "FrameBuffer::SetProtectionMode");
Markus Handell6deec382020-07-07 12:17:12 +0200355 MutexLock lock(&mutex_);
Henrik Boströmc680c4a2019-04-03 10:27:36 +0000356 protection_mode_ = mode;
philipel4f6cd6a2016-08-03 10:59:32 +0200357}
358
philipel504c47d2016-06-30 17:33:02 +0200359void FrameBuffer::Stop() {
tommidb23ea62017-03-03 07:21:18 -0800360 TRACE_EVENT0("webrtc", "FrameBuffer::Stop");
Markus Handell6deec382020-07-07 12:17:12 +0200361 MutexLock lock(&mutex_);
Tommi430951a2020-05-19 23:27:29 +0200362 if (stopped_)
363 return;
Henrik Boströmc680c4a2019-04-03 10:27:36 +0000364 stopped_ = true;
Tommi430951a2020-05-19 23:27:29 +0200365
Sebastian Jansson11d0d7b2019-04-11 12:39:34 +0200366 CancelCallback();
philipel504c47d2016-06-30 17:33:02 +0200367}
368
Ilya Nikolaevskiye6a2d942018-11-07 14:32:28 +0100369void FrameBuffer::Clear() {
Markus Handell6deec382020-07-07 12:17:12 +0200370 MutexLock lock(&mutex_);
Henrik Boströmc680c4a2019-04-03 10:27:36 +0000371 ClearFramesAndHistory();
Ilya Nikolaevskiye6a2d942018-11-07 14:32:28 +0100372}
373
Johannes Kron111e9812020-10-26 13:54:40 +0100374int FrameBuffer::Size() {
375 MutexLock lock(&mutex_);
376 return frames_.size();
377}
378
philipele21be1d2017-09-25 06:37:12 -0700379void FrameBuffer::UpdateRtt(int64_t rtt_ms) {
Markus Handell6deec382020-07-07 12:17:12 +0200380 MutexLock lock(&mutex_);
Niels Möllerd9c2d942019-04-30 09:16:36 +0200381 jitter_estimator_.UpdateRtt(rtt_ms);
philipele21be1d2017-09-25 06:37:12 -0700382}
383
philipele7c891f2018-02-22 14:35:06 +0100384bool FrameBuffer::ValidReferences(const EncodedFrame& frame) const {
philipel112adf92017-06-15 09:06:21 -0700385 for (size_t i = 0; i < frame.num_references; ++i) {
philipel9aa9b8d2021-02-15 13:31:29 +0100386 if (frame.references[i] >= frame.Id())
philipel112adf92017-06-15 09:06:21 -0700387 return false;
philipel3b3c9c42017-09-11 09:38:36 -0700388
philipel112adf92017-06-15 09:06:21 -0700389 for (size_t j = i + 1; j < frame.num_references; ++j) {
390 if (frame.references[i] == frame.references[j])
391 return false;
392 }
393 }
394
philipel112adf92017-06-15 09:06:21 -0700395 return true;
396}
397
Sebastian Jansson11d0d7b2019-04-11 12:39:34 +0200398void FrameBuffer::CancelCallback() {
Tommi430951a2020-05-19 23:27:29 +0200399 // Called from the callback queue or from within Stop().
Sebastian Jansson11d0d7b2019-04-11 12:39:34 +0200400 frame_handler_ = {};
401 callback_task_.Stop();
402 callback_queue_ = nullptr;
Tommi430951a2020-05-19 23:27:29 +0200403 callback_checker_.Detach();
Sebastian Jansson11d0d7b2019-04-11 12:39:34 +0200404}
405
philipele7c891f2018-02-22 14:35:06 +0100406int64_t FrameBuffer::InsertFrame(std::unique_ptr<EncodedFrame> frame) {
tommidb23ea62017-03-03 07:21:18 -0800407 TRACE_EVENT0("webrtc", "FrameBuffer::InsertFrame");
philipel93e451b2016-10-06 12:25:13 +0200408 RTC_DCHECK(frame);
tommi0a735642017-03-14 06:23:57 -0700409
Markus Handell6deec382020-07-07 12:17:12 +0200410 MutexLock lock(&mutex_);
philipel29f730e2017-03-15 08:10:08 -0700411
philipel9aa9b8d2021-02-15 13:31:29 +0100412 int64_t last_continuous_frame_id = last_continuous_frame_.value_or(-1);
philipele0b2f152016-09-28 10:23:49 +0200413
philipel112adf92017-06-15 09:06:21 -0700414 if (!ValidReferences(*frame)) {
philipel9aa9b8d2021-02-15 13:31:29 +0100415 RTC_LOG(LS_WARNING) << "Frame " << frame->Id()
philipelf1091932021-02-11 15:25:08 +0100416 << " has invalid frame references, dropping frame.";
philipel9aa9b8d2021-02-15 13:31:29 +0100417 return last_continuous_frame_id;
philipel112adf92017-06-15 09:06:21 -0700418 }
419
Ilya Nikolaevskiy6551faf2019-01-10 15:16:47 +0100420 if (frames_.size() >= kMaxFramesBuffered) {
philipel9771c502018-03-02 11:06:27 +0100421 if (frame->is_keyframe()) {
philipel9aa9b8d2021-02-15 13:31:29 +0100422 RTC_LOG(LS_WARNING) << "Inserting keyframe " << frame->Id()
philipelf1091932021-02-11 15:25:08 +0100423 << " but buffer is full, clearing"
Jonas Olssonb2b20312020-01-14 12:11:31 +0100424 " buffer and inserting the frame.";
philipel9771c502018-03-02 11:06:27 +0100425 ClearFramesAndHistory();
426 } else {
philipel9aa9b8d2021-02-15 13:31:29 +0100427 RTC_LOG(LS_WARNING) << "Frame " << frame->Id()
philipelf1091932021-02-11 15:25:08 +0100428 << " could not be inserted due to the frame "
Jonas Olssonb2b20312020-01-14 12:11:31 +0100429 "buffer being full, dropping frame.";
philipel9aa9b8d2021-02-15 13:31:29 +0100430 return last_continuous_frame_id;
philipel9771c502018-03-02 11:06:27 +0100431 }
philipele0b2f152016-09-28 10:23:49 +0200432 }
433
Ilya Nikolaevskiy13717842019-01-14 13:24:22 +0100434 auto last_decoded_frame = decoded_frames_history_.GetLastDecodedFrameId();
435 auto last_decoded_frame_timestamp =
436 decoded_frames_history_.GetLastDecodedFrameTimestamp();
philipel9aa9b8d2021-02-15 13:31:29 +0100437 if (last_decoded_frame && frame->Id() <= *last_decoded_frame) {
Ilya Nikolaevskiy13717842019-01-14 13:24:22 +0100438 if (AheadOf(frame->Timestamp(), *last_decoded_frame_timestamp) &&
philipel3042c2d2017-08-18 04:55:02 -0700439 frame->is_keyframe()) {
philipel9aa9b8d2021-02-15 13:31:29 +0100440 // If this frame has a newer timestamp but an earlier frame id then we
441 // assume there has been a jump in the frame id due to some encoder
philipelfcc60062017-01-18 05:35:20 -0800442 // reconfiguration or some other reason. Even though this is not according
443 // to spec we can still continue to decode from this frame if it is a
444 // keyframe.
Mirko Bonadei675513b2017-11-09 11:09:25 +0100445 RTC_LOG(LS_WARNING)
philipel9aa9b8d2021-02-15 13:31:29 +0100446 << "A jump in frame id was detected, clearing buffer.";
philipelfcc60062017-01-18 05:35:20 -0800447 ClearFramesAndHistory();
philipel9aa9b8d2021-02-15 13:31:29 +0100448 last_continuous_frame_id = -1;
philipelfcc60062017-01-18 05:35:20 -0800449 } else {
philipel9aa9b8d2021-02-15 13:31:29 +0100450 RTC_LOG(LS_WARNING) << "Frame " << frame->Id() << " inserted after frame "
451 << *last_decoded_frame
philipelf1091932021-02-11 15:25:08 +0100452 << " was handed off for decoding, dropping frame.";
philipel9aa9b8d2021-02-15 13:31:29 +0100453 return last_continuous_frame_id;
philipelfcc60062017-01-18 05:35:20 -0800454 }
philipele0b2f152016-09-28 10:23:49 +0200455 }
456
philipel146a48b2017-04-20 04:04:38 -0700457 // Test if inserting this frame would cause the order of the frames to become
458 // ambiguous (covering more than half the interval of 2^16). This can happen
philipel9aa9b8d2021-02-15 13:31:29 +0100459 // when the frame id make large jumps mid stream.
460 if (!frames_.empty() && frame->Id() < frames_.begin()->first &&
461 frames_.rbegin()->first < frame->Id()) {
462 RTC_LOG(LS_WARNING) << "A jump in frame id was detected, clearing buffer.";
philipel146a48b2017-04-20 04:04:38 -0700463 ClearFramesAndHistory();
philipel9aa9b8d2021-02-15 13:31:29 +0100464 last_continuous_frame_id = -1;
philipel146a48b2017-04-20 04:04:38 -0700465 }
466
philipel9aa9b8d2021-02-15 13:31:29 +0100467 auto info = frames_.emplace(frame->Id(), FrameInfo()).first;
philipele0b2f152016-09-28 10:23:49 +0200468
philipel93e451b2016-10-06 12:25:13 +0200469 if (info->second.frame) {
philipel9aa9b8d2021-02-15 13:31:29 +0100470 return last_continuous_frame_id;
philipele0b2f152016-09-28 10:23:49 +0200471 }
472
philipel93e451b2016-10-06 12:25:13 +0200473 if (!UpdateFrameInfoWithIncomingFrame(*frame, info))
philipel9aa9b8d2021-02-15 13:31:29 +0100474 return last_continuous_frame_id;
Ruslan Burakov493a6502019-02-27 15:32:48 +0100475
476 if (!frame->delayed_by_retransmission())
477 timing_->IncomingTimestamp(frame->Timestamp(), frame->ReceivedTime());
philipel0a9f6de2018-02-28 11:29:47 +0100478
philipel0cb73262020-12-08 17:36:53 +0100479 // It can happen that a frame will be reported as fully received even if a
480 // lower spatial layer frame is missing.
481 if (stats_callback_ && frame->is_last_spatial_layer) {
Johannes Kronb88b44e2019-08-22 13:16:44 +0200482 stats_callback_->OnCompleteFrame(frame->is_keyframe(), frame->size(),
483 frame->contentType());
484 }
485
philipele0b2f152016-09-28 10:23:49 +0200486 info->second.frame = std::move(frame);
philipele0b2f152016-09-28 10:23:49 +0200487
488 if (info->second.num_missing_continuous == 0) {
489 info->second.continuous = true;
490 PropagateContinuity(info);
philipel9aa9b8d2021-02-15 13:31:29 +0100491 last_continuous_frame_id = *last_continuous_frame_;
philipele0b2f152016-09-28 10:23:49 +0200492
493 // Since we now have new continuous frames there might be a better frame
Sebastian Jansson11d0d7b2019-04-11 12:39:34 +0200494 // to return from NextFrame.
Sebastian Jansson11d0d7b2019-04-11 12:39:34 +0200495 if (callback_queue_) {
496 callback_queue_->PostTask([this] {
Markus Handell6deec382020-07-07 12:17:12 +0200497 MutexLock lock(&mutex_);
Sebastian Jansson11d0d7b2019-04-11 12:39:34 +0200498 if (!callback_task_.Running())
499 return;
500 RTC_CHECK(frame_handler_);
501 callback_task_.Stop();
502 StartWaitForNextFrameOnQueue();
503 });
504 }
philipele0b2f152016-09-28 10:23:49 +0200505 }
506
philipel9aa9b8d2021-02-15 13:31:29 +0100507 return last_continuous_frame_id;
philipelbe7a9e52016-05-19 12:19:35 +0200508}
509
philipele0b2f152016-09-28 10:23:49 +0200510void FrameBuffer::PropagateContinuity(FrameMap::iterator start) {
tommidb23ea62017-03-03 07:21:18 -0800511 TRACE_EVENT0("webrtc", "FrameBuffer::PropagateContinuity");
philipele0b2f152016-09-28 10:23:49 +0200512 RTC_DCHECK(start->second.continuous);
philipele0b2f152016-09-28 10:23:49 +0200513
514 std::queue<FrameMap::iterator> continuous_frames;
515 continuous_frames.push(start);
516
517 // A simple BFS to traverse continuous frames.
518 while (!continuous_frames.empty()) {
519 auto frame = continuous_frames.front();
520 continuous_frames.pop();
521
Ilya Nikolaevskiy6551faf2019-01-10 15:16:47 +0100522 if (!last_continuous_frame_ || *last_continuous_frame_ < frame->first) {
523 last_continuous_frame_ = frame->first;
524 }
philipele0b2f152016-09-28 10:23:49 +0200525
526 // Loop through all dependent frames, and if that frame no longer has
527 // any unfulfilled dependencies then that frame is continuous as well.
Elad Alon69321dd2019-01-10 15:02:54 +0100528 for (size_t d = 0; d < frame->second.dependent_frames.size(); ++d) {
philipele0b2f152016-09-28 10:23:49 +0200529 auto frame_ref = frames_.find(frame->second.dependent_frames[d]);
philipel112adf92017-06-15 09:06:21 -0700530 RTC_DCHECK(frame_ref != frames_.end());
philipele0b2f152016-09-28 10:23:49 +0200531
philipel112adf92017-06-15 09:06:21 -0700532 // TODO(philipel): Look into why we've seen this happen.
533 if (frame_ref != frames_.end()) {
534 --frame_ref->second.num_missing_continuous;
535 if (frame_ref->second.num_missing_continuous == 0) {
536 frame_ref->second.continuous = true;
537 continuous_frames.push(frame_ref);
538 }
philipele0b2f152016-09-28 10:23:49 +0200539 }
540 }
541 }
542}
543
544void FrameBuffer::PropagateDecodability(const FrameInfo& info) {
tommidb23ea62017-03-03 07:21:18 -0800545 TRACE_EVENT0("webrtc", "FrameBuffer::PropagateDecodability");
Elad Alon69321dd2019-01-10 15:02:54 +0100546 for (size_t d = 0; d < info.dependent_frames.size(); ++d) {
philipele0b2f152016-09-28 10:23:49 +0200547 auto ref_info = frames_.find(info.dependent_frames[d]);
philipel93e451b2016-10-06 12:25:13 +0200548 RTC_DCHECK(ref_info != frames_.end());
tommie95b78b2017-05-14 07:23:11 -0700549 // TODO(philipel): Look into why we've seen this happen.
550 if (ref_info != frames_.end()) {
551 RTC_DCHECK_GT(ref_info->second.num_missing_decodable, 0U);
552 --ref_info->second.num_missing_decodable;
553 }
philipele0b2f152016-09-28 10:23:49 +0200554 }
555}
556
philipele7c891f2018-02-22 14:35:06 +0100557bool FrameBuffer::UpdateFrameInfoWithIncomingFrame(const EncodedFrame& frame,
philipele0b2f152016-09-28 10:23:49 +0200558 FrameMap::iterator info) {
tommidb23ea62017-03-03 07:21:18 -0800559 TRACE_EVENT0("webrtc", "FrameBuffer::UpdateFrameInfoWithIncomingFrame");
Ilya Nikolaevskiy13717842019-01-14 13:24:22 +0100560 auto last_decoded_frame = decoded_frames_history_.GetLastDecodedFrameId();
561 RTC_DCHECK(!last_decoded_frame || *last_decoded_frame < info->first);
philipele0b2f152016-09-28 10:23:49 +0200562
Artem Titovdcd7fc72021-08-09 13:02:57 +0200563 // In this function we determine how many missing dependencies this `frame`
564 // has to become continuous/decodable. If a frame that this `frame` depend
philipel798b2822018-06-11 13:10:14 +0200565 // on has already been decoded then we can ignore that dependency since it has
566 // already been fulfilled.
567 //
Artem Titovdcd7fc72021-08-09 13:02:57 +0200568 // For all other frames we will register a backwards reference to this `frame`
569 // so that `num_missing_continuous` and `num_missing_decodable` can be
philipel798b2822018-06-11 13:10:14 +0200570 // decremented as frames become continuous/are decoded.
571 struct Dependency {
philipel9aa9b8d2021-02-15 13:31:29 +0100572 int64_t frame_id;
philipel798b2822018-06-11 13:10:14 +0200573 bool continuous;
574 };
575 std::vector<Dependency> not_yet_fulfilled_dependencies;
576
577 // Find all dependencies that have not yet been fulfilled.
philipele0b2f152016-09-28 10:23:49 +0200578 for (size_t i = 0; i < frame.num_references; ++i) {
Artem Titovdcd7fc72021-08-09 13:02:57 +0200579 // Does `frame` depend on a frame earlier than the last decoded one?
philipel9aa9b8d2021-02-15 13:31:29 +0100580 if (last_decoded_frame && frame.references[i] <= *last_decoded_frame) {
Artem Titovdcd7fc72021-08-09 13:02:57 +0200581 // Was that frame decoded? If not, this `frame` will never become
philipel798b2822018-06-11 13:10:14 +0200582 // decodable.
philipel9aa9b8d2021-02-15 13:31:29 +0100583 if (!decoded_frames_history_.WasDecoded(frame.references[i])) {
philipel65e1f942017-07-24 08:26:53 -0700584 int64_t now_ms = clock_->TimeInMilliseconds();
585 if (last_log_non_decoded_ms_ + kLogNonDecodedIntervalMs < now_ms) {
Mirko Bonadei675513b2017-11-09 11:09:25 +0100586 RTC_LOG(LS_WARNING)
philipel9aa9b8d2021-02-15 13:31:29 +0100587 << "Frame " << frame.Id()
philipelf1091932021-02-11 15:25:08 +0100588 << " depends on a non-decoded frame more previous than the last "
589 "decoded frame, dropping frame.";
philipel65e1f942017-07-24 08:26:53 -0700590 last_log_non_decoded_ms_ = now_ms;
591 }
philipele0b2f152016-09-28 10:23:49 +0200592 return false;
593 }
philipele0b2f152016-09-28 10:23:49 +0200594 } else {
philipel9aa9b8d2021-02-15 13:31:29 +0100595 auto ref_info = frames_.find(frame.references[i]);
philipel798b2822018-06-11 13:10:14 +0200596 bool ref_continuous =
597 ref_info != frames_.end() && ref_info->second.continuous;
philipel9aa9b8d2021-02-15 13:31:29 +0100598 not_yet_fulfilled_dependencies.push_back(
599 {frame.references[i], ref_continuous});
philipele0b2f152016-09-28 10:23:49 +0200600 }
philipelbe7a9e52016-05-19 12:19:35 +0200601 }
602
philipel798b2822018-06-11 13:10:14 +0200603 info->second.num_missing_continuous = not_yet_fulfilled_dependencies.size();
604 info->second.num_missing_decodable = not_yet_fulfilled_dependencies.size();
605
606 for (const Dependency& dep : not_yet_fulfilled_dependencies) {
607 if (dep.continuous)
608 --info->second.num_missing_continuous;
609
philipel9aa9b8d2021-02-15 13:31:29 +0100610 frames_[dep.frame_id].dependent_frames.push_back(frame.Id());
philipel798b2822018-06-11 13:10:14 +0200611 }
philipel93e451b2016-10-06 12:25:13 +0200612
philipelbe7a9e52016-05-19 12:19:35 +0200613 return true;
614}
615
philipelbe742702016-11-30 01:31:40 -0800616void FrameBuffer::UpdateJitterDelay() {
tommidb23ea62017-03-03 07:21:18 -0800617 TRACE_EVENT0("webrtc", "FrameBuffer::UpdateJitterDelay");
philipela45102f2017-02-22 05:30:39 -0800618 if (!stats_callback_)
619 return;
philipelbe742702016-11-30 01:31:40 -0800620
philipela45102f2017-02-22 05:30:39 -0800621 int max_decode_ms;
622 int current_delay_ms;
623 int target_delay_ms;
624 int jitter_buffer_ms;
625 int min_playout_delay_ms;
626 int render_delay_ms;
Johannes Kronbfd343b2019-07-01 10:07:50 +0200627 if (timing_->GetTimings(&max_decode_ms, &current_delay_ms, &target_delay_ms,
628 &jitter_buffer_ms, &min_playout_delay_ms,
629 &render_delay_ms)) {
philipela45102f2017-02-22 05:30:39 -0800630 stats_callback_->OnFrameBufferTimingsUpdated(
Johannes Kronbfd343b2019-07-01 10:07:50 +0200631 max_decode_ms, current_delay_ms, target_delay_ms, jitter_buffer_ms,
632 min_playout_delay_ms, render_delay_ms);
philipelbe742702016-11-30 01:31:40 -0800633 }
philipel266f0a42016-11-28 08:49:07 -0800634}
635
ilnik2edc6842017-07-06 03:06:50 -0700636void FrameBuffer::UpdateTimingFrameInfo() {
637 TRACE_EVENT0("webrtc", "FrameBuffer::UpdateTimingFrameInfo");
Danil Chapovalov0040b662018-06-18 10:48:16 +0200638 absl::optional<TimingFrameInfo> info = timing_->GetTimingFrameInfo();
philipel97187112018-03-23 10:43:21 +0100639 if (info && stats_callback_)
ilnik2edc6842017-07-06 03:06:50 -0700640 stats_callback_->OnTimingFrameInfoUpdated(*info);
641}
642
philipelfcc60062017-01-18 05:35:20 -0800643void FrameBuffer::ClearFramesAndHistory() {
ilnik2edc6842017-07-06 03:06:50 -0700644 TRACE_EVENT0("webrtc", "FrameBuffer::ClearFramesAndHistory");
Johannes Kron0c141c52019-08-26 15:04:43 +0200645 if (stats_callback_) {
philipel9aa9b8d2021-02-15 13:31:29 +0100646 unsigned int dropped_frames =
647 std::count_if(frames_.begin(), frames_.end(),
648 [](const std::pair<const int64_t, FrameInfo>& frame) {
649 return frame.second.frame != nullptr;
650 });
Johannes Kron0c141c52019-08-26 15:04:43 +0200651 if (dropped_frames > 0) {
652 stats_callback_->OnDroppedFrames(dropped_frames);
653 }
654 }
philipelfcc60062017-01-18 05:35:20 -0800655 frames_.clear();
Ilya Nikolaevskiy6551faf2019-01-10 15:16:47 +0100656 last_continuous_frame_.reset();
Ilya Nikolaevskiy5546aef2018-12-04 15:54:52 +0100657 frames_to_decode_.clear();
Ilya Nikolaevskiy13717842019-01-14 13:24:22 +0100658 decoded_frames_history_.Clear();
philipelfcc60062017-01-18 05:35:20 -0800659}
660
Niels Möllerff2e2152019-09-27 10:29:30 +0200661// TODO(philipel): Avoid the concatenation of frames here, by replacing
662// NextFrame and GetNextFrame with methods returning multiple frames.
Ilya Nikolaevskiy5546aef2018-12-04 15:54:52 +0100663EncodedFrame* FrameBuffer::CombineAndDeleteFrames(
664 const std::vector<EncodedFrame*>& frames) const {
665 RTC_DCHECK(!frames.empty());
Sergey Silkin61832dd2018-12-20 14:32:14 +0100666 EncodedFrame* first_frame = frames[0];
667 EncodedFrame* last_frame = frames.back();
Ilya Nikolaevskiy5546aef2018-12-04 15:54:52 +0100668 size_t total_length = 0;
669 for (size_t i = 0; i < frames.size(); ++i) {
670 total_length += frames[i]->size();
671 }
Niels Möllerff2e2152019-09-27 10:29:30 +0200672 auto encoded_image_buffer = EncodedImageBuffer::Create(total_length);
673 uint8_t* buffer = encoded_image_buffer->data();
philipel0cb73262020-12-08 17:36:53 +0100674 first_frame->SetSpatialLayerFrameSize(first_frame->SpatialIndex().value_or(0),
Sergey Silkin2799e632019-05-17 09:51:39 +0200675 first_frame->size());
Niels Möllerff2e2152019-09-27 10:29:30 +0200676 memcpy(buffer, first_frame->data(), first_frame->size());
677 buffer += first_frame->size();
Sergey Silkin2799e632019-05-17 09:51:39 +0200678
Sergey Silkin61832dd2018-12-20 14:32:14 +0100679 // Spatial index of combined frame is set equal to spatial index of its top
680 // spatial layer.
philipel0cb73262020-12-08 17:36:53 +0100681 first_frame->SetSpatialIndex(last_frame->SpatialIndex().value_or(0));
Sergey Silkin61832dd2018-12-20 14:32:14 +0100682
683 first_frame->video_timing_mutable()->network2_timestamp_ms =
684 last_frame->video_timing().network2_timestamp_ms;
685 first_frame->video_timing_mutable()->receive_finish_ms =
686 last_frame->video_timing().receive_finish_ms;
687
Ilya Nikolaevskiy5546aef2018-12-04 15:54:52 +0100688 // Append all remaining frames to the first one.
Ilya Nikolaevskiy5546aef2018-12-04 15:54:52 +0100689 for (size_t i = 1; i < frames.size(); ++i) {
Sergey Silkin61832dd2018-12-20 14:32:14 +0100690 EncodedFrame* next_frame = frames[i];
philipel0cb73262020-12-08 17:36:53 +0100691 first_frame->SetSpatialLayerFrameSize(
692 next_frame->SpatialIndex().value_or(0), next_frame->size());
Niels Möller9c843902019-01-11 10:21:35 +0100693 memcpy(buffer, next_frame->data(), next_frame->size());
Sergey Silkin61832dd2018-12-20 14:32:14 +0100694 buffer += next_frame->size();
695 delete next_frame;
Ilya Nikolaevskiy5546aef2018-12-04 15:54:52 +0100696 }
Niels Möllerff2e2152019-09-27 10:29:30 +0200697 first_frame->SetEncodedData(encoded_image_buffer);
Sergey Silkin61832dd2018-12-20 14:32:14 +0100698 return first_frame;
Ilya Nikolaevskiy5546aef2018-12-04 15:54:52 +0100699}
700
Niels Möllerbe682d42018-03-27 08:31:45 +0200701FrameBuffer::FrameInfo::FrameInfo() = default;
702FrameBuffer::FrameInfo::FrameInfo(FrameInfo&&) = default;
703FrameBuffer::FrameInfo::~FrameInfo() = default;
704
philipelbe7a9e52016-05-19 12:19:35 +0200705} // namespace video_coding
706} // namespace webrtc