kjellander@webrtc.org | 35a1756 | 2011-10-06 06:44:54 +0000 | [diff] [blame] | 1 | /* |
pwestin@webrtc.org | ce33035 | 2012-04-12 06:59:14 +0000 | [diff] [blame] | 2 | * Copyright (c) 2012 The WebRTC project authors. All Rights Reserved. |
kjellander@webrtc.org | 35a1756 | 2011-10-06 06:44:54 +0000 | [diff] [blame] | 3 | * |
| 4 | * Use of this source code is governed by a BSD-style license |
| 5 | * that can be found in the LICENSE file in the root of the source |
| 6 | * tree. An additional intellectual property rights grant can be found |
| 7 | * in the file PATENTS. All contributing project authors may |
| 8 | * be found in the AUTHORS file in the root of the source tree. |
| 9 | */ |
kjellander@webrtc.org | 5b97b12 | 2011-12-08 07:42:18 +0000 | [diff] [blame] | 10 | |
Mirko Bonadei | 92ea95e | 2017-09-15 06:47:31 +0200 | [diff] [blame] | 11 | #include "modules/video_coding/codecs/test/videoprocessor.h" |
kjellander@webrtc.org | 35a1756 | 2011-10-06 06:44:54 +0000 | [diff] [blame] | 12 | |
ssilkin | 612f858 | 2017-09-28 09:23:17 -0700 | [diff] [blame] | 13 | #include <algorithm> |
kjellander@webrtc.org | 35a1756 | 2011-10-06 06:44:54 +0000 | [diff] [blame] | 14 | #include <limits> |
Erik Språng | 08127a9 | 2016-11-16 16:41:30 +0100 | [diff] [blame] | 15 | #include <utility> |
kjellander@webrtc.org | 35a1756 | 2011-10-06 06:44:54 +0000 | [diff] [blame] | 16 | |
Mirko Bonadei | 92ea95e | 2017-09-15 06:47:31 +0200 | [diff] [blame] | 17 | #include "api/video/i420_buffer.h" |
Mirko Bonadei | 7120742 | 2017-09-15 13:58:09 +0200 | [diff] [blame] | 18 | #include "common_types.h" // NOLINT(build/include) |
ssilkin | 612f858 | 2017-09-28 09:23:17 -0700 | [diff] [blame] | 19 | #include "common_video/h264/h264_common.h" |
Rasmus Brandt | 5f7a891 | 2018-02-28 17:17:15 +0100 | [diff] [blame] | 20 | #include "common_video/libyuv/include/webrtc_libyuv.h" |
Sergey Silkin | 3be2a55 | 2018-01-17 15:11:44 +0100 | [diff] [blame] | 21 | #include "modules/rtp_rtcp/include/rtp_rtcp_defines.h" |
Mirko Bonadei | 92ea95e | 2017-09-15 06:47:31 +0200 | [diff] [blame] | 22 | #include "modules/video_coding/codecs/vp8/simulcast_rate_allocator.h" |
| 23 | #include "modules/video_coding/include/video_codec_initializer.h" |
| 24 | #include "modules/video_coding/utility/default_video_bitrate_allocator.h" |
| 25 | #include "rtc_base/checks.h" |
Mirko Bonadei | 92ea95e | 2017-09-15 06:47:31 +0200 | [diff] [blame] | 26 | #include "rtc_base/timeutils.h" |
Mirko Bonadei | 92ea95e | 2017-09-15 06:47:31 +0200 | [diff] [blame] | 27 | #include "test/gtest.h" |
Sergey Silkin | 10d9d59 | 2018-02-01 13:25:17 +0100 | [diff] [blame] | 28 | #include "third_party/libyuv/include/libyuv/scale.h" |
kjellander@webrtc.org | 35a1756 | 2011-10-06 06:44:54 +0000 | [diff] [blame] | 29 | |
| 30 | namespace webrtc { |
| 31 | namespace test { |
| 32 | |
brandtr | b78bc75 | 2017-02-22 01:26:59 -0800 | [diff] [blame] | 33 | namespace { |
Åsa Persson | 91af24a | 2018-01-24 17:20:18 +0100 | [diff] [blame] | 34 | const int kMsToRtpTimestamp = kVideoPayloadTypeFrequency / 1000; |
brandtr | 17b958c | 2017-03-07 01:41:43 -0800 | [diff] [blame] | 35 | |
brandtr | aebc61e | 2017-02-28 07:13:47 -0800 | [diff] [blame] | 36 | std::unique_ptr<VideoBitrateAllocator> CreateBitrateAllocator( |
brandtr | 07734a5 | 2017-08-08 08:35:53 -0700 | [diff] [blame] | 37 | TestConfig* config) { |
brandtr | aebc61e | 2017-02-28 07:13:47 -0800 | [diff] [blame] | 38 | std::unique_ptr<TemporalLayersFactory> tl_factory; |
brandtr | 07734a5 | 2017-08-08 08:35:53 -0700 | [diff] [blame] | 39 | if (config->codec_settings.codecType == VideoCodecType::kVideoCodecVP8) { |
brandtr | aebc61e | 2017-02-28 07:13:47 -0800 | [diff] [blame] | 40 | tl_factory.reset(new TemporalLayersFactory()); |
brandtr | 07734a5 | 2017-08-08 08:35:53 -0700 | [diff] [blame] | 41 | config->codec_settings.VP8()->tl_factory = tl_factory.get(); |
brandtr | aebc61e | 2017-02-28 07:13:47 -0800 | [diff] [blame] | 42 | } |
| 43 | return std::unique_ptr<VideoBitrateAllocator>( |
brandtr | 07734a5 | 2017-08-08 08:35:53 -0700 | [diff] [blame] | 44 | VideoCodecInitializer::CreateBitrateAllocator(config->codec_settings, |
brandtr | aebc61e | 2017-02-28 07:13:47 -0800 | [diff] [blame] | 45 | std::move(tl_factory))); |
| 46 | } |
| 47 | |
Sergey Silkin | 3be2a55 | 2018-01-17 15:11:44 +0100 | [diff] [blame] | 48 | size_t GetMaxNaluSizeBytes(const EncodedImage& encoded_frame, |
| 49 | const TestConfig& config) { |
ssilkin | 612f858 | 2017-09-28 09:23:17 -0700 | [diff] [blame] | 50 | if (config.codec_settings.codecType != kVideoCodecH264) |
Sergey Silkin | 3be2a55 | 2018-01-17 15:11:44 +0100 | [diff] [blame] | 51 | return 0; |
ssilkin | 612f858 | 2017-09-28 09:23:17 -0700 | [diff] [blame] | 52 | |
| 53 | std::vector<webrtc::H264::NaluIndex> nalu_indices = |
| 54 | webrtc::H264::FindNaluIndices(encoded_frame._buffer, |
| 55 | encoded_frame._length); |
| 56 | |
| 57 | RTC_CHECK(!nalu_indices.empty()); |
| 58 | |
Sergey Silkin | 3be2a55 | 2018-01-17 15:11:44 +0100 | [diff] [blame] | 59 | size_t max_size = 0; |
ssilkin | 612f858 | 2017-09-28 09:23:17 -0700 | [diff] [blame] | 60 | for (const webrtc::H264::NaluIndex& index : nalu_indices) |
Sergey Silkin | 3be2a55 | 2018-01-17 15:11:44 +0100 | [diff] [blame] | 61 | max_size = std::max(max_size, index.payload_size); |
ssilkin | 612f858 | 2017-09-28 09:23:17 -0700 | [diff] [blame] | 62 | |
Sergey Silkin | 3be2a55 | 2018-01-17 15:11:44 +0100 | [diff] [blame] | 63 | return max_size; |
ssilkin | 612f858 | 2017-09-28 09:23:17 -0700 | [diff] [blame] | 64 | } |
| 65 | |
asapersson | ae9ba04 | 2017-03-07 00:25:38 -0800 | [diff] [blame] | 66 | int GetElapsedTimeMicroseconds(int64_t start_ns, int64_t stop_ns) { |
| 67 | int64_t diff_us = (stop_ns - start_ns) / rtc::kNumNanosecsPerMicrosec; |
| 68 | RTC_DCHECK_GE(diff_us, std::numeric_limits<int>::min()); |
| 69 | RTC_DCHECK_LE(diff_us, std::numeric_limits<int>::max()); |
| 70 | return static_cast<int>(diff_us); |
| 71 | } |
| 72 | |
Åsa Persson | f0c4467 | 2017-10-24 16:03:39 +0200 | [diff] [blame] | 73 | void ExtractBufferWithSize(const VideoFrame& image, |
| 74 | int width, |
| 75 | int height, |
| 76 | rtc::Buffer* buffer) { |
| 77 | if (image.width() != width || image.height() != height) { |
| 78 | EXPECT_DOUBLE_EQ(static_cast<double>(width) / height, |
| 79 | static_cast<double>(image.width()) / image.height()); |
| 80 | // Same aspect ratio, no cropping needed. |
| 81 | rtc::scoped_refptr<I420Buffer> scaled(I420Buffer::Create(width, height)); |
| 82 | scaled->ScaleFrom(*image.video_frame_buffer()->ToI420()); |
| 83 | |
| 84 | size_t length = |
| 85 | CalcBufferSize(VideoType::kI420, scaled->width(), scaled->height()); |
| 86 | buffer->SetSize(length); |
| 87 | RTC_CHECK_NE(ExtractBuffer(scaled, length, buffer->data()), -1); |
| 88 | return; |
| 89 | } |
| 90 | |
| 91 | // No resize. |
| 92 | size_t length = |
| 93 | CalcBufferSize(VideoType::kI420, image.width(), image.height()); |
| 94 | buffer->SetSize(length); |
| 95 | RTC_CHECK_NE(ExtractBuffer(image, length, buffer->data()), -1); |
| 96 | } |
| 97 | |
brandtr | b78bc75 | 2017-02-22 01:26:59 -0800 | [diff] [blame] | 98 | } // namespace |
| 99 | |
brandtr | c409552 | 2017-08-07 08:12:33 -0700 | [diff] [blame] | 100 | VideoProcessor::VideoProcessor(webrtc::VideoEncoder* encoder, |
Sergey Silkin | 10d9d59 | 2018-02-01 13:25:17 +0100 | [diff] [blame] | 101 | VideoDecoderList* decoders, |
| 102 | FrameReader* input_frame_reader, |
brandtr | c409552 | 2017-08-07 08:12:33 -0700 | [diff] [blame] | 103 | const TestConfig& config, |
Sergey Silkin | 06a8f30 | 2018-02-20 09:48:26 +0100 | [diff] [blame] | 104 | Stats* stats, |
Sergey Silkin | 10d9d59 | 2018-02-01 13:25:17 +0100 | [diff] [blame] | 105 | IvfFileWriterList* encoded_frame_writers, |
| 106 | FrameWriterList* decoded_frame_writers) |
Åsa Persson | f0c4467 | 2017-10-24 16:03:39 +0200 | [diff] [blame] | 107 | : config_(config), |
Sergey Silkin | 10d9d59 | 2018-02-01 13:25:17 +0100 | [diff] [blame] | 108 | num_simulcast_or_spatial_layers_( |
| 109 | std::max(config_.NumberOfSimulcastStreams(), |
| 110 | config_.NumberOfSpatialLayers())), |
brandtr | 07734a5 | 2017-08-08 08:35:53 -0700 | [diff] [blame] | 111 | encoder_(encoder), |
Sergey Silkin | 10d9d59 | 2018-02-01 13:25:17 +0100 | [diff] [blame] | 112 | decoders_(decoders), |
brandtr | 07734a5 | 2017-08-08 08:35:53 -0700 | [diff] [blame] | 113 | bitrate_allocator_(CreateBitrateAllocator(&config_)), |
brandtr | bdd555c | 2017-08-21 01:34:04 -0700 | [diff] [blame] | 114 | encode_callback_(this), |
| 115 | decode_callback_(this), |
Sergey Silkin | 10d9d59 | 2018-02-01 13:25:17 +0100 | [diff] [blame] | 116 | input_frame_reader_(input_frame_reader), |
| 117 | encoded_frame_writers_(encoded_frame_writers), |
| 118 | decoded_frame_writers_(decoded_frame_writers), |
Rasmus Brandt | 5f7a891 | 2018-02-28 17:17:15 +0100 | [diff] [blame] | 119 | first_encoded_frame(true), |
Sergey Silkin | 3be2a55 | 2018-01-17 15:11:44 +0100 | [diff] [blame] | 120 | last_inputed_frame_num_(0), |
| 121 | last_encoded_frame_num_(0), |
Sergey Silkin | 10d9d59 | 2018-02-01 13:25:17 +0100 | [diff] [blame] | 122 | last_encoded_simulcast_svc_idx_(0), |
Sergey Silkin | 3be2a55 | 2018-01-17 15:11:44 +0100 | [diff] [blame] | 123 | last_decoded_frame_num_(0), |
Sergey Silkin | 3be2a55 | 2018-01-17 15:11:44 +0100 | [diff] [blame] | 124 | stats_(stats) { |
Rasmus Brandt | 4b381af | 2018-02-07 13:56:16 +0100 | [diff] [blame] | 125 | RTC_CHECK(rtc::TaskQueue::Current()) |
| 126 | << "VideoProcessor must be run on a task queue."; |
Sergey Silkin | 10d9d59 | 2018-02-01 13:25:17 +0100 | [diff] [blame] | 127 | RTC_CHECK(encoder); |
| 128 | RTC_CHECK(decoders && decoders->size() == num_simulcast_or_spatial_layers_); |
| 129 | RTC_CHECK(input_frame_reader); |
| 130 | RTC_CHECK(stats); |
| 131 | RTC_CHECK(!encoded_frame_writers || |
| 132 | encoded_frame_writers->size() == num_simulcast_or_spatial_layers_); |
| 133 | RTC_CHECK(!decoded_frame_writers || |
| 134 | decoded_frame_writers->size() == num_simulcast_or_spatial_layers_); |
brandtr | 17b958c | 2017-03-07 01:41:43 -0800 | [diff] [blame] | 135 | |
Sergey Silkin | 10d9d59 | 2018-02-01 13:25:17 +0100 | [diff] [blame] | 136 | // Setup required callbacks for the encoder and decoder and initialize them. |
brandtr | bdd555c | 2017-08-21 01:34:04 -0700 | [diff] [blame] | 137 | RTC_CHECK_EQ(encoder_->RegisterEncodeCompleteCallback(&encode_callback_), |
Åsa Persson | f0c4467 | 2017-10-24 16:03:39 +0200 | [diff] [blame] | 138 | WEBRTC_VIDEO_CODEC_OK); |
asapersson | 654d54c | 2017-02-10 00:16:07 -0800 | [diff] [blame] | 139 | |
Sergey Silkin | 1723cf9 | 2018-01-22 15:49:55 +0100 | [diff] [blame] | 140 | RTC_CHECK_EQ(encoder_->InitEncode(&config_.codec_settings, |
| 141 | static_cast<int>(config_.NumberOfCores()), |
| 142 | config_.max_payload_size_bytes), |
| 143 | WEBRTC_VIDEO_CODEC_OK); |
Sergey Silkin | 10d9d59 | 2018-02-01 13:25:17 +0100 | [diff] [blame] | 144 | |
| 145 | for (auto& decoder : *decoders_) { |
| 146 | RTC_CHECK_EQ(decoder->InitDecode(&config_.codec_settings, |
| 147 | static_cast<int>(config_.NumberOfCores())), |
| 148 | WEBRTC_VIDEO_CODEC_OK); |
| 149 | RTC_CHECK_EQ(decoder->RegisterDecodeCompleteCallback(&decode_callback_), |
| 150 | WEBRTC_VIDEO_CODEC_OK); |
| 151 | } |
kjellander@webrtc.org | 35a1756 | 2011-10-06 06:44:54 +0000 | [diff] [blame] | 152 | } |
| 153 | |
Åsa Persson | f0c4467 | 2017-10-24 16:03:39 +0200 | [diff] [blame] | 154 | VideoProcessor::~VideoProcessor() { |
brandtr | c8c5905 | 2017-08-21 06:44:16 -0700 | [diff] [blame] | 155 | RTC_DCHECK_CALLED_SEQUENTIALLY(&sequence_checker_); |
| 156 | |
brandtr | 77920a4 | 2017-08-11 07:48:15 -0700 | [diff] [blame] | 157 | RTC_CHECK_EQ(encoder_->Release(), WEBRTC_VIDEO_CODEC_OK); |
brandtr | bdd555c | 2017-08-21 01:34:04 -0700 | [diff] [blame] | 158 | encoder_->RegisterEncodeCompleteCallback(nullptr); |
Sergey Silkin | 10d9d59 | 2018-02-01 13:25:17 +0100 | [diff] [blame] | 159 | |
| 160 | for (auto& decoder : *decoders_) { |
| 161 | RTC_CHECK_EQ(decoder->Release(), WEBRTC_VIDEO_CODEC_OK); |
| 162 | decoder->RegisterDecodeCompleteCallback(nullptr); |
| 163 | } |
| 164 | |
| 165 | RTC_CHECK(last_encoded_frames_.empty()); |
brandtr | 77920a4 | 2017-08-11 07:48:15 -0700 | [diff] [blame] | 166 | } |
| 167 | |
brandtr | 8935d97 | 2017-09-06 01:53:22 -0700 | [diff] [blame] | 168 | void VideoProcessor::ProcessFrame() { |
brandtr | c8c5905 | 2017-08-21 06:44:16 -0700 | [diff] [blame] | 169 | RTC_DCHECK_CALLED_SEQUENTIALLY(&sequence_checker_); |
Sergey Silkin | 3be2a55 | 2018-01-17 15:11:44 +0100 | [diff] [blame] | 170 | const size_t frame_number = last_inputed_frame_num_++; |
asapersson | 654d54c | 2017-02-10 00:16:07 -0800 | [diff] [blame] | 171 | |
brandtr | bdd555c | 2017-08-21 01:34:04 -0700 | [diff] [blame] | 172 | // Get frame from file. |
magjed | 3f07549 | 2017-06-01 10:02:26 -0700 | [diff] [blame] | 173 | rtc::scoped_refptr<I420BufferInterface> buffer( |
Sergey Silkin | 10d9d59 | 2018-02-01 13:25:17 +0100 | [diff] [blame] | 174 | input_frame_reader_->ReadFrame()); |
brandtr | bdd555c | 2017-08-21 01:34:04 -0700 | [diff] [blame] | 175 | RTC_CHECK(buffer) << "Tried to read too many frames from the file."; |
Åsa Persson | 91af24a | 2018-01-24 17:20:18 +0100 | [diff] [blame] | 176 | |
| 177 | size_t rtp_timestamp = |
| 178 | (frame_number > 0) ? input_frames_[frame_number - 1]->timestamp() : 0; |
| 179 | rtp_timestamp += |
| 180 | kVideoPayloadTypeFrequency / config_.codec_settings.maxFramerate; |
| 181 | |
| 182 | input_frames_[frame_number] = rtc::MakeUnique<VideoFrame>( |
| 183 | buffer, static_cast<uint32_t>(rtp_timestamp), |
| 184 | static_cast<int64_t>(rtp_timestamp / kMsToRtpTimestamp), |
| 185 | webrtc::kVideoRotation_0); |
brandtr | 17b958c | 2017-03-07 01:41:43 -0800 | [diff] [blame] | 186 | |
Sergey Silkin | 64eaa99 | 2017-11-17 14:47:32 +0100 | [diff] [blame] | 187 | std::vector<FrameType> frame_types = config_.FrameTypeForFrame(frame_number); |
brandtr | 17b958c | 2017-03-07 01:41:43 -0800 | [diff] [blame] | 188 | |
Rasmus Brandt | 5f7a891 | 2018-02-28 17:17:15 +0100 | [diff] [blame] | 189 | // Create frame statistics object for all simulcast/spatial layers. |
Sergey Silkin | 10d9d59 | 2018-02-01 13:25:17 +0100 | [diff] [blame] | 190 | for (size_t simulcast_svc_idx = 0; |
| 191 | simulcast_svc_idx < num_simulcast_or_spatial_layers_; |
| 192 | ++simulcast_svc_idx) { |
Sergey Silkin | 06a8f30 | 2018-02-20 09:48:26 +0100 | [diff] [blame] | 193 | stats_->AddFrame(rtp_timestamp, simulcast_svc_idx); |
Sergey Silkin | 10d9d59 | 2018-02-01 13:25:17 +0100 | [diff] [blame] | 194 | } |
brandtr | 17b958c | 2017-03-07 01:41:43 -0800 | [diff] [blame] | 195 | |
| 196 | // For the highest measurement accuracy of the encode time, the start/stop |
| 197 | // time recordings should wrap the Encode call as tightly as possible. |
Sergey Silkin | 10d9d59 | 2018-02-01 13:25:17 +0100 | [diff] [blame] | 198 | const int64_t encode_start_ns = rtc::TimeNanos(); |
| 199 | for (size_t simulcast_svc_idx = 0; |
| 200 | simulcast_svc_idx < num_simulcast_or_spatial_layers_; |
| 201 | ++simulcast_svc_idx) { |
Sergey Silkin | 06a8f30 | 2018-02-20 09:48:26 +0100 | [diff] [blame] | 202 | FrameStatistics* frame_stat = |
| 203 | stats_->GetFrame(frame_number, simulcast_svc_idx); |
Sergey Silkin | 10d9d59 | 2018-02-01 13:25:17 +0100 | [diff] [blame] | 204 | frame_stat->encode_start_ns = encode_start_ns; |
| 205 | } |
| 206 | |
| 207 | const int encode_return_code = |
Sergey Silkin | 64eaa99 | 2017-11-17 14:47:32 +0100 | [diff] [blame] | 208 | encoder_->Encode(*input_frames_[frame_number], nullptr, &frame_types); |
Sergey Silkin | 10d9d59 | 2018-02-01 13:25:17 +0100 | [diff] [blame] | 209 | |
| 210 | for (size_t simulcast_svc_idx = 0; |
| 211 | simulcast_svc_idx < num_simulcast_or_spatial_layers_; |
| 212 | ++simulcast_svc_idx) { |
Sergey Silkin | 06a8f30 | 2018-02-20 09:48:26 +0100 | [diff] [blame] | 213 | FrameStatistics* frame_stat = |
| 214 | stats_->GetFrame(frame_number, simulcast_svc_idx); |
Sergey Silkin | 10d9d59 | 2018-02-01 13:25:17 +0100 | [diff] [blame] | 215 | frame_stat->encode_return_code = encode_return_code; |
| 216 | } |
| 217 | |
| 218 | // For async codecs frame decoding is done in frame encode callback. |
| 219 | if (!config_.IsAsyncCodec()) { |
| 220 | for (size_t simulcast_svc_idx = 0; |
| 221 | simulcast_svc_idx < num_simulcast_or_spatial_layers_; |
| 222 | ++simulcast_svc_idx) { |
| 223 | if (last_encoded_frames_.find(simulcast_svc_idx) != |
| 224 | last_encoded_frames_.end()) { |
| 225 | EncodedImage& encoded_image = last_encoded_frames_[simulcast_svc_idx]; |
| 226 | |
Sergey Silkin | 06a8f30 | 2018-02-20 09:48:26 +0100 | [diff] [blame] | 227 | FrameStatistics* frame_stat = |
| 228 | stats_->GetFrame(frame_number, simulcast_svc_idx); |
Sergey Silkin | 10d9d59 | 2018-02-01 13:25:17 +0100 | [diff] [blame] | 229 | |
| 230 | if (encoded_frame_writers_) { |
| 231 | RTC_CHECK(encoded_frame_writers_->at(simulcast_svc_idx) |
| 232 | ->WriteFrame(encoded_image, |
| 233 | config_.codec_settings.codecType)); |
| 234 | } |
| 235 | |
| 236 | // For the highest measurement accuracy of the decode time, the |
| 237 | // start/stop time recordings should wrap the Decode call as tightly as |
| 238 | // possible. |
| 239 | frame_stat->decode_start_ns = rtc::TimeNanos(); |
| 240 | frame_stat->decode_return_code = |
| 241 | decoders_->at(simulcast_svc_idx) |
| 242 | ->Decode(encoded_image, false, nullptr); |
| 243 | |
| 244 | RTC_CHECK(encoded_image._buffer); |
| 245 | delete[] encoded_image._buffer; |
| 246 | encoded_image._buffer = nullptr; |
| 247 | |
| 248 | last_encoded_frames_.erase(simulcast_svc_idx); |
| 249 | } |
| 250 | } |
| 251 | } |
kjellander@webrtc.org | 35a1756 | 2011-10-06 06:44:54 +0000 | [diff] [blame] | 252 | } |
| 253 | |
Sergey Silkin | 3be2a55 | 2018-01-17 15:11:44 +0100 | [diff] [blame] | 254 | void VideoProcessor::SetRates(size_t bitrate_kbps, size_t framerate_fps) { |
brandtr | c8c5905 | 2017-08-21 06:44:16 -0700 | [diff] [blame] | 255 | RTC_DCHECK_CALLED_SEQUENTIALLY(&sequence_checker_); |
Sergey Silkin | 3be2a55 | 2018-01-17 15:11:44 +0100 | [diff] [blame] | 256 | config_.codec_settings.maxFramerate = static_cast<uint32_t>(framerate_fps); |
| 257 | bitrate_allocation_ = bitrate_allocator_->GetAllocation( |
| 258 | static_cast<uint32_t>(bitrate_kbps * 1000), |
| 259 | static_cast<uint32_t>(framerate_fps)); |
| 260 | const int set_rates_result = encoder_->SetRateAllocation( |
| 261 | bitrate_allocation_, static_cast<uint32_t>(framerate_fps)); |
brandtr | bea36fd | 2017-08-07 03:36:54 -0700 | [diff] [blame] | 262 | RTC_DCHECK_GE(set_rates_result, 0) |
brandtr | bdd555c | 2017-08-21 01:34:04 -0700 | [diff] [blame] | 263 | << "Failed to update encoder with new rate " << bitrate_kbps << "."; |
brandtr | bea36fd | 2017-08-07 03:36:54 -0700 | [diff] [blame] | 264 | } |
| 265 | |
Sergey Silkin | 10d9d59 | 2018-02-01 13:25:17 +0100 | [diff] [blame] | 266 | void VideoProcessor::FrameEncoded( |
| 267 | const webrtc::EncodedImage& encoded_image, |
| 268 | const webrtc::CodecSpecificInfo& codec_specific) { |
brandtr | c8c5905 | 2017-08-21 06:44:16 -0700 | [diff] [blame] | 269 | RTC_DCHECK_CALLED_SEQUENTIALLY(&sequence_checker_); |
| 270 | |
brandtr | 32e0d26 | 2017-02-15 05:29:38 -0800 | [diff] [blame] | 271 | // For the highest measurement accuracy of the encode time, the start/stop |
| 272 | // time recordings should wrap the Encode call as tightly as possible. |
Rasmus Brandt | 5f7a891 | 2018-02-28 17:17:15 +0100 | [diff] [blame] | 273 | const int64_t encode_stop_ns = rtc::TimeNanos(); |
brandtr | 32e0d26 | 2017-02-15 05:29:38 -0800 | [diff] [blame] | 274 | |
Sergey Silkin | 10d9d59 | 2018-02-01 13:25:17 +0100 | [diff] [blame] | 275 | const VideoCodecType codec = codec_specific.codecType; |
Rasmus Brandt | f7a3558 | 2017-10-24 10:16:33 +0200 | [diff] [blame] | 276 | if (config_.encoded_frame_checker) { |
| 277 | config_.encoded_frame_checker->CheckEncodedFrame(codec, encoded_image); |
| 278 | } |
brandtr | b78bc75 | 2017-02-22 01:26:59 -0800 | [diff] [blame] | 279 | |
Sergey Silkin | 10d9d59 | 2018-02-01 13:25:17 +0100 | [diff] [blame] | 280 | size_t simulcast_svc_idx = 0; |
| 281 | size_t temporal_idx = 0; |
Sergey Silkin | 64eaa99 | 2017-11-17 14:47:32 +0100 | [diff] [blame] | 282 | |
Sergey Silkin | 10d9d59 | 2018-02-01 13:25:17 +0100 | [diff] [blame] | 283 | if (codec == kVideoCodecVP8) { |
| 284 | simulcast_svc_idx = codec_specific.codecSpecific.VP8.simulcastIdx; |
| 285 | temporal_idx = codec_specific.codecSpecific.VP8.temporalIdx; |
| 286 | } else if (codec == kVideoCodecVP9) { |
| 287 | simulcast_svc_idx = codec_specific.codecSpecific.VP9.spatial_idx; |
| 288 | temporal_idx = codec_specific.codecSpecific.VP9.temporal_idx; |
| 289 | } |
| 290 | |
| 291 | if (simulcast_svc_idx == kNoSpatialIdx) { |
| 292 | simulcast_svc_idx = 0; |
| 293 | } |
| 294 | |
| 295 | if (temporal_idx == kNoTemporalIdx) { |
| 296 | temporal_idx = 0; |
| 297 | } |
| 298 | |
| 299 | const size_t frame_wxh = |
| 300 | encoded_image._encodedWidth * encoded_image._encodedHeight; |
| 301 | frame_wxh_to_simulcast_svc_idx_[frame_wxh] = simulcast_svc_idx; |
| 302 | |
Sergey Silkin | 06a8f30 | 2018-02-20 09:48:26 +0100 | [diff] [blame] | 303 | FrameStatistics* frame_stat = stats_->GetFrameWithTimestamp( |
| 304 | encoded_image._timeStamp, simulcast_svc_idx); |
Åsa Persson | a6e7b88 | 2018-01-19 14:57:10 +0100 | [diff] [blame] | 305 | const size_t frame_number = frame_stat->frame_number; |
Sergey Silkin | 10d9d59 | 2018-02-01 13:25:17 +0100 | [diff] [blame] | 306 | |
| 307 | // Reordering is unexpected. Frames of different layers have the same value |
| 308 | // of frame_number. VP8 multi-res delivers frames starting from hires layer. |
| 309 | RTC_CHECK_GE(frame_number, last_encoded_frame_num_); |
| 310 | |
| 311 | // Ensure SVC spatial layers are delivered in ascending order. |
| 312 | if (config_.NumberOfSpatialLayers() > 1) { |
Rasmus Brandt | 5f7a891 | 2018-02-28 17:17:15 +0100 | [diff] [blame] | 313 | RTC_CHECK(first_encoded_frame || frame_number >= last_encoded_frame_num_ || |
| 314 | simulcast_svc_idx > last_encoded_simulcast_svc_idx_); |
Sergey Silkin | 3be2a55 | 2018-01-17 15:11:44 +0100 | [diff] [blame] | 315 | } |
Rasmus Brandt | 5f7a891 | 2018-02-28 17:17:15 +0100 | [diff] [blame] | 316 | first_encoded_frame = false; |
brandtr | 17b958c | 2017-03-07 01:41:43 -0800 | [diff] [blame] | 317 | last_encoded_frame_num_ = frame_number; |
Sergey Silkin | 10d9d59 | 2018-02-01 13:25:17 +0100 | [diff] [blame] | 318 | last_encoded_simulcast_svc_idx_ = simulcast_svc_idx; |
brandtr | 17b958c | 2017-03-07 01:41:43 -0800 | [diff] [blame] | 319 | |
brandtr | 8935d97 | 2017-09-06 01:53:22 -0700 | [diff] [blame] | 320 | // Update frame statistics. |
Sergey Silkin | 10d9d59 | 2018-02-01 13:25:17 +0100 | [diff] [blame] | 321 | frame_stat->encoding_successful = true; |
brandtr | 8935d97 | 2017-09-06 01:53:22 -0700 | [diff] [blame] | 322 | frame_stat->encode_time_us = |
| 323 | GetElapsedTimeMicroseconds(frame_stat->encode_start_ns, encode_stop_ns); |
Sergey Silkin | 10d9d59 | 2018-02-01 13:25:17 +0100 | [diff] [blame] | 324 | |
Sergey Silkin | 06a8f30 | 2018-02-20 09:48:26 +0100 | [diff] [blame] | 325 | if (codec == kVideoCodecVP9) { |
| 326 | const CodecSpecificInfoVP9& vp9_info = codec_specific.codecSpecific.VP9; |
| 327 | frame_stat->inter_layer_predicted = vp9_info.inter_layer_predicted; |
| 328 | |
| 329 | // TODO(ssilkin): Implement bitrate allocation for VP9 SVC. For now set |
| 330 | // target for base layers equal to total target to avoid devision by zero |
| 331 | // at analysis. |
| 332 | frame_stat->target_bitrate_kbps = bitrate_allocation_.get_sum_kbps(); |
| 333 | } else { |
| 334 | frame_stat->target_bitrate_kbps = |
| 335 | (bitrate_allocation_.GetBitrate(simulcast_svc_idx, temporal_idx) + |
| 336 | 500) / |
| 337 | 1000; |
| 338 | } |
| 339 | |
brandtr | 8935d97 | 2017-09-06 01:53:22 -0700 | [diff] [blame] | 340 | frame_stat->encoded_frame_size_bytes = encoded_image._length; |
brandtr | 17b958c | 2017-03-07 01:41:43 -0800 | [diff] [blame] | 341 | frame_stat->frame_type = encoded_image._frameType; |
Sergey Silkin | 10d9d59 | 2018-02-01 13:25:17 +0100 | [diff] [blame] | 342 | frame_stat->temporal_layer_idx = temporal_idx; |
| 343 | frame_stat->simulcast_svc_idx = simulcast_svc_idx; |
Sergey Silkin | 3be2a55 | 2018-01-17 15:11:44 +0100 | [diff] [blame] | 344 | frame_stat->max_nalu_size_bytes = GetMaxNaluSizeBytes(encoded_image, config_); |
Sergey Silkin | 10d9d59 | 2018-02-01 13:25:17 +0100 | [diff] [blame] | 345 | frame_stat->qp = encoded_image.qp_; |
ssilkin | 612f858 | 2017-09-28 09:23:17 -0700 | [diff] [blame] | 346 | |
Sergey Silkin | 10d9d59 | 2018-02-01 13:25:17 +0100 | [diff] [blame] | 347 | if (!config_.IsAsyncCodec()) { |
| 348 | // Store encoded frame. It will be decoded after all layers are encoded. |
| 349 | CopyEncodedImage(encoded_image, codec, frame_number, simulcast_svc_idx); |
| 350 | } else { |
| 351 | const size_t simulcast_idx = |
| 352 | codec == kVideoCodecVP8 ? codec_specific.codecSpecific.VP8.simulcastIdx |
| 353 | : 0; |
| 354 | frame_stat->decode_start_ns = rtc::TimeNanos(); |
| 355 | frame_stat->decode_return_code = |
| 356 | decoders_->at(simulcast_idx)->Decode(encoded_image, false, nullptr); |
brandtr | 8935d97 | 2017-09-06 01:53:22 -0700 | [diff] [blame] | 357 | } |
kjellander@webrtc.org | 35a1756 | 2011-10-06 06:44:54 +0000 | [diff] [blame] | 358 | } |
| 359 | |
Sergey Silkin | 64eaa99 | 2017-11-17 14:47:32 +0100 | [diff] [blame] | 360 | void VideoProcessor::FrameDecoded(const VideoFrame& decoded_frame) { |
brandtr | c8c5905 | 2017-08-21 06:44:16 -0700 | [diff] [blame] | 361 | RTC_DCHECK_CALLED_SEQUENTIALLY(&sequence_checker_); |
| 362 | |
brandtr | 32e0d26 | 2017-02-15 05:29:38 -0800 | [diff] [blame] | 363 | // For the highest measurement accuracy of the decode time, the start/stop |
| 364 | // time recordings should wrap the Decode call as tightly as possible. |
Rasmus Brandt | 5f7a891 | 2018-02-28 17:17:15 +0100 | [diff] [blame] | 365 | const int64_t decode_stop_ns = rtc::TimeNanos(); |
brandtr | 8bc9385 | 2017-02-15 05:19:51 -0800 | [diff] [blame] | 366 | |
Sergey Silkin | 10d9d59 | 2018-02-01 13:25:17 +0100 | [diff] [blame] | 367 | RTC_CHECK(frame_wxh_to_simulcast_svc_idx_.find(decoded_frame.size()) != |
| 368 | frame_wxh_to_simulcast_svc_idx_.end()); |
| 369 | const size_t simulcast_svc_idx = |
| 370 | frame_wxh_to_simulcast_svc_idx_[decoded_frame.size()]; |
| 371 | |
Sergey Silkin | 06a8f30 | 2018-02-20 09:48:26 +0100 | [diff] [blame] | 372 | FrameStatistics* frame_stat = stats_->GetFrameWithTimestamp( |
| 373 | decoded_frame.timestamp(), simulcast_svc_idx); |
Åsa Persson | a6e7b88 | 2018-01-19 14:57:10 +0100 | [diff] [blame] | 374 | const size_t frame_number = frame_stat->frame_number; |
Sergey Silkin | 64eaa99 | 2017-11-17 14:47:32 +0100 | [diff] [blame] | 375 | |
Sergey Silkin | 10d9d59 | 2018-02-01 13:25:17 +0100 | [diff] [blame] | 376 | // Reordering is unexpected. Frames of different layers have the same value |
| 377 | // of frame_number. |
| 378 | RTC_CHECK_GE(frame_number, last_decoded_frame_num_); |
brandtr | 17b958c | 2017-03-07 01:41:43 -0800 | [diff] [blame] | 379 | last_decoded_frame_num_ = frame_number; |
| 380 | |
Sergey Silkin | 10d9d59 | 2018-02-01 13:25:17 +0100 | [diff] [blame] | 381 | // Update frame statistics. |
| 382 | frame_stat->decoding_successful = true; |
| 383 | frame_stat->decode_time_us = |
| 384 | GetElapsedTimeMicroseconds(frame_stat->decode_start_ns, decode_stop_ns); |
| 385 | frame_stat->decoded_width = decoded_frame.width(); |
| 386 | frame_stat->decoded_height = decoded_frame.height(); |
| 387 | |
Sergey Silkin | 64eaa99 | 2017-11-17 14:47:32 +0100 | [diff] [blame] | 388 | // Skip quality metrics calculation to not affect CPU usage. |
| 389 | if (!config_.measure_cpu) { |
Sergey Silkin | 10d9d59 | 2018-02-01 13:25:17 +0100 | [diff] [blame] | 390 | CalculateFrameQuality(*input_frames_[frame_number], decoded_frame, |
| 391 | frame_stat); |
Sergey Silkin | 64eaa99 | 2017-11-17 14:47:32 +0100 | [diff] [blame] | 392 | } |
Niels Möller | 718a763 | 2016-06-13 13:06:01 +0200 | [diff] [blame] | 393 | |
Sergey Silkin | 64eaa99 | 2017-11-17 14:47:32 +0100 | [diff] [blame] | 394 | // Delay erasing of input frames by one frame. The current frame might |
| 395 | // still be needed for other simulcast stream or spatial layer. |
Sergey Silkin | 3be2a55 | 2018-01-17 15:11:44 +0100 | [diff] [blame] | 396 | if (frame_number > 0) { |
| 397 | auto input_frame_erase_to = input_frames_.lower_bound(frame_number - 1); |
Sergey Silkin | 64eaa99 | 2017-11-17 14:47:32 +0100 | [diff] [blame] | 398 | input_frames_.erase(input_frames_.begin(), input_frame_erase_to); |
| 399 | } |
| 400 | |
Sergey Silkin | 10d9d59 | 2018-02-01 13:25:17 +0100 | [diff] [blame] | 401 | if (decoded_frame_writers_) { |
Sergey Silkin | 64eaa99 | 2017-11-17 14:47:32 +0100 | [diff] [blame] | 402 | ExtractBufferWithSize(decoded_frame, config_.codec_settings.width, |
| 403 | config_.codec_settings.height, |
Rasmus Brandt | 5f7a891 | 2018-02-28 17:17:15 +0100 | [diff] [blame] | 404 | &tmp_planar_i420_buffer_); |
| 405 | RTC_CHECK(simulcast_svc_idx < decoded_frame_writers_->size()); |
| 406 | RTC_CHECK_EQ(tmp_planar_i420_buffer_.size(), |
| 407 | decoded_frame_writers_->at(simulcast_svc_idx)->FrameLength()); |
| 408 | RTC_CHECK(decoded_frame_writers_->at(simulcast_svc_idx) |
| 409 | ->WriteFrame(tmp_planar_i420_buffer_.data())); |
Sergey Silkin | 64eaa99 | 2017-11-17 14:47:32 +0100 | [diff] [blame] | 410 | } |
Åsa Persson | f0c4467 | 2017-10-24 16:03:39 +0200 | [diff] [blame] | 411 | } |
brandtr | 17b958c | 2017-03-07 01:41:43 -0800 | [diff] [blame] | 412 | |
Sergey Silkin | 10d9d59 | 2018-02-01 13:25:17 +0100 | [diff] [blame] | 413 | void VideoProcessor::CopyEncodedImage(const EncodedImage& encoded_image, |
| 414 | const VideoCodecType codec, |
| 415 | size_t frame_number, |
| 416 | size_t simulcast_svc_idx) { |
| 417 | RTC_DCHECK_CALLED_SEQUENTIALLY(&sequence_checker_); |
| 418 | |
| 419 | EncodedImage base_image; |
| 420 | RTC_CHECK_EQ(base_image._length, 0); |
| 421 | |
| 422 | // Each SVC layer is decoded with dedicated decoder. Add data of base layers |
| 423 | // to current coded frame buffer. |
| 424 | if (config_.NumberOfSpatialLayers() > 1 && simulcast_svc_idx > 0) { |
| 425 | RTC_CHECK(last_encoded_frames_.find(simulcast_svc_idx - 1) != |
| 426 | last_encoded_frames_.end()); |
| 427 | base_image = last_encoded_frames_[simulcast_svc_idx - 1]; |
| 428 | } |
| 429 | |
| 430 | const size_t payload_size_bytes = base_image._length + encoded_image._length; |
| 431 | const size_t buffer_size_bytes = |
| 432 | payload_size_bytes + EncodedImage::GetBufferPaddingBytes(codec); |
| 433 | |
| 434 | uint8_t* copied_buffer = new uint8_t[buffer_size_bytes]; |
| 435 | RTC_CHECK(copied_buffer); |
| 436 | |
| 437 | if (base_image._length) { |
| 438 | memcpy(copied_buffer, base_image._buffer, base_image._length); |
| 439 | } |
| 440 | |
| 441 | memcpy(copied_buffer + base_image._length, encoded_image._buffer, |
| 442 | encoded_image._length); |
| 443 | |
| 444 | EncodedImage copied_image = encoded_image; |
| 445 | copied_image = encoded_image; |
| 446 | copied_image._buffer = copied_buffer; |
| 447 | copied_image._length = payload_size_bytes; |
| 448 | copied_image._size = buffer_size_bytes; |
| 449 | |
| 450 | last_encoded_frames_[simulcast_svc_idx] = copied_image; |
| 451 | } |
| 452 | |
| 453 | void VideoProcessor::CalculateFrameQuality(const VideoFrame& ref_frame, |
| 454 | const VideoFrame& dec_frame, |
Sergey Silkin | 06a8f30 | 2018-02-20 09:48:26 +0100 | [diff] [blame] | 455 | FrameStatistics* frame_stat) { |
Sergey Silkin | 10d9d59 | 2018-02-01 13:25:17 +0100 | [diff] [blame] | 456 | if (ref_frame.width() == dec_frame.width() || |
| 457 | ref_frame.height() == dec_frame.height()) { |
| 458 | frame_stat->psnr = I420PSNR(&ref_frame, &dec_frame); |
| 459 | frame_stat->ssim = I420SSIM(&ref_frame, &dec_frame); |
| 460 | } else { |
| 461 | RTC_CHECK_GE(ref_frame.width(), dec_frame.width()); |
| 462 | RTC_CHECK_GE(ref_frame.height(), dec_frame.height()); |
| 463 | // Downscale reference frame. Use bilinear interpolation since it is used |
| 464 | // to get lowres inputs for encoder at simulcasting. |
| 465 | // TODO(ssilkin): Sync with VP9 SVC which uses 8-taps polyphase. |
| 466 | rtc::scoped_refptr<I420Buffer> scaled_buffer = |
| 467 | I420Buffer::Create(dec_frame.width(), dec_frame.height()); |
| 468 | const I420BufferInterface& ref_buffer = |
| 469 | *ref_frame.video_frame_buffer()->ToI420(); |
| 470 | I420Scale(ref_buffer.DataY(), ref_buffer.StrideY(), ref_buffer.DataU(), |
| 471 | ref_buffer.StrideU(), ref_buffer.DataV(), ref_buffer.StrideV(), |
| 472 | ref_buffer.width(), ref_buffer.height(), |
| 473 | scaled_buffer->MutableDataY(), scaled_buffer->StrideY(), |
| 474 | scaled_buffer->MutableDataU(), scaled_buffer->StrideU(), |
| 475 | scaled_buffer->MutableDataV(), scaled_buffer->StrideV(), |
| 476 | scaled_buffer->width(), scaled_buffer->height(), |
Sergey Silkin | 06a8f30 | 2018-02-20 09:48:26 +0100 | [diff] [blame] | 477 | libyuv::kFilterBox); |
Sergey Silkin | 10d9d59 | 2018-02-01 13:25:17 +0100 | [diff] [blame] | 478 | frame_stat->psnr = |
| 479 | I420PSNR(*scaled_buffer, *dec_frame.video_frame_buffer()->ToI420()); |
| 480 | frame_stat->ssim = |
| 481 | I420SSIM(*scaled_buffer, *dec_frame.video_frame_buffer()->ToI420()); |
| 482 | } |
| 483 | } |
| 484 | |
kjellander@webrtc.org | 35a1756 | 2011-10-06 06:44:54 +0000 | [diff] [blame] | 485 | } // namespace test |
| 486 | } // namespace webrtc |