blob: aec29a457010362ee0579c54352c01be53047c8a [file] [log] [blame]
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +00001/*
2 * Copyright (c) 2011 The WebRTC project authors. All Rights Reserved.
3 *
4 * Use of this source code is governed by a BSD-style license
5 * that can be found in the LICENSE file in the root of the source
6 * tree. An additional intellectual property rights grant can be found
7 * in the file PATENTS. All contributing project authors may
8 * be found in the AUTHORS file in the root of the source tree.
9 */
10
Mirko Bonadei92ea95e2017-09-15 06:47:31 +020011#include "modules/audio_coding/neteq/include/neteq.h"
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +000012
pbos@webrtc.org3ecc1622014-03-07 15:23:34 +000013#include <math.h>
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +000014#include <stdlib.h>
15#include <string.h> // memset
16
turaj@webrtc.org8d1cdaa2014-04-11 18:47:55 +000017#include <algorithm>
kwiberg2d0c3322016-02-14 09:28:33 -080018#include <memory>
turaj@webrtc.org78b41a02013-11-22 20:27:07 +000019#include <set>
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +000020#include <string>
21#include <vector>
22
Fredrik Solenbergbbf21a32018-04-12 22:44:09 +020023#include "api/audio/audio_frame.h"
Mirko Bonadei92ea95e2017-09-15 06:47:31 +020024#include "api/audio_codecs/builtin_audio_decoder_factory.h"
Mirko Bonadei71207422017-09-15 13:58:09 +020025#include "common_types.h" // NOLINT(build/include)
Mirko Bonadei92ea95e2017-09-15 06:47:31 +020026#include "modules/audio_coding/codecs/pcm16b/pcm16b.h"
27#include "modules/audio_coding/neteq/tools/audio_loop.h"
Henrik Lundin7687ad52018-07-02 10:14:46 +020028#include "modules/audio_coding/neteq/tools/neteq_packet_source_input.h"
29#include "modules/audio_coding/neteq/tools/neteq_test.h"
Mirko Bonadei92ea95e2017-09-15 06:47:31 +020030#include "modules/audio_coding/neteq/tools/rtp_file_source.h"
Mirko Bonadei92ea95e2017-09-15 06:47:31 +020031#include "rtc_base/ignore_wundef.h"
Joachim Bauch4e909192017-12-19 22:27:51 +010032#include "rtc_base/messagedigest.h"
Karl Wiberge40468b2017-11-22 10:42:26 +010033#include "rtc_base/numerics/safe_conversions.h"
Mirko Bonadei92ea95e2017-09-15 06:47:31 +020034#include "rtc_base/protobuf_utils.h"
Mirko Bonadei92ea95e2017-09-15 06:47:31 +020035#include "rtc_base/stringencode.h"
Jonas Olsson366a50c2018-09-06 13:41:30 +020036#include "rtc_base/strings/string_builder.h"
Niels Möllera12c42a2018-07-25 16:05:48 +020037#include "rtc_base/system/arch.h"
Henrik Lundine9619f82017-11-27 14:05:27 +010038#include "test/field_trial.h"
Mirko Bonadei92ea95e2017-09-15 06:47:31 +020039#include "test/gtest.h"
40#include "test/testsupport/fileutils.h"
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +000041
Mirko Bonadei81ca3bf2018-01-09 09:40:39 +010042// This must come after test/gtest.h
43#include "rtc_base/flags.h" // NOLINT(build/include)
44
minyue5f026d02015-12-16 07:36:04 -080045#ifdef WEBRTC_NETEQ_UNITTEST_BITEXACT
kwiberg77eab702016-09-28 17:42:01 -070046RTC_PUSH_IGNORING_WUNDEF()
minyue5f026d02015-12-16 07:36:04 -080047#ifdef WEBRTC_ANDROID_PLATFORM_BUILD
48#include "external/webrtc/webrtc/modules/audio_coding/neteq/neteq_unittest.pb.h"
49#else
Mirko Bonadei92ea95e2017-09-15 06:47:31 +020050#include "modules/audio_coding/neteq/neteq_unittest.pb.h"
minyue5f026d02015-12-16 07:36:04 -080051#endif
kwiberg77eab702016-09-28 17:42:01 -070052RTC_POP_IGNORING_WUNDEF()
minyue5f026d02015-12-16 07:36:04 -080053#endif
54
Mirko Bonadei2dfa9982018-10-18 11:35:32 +020055WEBRTC_DEFINE_bool(gen_ref, false, "Generate reference files.");
turaj@webrtc.orga6101d72013-10-01 22:01:09 +000056
kwiberg5adaf732016-10-04 09:33:27 -070057namespace webrtc {
58
minyue5f026d02015-12-16 07:36:04 -080059namespace {
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +000060
minyue4f906772016-04-29 11:05:14 -070061const std::string& PlatformChecksum(const std::string& checksum_general,
Henrik Lundin8cd750d2017-10-12 13:07:11 +020062 const std::string& checksum_android_32,
63 const std::string& checksum_android_64,
minyue4f906772016-04-29 11:05:14 -070064 const std::string& checksum_win_32,
65 const std::string& checksum_win_64) {
kwiberg77eab702016-09-28 17:42:01 -070066#if defined(WEBRTC_ANDROID)
Yves Gerey665174f2018-06-19 15:03:05 +020067#ifdef WEBRTC_ARCH_64_BITS
68 return checksum_android_64;
69#else
70 return checksum_android_32;
71#endif // WEBRTC_ARCH_64_BITS
kwiberg77eab702016-09-28 17:42:01 -070072#elif defined(WEBRTC_WIN)
Yves Gerey665174f2018-06-19 15:03:05 +020073#ifdef WEBRTC_ARCH_64_BITS
74 return checksum_win_64;
75#else
76 return checksum_win_32;
77#endif // WEBRTC_ARCH_64_BITS
minyue4f906772016-04-29 11:05:14 -070078#else
79 return checksum_general;
80#endif // WEBRTC_WIN
81}
82
minyue5f026d02015-12-16 07:36:04 -080083#ifdef WEBRTC_NETEQ_UNITTEST_BITEXACT
84void Convert(const webrtc::NetEqNetworkStatistics& stats_raw,
85 webrtc::neteq_unittest::NetEqNetworkStatistics* stats) {
86 stats->set_current_buffer_size_ms(stats_raw.current_buffer_size_ms);
87 stats->set_preferred_buffer_size_ms(stats_raw.preferred_buffer_size_ms);
88 stats->set_jitter_peaks_found(stats_raw.jitter_peaks_found);
89 stats->set_packet_loss_rate(stats_raw.packet_loss_rate);
minyue5f026d02015-12-16 07:36:04 -080090 stats->set_expand_rate(stats_raw.expand_rate);
91 stats->set_speech_expand_rate(stats_raw.speech_expand_rate);
92 stats->set_preemptive_rate(stats_raw.preemptive_rate);
93 stats->set_accelerate_rate(stats_raw.accelerate_rate);
94 stats->set_secondary_decoded_rate(stats_raw.secondary_decoded_rate);
minyue-webrtc0c3ca752017-08-23 15:59:38 +020095 stats->set_secondary_discarded_rate(stats_raw.secondary_discarded_rate);
minyue5f026d02015-12-16 07:36:04 -080096 stats->set_clockdrift_ppm(stats_raw.clockdrift_ppm);
97 stats->set_added_zero_samples(stats_raw.added_zero_samples);
98 stats->set_mean_waiting_time_ms(stats_raw.mean_waiting_time_ms);
99 stats->set_median_waiting_time_ms(stats_raw.median_waiting_time_ms);
100 stats->set_min_waiting_time_ms(stats_raw.min_waiting_time_ms);
101 stats->set_max_waiting_time_ms(stats_raw.max_waiting_time_ms);
102}
103
104void Convert(const webrtc::RtcpStatistics& stats_raw,
105 webrtc::neteq_unittest::RtcpStatistics* stats) {
106 stats->set_fraction_lost(stats_raw.fraction_lost);
srte186d9c32017-08-04 05:03:53 -0700107 stats->set_cumulative_lost(stats_raw.packets_lost);
minyue5f026d02015-12-16 07:36:04 -0800108 stats->set_extended_max_sequence_number(
srte186d9c32017-08-04 05:03:53 -0700109 stats_raw.extended_highest_sequence_number);
minyue5f026d02015-12-16 07:36:04 -0800110 stats->set_jitter(stats_raw.jitter);
111}
112
Yves Gerey665174f2018-06-19 15:03:05 +0200113void AddMessage(FILE* file,
114 rtc::MessageDigest* digest,
minyue4f906772016-04-29 11:05:14 -0700115 const std::string& message) {
minyue5f026d02015-12-16 07:36:04 -0800116 int32_t size = message.length();
minyue4f906772016-04-29 11:05:14 -0700117 if (file)
118 ASSERT_EQ(1u, fwrite(&size, sizeof(size), 1, file));
119 digest->Update(&size, sizeof(size));
120
121 if (file)
122 ASSERT_EQ(static_cast<size_t>(size),
123 fwrite(message.data(), sizeof(char), size, file));
124 digest->Update(message.data(), sizeof(char) * size);
minyue5f026d02015-12-16 07:36:04 -0800125}
126
minyue5f026d02015-12-16 07:36:04 -0800127#endif // WEBRTC_NETEQ_UNITTEST_BITEXACT
128
henrik.lundin7a926812016-05-12 13:51:28 -0700129void LoadDecoders(webrtc::NetEq* neteq) {
kwiberg5adaf732016-10-04 09:33:27 -0700130 ASSERT_EQ(true,
131 neteq->RegisterPayloadType(0, SdpAudioFormat("pcmu", 8000, 1)));
132 // Use non-SdpAudioFormat argument when registering PCMa, so that we get test
133 // coverage for that as well.
henrik.lundin7a926812016-05-12 13:51:28 -0700134 ASSERT_EQ(0, neteq->RegisterPayloadType(webrtc::NetEqDecoder::kDecoderPCMa,
135 "pcma", 8));
136#ifdef WEBRTC_CODEC_ILBC
kwiberg5adaf732016-10-04 09:33:27 -0700137 ASSERT_EQ(true,
138 neteq->RegisterPayloadType(102, SdpAudioFormat("ilbc", 8000, 1)));
henrik.lundin7a926812016-05-12 13:51:28 -0700139#endif
140#if defined(WEBRTC_CODEC_ISAC) || defined(WEBRTC_CODEC_ISACFX)
kwiberg5adaf732016-10-04 09:33:27 -0700141 ASSERT_EQ(true,
142 neteq->RegisterPayloadType(103, SdpAudioFormat("isac", 16000, 1)));
henrik.lundin7a926812016-05-12 13:51:28 -0700143#endif
144#ifdef WEBRTC_CODEC_ISAC
kwiberg5adaf732016-10-04 09:33:27 -0700145 ASSERT_EQ(true,
146 neteq->RegisterPayloadType(104, SdpAudioFormat("isac", 32000, 1)));
henrik.lundin7a926812016-05-12 13:51:28 -0700147#endif
148#ifdef WEBRTC_CODEC_OPUS
kwiberg5adaf732016-10-04 09:33:27 -0700149 ASSERT_EQ(true,
150 neteq->RegisterPayloadType(
151 111, SdpAudioFormat("opus", 48000, 2, {{"stereo", "0"}})));
henrik.lundin7a926812016-05-12 13:51:28 -0700152#endif
kwiberg5adaf732016-10-04 09:33:27 -0700153 ASSERT_EQ(true,
154 neteq->RegisterPayloadType(93, SdpAudioFormat("L16", 8000, 1)));
155 ASSERT_EQ(true,
156 neteq->RegisterPayloadType(94, SdpAudioFormat("L16", 16000, 1)));
157 ASSERT_EQ(true,
158 neteq->RegisterPayloadType(95, SdpAudioFormat("L16", 32000, 1)));
159 ASSERT_EQ(true,
160 neteq->RegisterPayloadType(13, SdpAudioFormat("cn", 8000, 1)));
161 ASSERT_EQ(true,
162 neteq->RegisterPayloadType(98, SdpAudioFormat("cn", 16000, 1)));
henrik.lundin7a926812016-05-12 13:51:28 -0700163}
minyue5f026d02015-12-16 07:36:04 -0800164} // namespace
165
minyue4f906772016-04-29 11:05:14 -0700166class ResultSink {
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000167 public:
minyue4f906772016-04-29 11:05:14 -0700168 explicit ResultSink(const std::string& output_file);
169 ~ResultSink();
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000170
Yves Gerey665174f2018-06-19 15:03:05 +0200171 template <typename T>
172 void AddResult(const T* test_results, size_t length);
minyue4f906772016-04-29 11:05:14 -0700173
174 void AddResult(const NetEqNetworkStatistics& stats);
175 void AddResult(const RtcpStatistics& stats);
176
177 void VerifyChecksum(const std::string& ref_check_sum);
178
179 private:
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000180 FILE* output_fp_;
minyue4f906772016-04-29 11:05:14 -0700181 std::unique_ptr<rtc::MessageDigest> digest_;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000182};
183
Joachim Bauch4e909192017-12-19 22:27:51 +0100184ResultSink::ResultSink(const std::string& output_file)
minyue4f906772016-04-29 11:05:14 -0700185 : output_fp_(nullptr),
Joachim Bauch4e909192017-12-19 22:27:51 +0100186 digest_(rtc::MessageDigestFactory::Create(rtc::DIGEST_SHA_1)) {
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000187 if (!output_file.empty()) {
188 output_fp_ = fopen(output_file.c_str(), "wb");
189 EXPECT_TRUE(output_fp_ != NULL);
190 }
191}
192
minyue4f906772016-04-29 11:05:14 -0700193ResultSink::~ResultSink() {
194 if (output_fp_)
195 fclose(output_fp_);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000196}
197
Yves Gerey665174f2018-06-19 15:03:05 +0200198template <typename T>
yujo36b1a5f2017-06-12 12:45:32 -0700199void ResultSink::AddResult(const T* test_results, size_t length) {
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000200 if (output_fp_) {
yujo36b1a5f2017-06-12 12:45:32 -0700201 ASSERT_EQ(length, fwrite(test_results, sizeof(T), length, output_fp_));
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000202 }
yujo36b1a5f2017-06-12 12:45:32 -0700203 digest_->Update(test_results, sizeof(T) * length);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000204}
205
minyue4f906772016-04-29 11:05:14 -0700206void ResultSink::AddResult(const NetEqNetworkStatistics& stats_raw) {
minyue5f026d02015-12-16 07:36:04 -0800207#ifdef WEBRTC_NETEQ_UNITTEST_BITEXACT
minyue5f026d02015-12-16 07:36:04 -0800208 neteq_unittest::NetEqNetworkStatistics stats;
209 Convert(stats_raw, &stats);
210
mbonadei7c2c8432017-04-07 00:59:12 -0700211 ProtoString stats_string;
minyue5f026d02015-12-16 07:36:04 -0800212 ASSERT_TRUE(stats.SerializeToString(&stats_string));
minyue4f906772016-04-29 11:05:14 -0700213 AddMessage(output_fp_, digest_.get(), stats_string);
minyue5f026d02015-12-16 07:36:04 -0800214#else
215 FAIL() << "Writing to reference file requires Proto Buffer.";
216#endif // WEBRTC_NETEQ_UNITTEST_BITEXACT
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000217}
218
minyue4f906772016-04-29 11:05:14 -0700219void ResultSink::AddResult(const RtcpStatistics& stats_raw) {
minyue5f026d02015-12-16 07:36:04 -0800220#ifdef WEBRTC_NETEQ_UNITTEST_BITEXACT
minyue5f026d02015-12-16 07:36:04 -0800221 neteq_unittest::RtcpStatistics stats;
222 Convert(stats_raw, &stats);
223
mbonadei7c2c8432017-04-07 00:59:12 -0700224 ProtoString stats_string;
minyue5f026d02015-12-16 07:36:04 -0800225 ASSERT_TRUE(stats.SerializeToString(&stats_string));
minyue4f906772016-04-29 11:05:14 -0700226 AddMessage(output_fp_, digest_.get(), stats_string);
minyue5f026d02015-12-16 07:36:04 -0800227#else
228 FAIL() << "Writing to reference file requires Proto Buffer.";
229#endif // WEBRTC_NETEQ_UNITTEST_BITEXACT
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000230}
231
minyue4f906772016-04-29 11:05:14 -0700232void ResultSink::VerifyChecksum(const std::string& checksum) {
233 std::vector<char> buffer;
234 buffer.resize(digest_->Size());
235 digest_->Finish(&buffer[0], buffer.size());
236 const std::string result = rtc::hex_encode(&buffer[0], digest_->Size());
237 EXPECT_EQ(checksum, result);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000238}
239
240class NetEqDecodingTest : public ::testing::Test {
241 protected:
242 // NetEQ must be polled for data once every 10 ms. Thus, neither of the
243 // constants below can be changed.
244 static const int kTimeStepMs = 10;
Peter Kastingdce40cf2015-08-24 14:52:23 -0700245 static const size_t kBlockSize8kHz = kTimeStepMs * 8;
246 static const size_t kBlockSize16kHz = kTimeStepMs * 16;
247 static const size_t kBlockSize32kHz = kTimeStepMs * 32;
minyue93c08b72015-12-22 09:57:41 -0800248 static const size_t kBlockSize48kHz = kTimeStepMs * 48;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000249 static const int kInitSampleRateHz = 8000;
250
251 NetEqDecodingTest();
252 virtual void SetUp();
253 virtual void TearDown();
254 void SelectDecoders(NetEqDecoder* used_codec);
Yves Gerey665174f2018-06-19 15:03:05 +0200255 void OpenInputFile(const std::string& rtp_file);
henrik.lundin6d8e0112016-03-04 10:34:21 -0800256 void Process();
minyue5f026d02015-12-16 07:36:04 -0800257
henrik.lundin@webrtc.org4e4b0982014-08-11 14:48:49 +0000258 void DecodeAndCompare(const std::string& rtp_file,
minyue4f906772016-04-29 11:05:14 -0700259 const std::string& output_checksum,
260 const std::string& network_stats_checksum,
261 const std::string& rtcp_stats_checksum,
262 bool gen_ref);
minyue5f026d02015-12-16 07:36:04 -0800263
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000264 static void PopulateRtpInfo(int frame_index,
265 int timestamp,
henrik.lundin246ef3e2017-04-24 09:14:32 -0700266 RTPHeader* rtp_info);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000267 static void PopulateCng(int frame_index,
268 int timestamp,
henrik.lundin246ef3e2017-04-24 09:14:32 -0700269 RTPHeader* rtp_info,
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000270 uint8_t* payload,
pkasting@chromium.org4591fbd2014-11-20 22:28:14 +0000271 size_t* payload_len);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000272
Yves Gerey665174f2018-06-19 15:03:05 +0200273 void WrapTest(uint16_t start_seq_no,
274 uint32_t start_timestamp,
turaj@webrtc.org78b41a02013-11-22 20:27:07 +0000275 const std::set<uint16_t>& drop_seq_numbers,
Yves Gerey665174f2018-06-19 15:03:05 +0200276 bool expect_seq_no_wrap,
277 bool expect_timestamp_wrap);
turaj@webrtc.org78b41a02013-11-22 20:27:07 +0000278
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000279 void LongCngWithClockDrift(double drift_factor,
280 double network_freeze_ms,
281 bool pull_audio_during_freeze,
282 int delay_tolerance_ms,
283 int max_time_to_speech_ms);
284
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +0000285 void DuplicateCng();
henrik.lundin@webrtc.orgfcfc6a92014-02-13 11:42:28 +0000286
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000287 NetEq* neteq_;
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +0000288 NetEq::Config config_;
kwiberg2d0c3322016-02-14 09:28:33 -0800289 std::unique_ptr<test::RtpFileSource> rtp_source_;
290 std::unique_ptr<test::Packet> packet_;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000291 unsigned int sim_clock_;
henrik.lundin6d8e0112016-03-04 10:34:21 -0800292 AudioFrame out_frame_;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000293 int output_sample_rate_;
turaj@webrtc.org8d1cdaa2014-04-11 18:47:55 +0000294 int algorithmic_delay_ms_;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000295};
296
297// Allocating the static const so that it can be passed by reference.
298const int NetEqDecodingTest::kTimeStepMs;
Peter Kastingdce40cf2015-08-24 14:52:23 -0700299const size_t NetEqDecodingTest::kBlockSize8kHz;
300const size_t NetEqDecodingTest::kBlockSize16kHz;
301const size_t NetEqDecodingTest::kBlockSize32kHz;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000302const int NetEqDecodingTest::kInitSampleRateHz;
303
304NetEqDecodingTest::NetEqDecodingTest()
305 : neteq_(NULL),
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +0000306 config_(),
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000307 sim_clock_(0),
turaj@webrtc.org8d1cdaa2014-04-11 18:47:55 +0000308 output_sample_rate_(kInitSampleRateHz),
309 algorithmic_delay_ms_(0) {
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +0000310 config_.sample_rate_hz = kInitSampleRateHz;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000311}
312
313void NetEqDecodingTest::SetUp() {
ossue3525782016-05-25 07:37:43 -0700314 neteq_ = NetEq::Create(config_, CreateBuiltinAudioDecoderFactory());
turaj@webrtc.org8d1cdaa2014-04-11 18:47:55 +0000315 NetEqNetworkStatistics stat;
316 ASSERT_EQ(0, neteq_->NetworkStatistics(&stat));
317 algorithmic_delay_ms_ = stat.current_buffer_size_ms;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000318 ASSERT_TRUE(neteq_);
henrik.lundin7a926812016-05-12 13:51:28 -0700319 LoadDecoders(neteq_);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000320}
321
322void NetEqDecodingTest::TearDown() {
323 delete neteq_;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000324}
325
Yves Gerey665174f2018-06-19 15:03:05 +0200326void NetEqDecodingTest::OpenInputFile(const std::string& rtp_file) {
henrik.lundin@webrtc.org966a7082014-11-17 09:08:38 +0000327 rtp_source_.reset(test::RtpFileSource::Create(rtp_file));
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000328}
329
henrik.lundin6d8e0112016-03-04 10:34:21 -0800330void NetEqDecodingTest::Process() {
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000331 // Check if time to receive.
henrik.lundin@webrtc.org966a7082014-11-17 09:08:38 +0000332 while (packet_ && sim_clock_ >= packet_->time_ms()) {
333 if (packet_->payload_length_bytes() > 0) {
ivoc72c08ed2016-01-20 07:26:24 -0800334#ifndef WEBRTC_CODEC_ISAC
335 // Ignore payload type 104 (iSAC-swb) if ISAC is not supported.
henrik.lundin246ef3e2017-04-24 09:14:32 -0700336 if (packet_->header().payloadType != 104)
ivoc72c08ed2016-01-20 07:26:24 -0800337#endif
Henrik Lundin70c09bd2017-04-24 15:56:56 +0200338 ASSERT_EQ(0,
339 neteq_->InsertPacket(
henrik.lundin246ef3e2017-04-24 09:14:32 -0700340 packet_->header(),
Henrik Lundin70c09bd2017-04-24 15:56:56 +0200341 rtc::ArrayView<const uint8_t>(
342 packet_->payload(), packet_->payload_length_bytes()),
343 static_cast<uint32_t>(packet_->time_ms() *
344 (output_sample_rate_ / 1000))));
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000345 }
346 // Get next packet.
henrik.lundin46ba49c2016-05-24 22:50:47 -0700347 packet_ = rtp_source_->NextPacket();
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000348 }
349
henrik.lundin@webrtc.orge1d468c2013-01-30 07:37:20 +0000350 // Get audio from NetEq.
henrik.lundin7a926812016-05-12 13:51:28 -0700351 bool muted;
352 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
353 ASSERT_FALSE(muted);
henrik.lundin6d8e0112016-03-04 10:34:21 -0800354 ASSERT_TRUE((out_frame_.samples_per_channel_ == kBlockSize8kHz) ||
355 (out_frame_.samples_per_channel_ == kBlockSize16kHz) ||
356 (out_frame_.samples_per_channel_ == kBlockSize32kHz) ||
357 (out_frame_.samples_per_channel_ == kBlockSize48kHz));
358 output_sample_rate_ = out_frame_.sample_rate_hz_;
henrik.lundind89814b2015-11-23 06:49:25 -0800359 EXPECT_EQ(output_sample_rate_, neteq_->last_output_sample_rate_hz());
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000360
361 // Increase time.
362 sim_clock_ += kTimeStepMs;
363}
364
minyue4f906772016-04-29 11:05:14 -0700365void NetEqDecodingTest::DecodeAndCompare(
366 const std::string& rtp_file,
367 const std::string& output_checksum,
368 const std::string& network_stats_checksum,
369 const std::string& rtcp_stats_checksum,
370 bool gen_ref) {
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000371 OpenInputFile(rtp_file);
372
minyue4f906772016-04-29 11:05:14 -0700373 std::string ref_out_file =
374 gen_ref ? webrtc::test::OutputPath() + "neteq_universal_ref.pcm" : "";
375 ResultSink output(ref_out_file);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000376
minyue4f906772016-04-29 11:05:14 -0700377 std::string stat_out_file =
378 gen_ref ? webrtc::test::OutputPath() + "neteq_network_stats.dat" : "";
379 ResultSink network_stats(stat_out_file);
henrik.lundin@webrtc.org4e4b0982014-08-11 14:48:49 +0000380
minyue4f906772016-04-29 11:05:14 -0700381 std::string rtcp_out_file =
382 gen_ref ? webrtc::test::OutputPath() + "neteq_rtcp_stats.dat" : "";
383 ResultSink rtcp_stats(rtcp_out_file);
henrik.lundin@webrtc.org4e4b0982014-08-11 14:48:49 +0000384
henrik.lundin46ba49c2016-05-24 22:50:47 -0700385 packet_ = rtp_source_->NextPacket();
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000386 int i = 0;
Henrik Lundinac0a5032017-09-25 12:22:46 +0200387 uint64_t last_concealed_samples = 0;
388 uint64_t last_total_samples_received = 0;
henrik.lundin@webrtc.org966a7082014-11-17 09:08:38 +0000389 while (packet_) {
Jonas Olsson366a50c2018-09-06 13:41:30 +0200390 rtc::StringBuilder ss;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000391 ss << "Lap number " << i++ << " in DecodeAndCompare while loop";
392 SCOPED_TRACE(ss.str()); // Print out the parameter values on failure.
henrik.lundin6d8e0112016-03-04 10:34:21 -0800393 ASSERT_NO_FATAL_FAILURE(Process());
Yves Gerey665174f2018-06-19 15:03:05 +0200394 ASSERT_NO_FATAL_FAILURE(
395 output.AddResult(out_frame_.data(), out_frame_.samples_per_channel_));
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000396
397 // Query the network statistics API once per second
398 if (sim_clock_ % 1000 == 0) {
399 // Process NetworkStatistics.
minyue4f906772016-04-29 11:05:14 -0700400 NetEqNetworkStatistics current_network_stats;
401 ASSERT_EQ(0, neteq_->NetworkStatistics(&current_network_stats));
402 ASSERT_NO_FATAL_FAILURE(network_stats.AddResult(current_network_stats));
403
henrik.lundin9c3efd02015-08-27 13:12:22 -0700404 // Compare with CurrentDelay, which should be identical.
minyue4f906772016-04-29 11:05:14 -0700405 EXPECT_EQ(current_network_stats.current_buffer_size_ms,
406 neteq_->CurrentDelayMs());
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000407
Henrik Lundinac0a5032017-09-25 12:22:46 +0200408 // Verify that liftime stats and network stats report similar loss
409 // concealment rates.
410 auto lifetime_stats = neteq_->GetLifetimeStatistics();
411 const uint64_t delta_concealed_samples =
412 lifetime_stats.concealed_samples - last_concealed_samples;
413 last_concealed_samples = lifetime_stats.concealed_samples;
414 const uint64_t delta_total_samples_received =
415 lifetime_stats.total_samples_received - last_total_samples_received;
416 last_total_samples_received = lifetime_stats.total_samples_received;
417 // The tolerance is 1% but expressed in Q14.
418 EXPECT_NEAR(
419 (delta_concealed_samples << 14) / delta_total_samples_received,
420 current_network_stats.expand_rate, (2 << 14) / 100.0);
421
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000422 // Process RTCPstat.
minyue4f906772016-04-29 11:05:14 -0700423 RtcpStatistics current_rtcp_stats;
424 neteq_->GetRtcpStatistics(&current_rtcp_stats);
425 ASSERT_NO_FATAL_FAILURE(rtcp_stats.AddResult(current_rtcp_stats));
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000426 }
427 }
minyue4f906772016-04-29 11:05:14 -0700428
429 SCOPED_TRACE("Check output audio.");
430 output.VerifyChecksum(output_checksum);
431 SCOPED_TRACE("Check network stats.");
432 network_stats.VerifyChecksum(network_stats_checksum);
433 SCOPED_TRACE("Check rtcp stats.");
434 rtcp_stats.VerifyChecksum(rtcp_stats_checksum);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000435}
436
437void NetEqDecodingTest::PopulateRtpInfo(int frame_index,
438 int timestamp,
henrik.lundin246ef3e2017-04-24 09:14:32 -0700439 RTPHeader* rtp_info) {
440 rtp_info->sequenceNumber = frame_index;
441 rtp_info->timestamp = timestamp;
442 rtp_info->ssrc = 0x1234; // Just an arbitrary SSRC.
443 rtp_info->payloadType = 94; // PCM16b WB codec.
444 rtp_info->markerBit = 0;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000445}
446
447void NetEqDecodingTest::PopulateCng(int frame_index,
448 int timestamp,
henrik.lundin246ef3e2017-04-24 09:14:32 -0700449 RTPHeader* rtp_info,
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000450 uint8_t* payload,
pkasting@chromium.org4591fbd2014-11-20 22:28:14 +0000451 size_t* payload_len) {
henrik.lundin246ef3e2017-04-24 09:14:32 -0700452 rtp_info->sequenceNumber = frame_index;
453 rtp_info->timestamp = timestamp;
454 rtp_info->ssrc = 0x1234; // Just an arbitrary SSRC.
455 rtp_info->payloadType = 98; // WB CNG.
456 rtp_info->markerBit = 0;
Yves Gerey665174f2018-06-19 15:03:05 +0200457 payload[0] = 64; // Noise level -64 dBov, quite arbitrarily chosen.
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000458 *payload_len = 1; // Only noise level, no spectral parameters.
459}
460
ivoc72c08ed2016-01-20 07:26:24 -0800461#if !defined(WEBRTC_IOS) && defined(WEBRTC_NETEQ_UNITTEST_BITEXACT) && \
462 (defined(WEBRTC_CODEC_ISAC) || defined(WEBRTC_CODEC_ISACFX)) && \
Karl Wibergeb254b42017-11-01 15:08:12 +0100463 defined(WEBRTC_CODEC_ILBC) && !defined(WEBRTC_ARCH_ARM64)
minyue5f026d02015-12-16 07:36:04 -0800464#define MAYBE_TestBitExactness TestBitExactness
kwiberg98ab3a42015-09-30 21:54:21 -0700465#else
minyue5f026d02015-12-16 07:36:04 -0800466#define MAYBE_TestBitExactness DISABLED_TestBitExactness
kwiberg98ab3a42015-09-30 21:54:21 -0700467#endif
minyue5f026d02015-12-16 07:36:04 -0800468TEST_F(NetEqDecodingTest, MAYBE_TestBitExactness) {
minyue49c454e2016-01-08 11:30:14 -0800469 const std::string input_rtp_file =
470 webrtc::test::ResourcePath("audio_coding/neteq_universal_new", "rtp");
henrik.lundin@webrtc.org4e4b0982014-08-11 14:48:49 +0000471
Yves Gerey665174f2018-06-19 15:03:05 +0200472 const std::string output_checksum =
473 PlatformChecksum("0c6dc227f781c81a229970f8fceda1a012498cba",
474 "15c4a2202877a414515e218bdb7992f0ad53e5af", "not used",
475 "0c6dc227f781c81a229970f8fceda1a012498cba",
476 "25fc4c863caa499aa447a5b8d059f5452cbcc500");
minyue4f906772016-04-29 11:05:14 -0700477
henrik.lundin2979f552017-05-05 05:04:16 -0700478 const std::string network_stats_checksum =
Henrik Lundin6dc82e82018-05-22 10:40:23 +0200479 PlatformChecksum("4b2370f5c794741d2a46be5c7935c66ef3fb53e9",
Yves Gerey665174f2018-06-19 15:03:05 +0200480 "e339cb2adf5ab3dfc21cb7205d670a34751e8336", "not used",
Henrik Lundin6dc82e82018-05-22 10:40:23 +0200481 "4b2370f5c794741d2a46be5c7935c66ef3fb53e9",
482 "4b2370f5c794741d2a46be5c7935c66ef3fb53e9");
minyue4f906772016-04-29 11:05:14 -0700483
Yves Gerey665174f2018-06-19 15:03:05 +0200484 const std::string rtcp_stats_checksum =
485 PlatformChecksum("b8880bf9fed2487efbddcb8d94b9937a29ae521d",
486 "f3f7b3d3e71d7e635240b5373b57df6a7e4ce9d4", "not used",
487 "b8880bf9fed2487efbddcb8d94b9937a29ae521d",
488 "b8880bf9fed2487efbddcb8d94b9937a29ae521d");
minyue4f906772016-04-29 11:05:14 -0700489
Yves Gerey665174f2018-06-19 15:03:05 +0200490 DecodeAndCompare(input_rtp_file, output_checksum, network_stats_checksum,
491 rtcp_stats_checksum, FLAG_gen_ref);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000492}
493
Yves Gerey665174f2018-06-19 15:03:05 +0200494#if !defined(WEBRTC_IOS) && defined(WEBRTC_NETEQ_UNITTEST_BITEXACT) && \
minyue-webrtc516711c2017-07-27 17:45:49 +0200495 defined(WEBRTC_CODEC_OPUS)
minyue93c08b72015-12-22 09:57:41 -0800496#define MAYBE_TestOpusBitExactness TestOpusBitExactness
497#else
498#define MAYBE_TestOpusBitExactness DISABLED_TestOpusBitExactness
499#endif
minyue-webrtcadb58b82017-07-26 17:59:59 +0200500TEST_F(NetEqDecodingTest, MAYBE_TestOpusBitExactness) {
minyue93c08b72015-12-22 09:57:41 -0800501 const std::string input_rtp_file =
502 webrtc::test::ResourcePath("audio_coding/neteq_opus", "rtp");
minyue93c08b72015-12-22 09:57:41 -0800503
Yves Gerey665174f2018-06-19 15:03:05 +0200504 const std::string output_checksum =
Yves Gerey466620b2018-11-09 21:15:33 +0100505 PlatformChecksum("2c05677daa968d6c68b92adf4affb7cd9bb4d363",
Yves Gerey665174f2018-06-19 15:03:05 +0200506 "b7b7ed802b0e18ee416973bf3b9ae98599b0181d",
507 "5876e52dda90d5ca433c3726555b907b97c86374",
Yves Gerey466620b2018-11-09 21:15:33 +0100508 "2c05677daa968d6c68b92adf4affb7cd9bb4d363",
509 "2c05677daa968d6c68b92adf4affb7cd9bb4d363");
minyue4f906772016-04-29 11:05:14 -0700510
henrik.lundin2979f552017-05-05 05:04:16 -0700511 const std::string network_stats_checksum =
Henrik Lundin6dc82e82018-05-22 10:40:23 +0200512 PlatformChecksum("adb3272498e436d1c019cbfd71610e9510c54497",
513 "fa935a91abc7291db47428a2d7c5361b98713a92",
514 "42106aa5267300f709f63737707ef07afd9dac61",
515 "adb3272498e436d1c019cbfd71610e9510c54497",
516 "adb3272498e436d1c019cbfd71610e9510c54497");
minyue4f906772016-04-29 11:05:14 -0700517
Yves Gerey665174f2018-06-19 15:03:05 +0200518 const std::string rtcp_stats_checksum =
519 PlatformChecksum("e37c797e3de6a64dda88c9ade7a013d022a2e1e0",
520 "e37c797e3de6a64dda88c9ade7a013d022a2e1e0",
521 "e37c797e3de6a64dda88c9ade7a013d022a2e1e0",
522 "e37c797e3de6a64dda88c9ade7a013d022a2e1e0",
523 "e37c797e3de6a64dda88c9ade7a013d022a2e1e0");
minyue4f906772016-04-29 11:05:14 -0700524
Yves Gerey665174f2018-06-19 15:03:05 +0200525 DecodeAndCompare(input_rtp_file, output_checksum, network_stats_checksum,
526 rtcp_stats_checksum, FLAG_gen_ref);
minyue93c08b72015-12-22 09:57:41 -0800527}
528
Yves Gerey665174f2018-06-19 15:03:05 +0200529#if !defined(WEBRTC_IOS) && defined(WEBRTC_NETEQ_UNITTEST_BITEXACT) && \
Henrik Lundine9619f82017-11-27 14:05:27 +0100530 defined(WEBRTC_CODEC_OPUS)
531#define MAYBE_TestOpusDtxBitExactness TestOpusDtxBitExactness
532#else
533#define MAYBE_TestOpusDtxBitExactness DISABLED_TestOpusDtxBitExactness
534#endif
Henrik Lundin4f2a4a12018-01-26 17:32:56 +0100535TEST_F(NetEqDecodingTest, MAYBE_TestOpusDtxBitExactness) {
Henrik Lundine9619f82017-11-27 14:05:27 +0100536 const std::string input_rtp_file =
537 webrtc::test::ResourcePath("audio_coding/neteq_opus_dtx", "rtp");
538
539 const std::string output_checksum =
Yves Gerey466620b2018-11-09 21:15:33 +0100540 PlatformChecksum("2ac10c4e79aeedd0df2863b079da5848b40f00b5",
Henrik Lundine9619f82017-11-27 14:05:27 +0100541 "3ec991b96872123f1554c03c543ca5d518431e46",
542 "da9f9a2d94e0c2d67342fad4965d7b91cda50b25",
Yves Gerey466620b2018-11-09 21:15:33 +0100543 "2ac10c4e79aeedd0df2863b079da5848b40f00b5",
544 "2ac10c4e79aeedd0df2863b079da5848b40f00b5");
Henrik Lundine9619f82017-11-27 14:05:27 +0100545
546 const std::string network_stats_checksum =
547 "bab58dc587d956f326056d7340c96eb9d2d3cc21";
548
549 const std::string rtcp_stats_checksum =
550 "ac27a7f305efb58b39bf123dccee25dee5758e63";
551
552 DecodeAndCompare(input_rtp_file, output_checksum, network_stats_checksum,
553 rtcp_stats_checksum, FLAG_gen_ref);
554}
555
henrik.lundin@webrtc.org7cbc4f92014-10-07 06:37:39 +0000556// Use fax mode to avoid time-scaling. This is to simplify the testing of
557// packet waiting times in the packet buffer.
558class NetEqDecodingTestFaxMode : public NetEqDecodingTest {
559 protected:
560 NetEqDecodingTestFaxMode() : NetEqDecodingTest() {
Henrik Lundin7687ad52018-07-02 10:14:46 +0200561 config_.for_test_no_time_stretching = true;
henrik.lundin@webrtc.org7cbc4f92014-10-07 06:37:39 +0000562 }
Gustaf Ullbergb0a02072017-10-02 12:00:34 +0200563 void TestJitterBufferDelay(bool apply_packet_loss);
henrik.lundin@webrtc.org7cbc4f92014-10-07 06:37:39 +0000564};
565
566TEST_F(NetEqDecodingTestFaxMode, TestFrameWaitingTimeStatistics) {
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000567 // Insert 30 dummy packets at once. Each packet contains 10 ms 16 kHz audio.
568 size_t num_frames = 30;
pkasting@chromium.org4591fbd2014-11-20 22:28:14 +0000569 const size_t kSamples = 10 * 16;
570 const size_t kPayloadBytes = kSamples * 2;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000571 for (size_t i = 0; i < num_frames; ++i) {
kwibergee2bac22015-11-11 10:34:00 -0800572 const uint8_t payload[kPayloadBytes] = {0};
henrik.lundin246ef3e2017-04-24 09:14:32 -0700573 RTPHeader rtp_info;
Mirko Bonadeia8110272017-10-18 14:22:50 +0200574 rtp_info.sequenceNumber = rtc::checked_cast<uint16_t>(i);
575 rtp_info.timestamp = rtc::checked_cast<uint32_t>(i * kSamples);
henrik.lundin246ef3e2017-04-24 09:14:32 -0700576 rtp_info.ssrc = 0x1234; // Just an arbitrary SSRC.
577 rtp_info.payloadType = 94; // PCM16b WB codec.
578 rtp_info.markerBit = 0;
579 ASSERT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000580 }
581 // Pull out all data.
582 for (size_t i = 0; i < num_frames; ++i) {
henrik.lundin7a926812016-05-12 13:51:28 -0700583 bool muted;
584 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -0800585 ASSERT_EQ(kBlockSize16kHz, out_frame_.samples_per_channel_);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000586 }
587
Henrik Lundin1bb8cf82015-08-25 13:08:04 +0200588 NetEqNetworkStatistics stats;
589 EXPECT_EQ(0, neteq_->NetworkStatistics(&stats));
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000590 // Since all frames are dumped into NetEQ at once, but pulled out with 10 ms
591 // spacing (per definition), we expect the delay to increase with 10 ms for
Henrik Lundin1bb8cf82015-08-25 13:08:04 +0200592 // each packet. Thus, we are calculating the statistics for a series from 10
593 // to 300, in steps of 10 ms.
594 EXPECT_EQ(155, stats.mean_waiting_time_ms);
595 EXPECT_EQ(155, stats.median_waiting_time_ms);
596 EXPECT_EQ(10, stats.min_waiting_time_ms);
597 EXPECT_EQ(300, stats.max_waiting_time_ms);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000598
599 // Check statistics again and make sure it's been reset.
Henrik Lundin1bb8cf82015-08-25 13:08:04 +0200600 EXPECT_EQ(0, neteq_->NetworkStatistics(&stats));
601 EXPECT_EQ(-1, stats.mean_waiting_time_ms);
602 EXPECT_EQ(-1, stats.median_waiting_time_ms);
603 EXPECT_EQ(-1, stats.min_waiting_time_ms);
604 EXPECT_EQ(-1, stats.max_waiting_time_ms);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000605}
606
henrik.lundin@webrtc.orgb4e80e02014-05-15 07:14:00 +0000607TEST_F(NetEqDecodingTest, TestAverageInterArrivalTimeNegative) {
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000608 const int kNumFrames = 3000; // Needed for convergence.
609 int frame_index = 0;
pkasting@chromium.org4591fbd2014-11-20 22:28:14 +0000610 const size_t kSamples = 10 * 16;
611 const size_t kPayloadBytes = kSamples * 2;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000612 while (frame_index < kNumFrames) {
613 // Insert one packet each time, except every 10th time where we insert two
614 // packets at once. This will create a negative clock-drift of approx. 10%.
615 int num_packets = (frame_index % 10 == 0 ? 2 : 1);
616 for (int n = 0; n < num_packets; ++n) {
617 uint8_t payload[kPayloadBytes] = {0};
henrik.lundin246ef3e2017-04-24 09:14:32 -0700618 RTPHeader rtp_info;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000619 PopulateRtpInfo(frame_index, frame_index * kSamples, &rtp_info);
henrik.lundin246ef3e2017-04-24 09:14:32 -0700620 ASSERT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000621 ++frame_index;
622 }
623
624 // Pull out data once.
henrik.lundin7a926812016-05-12 13:51:28 -0700625 bool muted;
626 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -0800627 ASSERT_EQ(kBlockSize16kHz, out_frame_.samples_per_channel_);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000628 }
629
630 NetEqNetworkStatistics network_stats;
631 ASSERT_EQ(0, neteq_->NetworkStatistics(&network_stats));
henrik.lundin0d838572016-10-13 03:35:55 -0700632 EXPECT_EQ(-103192, network_stats.clockdrift_ppm);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000633}
634
henrik.lundin@webrtc.orgb4e80e02014-05-15 07:14:00 +0000635TEST_F(NetEqDecodingTest, TestAverageInterArrivalTimePositive) {
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000636 const int kNumFrames = 5000; // Needed for convergence.
637 int frame_index = 0;
pkasting@chromium.org4591fbd2014-11-20 22:28:14 +0000638 const size_t kSamples = 10 * 16;
639 const size_t kPayloadBytes = kSamples * 2;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000640 for (int i = 0; i < kNumFrames; ++i) {
641 // Insert one packet each time, except every 10th time where we don't insert
642 // any packet. This will create a positive clock-drift of approx. 11%.
643 int num_packets = (i % 10 == 9 ? 0 : 1);
644 for (int n = 0; n < num_packets; ++n) {
645 uint8_t payload[kPayloadBytes] = {0};
henrik.lundin246ef3e2017-04-24 09:14:32 -0700646 RTPHeader rtp_info;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000647 PopulateRtpInfo(frame_index, frame_index * kSamples, &rtp_info);
henrik.lundin246ef3e2017-04-24 09:14:32 -0700648 ASSERT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000649 ++frame_index;
650 }
651
652 // Pull out data once.
henrik.lundin7a926812016-05-12 13:51:28 -0700653 bool muted;
654 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -0800655 ASSERT_EQ(kBlockSize16kHz, out_frame_.samples_per_channel_);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000656 }
657
658 NetEqNetworkStatistics network_stats;
659 ASSERT_EQ(0, neteq_->NetworkStatistics(&network_stats));
henrik.lundin0d838572016-10-13 03:35:55 -0700660 EXPECT_EQ(110953, network_stats.clockdrift_ppm);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000661}
662
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000663void NetEqDecodingTest::LongCngWithClockDrift(double drift_factor,
664 double network_freeze_ms,
665 bool pull_audio_during_freeze,
666 int delay_tolerance_ms,
667 int max_time_to_speech_ms) {
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000668 uint16_t seq_no = 0;
669 uint32_t timestamp = 0;
670 const int kFrameSizeMs = 30;
pkasting@chromium.org4591fbd2014-11-20 22:28:14 +0000671 const size_t kSamples = kFrameSizeMs * 16;
672 const size_t kPayloadBytes = kSamples * 2;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000673 double next_input_time_ms = 0.0;
674 double t_ms;
henrik.lundin7a926812016-05-12 13:51:28 -0700675 bool muted;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000676
677 // Insert speech for 5 seconds.
678 const int kSpeechDurationMs = 5000;
679 for (t_ms = 0; t_ms < kSpeechDurationMs; t_ms += 10) {
680 // Each turn in this for loop is 10 ms.
681 while (next_input_time_ms <= t_ms) {
682 // Insert one 30 ms speech frame.
683 uint8_t payload[kPayloadBytes] = {0};
henrik.lundin246ef3e2017-04-24 09:14:32 -0700684 RTPHeader rtp_info;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000685 PopulateRtpInfo(seq_no, timestamp, &rtp_info);
henrik.lundin246ef3e2017-04-24 09:14:32 -0700686 ASSERT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000687 ++seq_no;
688 timestamp += kSamples;
henrik.lundin@webrtc.orgfcfc6a92014-02-13 11:42:28 +0000689 next_input_time_ms += static_cast<double>(kFrameSizeMs) * drift_factor;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000690 }
691 // Pull out data once.
henrik.lundin7a926812016-05-12 13:51:28 -0700692 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -0800693 ASSERT_EQ(kBlockSize16kHz, out_frame_.samples_per_channel_);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000694 }
695
henrik.lundin55480f52016-03-08 02:37:57 -0800696 EXPECT_EQ(AudioFrame::kNormalSpeech, out_frame_.speech_type_);
Danil Chapovalovb6021232018-06-19 13:26:36 +0200697 absl::optional<uint32_t> playout_timestamp = neteq_->GetPlayoutTimestamp();
henrik.lundin0d96ab72016-04-06 12:28:26 -0700698 ASSERT_TRUE(playout_timestamp);
699 int32_t delay_before = timestamp - *playout_timestamp;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000700
701 // Insert CNG for 1 minute (= 60000 ms).
702 const int kCngPeriodMs = 100;
703 const int kCngPeriodSamples = kCngPeriodMs * 16; // Period in 16 kHz samples.
704 const int kCngDurationMs = 60000;
705 for (; t_ms < kSpeechDurationMs + kCngDurationMs; t_ms += 10) {
706 // Each turn in this for loop is 10 ms.
707 while (next_input_time_ms <= t_ms) {
708 // Insert one CNG frame each 100 ms.
709 uint8_t payload[kPayloadBytes];
pkasting@chromium.org4591fbd2014-11-20 22:28:14 +0000710 size_t payload_len;
henrik.lundin246ef3e2017-04-24 09:14:32 -0700711 RTPHeader rtp_info;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000712 PopulateCng(seq_no, timestamp, &rtp_info, payload, &payload_len);
kwibergee2bac22015-11-11 10:34:00 -0800713 ASSERT_EQ(0, neteq_->InsertPacket(
henrik.lundin246ef3e2017-04-24 09:14:32 -0700714 rtp_info,
kwibergee2bac22015-11-11 10:34:00 -0800715 rtc::ArrayView<const uint8_t>(payload, payload_len), 0));
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000716 ++seq_no;
717 timestamp += kCngPeriodSamples;
henrik.lundin@webrtc.orgfcfc6a92014-02-13 11:42:28 +0000718 next_input_time_ms += static_cast<double>(kCngPeriodMs) * drift_factor;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000719 }
720 // Pull out data once.
henrik.lundin7a926812016-05-12 13:51:28 -0700721 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -0800722 ASSERT_EQ(kBlockSize16kHz, out_frame_.samples_per_channel_);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000723 }
724
henrik.lundin55480f52016-03-08 02:37:57 -0800725 EXPECT_EQ(AudioFrame::kCNG, out_frame_.speech_type_);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000726
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000727 if (network_freeze_ms > 0) {
728 // First keep pulling audio for |network_freeze_ms| without inserting
729 // any data, then insert CNG data corresponding to |network_freeze_ms|
730 // without pulling any output audio.
731 const double loop_end_time = t_ms + network_freeze_ms;
732 for (; t_ms < loop_end_time; t_ms += 10) {
733 // Pull out data once.
henrik.lundin7a926812016-05-12 13:51:28 -0700734 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -0800735 ASSERT_EQ(kBlockSize16kHz, out_frame_.samples_per_channel_);
henrik.lundin55480f52016-03-08 02:37:57 -0800736 EXPECT_EQ(AudioFrame::kCNG, out_frame_.speech_type_);
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000737 }
738 bool pull_once = pull_audio_during_freeze;
739 // If |pull_once| is true, GetAudio will be called once half-way through
740 // the network recovery period.
741 double pull_time_ms = (t_ms + next_input_time_ms) / 2;
742 while (next_input_time_ms <= t_ms) {
743 if (pull_once && next_input_time_ms >= pull_time_ms) {
744 pull_once = false;
745 // Pull out data once.
henrik.lundin7a926812016-05-12 13:51:28 -0700746 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -0800747 ASSERT_EQ(kBlockSize16kHz, out_frame_.samples_per_channel_);
henrik.lundin55480f52016-03-08 02:37:57 -0800748 EXPECT_EQ(AudioFrame::kCNG, out_frame_.speech_type_);
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000749 t_ms += 10;
750 }
751 // Insert one CNG frame each 100 ms.
752 uint8_t payload[kPayloadBytes];
pkasting@chromium.org4591fbd2014-11-20 22:28:14 +0000753 size_t payload_len;
henrik.lundin246ef3e2017-04-24 09:14:32 -0700754 RTPHeader rtp_info;
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000755 PopulateCng(seq_no, timestamp, &rtp_info, payload, &payload_len);
kwibergee2bac22015-11-11 10:34:00 -0800756 ASSERT_EQ(0, neteq_->InsertPacket(
henrik.lundin246ef3e2017-04-24 09:14:32 -0700757 rtp_info,
kwibergee2bac22015-11-11 10:34:00 -0800758 rtc::ArrayView<const uint8_t>(payload, payload_len), 0));
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000759 ++seq_no;
760 timestamp += kCngPeriodSamples;
761 next_input_time_ms += kCngPeriodMs * drift_factor;
762 }
763 }
764
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000765 // Insert speech again until output type is speech.
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000766 double speech_restart_time_ms = t_ms;
henrik.lundin55480f52016-03-08 02:37:57 -0800767 while (out_frame_.speech_type_ != AudioFrame::kNormalSpeech) {
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000768 // Each turn in this for loop is 10 ms.
769 while (next_input_time_ms <= t_ms) {
770 // Insert one 30 ms speech frame.
771 uint8_t payload[kPayloadBytes] = {0};
henrik.lundin246ef3e2017-04-24 09:14:32 -0700772 RTPHeader rtp_info;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000773 PopulateRtpInfo(seq_no, timestamp, &rtp_info);
henrik.lundin246ef3e2017-04-24 09:14:32 -0700774 ASSERT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000775 ++seq_no;
776 timestamp += kSamples;
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000777 next_input_time_ms += kFrameSizeMs * drift_factor;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000778 }
779 // Pull out data once.
henrik.lundin7a926812016-05-12 13:51:28 -0700780 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -0800781 ASSERT_EQ(kBlockSize16kHz, out_frame_.samples_per_channel_);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000782 // Increase clock.
783 t_ms += 10;
784 }
785
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000786 // Check that the speech starts again within reasonable time.
787 double time_until_speech_returns_ms = t_ms - speech_restart_time_ms;
788 EXPECT_LT(time_until_speech_returns_ms, max_time_to_speech_ms);
henrik.lundin114c1b32017-04-26 07:47:32 -0700789 playout_timestamp = neteq_->GetPlayoutTimestamp();
henrik.lundin0d96ab72016-04-06 12:28:26 -0700790 ASSERT_TRUE(playout_timestamp);
791 int32_t delay_after = timestamp - *playout_timestamp;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000792 // Compare delay before and after, and make sure it differs less than 20 ms.
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000793 EXPECT_LE(delay_after, delay_before + delay_tolerance_ms * 16);
794 EXPECT_GE(delay_after, delay_before - delay_tolerance_ms * 16);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000795}
796
henrik.lundin@webrtc.orgb4e80e02014-05-15 07:14:00 +0000797TEST_F(NetEqDecodingTest, LongCngWithNegativeClockDrift) {
henrik.lundin@webrtc.orgfcfc6a92014-02-13 11:42:28 +0000798 // Apply a clock drift of -25 ms / s (sender faster than receiver).
799 const double kDriftFactor = 1000.0 / (1000.0 + 25.0);
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000800 const double kNetworkFreezeTimeMs = 0.0;
801 const bool kGetAudioDuringFreezeRecovery = false;
802 const int kDelayToleranceMs = 20;
803 const int kMaxTimeToSpeechMs = 100;
Yves Gerey665174f2018-06-19 15:03:05 +0200804 LongCngWithClockDrift(kDriftFactor, kNetworkFreezeTimeMs,
805 kGetAudioDuringFreezeRecovery, kDelayToleranceMs,
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000806 kMaxTimeToSpeechMs);
henrik.lundin@webrtc.orgfcfc6a92014-02-13 11:42:28 +0000807}
808
henrik.lundin@webrtc.orgb4e80e02014-05-15 07:14:00 +0000809TEST_F(NetEqDecodingTest, LongCngWithPositiveClockDrift) {
henrik.lundin@webrtc.orgfcfc6a92014-02-13 11:42:28 +0000810 // Apply a clock drift of +25 ms / s (sender slower than receiver).
811 const double kDriftFactor = 1000.0 / (1000.0 - 25.0);
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000812 const double kNetworkFreezeTimeMs = 0.0;
813 const bool kGetAudioDuringFreezeRecovery = false;
814 const int kDelayToleranceMs = 20;
815 const int kMaxTimeToSpeechMs = 100;
Yves Gerey665174f2018-06-19 15:03:05 +0200816 LongCngWithClockDrift(kDriftFactor, kNetworkFreezeTimeMs,
817 kGetAudioDuringFreezeRecovery, kDelayToleranceMs,
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000818 kMaxTimeToSpeechMs);
819}
820
henrik.lundin@webrtc.orgb4e80e02014-05-15 07:14:00 +0000821TEST_F(NetEqDecodingTest, LongCngWithNegativeClockDriftNetworkFreeze) {
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000822 // Apply a clock drift of -25 ms / s (sender faster than receiver).
823 const double kDriftFactor = 1000.0 / (1000.0 + 25.0);
824 const double kNetworkFreezeTimeMs = 5000.0;
825 const bool kGetAudioDuringFreezeRecovery = false;
826 const int kDelayToleranceMs = 50;
827 const int kMaxTimeToSpeechMs = 200;
Yves Gerey665174f2018-06-19 15:03:05 +0200828 LongCngWithClockDrift(kDriftFactor, kNetworkFreezeTimeMs,
829 kGetAudioDuringFreezeRecovery, kDelayToleranceMs,
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000830 kMaxTimeToSpeechMs);
831}
832
henrik.lundin@webrtc.orgb4e80e02014-05-15 07:14:00 +0000833TEST_F(NetEqDecodingTest, LongCngWithPositiveClockDriftNetworkFreeze) {
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000834 // Apply a clock drift of +25 ms / s (sender slower than receiver).
835 const double kDriftFactor = 1000.0 / (1000.0 - 25.0);
836 const double kNetworkFreezeTimeMs = 5000.0;
837 const bool kGetAudioDuringFreezeRecovery = false;
838 const int kDelayToleranceMs = 20;
839 const int kMaxTimeToSpeechMs = 100;
Yves Gerey665174f2018-06-19 15:03:05 +0200840 LongCngWithClockDrift(kDriftFactor, kNetworkFreezeTimeMs,
841 kGetAudioDuringFreezeRecovery, kDelayToleranceMs,
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000842 kMaxTimeToSpeechMs);
843}
844
henrik.lundin@webrtc.orgb4e80e02014-05-15 07:14:00 +0000845TEST_F(NetEqDecodingTest, LongCngWithPositiveClockDriftNetworkFreezeExtraPull) {
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000846 // Apply a clock drift of +25 ms / s (sender slower than receiver).
847 const double kDriftFactor = 1000.0 / (1000.0 - 25.0);
848 const double kNetworkFreezeTimeMs = 5000.0;
849 const bool kGetAudioDuringFreezeRecovery = true;
850 const int kDelayToleranceMs = 20;
851 const int kMaxTimeToSpeechMs = 100;
Yves Gerey665174f2018-06-19 15:03:05 +0200852 LongCngWithClockDrift(kDriftFactor, kNetworkFreezeTimeMs,
853 kGetAudioDuringFreezeRecovery, kDelayToleranceMs,
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000854 kMaxTimeToSpeechMs);
855}
856
henrik.lundin@webrtc.orgb4e80e02014-05-15 07:14:00 +0000857TEST_F(NetEqDecodingTest, LongCngWithoutClockDrift) {
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000858 const double kDriftFactor = 1.0; // No drift.
859 const double kNetworkFreezeTimeMs = 0.0;
860 const bool kGetAudioDuringFreezeRecovery = false;
861 const int kDelayToleranceMs = 10;
862 const int kMaxTimeToSpeechMs = 50;
Yves Gerey665174f2018-06-19 15:03:05 +0200863 LongCngWithClockDrift(kDriftFactor, kNetworkFreezeTimeMs,
864 kGetAudioDuringFreezeRecovery, kDelayToleranceMs,
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000865 kMaxTimeToSpeechMs);
henrik.lundin@webrtc.orgfcfc6a92014-02-13 11:42:28 +0000866}
867
henrik.lundin@webrtc.orgb4e80e02014-05-15 07:14:00 +0000868TEST_F(NetEqDecodingTest, UnknownPayloadType) {
pkasting@chromium.org4591fbd2014-11-20 22:28:14 +0000869 const size_t kPayloadBytes = 100;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000870 uint8_t payload[kPayloadBytes] = {0};
henrik.lundin246ef3e2017-04-24 09:14:32 -0700871 RTPHeader rtp_info;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000872 PopulateRtpInfo(0, 0, &rtp_info);
henrik.lundin246ef3e2017-04-24 09:14:32 -0700873 rtp_info.payloadType = 1; // Not registered as a decoder.
874 EXPECT_EQ(NetEq::kFail, neteq_->InsertPacket(rtp_info, payload, 0));
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000875}
876
Peter Boströme2976c82016-01-04 22:44:05 +0100877#if defined(WEBRTC_CODEC_ISAC) || defined(WEBRTC_CODEC_ISACFX)
ivoc72c08ed2016-01-20 07:26:24 -0800878#define MAYBE_DecoderError DecoderError
879#else
880#define MAYBE_DecoderError DISABLED_DecoderError
881#endif
882
Peter Boströme2976c82016-01-04 22:44:05 +0100883TEST_F(NetEqDecodingTest, MAYBE_DecoderError) {
pkasting@chromium.org4591fbd2014-11-20 22:28:14 +0000884 const size_t kPayloadBytes = 100;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000885 uint8_t payload[kPayloadBytes] = {0};
henrik.lundin246ef3e2017-04-24 09:14:32 -0700886 RTPHeader rtp_info;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000887 PopulateRtpInfo(0, 0, &rtp_info);
henrik.lundin246ef3e2017-04-24 09:14:32 -0700888 rtp_info.payloadType = 103; // iSAC, but the payload is invalid.
889 EXPECT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000890 // Set all of |out_data_| to 1, and verify that it was set to 0 by the call
891 // to GetAudio.
yujo36b1a5f2017-06-12 12:45:32 -0700892 int16_t* out_frame_data = out_frame_.mutable_data();
henrik.lundin6d8e0112016-03-04 10:34:21 -0800893 for (size_t i = 0; i < AudioFrame::kMaxDataSizeSamples; ++i) {
yujo36b1a5f2017-06-12 12:45:32 -0700894 out_frame_data[i] = 1;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000895 }
henrik.lundin7a926812016-05-12 13:51:28 -0700896 bool muted;
897 EXPECT_EQ(NetEq::kFail, neteq_->GetAudio(&out_frame_, &muted));
898 ASSERT_FALSE(muted);
ivoc72c08ed2016-01-20 07:26:24 -0800899
yujo36b1a5f2017-06-12 12:45:32 -0700900 // Verify that the first 160 samples are set to 0.
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000901 static const int kExpectedOutputLength = 160; // 10 ms at 16 kHz sample rate.
yujo36b1a5f2017-06-12 12:45:32 -0700902 const int16_t* const_out_frame_data = out_frame_.data();
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000903 for (int i = 0; i < kExpectedOutputLength; ++i) {
Jonas Olsson366a50c2018-09-06 13:41:30 +0200904 rtc::StringBuilder ss;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000905 ss << "i = " << i;
906 SCOPED_TRACE(ss.str()); // Print out the parameter values on failure.
yujo36b1a5f2017-06-12 12:45:32 -0700907 EXPECT_EQ(0, const_out_frame_data[i]);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000908 }
909}
910
henrik.lundin@webrtc.orgb4e80e02014-05-15 07:14:00 +0000911TEST_F(NetEqDecodingTest, GetAudioBeforeInsertPacket) {
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000912 // Set all of |out_data_| to 1, and verify that it was set to 0 by the call
913 // to GetAudio.
yujo36b1a5f2017-06-12 12:45:32 -0700914 int16_t* out_frame_data = out_frame_.mutable_data();
henrik.lundin6d8e0112016-03-04 10:34:21 -0800915 for (size_t i = 0; i < AudioFrame::kMaxDataSizeSamples; ++i) {
yujo36b1a5f2017-06-12 12:45:32 -0700916 out_frame_data[i] = 1;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000917 }
henrik.lundin7a926812016-05-12 13:51:28 -0700918 bool muted;
919 EXPECT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
920 ASSERT_FALSE(muted);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000921 // Verify that the first block of samples is set to 0.
922 static const int kExpectedOutputLength =
923 kInitSampleRateHz / 100; // 10 ms at initial sample rate.
yujo36b1a5f2017-06-12 12:45:32 -0700924 const int16_t* const_out_frame_data = out_frame_.data();
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000925 for (int i = 0; i < kExpectedOutputLength; ++i) {
Jonas Olsson366a50c2018-09-06 13:41:30 +0200926 rtc::StringBuilder ss;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000927 ss << "i = " << i;
928 SCOPED_TRACE(ss.str()); // Print out the parameter values on failure.
yujo36b1a5f2017-06-12 12:45:32 -0700929 EXPECT_EQ(0, const_out_frame_data[i]);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000930 }
henrik.lundind89814b2015-11-23 06:49:25 -0800931 // Verify that the sample rate did not change from the initial configuration.
932 EXPECT_EQ(config_.sample_rate_hz, neteq_->last_output_sample_rate_hz());
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000933}
turaj@webrtc.orgff43c852013-09-25 00:07:27 +0000934
henrik.lundin@webrtc.org9b8102c2014-08-21 08:27:44 +0000935class NetEqBgnTest : public NetEqDecodingTest {
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +0000936 protected:
henrik.lundin@webrtc.org9b8102c2014-08-21 08:27:44 +0000937 void CheckBgn(int sampling_rate_hz) {
Peter Kastingdce40cf2015-08-24 14:52:23 -0700938 size_t expected_samples_per_channel = 0;
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +0000939 uint8_t payload_type = 0xFF; // Invalid.
940 if (sampling_rate_hz == 8000) {
941 expected_samples_per_channel = kBlockSize8kHz;
942 payload_type = 93; // PCM 16, 8 kHz.
943 } else if (sampling_rate_hz == 16000) {
944 expected_samples_per_channel = kBlockSize16kHz;
945 payload_type = 94; // PCM 16, 16 kHZ.
946 } else if (sampling_rate_hz == 32000) {
947 expected_samples_per_channel = kBlockSize32kHz;
948 payload_type = 95; // PCM 16, 32 kHz.
949 } else {
950 ASSERT_TRUE(false); // Unsupported test case.
951 }
turaj@webrtc.orgff43c852013-09-25 00:07:27 +0000952
henrik.lundin6d8e0112016-03-04 10:34:21 -0800953 AudioFrame output;
henrik.lundin@webrtc.org9b8102c2014-08-21 08:27:44 +0000954 test::AudioLoop input;
955 // We are using the same 32 kHz input file for all tests, regardless of
956 // |sampling_rate_hz|. The output may sound weird, but the test is still
957 // valid.
958 ASSERT_TRUE(input.Init(
959 webrtc::test::ResourcePath("audio_coding/testfile32kHz", "pcm"),
960 10 * sampling_rate_hz, // Max 10 seconds loop length.
Peter Kastingdce40cf2015-08-24 14:52:23 -0700961 expected_samples_per_channel));
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +0000962
963 // Payload of 10 ms of PCM16 32 kHz.
964 uint8_t payload[kBlockSize32kHz * sizeof(int16_t)];
henrik.lundin246ef3e2017-04-24 09:14:32 -0700965 RTPHeader rtp_info;
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +0000966 PopulateRtpInfo(0, 0, &rtp_info);
henrik.lundin246ef3e2017-04-24 09:14:32 -0700967 rtp_info.payloadType = payload_type;
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +0000968
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +0000969 uint32_t receive_timestamp = 0;
henrik.lundin7a926812016-05-12 13:51:28 -0700970 bool muted;
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +0000971 for (int n = 0; n < 10; ++n) { // Insert few packets and get audio.
kwiberg288886b2015-11-06 01:21:35 -0800972 auto block = input.GetNextBlock();
973 ASSERT_EQ(expected_samples_per_channel, block.size());
974 size_t enc_len_bytes =
975 WebRtcPcm16b_Encode(block.data(), block.size(), payload);
henrik.lundin@webrtc.org9b8102c2014-08-21 08:27:44 +0000976 ASSERT_EQ(enc_len_bytes, expected_samples_per_channel * 2);
977
Henrik Lundin70c09bd2017-04-24 15:56:56 +0200978 ASSERT_EQ(0, neteq_->InsertPacket(
henrik.lundin246ef3e2017-04-24 09:14:32 -0700979 rtp_info,
Henrik Lundin70c09bd2017-04-24 15:56:56 +0200980 rtc::ArrayView<const uint8_t>(payload, enc_len_bytes),
981 receive_timestamp));
henrik.lundin6d8e0112016-03-04 10:34:21 -0800982 output.Reset();
henrik.lundin7a926812016-05-12 13:51:28 -0700983 ASSERT_EQ(0, neteq_->GetAudio(&output, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -0800984 ASSERT_EQ(1u, output.num_channels_);
985 ASSERT_EQ(expected_samples_per_channel, output.samples_per_channel_);
henrik.lundin55480f52016-03-08 02:37:57 -0800986 ASSERT_EQ(AudioFrame::kNormalSpeech, output.speech_type_);
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +0000987
988 // Next packet.
Yves Gerey665174f2018-06-19 15:03:05 +0200989 rtp_info.timestamp +=
990 rtc::checked_cast<uint32_t>(expected_samples_per_channel);
henrik.lundin246ef3e2017-04-24 09:14:32 -0700991 rtp_info.sequenceNumber++;
Yves Gerey665174f2018-06-19 15:03:05 +0200992 receive_timestamp +=
993 rtc::checked_cast<uint32_t>(expected_samples_per_channel);
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +0000994 }
995
henrik.lundin6d8e0112016-03-04 10:34:21 -0800996 output.Reset();
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +0000997
998 // Get audio without inserting packets, expecting PLC and PLC-to-CNG. Pull
999 // one frame without checking speech-type. This is the first frame pulled
1000 // without inserting any packet, and might not be labeled as PLC.
henrik.lundin7a926812016-05-12 13:51:28 -07001001 ASSERT_EQ(0, neteq_->GetAudio(&output, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -08001002 ASSERT_EQ(1u, output.num_channels_);
1003 ASSERT_EQ(expected_samples_per_channel, output.samples_per_channel_);
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +00001004
1005 // To be able to test the fading of background noise we need at lease to
1006 // pull 611 frames.
1007 const int kFadingThreshold = 611;
1008
1009 // Test several CNG-to-PLC packet for the expected behavior. The number 20
1010 // is arbitrary, but sufficiently large to test enough number of frames.
1011 const int kNumPlcToCngTestFrames = 20;
1012 bool plc_to_cng = false;
1013 for (int n = 0; n < kFadingThreshold + kNumPlcToCngTestFrames; ++n) {
henrik.lundin6d8e0112016-03-04 10:34:21 -08001014 output.Reset();
yujo36b1a5f2017-06-12 12:45:32 -07001015 // Set to non-zero.
1016 memset(output.mutable_data(), 1, AudioFrame::kMaxDataSizeBytes);
henrik.lundin7a926812016-05-12 13:51:28 -07001017 ASSERT_EQ(0, neteq_->GetAudio(&output, &muted));
1018 ASSERT_FALSE(muted);
henrik.lundin6d8e0112016-03-04 10:34:21 -08001019 ASSERT_EQ(1u, output.num_channels_);
1020 ASSERT_EQ(expected_samples_per_channel, output.samples_per_channel_);
henrik.lundin55480f52016-03-08 02:37:57 -08001021 if (output.speech_type_ == AudioFrame::kPLCCNG) {
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +00001022 plc_to_cng = true;
1023 double sum_squared = 0;
yujo36b1a5f2017-06-12 12:45:32 -07001024 const int16_t* output_data = output.data();
henrik.lundin6d8e0112016-03-04 10:34:21 -08001025 for (size_t k = 0;
1026 k < output.num_channels_ * output.samples_per_channel_; ++k)
yujo36b1a5f2017-06-12 12:45:32 -07001027 sum_squared += output_data[k] * output_data[k];
Henrik Lundin67190172018-04-20 15:34:48 +02001028 EXPECT_EQ(0, sum_squared);
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +00001029 } else {
henrik.lundin55480f52016-03-08 02:37:57 -08001030 EXPECT_EQ(AudioFrame::kPLC, output.speech_type_);
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +00001031 }
1032 }
1033 EXPECT_TRUE(plc_to_cng); // Just to be sure that PLC-to-CNG has occurred.
1034 }
1035};
1036
Henrik Lundin67190172018-04-20 15:34:48 +02001037TEST_F(NetEqBgnTest, RunTest) {
henrik.lundin@webrtc.org9b8102c2014-08-21 08:27:44 +00001038 CheckBgn(8000);
1039 CheckBgn(16000);
1040 CheckBgn(32000);
1041}
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +00001042
turaj@webrtc.org78b41a02013-11-22 20:27:07 +00001043void NetEqDecodingTest::WrapTest(uint16_t start_seq_no,
1044 uint32_t start_timestamp,
1045 const std::set<uint16_t>& drop_seq_numbers,
1046 bool expect_seq_no_wrap,
1047 bool expect_timestamp_wrap) {
1048 uint16_t seq_no = start_seq_no;
1049 uint32_t timestamp = start_timestamp;
1050 const int kBlocksPerFrame = 3; // Number of 10 ms blocks per frame.
1051 const int kFrameSizeMs = kBlocksPerFrame * kTimeStepMs;
1052 const int kSamples = kBlockSize16kHz * kBlocksPerFrame;
pkasting@chromium.org4591fbd2014-11-20 22:28:14 +00001053 const size_t kPayloadBytes = kSamples * sizeof(int16_t);
turaj@webrtc.org78b41a02013-11-22 20:27:07 +00001054 double next_input_time_ms = 0.0;
turaj@webrtc.org78b41a02013-11-22 20:27:07 +00001055 uint32_t receive_timestamp = 0;
1056
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001057 // Insert speech for 2 seconds.
turaj@webrtc.org78b41a02013-11-22 20:27:07 +00001058 const int kSpeechDurationMs = 2000;
1059 int packets_inserted = 0;
1060 uint16_t last_seq_no;
1061 uint32_t last_timestamp;
1062 bool timestamp_wrapped = false;
1063 bool seq_no_wrapped = false;
1064 for (double t_ms = 0; t_ms < kSpeechDurationMs; t_ms += 10) {
1065 // Each turn in this for loop is 10 ms.
1066 while (next_input_time_ms <= t_ms) {
1067 // Insert one 30 ms speech frame.
1068 uint8_t payload[kPayloadBytes] = {0};
henrik.lundin246ef3e2017-04-24 09:14:32 -07001069 RTPHeader rtp_info;
turaj@webrtc.org78b41a02013-11-22 20:27:07 +00001070 PopulateRtpInfo(seq_no, timestamp, &rtp_info);
1071 if (drop_seq_numbers.find(seq_no) == drop_seq_numbers.end()) {
1072 // This sequence number was not in the set to drop. Insert it.
henrik.lundin246ef3e2017-04-24 09:14:32 -07001073 ASSERT_EQ(0,
1074 neteq_->InsertPacket(rtp_info, payload, receive_timestamp));
turaj@webrtc.org78b41a02013-11-22 20:27:07 +00001075 ++packets_inserted;
1076 }
1077 NetEqNetworkStatistics network_stats;
1078 ASSERT_EQ(0, neteq_->NetworkStatistics(&network_stats));
1079
1080 // Due to internal NetEq logic, preferred buffer-size is about 4 times the
1081 // packet size for first few packets. Therefore we refrain from checking
1082 // the criteria.
1083 if (packets_inserted > 4) {
1084 // Expect preferred and actual buffer size to be no more than 2 frames.
1085 EXPECT_LE(network_stats.preferred_buffer_size_ms, kFrameSizeMs * 2);
Yves Gerey665174f2018-06-19 15:03:05 +02001086 EXPECT_LE(network_stats.current_buffer_size_ms,
1087 kFrameSizeMs * 2 + algorithmic_delay_ms_);
turaj@webrtc.org78b41a02013-11-22 20:27:07 +00001088 }
1089 last_seq_no = seq_no;
1090 last_timestamp = timestamp;
1091
1092 ++seq_no;
1093 timestamp += kSamples;
1094 receive_timestamp += kSamples;
1095 next_input_time_ms += static_cast<double>(kFrameSizeMs);
1096
1097 seq_no_wrapped |= seq_no < last_seq_no;
1098 timestamp_wrapped |= timestamp < last_timestamp;
1099 }
1100 // Pull out data once.
henrik.lundin6d8e0112016-03-04 10:34:21 -08001101 AudioFrame output;
henrik.lundin7a926812016-05-12 13:51:28 -07001102 bool muted;
1103 ASSERT_EQ(0, neteq_->GetAudio(&output, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -08001104 ASSERT_EQ(kBlockSize16kHz, output.samples_per_channel_);
1105 ASSERT_EQ(1u, output.num_channels_);
turaj@webrtc.org78b41a02013-11-22 20:27:07 +00001106
1107 // Expect delay (in samples) to be less than 2 packets.
Danil Chapovalovb6021232018-06-19 13:26:36 +02001108 absl::optional<uint32_t> playout_timestamp = neteq_->GetPlayoutTimestamp();
henrik.lundin0d96ab72016-04-06 12:28:26 -07001109 ASSERT_TRUE(playout_timestamp);
1110 EXPECT_LE(timestamp - *playout_timestamp,
turaj@webrtc.org78b41a02013-11-22 20:27:07 +00001111 static_cast<uint32_t>(kSamples * 2));
turaj@webrtc.org78b41a02013-11-22 20:27:07 +00001112 }
1113 // Make sure we have actually tested wrap-around.
1114 ASSERT_EQ(expect_seq_no_wrap, seq_no_wrapped);
1115 ASSERT_EQ(expect_timestamp_wrap, timestamp_wrapped);
1116}
1117
1118TEST_F(NetEqDecodingTest, SequenceNumberWrap) {
1119 // Start with a sequence number that will soon wrap.
1120 std::set<uint16_t> drop_seq_numbers; // Don't drop any packets.
1121 WrapTest(0xFFFF - 10, 0, drop_seq_numbers, true, false);
1122}
1123
1124TEST_F(NetEqDecodingTest, SequenceNumberWrapAndDrop) {
1125 // Start with a sequence number that will soon wrap.
1126 std::set<uint16_t> drop_seq_numbers;
1127 drop_seq_numbers.insert(0xFFFF);
1128 drop_seq_numbers.insert(0x0);
1129 WrapTest(0xFFFF - 10, 0, drop_seq_numbers, true, false);
1130}
1131
1132TEST_F(NetEqDecodingTest, TimestampWrap) {
1133 // Start with a timestamp that will soon wrap.
1134 std::set<uint16_t> drop_seq_numbers;
1135 WrapTest(0, 0xFFFFFFFF - 3000, drop_seq_numbers, false, true);
1136}
1137
1138TEST_F(NetEqDecodingTest, TimestampAndSequenceNumberWrap) {
1139 // Start with a timestamp and a sequence number that will wrap at the same
1140 // time.
1141 std::set<uint16_t> drop_seq_numbers;
1142 WrapTest(0xFFFF - 10, 0xFFFFFFFF - 5000, drop_seq_numbers, true, true);
1143}
1144
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001145void NetEqDecodingTest::DuplicateCng() {
1146 uint16_t seq_no = 0;
1147 uint32_t timestamp = 0;
1148 const int kFrameSizeMs = 10;
1149 const int kSampleRateKhz = 16;
1150 const int kSamples = kFrameSizeMs * kSampleRateKhz;
pkasting@chromium.org4591fbd2014-11-20 22:28:14 +00001151 const size_t kPayloadBytes = kSamples * 2;
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001152
Yves Gerey665174f2018-06-19 15:03:05 +02001153 const int algorithmic_delay_samples =
1154 std::max(algorithmic_delay_ms_ * kSampleRateKhz, 5 * kSampleRateKhz / 8);
henrik.lundin@webrtc.orgc93437e2014-12-01 11:42:42 +00001155 // Insert three speech packets. Three are needed to get the frame length
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001156 // correct.
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001157 uint8_t payload[kPayloadBytes] = {0};
henrik.lundin246ef3e2017-04-24 09:14:32 -07001158 RTPHeader rtp_info;
henrik.lundin7a926812016-05-12 13:51:28 -07001159 bool muted;
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001160 for (int i = 0; i < 3; ++i) {
1161 PopulateRtpInfo(seq_no, timestamp, &rtp_info);
henrik.lundin246ef3e2017-04-24 09:14:32 -07001162 ASSERT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001163 ++seq_no;
1164 timestamp += kSamples;
1165
1166 // Pull audio once.
henrik.lundin7a926812016-05-12 13:51:28 -07001167 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -08001168 ASSERT_EQ(kBlockSize16kHz, out_frame_.samples_per_channel_);
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001169 }
1170 // Verify speech output.
henrik.lundin55480f52016-03-08 02:37:57 -08001171 EXPECT_EQ(AudioFrame::kNormalSpeech, out_frame_.speech_type_);
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001172
1173 // Insert same CNG packet twice.
1174 const int kCngPeriodMs = 100;
1175 const int kCngPeriodSamples = kCngPeriodMs * kSampleRateKhz;
pkasting@chromium.org4591fbd2014-11-20 22:28:14 +00001176 size_t payload_len;
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001177 PopulateCng(seq_no, timestamp, &rtp_info, payload, &payload_len);
1178 // This is the first time this CNG packet is inserted.
henrik.lundin246ef3e2017-04-24 09:14:32 -07001179 ASSERT_EQ(
1180 0, neteq_->InsertPacket(
1181 rtp_info, rtc::ArrayView<const uint8_t>(payload, payload_len), 0));
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001182
1183 // Pull audio once and make sure CNG is played.
henrik.lundin7a926812016-05-12 13:51:28 -07001184 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -08001185 ASSERT_EQ(kBlockSize16kHz, out_frame_.samples_per_channel_);
henrik.lundin55480f52016-03-08 02:37:57 -08001186 EXPECT_EQ(AudioFrame::kCNG, out_frame_.speech_type_);
henrik.lundin114c1b32017-04-26 07:47:32 -07001187 EXPECT_FALSE(
1188 neteq_->GetPlayoutTimestamp()); // Returns empty value during CNG.
henrik.lundin0d96ab72016-04-06 12:28:26 -07001189 EXPECT_EQ(timestamp - algorithmic_delay_samples,
1190 out_frame_.timestamp_ + out_frame_.samples_per_channel_);
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001191
1192 // Insert the same CNG packet again. Note that at this point it is old, since
1193 // we have already decoded the first copy of it.
henrik.lundin246ef3e2017-04-24 09:14:32 -07001194 ASSERT_EQ(
1195 0, neteq_->InsertPacket(
1196 rtp_info, rtc::ArrayView<const uint8_t>(payload, payload_len), 0));
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001197
1198 // Pull audio until we have played |kCngPeriodMs| of CNG. Start at 10 ms since
1199 // we have already pulled out CNG once.
1200 for (int cng_time_ms = 10; cng_time_ms < kCngPeriodMs; cng_time_ms += 10) {
henrik.lundin7a926812016-05-12 13:51:28 -07001201 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -08001202 ASSERT_EQ(kBlockSize16kHz, out_frame_.samples_per_channel_);
henrik.lundin55480f52016-03-08 02:37:57 -08001203 EXPECT_EQ(AudioFrame::kCNG, out_frame_.speech_type_);
henrik.lundin114c1b32017-04-26 07:47:32 -07001204 EXPECT_FALSE(
1205 neteq_->GetPlayoutTimestamp()); // Returns empty value during CNG.
turaj@webrtc.org8d1cdaa2014-04-11 18:47:55 +00001206 EXPECT_EQ(timestamp - algorithmic_delay_samples,
henrik.lundin0d96ab72016-04-06 12:28:26 -07001207 out_frame_.timestamp_ + out_frame_.samples_per_channel_);
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001208 }
1209
1210 // Insert speech again.
1211 ++seq_no;
1212 timestamp += kCngPeriodSamples;
1213 PopulateRtpInfo(seq_no, timestamp, &rtp_info);
henrik.lundin246ef3e2017-04-24 09:14:32 -07001214 ASSERT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001215
1216 // Pull audio once and verify that the output is speech again.
henrik.lundin7a926812016-05-12 13:51:28 -07001217 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -08001218 ASSERT_EQ(kBlockSize16kHz, out_frame_.samples_per_channel_);
henrik.lundin55480f52016-03-08 02:37:57 -08001219 EXPECT_EQ(AudioFrame::kNormalSpeech, out_frame_.speech_type_);
Danil Chapovalovb6021232018-06-19 13:26:36 +02001220 absl::optional<uint32_t> playout_timestamp = neteq_->GetPlayoutTimestamp();
henrik.lundin0d96ab72016-04-06 12:28:26 -07001221 ASSERT_TRUE(playout_timestamp);
turaj@webrtc.org8d1cdaa2014-04-11 18:47:55 +00001222 EXPECT_EQ(timestamp + kSamples - algorithmic_delay_samples,
henrik.lundin0d96ab72016-04-06 12:28:26 -07001223 *playout_timestamp);
wu@webrtc.org94454b72014-06-05 20:34:08 +00001224}
1225
Yves Gerey665174f2018-06-19 15:03:05 +02001226TEST_F(NetEqDecodingTest, DiscardDuplicateCng) {
1227 DuplicateCng();
1228}
henrik.lundin@webrtc.orgc93437e2014-12-01 11:42:42 +00001229
1230TEST_F(NetEqDecodingTest, CngFirst) {
1231 uint16_t seq_no = 0;
1232 uint32_t timestamp = 0;
1233 const int kFrameSizeMs = 10;
1234 const int kSampleRateKhz = 16;
1235 const int kSamples = kFrameSizeMs * kSampleRateKhz;
1236 const int kPayloadBytes = kSamples * 2;
1237 const int kCngPeriodMs = 100;
1238 const int kCngPeriodSamples = kCngPeriodMs * kSampleRateKhz;
1239 size_t payload_len;
1240
1241 uint8_t payload[kPayloadBytes] = {0};
henrik.lundin246ef3e2017-04-24 09:14:32 -07001242 RTPHeader rtp_info;
henrik.lundin@webrtc.orgc93437e2014-12-01 11:42:42 +00001243
1244 PopulateCng(seq_no, timestamp, &rtp_info, payload, &payload_len);
henrik.lundin246ef3e2017-04-24 09:14:32 -07001245 ASSERT_EQ(
1246 NetEq::kOK,
1247 neteq_->InsertPacket(
1248 rtp_info, rtc::ArrayView<const uint8_t>(payload, payload_len), 0));
henrik.lundin@webrtc.orgc93437e2014-12-01 11:42:42 +00001249 ++seq_no;
1250 timestamp += kCngPeriodSamples;
1251
1252 // Pull audio once and make sure CNG is played.
henrik.lundin7a926812016-05-12 13:51:28 -07001253 bool muted;
1254 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -08001255 ASSERT_EQ(kBlockSize16kHz, out_frame_.samples_per_channel_);
henrik.lundin55480f52016-03-08 02:37:57 -08001256 EXPECT_EQ(AudioFrame::kCNG, out_frame_.speech_type_);
henrik.lundin@webrtc.orgc93437e2014-12-01 11:42:42 +00001257
1258 // Insert some speech packets.
henrik.lundin549d80b2016-08-25 00:44:24 -07001259 const uint32_t first_speech_timestamp = timestamp;
1260 int timeout_counter = 0;
1261 do {
1262 ASSERT_LT(timeout_counter++, 20) << "Test timed out";
henrik.lundin@webrtc.orgc93437e2014-12-01 11:42:42 +00001263 PopulateRtpInfo(seq_no, timestamp, &rtp_info);
henrik.lundin246ef3e2017-04-24 09:14:32 -07001264 ASSERT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
henrik.lundin@webrtc.orgc93437e2014-12-01 11:42:42 +00001265 ++seq_no;
1266 timestamp += kSamples;
1267
1268 // Pull audio once.
henrik.lundin7a926812016-05-12 13:51:28 -07001269 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -08001270 ASSERT_EQ(kBlockSize16kHz, out_frame_.samples_per_channel_);
henrik.lundin549d80b2016-08-25 00:44:24 -07001271 } while (!IsNewerTimestamp(out_frame_.timestamp_, first_speech_timestamp));
henrik.lundin@webrtc.orgc93437e2014-12-01 11:42:42 +00001272 // Verify speech output.
henrik.lundin55480f52016-03-08 02:37:57 -08001273 EXPECT_EQ(AudioFrame::kNormalSpeech, out_frame_.speech_type_);
henrik.lundin@webrtc.orgc93437e2014-12-01 11:42:42 +00001274}
henrik.lundin7a926812016-05-12 13:51:28 -07001275
1276class NetEqDecodingTestWithMutedState : public NetEqDecodingTest {
1277 public:
1278 NetEqDecodingTestWithMutedState() : NetEqDecodingTest() {
1279 config_.enable_muted_state = true;
1280 }
1281
1282 protected:
1283 static constexpr size_t kSamples = 10 * 16;
1284 static constexpr size_t kPayloadBytes = kSamples * 2;
1285
1286 void InsertPacket(uint32_t rtp_timestamp) {
1287 uint8_t payload[kPayloadBytes] = {0};
henrik.lundin246ef3e2017-04-24 09:14:32 -07001288 RTPHeader rtp_info;
henrik.lundin7a926812016-05-12 13:51:28 -07001289 PopulateRtpInfo(0, rtp_timestamp, &rtp_info);
henrik.lundin246ef3e2017-04-24 09:14:32 -07001290 EXPECT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
henrik.lundin7a926812016-05-12 13:51:28 -07001291 }
1292
henrik.lundin42feb512016-09-20 06:51:40 -07001293 void InsertCngPacket(uint32_t rtp_timestamp) {
1294 uint8_t payload[kPayloadBytes] = {0};
henrik.lundin246ef3e2017-04-24 09:14:32 -07001295 RTPHeader rtp_info;
henrik.lundin42feb512016-09-20 06:51:40 -07001296 size_t payload_len;
1297 PopulateCng(0, rtp_timestamp, &rtp_info, payload, &payload_len);
henrik.lundin246ef3e2017-04-24 09:14:32 -07001298 EXPECT_EQ(
1299 NetEq::kOK,
1300 neteq_->InsertPacket(
1301 rtp_info, rtc::ArrayView<const uint8_t>(payload, payload_len), 0));
henrik.lundin42feb512016-09-20 06:51:40 -07001302 }
1303
henrik.lundin7a926812016-05-12 13:51:28 -07001304 bool GetAudioReturnMuted() {
1305 bool muted;
1306 EXPECT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
1307 return muted;
1308 }
1309
1310 void GetAudioUntilMuted() {
1311 while (!GetAudioReturnMuted()) {
1312 ASSERT_LT(counter_++, 1000) << "Test timed out";
1313 }
1314 }
1315
1316 void GetAudioUntilNormal() {
1317 bool muted = false;
1318 while (out_frame_.speech_type_ != AudioFrame::kNormalSpeech) {
1319 EXPECT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
1320 ASSERT_LT(counter_++, 1000) << "Test timed out";
1321 }
1322 EXPECT_FALSE(muted);
1323 }
1324
1325 int counter_ = 0;
1326};
1327
1328// Verifies that NetEq goes in and out of muted state as expected.
1329TEST_F(NetEqDecodingTestWithMutedState, MutedState) {
1330 // Insert one speech packet.
1331 InsertPacket(0);
1332 // Pull out audio once and expect it not to be muted.
1333 EXPECT_FALSE(GetAudioReturnMuted());
1334 // Pull data until faded out.
1335 GetAudioUntilMuted();
henrik.lundina4491072017-07-06 05:23:53 -07001336 EXPECT_TRUE(out_frame_.muted());
henrik.lundin7a926812016-05-12 13:51:28 -07001337
1338 // Verify that output audio is not written during muted mode. Other parameters
1339 // should be correct, though.
1340 AudioFrame new_frame;
yujo36b1a5f2017-06-12 12:45:32 -07001341 int16_t* frame_data = new_frame.mutable_data();
1342 for (size_t i = 0; i < AudioFrame::kMaxDataSizeSamples; i++) {
1343 frame_data[i] = 17;
henrik.lundin7a926812016-05-12 13:51:28 -07001344 }
1345 bool muted;
1346 EXPECT_EQ(0, neteq_->GetAudio(&new_frame, &muted));
1347 EXPECT_TRUE(muted);
henrik.lundina4491072017-07-06 05:23:53 -07001348 EXPECT_TRUE(out_frame_.muted());
yujo36b1a5f2017-06-12 12:45:32 -07001349 for (size_t i = 0; i < AudioFrame::kMaxDataSizeSamples; i++) {
1350 EXPECT_EQ(17, frame_data[i]);
henrik.lundin7a926812016-05-12 13:51:28 -07001351 }
1352 EXPECT_EQ(out_frame_.timestamp_ + out_frame_.samples_per_channel_,
1353 new_frame.timestamp_);
1354 EXPECT_EQ(out_frame_.samples_per_channel_, new_frame.samples_per_channel_);
1355 EXPECT_EQ(out_frame_.sample_rate_hz_, new_frame.sample_rate_hz_);
1356 EXPECT_EQ(out_frame_.num_channels_, new_frame.num_channels_);
1357 EXPECT_EQ(out_frame_.speech_type_, new_frame.speech_type_);
1358 EXPECT_EQ(out_frame_.vad_activity_, new_frame.vad_activity_);
1359
1360 // Insert new data. Timestamp is corrected for the time elapsed since the last
1361 // packet. Verify that normal operation resumes.
1362 InsertPacket(kSamples * counter_);
1363 GetAudioUntilNormal();
henrik.lundina4491072017-07-06 05:23:53 -07001364 EXPECT_FALSE(out_frame_.muted());
henrik.lundin612c25e2016-05-25 08:21:04 -07001365
1366 NetEqNetworkStatistics stats;
1367 EXPECT_EQ(0, neteq_->NetworkStatistics(&stats));
1368 // NetEqNetworkStatistics::expand_rate tells the fraction of samples that were
1369 // concealment samples, in Q14 (16384 = 100%) .The vast majority should be
1370 // concealment samples in this test.
1371 EXPECT_GT(stats.expand_rate, 14000);
1372 // And, it should be greater than the speech_expand_rate.
1373 EXPECT_GT(stats.expand_rate, stats.speech_expand_rate);
henrik.lundin7a926812016-05-12 13:51:28 -07001374}
1375
1376// Verifies that NetEq goes out of muted state when given a delayed packet.
1377TEST_F(NetEqDecodingTestWithMutedState, MutedStateDelayedPacket) {
1378 // Insert one speech packet.
1379 InsertPacket(0);
1380 // Pull out audio once and expect it not to be muted.
1381 EXPECT_FALSE(GetAudioReturnMuted());
1382 // Pull data until faded out.
1383 GetAudioUntilMuted();
1384 // Insert new data. Timestamp is only corrected for the half of the time
1385 // elapsed since the last packet. That is, the new packet is delayed. Verify
1386 // that normal operation resumes.
1387 InsertPacket(kSamples * counter_ / 2);
1388 GetAudioUntilNormal();
1389}
1390
1391// Verifies that NetEq goes out of muted state when given a future packet.
1392TEST_F(NetEqDecodingTestWithMutedState, MutedStateFuturePacket) {
1393 // Insert one speech packet.
1394 InsertPacket(0);
1395 // Pull out audio once and expect it not to be muted.
1396 EXPECT_FALSE(GetAudioReturnMuted());
1397 // Pull data until faded out.
1398 GetAudioUntilMuted();
1399 // Insert new data. Timestamp is over-corrected for the time elapsed since the
1400 // last packet. That is, the new packet is too early. Verify that normal
1401 // operation resumes.
1402 InsertPacket(kSamples * counter_ * 2);
1403 GetAudioUntilNormal();
1404}
1405
1406// Verifies that NetEq goes out of muted state when given an old packet.
1407TEST_F(NetEqDecodingTestWithMutedState, MutedStateOldPacket) {
1408 // Insert one speech packet.
1409 InsertPacket(0);
1410 // Pull out audio once and expect it not to be muted.
1411 EXPECT_FALSE(GetAudioReturnMuted());
1412 // Pull data until faded out.
1413 GetAudioUntilMuted();
1414
1415 EXPECT_NE(AudioFrame::kNormalSpeech, out_frame_.speech_type_);
1416 // Insert packet which is older than the first packet.
1417 InsertPacket(kSamples * (counter_ - 1000));
1418 EXPECT_FALSE(GetAudioReturnMuted());
1419 EXPECT_EQ(AudioFrame::kNormalSpeech, out_frame_.speech_type_);
1420}
1421
henrik.lundin42feb512016-09-20 06:51:40 -07001422// Verifies that NetEq doesn't enter muted state when CNG mode is active and the
1423// packet stream is suspended for a long time.
1424TEST_F(NetEqDecodingTestWithMutedState, DoNotMuteExtendedCngWithoutPackets) {
1425 // Insert one CNG packet.
1426 InsertCngPacket(0);
1427
1428 // Pull 10 seconds of audio (10 ms audio generated per lap).
1429 for (int i = 0; i < 1000; ++i) {
1430 bool muted;
1431 EXPECT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
1432 ASSERT_FALSE(muted);
1433 }
1434 EXPECT_EQ(AudioFrame::kCNG, out_frame_.speech_type_);
1435}
1436
1437// Verifies that NetEq goes back to normal after a long CNG period with the
1438// packet stream suspended.
1439TEST_F(NetEqDecodingTestWithMutedState, RecoverAfterExtendedCngWithoutPackets) {
1440 // Insert one CNG packet.
1441 InsertCngPacket(0);
1442
1443 // Pull 10 seconds of audio (10 ms audio generated per lap).
1444 for (int i = 0; i < 1000; ++i) {
1445 bool muted;
1446 EXPECT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
1447 }
1448
1449 // Insert new data. Timestamp is corrected for the time elapsed since the last
1450 // packet. Verify that normal operation resumes.
1451 InsertPacket(kSamples * counter_);
1452 GetAudioUntilNormal();
1453}
1454
henrik.lundin7a926812016-05-12 13:51:28 -07001455class NetEqDecodingTestTwoInstances : public NetEqDecodingTest {
1456 public:
1457 NetEqDecodingTestTwoInstances() : NetEqDecodingTest() {}
1458
1459 void SetUp() override {
1460 NetEqDecodingTest::SetUp();
1461 config2_ = config_;
1462 }
1463
1464 void CreateSecondInstance() {
ossue3525782016-05-25 07:37:43 -07001465 neteq2_.reset(NetEq::Create(config2_, CreateBuiltinAudioDecoderFactory()));
henrik.lundin7a926812016-05-12 13:51:28 -07001466 ASSERT_TRUE(neteq2_);
1467 LoadDecoders(neteq2_.get());
1468 }
1469
1470 protected:
1471 std::unique_ptr<NetEq> neteq2_;
1472 NetEq::Config config2_;
1473};
1474
1475namespace {
1476::testing::AssertionResult AudioFramesEqualExceptData(const AudioFrame& a,
1477 const AudioFrame& b) {
1478 if (a.timestamp_ != b.timestamp_)
1479 return ::testing::AssertionFailure() << "timestamp_ diff (" << a.timestamp_
1480 << " != " << b.timestamp_ << ")";
1481 if (a.sample_rate_hz_ != b.sample_rate_hz_)
Yves Gerey665174f2018-06-19 15:03:05 +02001482 return ::testing::AssertionFailure()
1483 << "sample_rate_hz_ diff (" << a.sample_rate_hz_
1484 << " != " << b.sample_rate_hz_ << ")";
henrik.lundin7a926812016-05-12 13:51:28 -07001485 if (a.samples_per_channel_ != b.samples_per_channel_)
1486 return ::testing::AssertionFailure()
1487 << "samples_per_channel_ diff (" << a.samples_per_channel_
1488 << " != " << b.samples_per_channel_ << ")";
1489 if (a.num_channels_ != b.num_channels_)
Yves Gerey665174f2018-06-19 15:03:05 +02001490 return ::testing::AssertionFailure()
1491 << "num_channels_ diff (" << a.num_channels_
1492 << " != " << b.num_channels_ << ")";
henrik.lundin7a926812016-05-12 13:51:28 -07001493 if (a.speech_type_ != b.speech_type_)
Yves Gerey665174f2018-06-19 15:03:05 +02001494 return ::testing::AssertionFailure()
1495 << "speech_type_ diff (" << a.speech_type_
1496 << " != " << b.speech_type_ << ")";
henrik.lundin7a926812016-05-12 13:51:28 -07001497 if (a.vad_activity_ != b.vad_activity_)
Yves Gerey665174f2018-06-19 15:03:05 +02001498 return ::testing::AssertionFailure()
1499 << "vad_activity_ diff (" << a.vad_activity_
1500 << " != " << b.vad_activity_ << ")";
henrik.lundin7a926812016-05-12 13:51:28 -07001501 return ::testing::AssertionSuccess();
1502}
1503
1504::testing::AssertionResult AudioFramesEqual(const AudioFrame& a,
1505 const AudioFrame& b) {
1506 ::testing::AssertionResult res = AudioFramesEqualExceptData(a, b);
1507 if (!res)
1508 return res;
Yves Gerey665174f2018-06-19 15:03:05 +02001509 if (memcmp(a.data(), b.data(),
1510 a.samples_per_channel_ * a.num_channels_ * sizeof(*a.data())) !=
1511 0) {
henrik.lundin7a926812016-05-12 13:51:28 -07001512 return ::testing::AssertionFailure() << "data_ diff";
1513 }
1514 return ::testing::AssertionSuccess();
1515}
1516
1517} // namespace
1518
1519TEST_F(NetEqDecodingTestTwoInstances, CompareMutedStateOnOff) {
1520 ASSERT_FALSE(config_.enable_muted_state);
1521 config2_.enable_muted_state = true;
1522 CreateSecondInstance();
1523
1524 // Insert one speech packet into both NetEqs.
1525 const size_t kSamples = 10 * 16;
1526 const size_t kPayloadBytes = kSamples * 2;
1527 uint8_t payload[kPayloadBytes] = {0};
henrik.lundin246ef3e2017-04-24 09:14:32 -07001528 RTPHeader rtp_info;
henrik.lundin7a926812016-05-12 13:51:28 -07001529 PopulateRtpInfo(0, 0, &rtp_info);
henrik.lundin246ef3e2017-04-24 09:14:32 -07001530 EXPECT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
1531 EXPECT_EQ(0, neteq2_->InsertPacket(rtp_info, payload, 0));
henrik.lundin7a926812016-05-12 13:51:28 -07001532
1533 AudioFrame out_frame1, out_frame2;
1534 bool muted;
1535 for (int i = 0; i < 1000; ++i) {
Jonas Olsson366a50c2018-09-06 13:41:30 +02001536 rtc::StringBuilder ss;
henrik.lundin7a926812016-05-12 13:51:28 -07001537 ss << "i = " << i;
1538 SCOPED_TRACE(ss.str()); // Print out the loop iterator on failure.
1539 EXPECT_EQ(0, neteq_->GetAudio(&out_frame1, &muted));
1540 EXPECT_FALSE(muted);
1541 EXPECT_EQ(0, neteq2_->GetAudio(&out_frame2, &muted));
1542 if (muted) {
1543 EXPECT_TRUE(AudioFramesEqualExceptData(out_frame1, out_frame2));
1544 } else {
1545 EXPECT_TRUE(AudioFramesEqual(out_frame1, out_frame2));
1546 }
1547 }
1548 EXPECT_TRUE(muted);
1549
1550 // Insert new data. Timestamp is corrected for the time elapsed since the last
1551 // packet.
1552 PopulateRtpInfo(0, kSamples * 1000, &rtp_info);
henrik.lundin246ef3e2017-04-24 09:14:32 -07001553 EXPECT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
1554 EXPECT_EQ(0, neteq2_->InsertPacket(rtp_info, payload, 0));
henrik.lundin7a926812016-05-12 13:51:28 -07001555
1556 int counter = 0;
1557 while (out_frame1.speech_type_ != AudioFrame::kNormalSpeech) {
1558 ASSERT_LT(counter++, 1000) << "Test timed out";
Jonas Olsson366a50c2018-09-06 13:41:30 +02001559 rtc::StringBuilder ss;
henrik.lundin7a926812016-05-12 13:51:28 -07001560 ss << "counter = " << counter;
1561 SCOPED_TRACE(ss.str()); // Print out the loop iterator on failure.
1562 EXPECT_EQ(0, neteq_->GetAudio(&out_frame1, &muted));
1563 EXPECT_FALSE(muted);
1564 EXPECT_EQ(0, neteq2_->GetAudio(&out_frame2, &muted));
1565 if (muted) {
1566 EXPECT_TRUE(AudioFramesEqualExceptData(out_frame1, out_frame2));
1567 } else {
1568 EXPECT_TRUE(AudioFramesEqual(out_frame1, out_frame2));
1569 }
1570 }
1571 EXPECT_FALSE(muted);
1572}
1573
henrik.lundin114c1b32017-04-26 07:47:32 -07001574TEST_F(NetEqDecodingTest, LastDecodedTimestampsEmpty) {
1575 EXPECT_TRUE(neteq_->LastDecodedTimestamps().empty());
1576
1577 // Pull out data once.
1578 AudioFrame output;
1579 bool muted;
1580 ASSERT_EQ(0, neteq_->GetAudio(&output, &muted));
1581
1582 EXPECT_TRUE(neteq_->LastDecodedTimestamps().empty());
1583}
1584
1585TEST_F(NetEqDecodingTest, LastDecodedTimestampsOneDecoded) {
1586 // Insert one packet with PCM16b WB data (this is what PopulateRtpInfo does by
1587 // default). Make the length 10 ms.
1588 constexpr size_t kPayloadSamples = 16 * 10;
1589 constexpr size_t kPayloadBytes = 2 * kPayloadSamples;
1590 uint8_t payload[kPayloadBytes] = {0};
1591
1592 RTPHeader rtp_info;
1593 constexpr uint32_t kRtpTimestamp = 0x1234;
1594 PopulateRtpInfo(0, kRtpTimestamp, &rtp_info);
1595 EXPECT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
1596
1597 // Pull out data once.
1598 AudioFrame output;
1599 bool muted;
1600 ASSERT_EQ(0, neteq_->GetAudio(&output, &muted));
1601
1602 EXPECT_EQ(std::vector<uint32_t>({kRtpTimestamp}),
1603 neteq_->LastDecodedTimestamps());
1604
1605 // Nothing decoded on the second call.
1606 ASSERT_EQ(0, neteq_->GetAudio(&output, &muted));
1607 EXPECT_TRUE(neteq_->LastDecodedTimestamps().empty());
1608}
1609
1610TEST_F(NetEqDecodingTest, LastDecodedTimestampsTwoDecoded) {
1611 // Insert two packets with PCM16b WB data (this is what PopulateRtpInfo does
1612 // by default). Make the length 5 ms so that NetEq must decode them both in
1613 // the same GetAudio call.
1614 constexpr size_t kPayloadSamples = 16 * 5;
1615 constexpr size_t kPayloadBytes = 2 * kPayloadSamples;
1616 uint8_t payload[kPayloadBytes] = {0};
1617
1618 RTPHeader rtp_info;
1619 constexpr uint32_t kRtpTimestamp1 = 0x1234;
1620 PopulateRtpInfo(0, kRtpTimestamp1, &rtp_info);
1621 EXPECT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
1622 constexpr uint32_t kRtpTimestamp2 = kRtpTimestamp1 + kPayloadSamples;
1623 PopulateRtpInfo(1, kRtpTimestamp2, &rtp_info);
1624 EXPECT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
1625
1626 // Pull out data once.
1627 AudioFrame output;
1628 bool muted;
1629 ASSERT_EQ(0, neteq_->GetAudio(&output, &muted));
1630
1631 EXPECT_EQ(std::vector<uint32_t>({kRtpTimestamp1, kRtpTimestamp2}),
1632 neteq_->LastDecodedTimestamps());
1633}
1634
Gustaf Ullberg9a2e9062017-09-18 09:28:20 +02001635TEST_F(NetEqDecodingTest, TestConcealmentEvents) {
1636 const int kNumConcealmentEvents = 19;
1637 const size_t kSamples = 10 * 16;
1638 const size_t kPayloadBytes = kSamples * 2;
1639 int seq_no = 0;
1640 RTPHeader rtp_info;
1641 rtp_info.ssrc = 0x1234; // Just an arbitrary SSRC.
1642 rtp_info.payloadType = 94; // PCM16b WB codec.
1643 rtp_info.markerBit = 0;
1644 const uint8_t payload[kPayloadBytes] = {0};
1645 bool muted;
1646
1647 for (int i = 0; i < kNumConcealmentEvents; i++) {
1648 // Insert some packets of 10 ms size.
1649 for (int j = 0; j < 10; j++) {
1650 rtp_info.sequenceNumber = seq_no++;
1651 rtp_info.timestamp = rtp_info.sequenceNumber * kSamples;
1652 neteq_->InsertPacket(rtp_info, payload, 0);
1653 neteq_->GetAudio(&out_frame_, &muted);
1654 }
1655
1656 // Lose a number of packets.
1657 int num_lost = 1 + i;
1658 for (int j = 0; j < num_lost; j++) {
1659 seq_no++;
1660 neteq_->GetAudio(&out_frame_, &muted);
1661 }
1662 }
1663
1664 // Check number of concealment events.
1665 NetEqLifetimeStatistics stats = neteq_->GetLifetimeStatistics();
1666 EXPECT_EQ(kNumConcealmentEvents, static_cast<int>(stats.concealment_events));
1667}
1668
Gustaf Ullbergb0a02072017-10-02 12:00:34 +02001669// Test that the jitter buffer delay stat is computed correctly.
1670void NetEqDecodingTestFaxMode::TestJitterBufferDelay(bool apply_packet_loss) {
1671 const int kNumPackets = 10;
1672 const int kDelayInNumPackets = 2;
1673 const int kPacketLenMs = 10; // All packets are of 10 ms size.
1674 const size_t kSamples = kPacketLenMs * 16;
1675 const size_t kPayloadBytes = kSamples * 2;
1676 RTPHeader rtp_info;
1677 rtp_info.ssrc = 0x1234; // Just an arbitrary SSRC.
1678 rtp_info.payloadType = 94; // PCM16b WB codec.
1679 rtp_info.markerBit = 0;
1680 const uint8_t payload[kPayloadBytes] = {0};
1681 bool muted;
1682 int packets_sent = 0;
1683 int packets_received = 0;
1684 int expected_delay = 0;
1685 while (packets_received < kNumPackets) {
1686 // Insert packet.
1687 if (packets_sent < kNumPackets) {
1688 rtp_info.sequenceNumber = packets_sent++;
1689 rtp_info.timestamp = rtp_info.sequenceNumber * kSamples;
1690 neteq_->InsertPacket(rtp_info, payload, 0);
1691 }
1692
1693 // Get packet.
1694 if (packets_sent > kDelayInNumPackets) {
1695 neteq_->GetAudio(&out_frame_, &muted);
1696 packets_received++;
1697
1698 // The delay reported by the jitter buffer never exceeds
1699 // the number of samples previously fetched with GetAudio
1700 // (hence the min()).
1701 int packets_delay = std::min(packets_received, kDelayInNumPackets + 1);
1702
1703 // The increase of the expected delay is the product of
1704 // the current delay of the jitter buffer in ms * the
1705 // number of samples that are sent for play out.
1706 int current_delay_ms = packets_delay * kPacketLenMs;
1707 expected_delay += current_delay_ms * kSamples;
1708 }
1709 }
1710
1711 if (apply_packet_loss) {
1712 // Extra call to GetAudio to cause concealment.
1713 neteq_->GetAudio(&out_frame_, &muted);
1714 }
1715
1716 // Check jitter buffer delay.
1717 NetEqLifetimeStatistics stats = neteq_->GetLifetimeStatistics();
1718 EXPECT_EQ(expected_delay, static_cast<int>(stats.jitter_buffer_delay_ms));
1719}
1720
1721TEST_F(NetEqDecodingTestFaxMode, TestJitterBufferDelayWithoutLoss) {
1722 TestJitterBufferDelay(false);
1723}
1724
1725TEST_F(NetEqDecodingTestFaxMode, TestJitterBufferDelayWithLoss) {
1726 TestJitterBufferDelay(true);
1727}
1728
Henrik Lundin7687ad52018-07-02 10:14:46 +02001729namespace test {
Henrik Lundin7687ad52018-07-02 10:14:46 +02001730TEST(NetEqNoTimeStretchingMode, RunTest) {
1731 NetEq::Config config;
1732 config.for_test_no_time_stretching = true;
1733 auto codecs = NetEqTest::StandardDecoderMap();
1734 NetEqTest::ExtDecoderMap ext_codecs;
1735 NetEqPacketSourceInput::RtpHeaderExtensionMap rtp_ext_map = {
1736 {1, kRtpExtensionAudioLevel},
1737 {3, kRtpExtensionAbsoluteSendTime},
1738 {5, kRtpExtensionTransportSequenceNumber},
1739 {7, kRtpExtensionVideoContentType},
1740 {8, kRtpExtensionVideoTiming}};
1741 std::unique_ptr<NetEqInput> input(new NetEqRtpDumpInput(
1742 webrtc::test::ResourcePath("audio_coding/neteq_universal_new", "rtp"),
Bjorn Terelius5350d1c2018-10-11 16:51:23 +02001743 rtp_ext_map, absl::nullopt /*No SSRC filter*/));
Henrik Lundin7687ad52018-07-02 10:14:46 +02001744 std::unique_ptr<TimeLimitedNetEqInput> input_time_limit(
1745 new TimeLimitedNetEqInput(std::move(input), 20000));
1746 std::unique_ptr<AudioSink> output(new VoidAudioSink);
1747 NetEqTest::Callbacks callbacks;
1748 NetEqTest test(config, codecs, ext_codecs, std::move(input_time_limit),
1749 std::move(output), callbacks);
1750 test.Run();
1751 const auto stats = test.SimulationStats();
1752 EXPECT_EQ(0, stats.accelerate_rate);
1753 EXPECT_EQ(0, stats.preemptive_rate);
1754}
Henrik Lundin7687ad52018-07-02 10:14:46 +02001755
1756} // namespace test
henrik.lundin@webrtc.orge7ce4372014-01-09 14:01:55 +00001757} // namespace webrtc