blob: 9f7d04d4c1565365abc696ca84e050d1d4a280f5 [file] [log] [blame]
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +00001/*
2 * Copyright (c) 2011 The WebRTC project authors. All Rights Reserved.
3 *
4 * Use of this source code is governed by a BSD-style license
5 * that can be found in the LICENSE file in the root of the source
6 * tree. An additional intellectual property rights grant can be found
7 * in the file PATENTS. All contributing project authors may
8 * be found in the AUTHORS file in the root of the source tree.
9 */
10
Mirko Bonadei92ea95e2017-09-15 06:47:31 +020011#include "modules/audio_coding/neteq/include/neteq.h"
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +000012
pbos@webrtc.org3ecc1622014-03-07 15:23:34 +000013#include <math.h>
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +000014#include <stdlib.h>
15#include <string.h> // memset
16
turaj@webrtc.org8d1cdaa2014-04-11 18:47:55 +000017#include <algorithm>
kwiberg2d0c3322016-02-14 09:28:33 -080018#include <memory>
turaj@webrtc.org78b41a02013-11-22 20:27:07 +000019#include <set>
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +000020#include <string>
21#include <vector>
22
Fredrik Solenbergbbf21a32018-04-12 22:44:09 +020023#include "api/audio/audio_frame.h"
Mirko Bonadei92ea95e2017-09-15 06:47:31 +020024#include "api/audio_codecs/builtin_audio_decoder_factory.h"
Mirko Bonadei92ea95e2017-09-15 06:47:31 +020025#include "modules/audio_coding/codecs/pcm16b/pcm16b.h"
26#include "modules/audio_coding/neteq/tools/audio_loop.h"
Henrik Lundin7687ad52018-07-02 10:14:46 +020027#include "modules/audio_coding/neteq/tools/neteq_packet_source_input.h"
28#include "modules/audio_coding/neteq/tools/neteq_test.h"
Mirko Bonadei92ea95e2017-09-15 06:47:31 +020029#include "modules/audio_coding/neteq/tools/rtp_file_source.h"
Yves Gerey3e707812018-11-28 16:47:49 +010030#include "modules/include/module_common_types_public.h"
Niels Möller53382cb2018-11-27 14:05:08 +010031#include "modules/rtp_rtcp/include/rtcp_statistics.h"
Yves Gerey3e707812018-11-28 16:47:49 +010032#include "modules/rtp_rtcp/include/rtp_rtcp_defines.h"
Mirko Bonadei92ea95e2017-09-15 06:47:31 +020033#include "rtc_base/ignore_wundef.h"
Steve Anton10542f22019-01-11 09:11:00 -080034#include "rtc_base/message_digest.h"
Karl Wiberge40468b2017-11-22 10:42:26 +010035#include "rtc_base/numerics/safe_conversions.h"
Steve Anton10542f22019-01-11 09:11:00 -080036#include "rtc_base/string_encode.h"
Jonas Olsson366a50c2018-09-06 13:41:30 +020037#include "rtc_base/strings/string_builder.h"
Niels Möllera12c42a2018-07-25 16:05:48 +020038#include "rtc_base/system/arch.h"
Henrik Lundine9619f82017-11-27 14:05:27 +010039#include "test/field_trial.h"
Mirko Bonadei92ea95e2017-09-15 06:47:31 +020040#include "test/gtest.h"
Steve Anton10542f22019-01-11 09:11:00 -080041#include "test/testsupport/file_utils.h"
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +000042
Mirko Bonadei81ca3bf2018-01-09 09:40:39 +010043// This must come after test/gtest.h
44#include "rtc_base/flags.h" // NOLINT(build/include)
45
minyue5f026d02015-12-16 07:36:04 -080046#ifdef WEBRTC_NETEQ_UNITTEST_BITEXACT
kwiberg77eab702016-09-28 17:42:01 -070047RTC_PUSH_IGNORING_WUNDEF()
minyue5f026d02015-12-16 07:36:04 -080048#ifdef WEBRTC_ANDROID_PLATFORM_BUILD
49#include "external/webrtc/webrtc/modules/audio_coding/neteq/neteq_unittest.pb.h"
50#else
Mirko Bonadei92ea95e2017-09-15 06:47:31 +020051#include "modules/audio_coding/neteq/neteq_unittest.pb.h"
minyue5f026d02015-12-16 07:36:04 -080052#endif
kwiberg77eab702016-09-28 17:42:01 -070053RTC_POP_IGNORING_WUNDEF()
minyue5f026d02015-12-16 07:36:04 -080054#endif
55
Mirko Bonadei2dfa9982018-10-18 11:35:32 +020056WEBRTC_DEFINE_bool(gen_ref, false, "Generate reference files.");
turaj@webrtc.orga6101d72013-10-01 22:01:09 +000057
kwiberg5adaf732016-10-04 09:33:27 -070058namespace webrtc {
59
minyue5f026d02015-12-16 07:36:04 -080060namespace {
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +000061
minyue4f906772016-04-29 11:05:14 -070062const std::string& PlatformChecksum(const std::string& checksum_general,
Henrik Lundin8cd750d2017-10-12 13:07:11 +020063 const std::string& checksum_android_32,
64 const std::string& checksum_android_64,
minyue4f906772016-04-29 11:05:14 -070065 const std::string& checksum_win_32,
66 const std::string& checksum_win_64) {
kwiberg77eab702016-09-28 17:42:01 -070067#if defined(WEBRTC_ANDROID)
Yves Gerey665174f2018-06-19 15:03:05 +020068#ifdef WEBRTC_ARCH_64_BITS
69 return checksum_android_64;
70#else
71 return checksum_android_32;
72#endif // WEBRTC_ARCH_64_BITS
kwiberg77eab702016-09-28 17:42:01 -070073#elif defined(WEBRTC_WIN)
Yves Gerey665174f2018-06-19 15:03:05 +020074#ifdef WEBRTC_ARCH_64_BITS
75 return checksum_win_64;
76#else
77 return checksum_win_32;
78#endif // WEBRTC_ARCH_64_BITS
minyue4f906772016-04-29 11:05:14 -070079#else
80 return checksum_general;
81#endif // WEBRTC_WIN
82}
83
minyue5f026d02015-12-16 07:36:04 -080084#ifdef WEBRTC_NETEQ_UNITTEST_BITEXACT
85void Convert(const webrtc::NetEqNetworkStatistics& stats_raw,
86 webrtc::neteq_unittest::NetEqNetworkStatistics* stats) {
87 stats->set_current_buffer_size_ms(stats_raw.current_buffer_size_ms);
88 stats->set_preferred_buffer_size_ms(stats_raw.preferred_buffer_size_ms);
89 stats->set_jitter_peaks_found(stats_raw.jitter_peaks_found);
90 stats->set_packet_loss_rate(stats_raw.packet_loss_rate);
minyue5f026d02015-12-16 07:36:04 -080091 stats->set_expand_rate(stats_raw.expand_rate);
92 stats->set_speech_expand_rate(stats_raw.speech_expand_rate);
93 stats->set_preemptive_rate(stats_raw.preemptive_rate);
94 stats->set_accelerate_rate(stats_raw.accelerate_rate);
95 stats->set_secondary_decoded_rate(stats_raw.secondary_decoded_rate);
minyue-webrtc0c3ca752017-08-23 15:59:38 +020096 stats->set_secondary_discarded_rate(stats_raw.secondary_discarded_rate);
minyue5f026d02015-12-16 07:36:04 -080097 stats->set_clockdrift_ppm(stats_raw.clockdrift_ppm);
98 stats->set_added_zero_samples(stats_raw.added_zero_samples);
99 stats->set_mean_waiting_time_ms(stats_raw.mean_waiting_time_ms);
100 stats->set_median_waiting_time_ms(stats_raw.median_waiting_time_ms);
101 stats->set_min_waiting_time_ms(stats_raw.min_waiting_time_ms);
102 stats->set_max_waiting_time_ms(stats_raw.max_waiting_time_ms);
103}
104
105void Convert(const webrtc::RtcpStatistics& stats_raw,
106 webrtc::neteq_unittest::RtcpStatistics* stats) {
107 stats->set_fraction_lost(stats_raw.fraction_lost);
srte186d9c32017-08-04 05:03:53 -0700108 stats->set_cumulative_lost(stats_raw.packets_lost);
minyue5f026d02015-12-16 07:36:04 -0800109 stats->set_extended_max_sequence_number(
srte186d9c32017-08-04 05:03:53 -0700110 stats_raw.extended_highest_sequence_number);
minyue5f026d02015-12-16 07:36:04 -0800111 stats->set_jitter(stats_raw.jitter);
112}
113
Yves Gerey665174f2018-06-19 15:03:05 +0200114void AddMessage(FILE* file,
115 rtc::MessageDigest* digest,
minyue4f906772016-04-29 11:05:14 -0700116 const std::string& message) {
minyue5f026d02015-12-16 07:36:04 -0800117 int32_t size = message.length();
minyue4f906772016-04-29 11:05:14 -0700118 if (file)
119 ASSERT_EQ(1u, fwrite(&size, sizeof(size), 1, file));
120 digest->Update(&size, sizeof(size));
121
122 if (file)
123 ASSERT_EQ(static_cast<size_t>(size),
124 fwrite(message.data(), sizeof(char), size, file));
125 digest->Update(message.data(), sizeof(char) * size);
minyue5f026d02015-12-16 07:36:04 -0800126}
127
minyue5f026d02015-12-16 07:36:04 -0800128#endif // WEBRTC_NETEQ_UNITTEST_BITEXACT
129
henrik.lundin7a926812016-05-12 13:51:28 -0700130void LoadDecoders(webrtc::NetEq* neteq) {
kwiberg5adaf732016-10-04 09:33:27 -0700131 ASSERT_EQ(true,
132 neteq->RegisterPayloadType(0, SdpAudioFormat("pcmu", 8000, 1)));
Niels Möller05543682019-01-10 16:55:06 +0100133 ASSERT_EQ(true,
134 neteq->RegisterPayloadType(8, SdpAudioFormat("pcma", 8000, 1)));
henrik.lundin7a926812016-05-12 13:51:28 -0700135#ifdef WEBRTC_CODEC_ILBC
kwiberg5adaf732016-10-04 09:33:27 -0700136 ASSERT_EQ(true,
137 neteq->RegisterPayloadType(102, SdpAudioFormat("ilbc", 8000, 1)));
henrik.lundin7a926812016-05-12 13:51:28 -0700138#endif
139#if defined(WEBRTC_CODEC_ISAC) || defined(WEBRTC_CODEC_ISACFX)
kwiberg5adaf732016-10-04 09:33:27 -0700140 ASSERT_EQ(true,
141 neteq->RegisterPayloadType(103, SdpAudioFormat("isac", 16000, 1)));
henrik.lundin7a926812016-05-12 13:51:28 -0700142#endif
143#ifdef WEBRTC_CODEC_ISAC
kwiberg5adaf732016-10-04 09:33:27 -0700144 ASSERT_EQ(true,
145 neteq->RegisterPayloadType(104, SdpAudioFormat("isac", 32000, 1)));
henrik.lundin7a926812016-05-12 13:51:28 -0700146#endif
147#ifdef WEBRTC_CODEC_OPUS
kwiberg5adaf732016-10-04 09:33:27 -0700148 ASSERT_EQ(true,
149 neteq->RegisterPayloadType(
150 111, SdpAudioFormat("opus", 48000, 2, {{"stereo", "0"}})));
henrik.lundin7a926812016-05-12 13:51:28 -0700151#endif
kwiberg5adaf732016-10-04 09:33:27 -0700152 ASSERT_EQ(true,
153 neteq->RegisterPayloadType(93, SdpAudioFormat("L16", 8000, 1)));
154 ASSERT_EQ(true,
155 neteq->RegisterPayloadType(94, SdpAudioFormat("L16", 16000, 1)));
156 ASSERT_EQ(true,
157 neteq->RegisterPayloadType(95, SdpAudioFormat("L16", 32000, 1)));
158 ASSERT_EQ(true,
159 neteq->RegisterPayloadType(13, SdpAudioFormat("cn", 8000, 1)));
160 ASSERT_EQ(true,
161 neteq->RegisterPayloadType(98, SdpAudioFormat("cn", 16000, 1)));
henrik.lundin7a926812016-05-12 13:51:28 -0700162}
minyue5f026d02015-12-16 07:36:04 -0800163} // namespace
164
minyue4f906772016-04-29 11:05:14 -0700165class ResultSink {
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000166 public:
minyue4f906772016-04-29 11:05:14 -0700167 explicit ResultSink(const std::string& output_file);
168 ~ResultSink();
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000169
Yves Gerey665174f2018-06-19 15:03:05 +0200170 template <typename T>
171 void AddResult(const T* test_results, size_t length);
minyue4f906772016-04-29 11:05:14 -0700172
173 void AddResult(const NetEqNetworkStatistics& stats);
174 void AddResult(const RtcpStatistics& stats);
175
176 void VerifyChecksum(const std::string& ref_check_sum);
177
178 private:
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000179 FILE* output_fp_;
minyue4f906772016-04-29 11:05:14 -0700180 std::unique_ptr<rtc::MessageDigest> digest_;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000181};
182
Joachim Bauch4e909192017-12-19 22:27:51 +0100183ResultSink::ResultSink(const std::string& output_file)
minyue4f906772016-04-29 11:05:14 -0700184 : output_fp_(nullptr),
Joachim Bauch4e909192017-12-19 22:27:51 +0100185 digest_(rtc::MessageDigestFactory::Create(rtc::DIGEST_SHA_1)) {
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000186 if (!output_file.empty()) {
187 output_fp_ = fopen(output_file.c_str(), "wb");
188 EXPECT_TRUE(output_fp_ != NULL);
189 }
190}
191
minyue4f906772016-04-29 11:05:14 -0700192ResultSink::~ResultSink() {
193 if (output_fp_)
194 fclose(output_fp_);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000195}
196
Yves Gerey665174f2018-06-19 15:03:05 +0200197template <typename T>
yujo36b1a5f2017-06-12 12:45:32 -0700198void ResultSink::AddResult(const T* test_results, size_t length) {
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000199 if (output_fp_) {
yujo36b1a5f2017-06-12 12:45:32 -0700200 ASSERT_EQ(length, fwrite(test_results, sizeof(T), length, output_fp_));
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000201 }
yujo36b1a5f2017-06-12 12:45:32 -0700202 digest_->Update(test_results, sizeof(T) * length);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000203}
204
minyue4f906772016-04-29 11:05:14 -0700205void ResultSink::AddResult(const NetEqNetworkStatistics& stats_raw) {
minyue5f026d02015-12-16 07:36:04 -0800206#ifdef WEBRTC_NETEQ_UNITTEST_BITEXACT
minyue5f026d02015-12-16 07:36:04 -0800207 neteq_unittest::NetEqNetworkStatistics stats;
208 Convert(stats_raw, &stats);
209
Mirko Bonadeie45c6882019-02-16 09:59:29 +0100210 std::string stats_string;
minyue5f026d02015-12-16 07:36:04 -0800211 ASSERT_TRUE(stats.SerializeToString(&stats_string));
minyue4f906772016-04-29 11:05:14 -0700212 AddMessage(output_fp_, digest_.get(), stats_string);
minyue5f026d02015-12-16 07:36:04 -0800213#else
214 FAIL() << "Writing to reference file requires Proto Buffer.";
215#endif // WEBRTC_NETEQ_UNITTEST_BITEXACT
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000216}
217
minyue4f906772016-04-29 11:05:14 -0700218void ResultSink::AddResult(const RtcpStatistics& stats_raw) {
minyue5f026d02015-12-16 07:36:04 -0800219#ifdef WEBRTC_NETEQ_UNITTEST_BITEXACT
minyue5f026d02015-12-16 07:36:04 -0800220 neteq_unittest::RtcpStatistics stats;
221 Convert(stats_raw, &stats);
222
Mirko Bonadeie45c6882019-02-16 09:59:29 +0100223 std::string stats_string;
minyue5f026d02015-12-16 07:36:04 -0800224 ASSERT_TRUE(stats.SerializeToString(&stats_string));
minyue4f906772016-04-29 11:05:14 -0700225 AddMessage(output_fp_, digest_.get(), stats_string);
minyue5f026d02015-12-16 07:36:04 -0800226#else
227 FAIL() << "Writing to reference file requires Proto Buffer.";
228#endif // WEBRTC_NETEQ_UNITTEST_BITEXACT
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000229}
230
minyue4f906772016-04-29 11:05:14 -0700231void ResultSink::VerifyChecksum(const std::string& checksum) {
232 std::vector<char> buffer;
233 buffer.resize(digest_->Size());
234 digest_->Finish(&buffer[0], buffer.size());
235 const std::string result = rtc::hex_encode(&buffer[0], digest_->Size());
Yves Gereya038e712018-11-14 10:45:50 +0100236 if (checksum.size() == result.size()) {
237 EXPECT_EQ(checksum, result);
238 } else {
239 // Check result is one the '|'-separated checksums.
240 EXPECT_NE(checksum.find(result), std::string::npos)
241 << result << " should be one of these:\n"
242 << checksum;
243 }
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000244}
245
246class NetEqDecodingTest : public ::testing::Test {
247 protected:
248 // NetEQ must be polled for data once every 10 ms. Thus, neither of the
249 // constants below can be changed.
250 static const int kTimeStepMs = 10;
Peter Kastingdce40cf2015-08-24 14:52:23 -0700251 static const size_t kBlockSize8kHz = kTimeStepMs * 8;
252 static const size_t kBlockSize16kHz = kTimeStepMs * 16;
253 static const size_t kBlockSize32kHz = kTimeStepMs * 32;
minyue93c08b72015-12-22 09:57:41 -0800254 static const size_t kBlockSize48kHz = kTimeStepMs * 48;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000255 static const int kInitSampleRateHz = 8000;
256
257 NetEqDecodingTest();
258 virtual void SetUp();
259 virtual void TearDown();
Yves Gerey665174f2018-06-19 15:03:05 +0200260 void OpenInputFile(const std::string& rtp_file);
henrik.lundin6d8e0112016-03-04 10:34:21 -0800261 void Process();
minyue5f026d02015-12-16 07:36:04 -0800262
henrik.lundin@webrtc.org4e4b0982014-08-11 14:48:49 +0000263 void DecodeAndCompare(const std::string& rtp_file,
minyue4f906772016-04-29 11:05:14 -0700264 const std::string& output_checksum,
265 const std::string& network_stats_checksum,
minyue4f906772016-04-29 11:05:14 -0700266 bool gen_ref);
minyue5f026d02015-12-16 07:36:04 -0800267
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000268 static void PopulateRtpInfo(int frame_index,
269 int timestamp,
henrik.lundin246ef3e2017-04-24 09:14:32 -0700270 RTPHeader* rtp_info);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000271 static void PopulateCng(int frame_index,
272 int timestamp,
henrik.lundin246ef3e2017-04-24 09:14:32 -0700273 RTPHeader* rtp_info,
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000274 uint8_t* payload,
pkasting@chromium.org4591fbd2014-11-20 22:28:14 +0000275 size_t* payload_len);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000276
Yves Gerey665174f2018-06-19 15:03:05 +0200277 void WrapTest(uint16_t start_seq_no,
278 uint32_t start_timestamp,
turaj@webrtc.org78b41a02013-11-22 20:27:07 +0000279 const std::set<uint16_t>& drop_seq_numbers,
Yves Gerey665174f2018-06-19 15:03:05 +0200280 bool expect_seq_no_wrap,
281 bool expect_timestamp_wrap);
turaj@webrtc.org78b41a02013-11-22 20:27:07 +0000282
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000283 void LongCngWithClockDrift(double drift_factor,
284 double network_freeze_ms,
285 bool pull_audio_during_freeze,
286 int delay_tolerance_ms,
287 int max_time_to_speech_ms);
288
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +0000289 void DuplicateCng();
henrik.lundin@webrtc.orgfcfc6a92014-02-13 11:42:28 +0000290
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000291 NetEq* neteq_;
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +0000292 NetEq::Config config_;
kwiberg2d0c3322016-02-14 09:28:33 -0800293 std::unique_ptr<test::RtpFileSource> rtp_source_;
294 std::unique_ptr<test::Packet> packet_;
Ivo Creusen24192c22019-07-12 17:00:25 +0200295 unsigned int sim_clock_;
henrik.lundin6d8e0112016-03-04 10:34:21 -0800296 AudioFrame out_frame_;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000297 int output_sample_rate_;
turaj@webrtc.org8d1cdaa2014-04-11 18:47:55 +0000298 int algorithmic_delay_ms_;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000299};
300
301// Allocating the static const so that it can be passed by reference.
302const int NetEqDecodingTest::kTimeStepMs;
Peter Kastingdce40cf2015-08-24 14:52:23 -0700303const size_t NetEqDecodingTest::kBlockSize8kHz;
304const size_t NetEqDecodingTest::kBlockSize16kHz;
305const size_t NetEqDecodingTest::kBlockSize32kHz;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000306const int NetEqDecodingTest::kInitSampleRateHz;
307
308NetEqDecodingTest::NetEqDecodingTest()
Ivo Creusen24192c22019-07-12 17:00:25 +0200309 : neteq_(NULL),
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +0000310 config_(),
Ivo Creusen24192c22019-07-12 17:00:25 +0200311 sim_clock_(0),
turaj@webrtc.org8d1cdaa2014-04-11 18:47:55 +0000312 output_sample_rate_(kInitSampleRateHz),
313 algorithmic_delay_ms_(0) {
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +0000314 config_.sample_rate_hz = kInitSampleRateHz;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000315}
316
317void NetEqDecodingTest::SetUp() {
Ivo Creusen24192c22019-07-12 17:00:25 +0200318 neteq_ = NetEq::Create(config_, CreateBuiltinAudioDecoderFactory());
turaj@webrtc.org8d1cdaa2014-04-11 18:47:55 +0000319 NetEqNetworkStatistics stat;
320 ASSERT_EQ(0, neteq_->NetworkStatistics(&stat));
321 algorithmic_delay_ms_ = stat.current_buffer_size_ms;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000322 ASSERT_TRUE(neteq_);
henrik.lundin7a926812016-05-12 13:51:28 -0700323 LoadDecoders(neteq_);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000324}
325
326void NetEqDecodingTest::TearDown() {
327 delete neteq_;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000328}
329
Yves Gerey665174f2018-06-19 15:03:05 +0200330void NetEqDecodingTest::OpenInputFile(const std::string& rtp_file) {
henrik.lundin@webrtc.org966a7082014-11-17 09:08:38 +0000331 rtp_source_.reset(test::RtpFileSource::Create(rtp_file));
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000332}
333
henrik.lundin6d8e0112016-03-04 10:34:21 -0800334void NetEqDecodingTest::Process() {
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000335 // Check if time to receive.
Ivo Creusen24192c22019-07-12 17:00:25 +0200336 while (packet_ && sim_clock_ >= packet_->time_ms()) {
henrik.lundin@webrtc.org966a7082014-11-17 09:08:38 +0000337 if (packet_->payload_length_bytes() > 0) {
ivoc72c08ed2016-01-20 07:26:24 -0800338#ifndef WEBRTC_CODEC_ISAC
339 // Ignore payload type 104 (iSAC-swb) if ISAC is not supported.
henrik.lundin246ef3e2017-04-24 09:14:32 -0700340 if (packet_->header().payloadType != 104)
ivoc72c08ed2016-01-20 07:26:24 -0800341#endif
Henrik Lundin70c09bd2017-04-24 15:56:56 +0200342 ASSERT_EQ(0,
343 neteq_->InsertPacket(
henrik.lundin246ef3e2017-04-24 09:14:32 -0700344 packet_->header(),
Henrik Lundin70c09bd2017-04-24 15:56:56 +0200345 rtc::ArrayView<const uint8_t>(
346 packet_->payload(), packet_->payload_length_bytes()),
347 static_cast<uint32_t>(packet_->time_ms() *
348 (output_sample_rate_ / 1000))));
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000349 }
350 // Get next packet.
henrik.lundin46ba49c2016-05-24 22:50:47 -0700351 packet_ = rtp_source_->NextPacket();
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000352 }
353
henrik.lundin@webrtc.orge1d468c2013-01-30 07:37:20 +0000354 // Get audio from NetEq.
henrik.lundin7a926812016-05-12 13:51:28 -0700355 bool muted;
356 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
357 ASSERT_FALSE(muted);
henrik.lundin6d8e0112016-03-04 10:34:21 -0800358 ASSERT_TRUE((out_frame_.samples_per_channel_ == kBlockSize8kHz) ||
359 (out_frame_.samples_per_channel_ == kBlockSize16kHz) ||
360 (out_frame_.samples_per_channel_ == kBlockSize32kHz) ||
361 (out_frame_.samples_per_channel_ == kBlockSize48kHz));
362 output_sample_rate_ = out_frame_.sample_rate_hz_;
henrik.lundind89814b2015-11-23 06:49:25 -0800363 EXPECT_EQ(output_sample_rate_, neteq_->last_output_sample_rate_hz());
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000364
365 // Increase time.
Ivo Creusen24192c22019-07-12 17:00:25 +0200366 sim_clock_ += kTimeStepMs;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000367}
368
minyue4f906772016-04-29 11:05:14 -0700369void NetEqDecodingTest::DecodeAndCompare(
370 const std::string& rtp_file,
371 const std::string& output_checksum,
372 const std::string& network_stats_checksum,
minyue4f906772016-04-29 11:05:14 -0700373 bool gen_ref) {
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000374 OpenInputFile(rtp_file);
375
minyue4f906772016-04-29 11:05:14 -0700376 std::string ref_out_file =
377 gen_ref ? webrtc::test::OutputPath() + "neteq_universal_ref.pcm" : "";
378 ResultSink output(ref_out_file);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000379
minyue4f906772016-04-29 11:05:14 -0700380 std::string stat_out_file =
381 gen_ref ? webrtc::test::OutputPath() + "neteq_network_stats.dat" : "";
382 ResultSink network_stats(stat_out_file);
henrik.lundin@webrtc.org4e4b0982014-08-11 14:48:49 +0000383
henrik.lundin46ba49c2016-05-24 22:50:47 -0700384 packet_ = rtp_source_->NextPacket();
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000385 int i = 0;
Henrik Lundinac0a5032017-09-25 12:22:46 +0200386 uint64_t last_concealed_samples = 0;
387 uint64_t last_total_samples_received = 0;
henrik.lundin@webrtc.org966a7082014-11-17 09:08:38 +0000388 while (packet_) {
Jonas Olsson366a50c2018-09-06 13:41:30 +0200389 rtc::StringBuilder ss;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000390 ss << "Lap number " << i++ << " in DecodeAndCompare while loop";
391 SCOPED_TRACE(ss.str()); // Print out the parameter values on failure.
henrik.lundin6d8e0112016-03-04 10:34:21 -0800392 ASSERT_NO_FATAL_FAILURE(Process());
Yves Gerey665174f2018-06-19 15:03:05 +0200393 ASSERT_NO_FATAL_FAILURE(
394 output.AddResult(out_frame_.data(), out_frame_.samples_per_channel_));
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000395
396 // Query the network statistics API once per second
Ivo Creusen24192c22019-07-12 17:00:25 +0200397 if (sim_clock_ % 1000 == 0) {
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000398 // Process NetworkStatistics.
minyue4f906772016-04-29 11:05:14 -0700399 NetEqNetworkStatistics current_network_stats;
400 ASSERT_EQ(0, neteq_->NetworkStatistics(&current_network_stats));
401 ASSERT_NO_FATAL_FAILURE(network_stats.AddResult(current_network_stats));
402
Henrik Lundinac0a5032017-09-25 12:22:46 +0200403 // Verify that liftime stats and network stats report similar loss
404 // concealment rates.
405 auto lifetime_stats = neteq_->GetLifetimeStatistics();
406 const uint64_t delta_concealed_samples =
407 lifetime_stats.concealed_samples - last_concealed_samples;
408 last_concealed_samples = lifetime_stats.concealed_samples;
409 const uint64_t delta_total_samples_received =
410 lifetime_stats.total_samples_received - last_total_samples_received;
411 last_total_samples_received = lifetime_stats.total_samples_received;
412 // The tolerance is 1% but expressed in Q14.
413 EXPECT_NEAR(
414 (delta_concealed_samples << 14) / delta_total_samples_received,
415 current_network_stats.expand_rate, (2 << 14) / 100.0);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000416 }
417 }
minyue4f906772016-04-29 11:05:14 -0700418
419 SCOPED_TRACE("Check output audio.");
420 output.VerifyChecksum(output_checksum);
421 SCOPED_TRACE("Check network stats.");
422 network_stats.VerifyChecksum(network_stats_checksum);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000423}
424
425void NetEqDecodingTest::PopulateRtpInfo(int frame_index,
426 int timestamp,
henrik.lundin246ef3e2017-04-24 09:14:32 -0700427 RTPHeader* rtp_info) {
428 rtp_info->sequenceNumber = frame_index;
429 rtp_info->timestamp = timestamp;
430 rtp_info->ssrc = 0x1234; // Just an arbitrary SSRC.
431 rtp_info->payloadType = 94; // PCM16b WB codec.
432 rtp_info->markerBit = 0;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000433}
434
435void NetEqDecodingTest::PopulateCng(int frame_index,
436 int timestamp,
henrik.lundin246ef3e2017-04-24 09:14:32 -0700437 RTPHeader* rtp_info,
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000438 uint8_t* payload,
pkasting@chromium.org4591fbd2014-11-20 22:28:14 +0000439 size_t* payload_len) {
henrik.lundin246ef3e2017-04-24 09:14:32 -0700440 rtp_info->sequenceNumber = frame_index;
441 rtp_info->timestamp = timestamp;
442 rtp_info->ssrc = 0x1234; // Just an arbitrary SSRC.
443 rtp_info->payloadType = 98; // WB CNG.
444 rtp_info->markerBit = 0;
Yves Gerey665174f2018-06-19 15:03:05 +0200445 payload[0] = 64; // Noise level -64 dBov, quite arbitrarily chosen.
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000446 *payload_len = 1; // Only noise level, no spectral parameters.
447}
448
ivoc72c08ed2016-01-20 07:26:24 -0800449#if !defined(WEBRTC_IOS) && defined(WEBRTC_NETEQ_UNITTEST_BITEXACT) && \
450 (defined(WEBRTC_CODEC_ISAC) || defined(WEBRTC_CODEC_ISACFX)) && \
Karl Wibergeb254b42017-11-01 15:08:12 +0100451 defined(WEBRTC_CODEC_ILBC) && !defined(WEBRTC_ARCH_ARM64)
minyue5f026d02015-12-16 07:36:04 -0800452#define MAYBE_TestBitExactness TestBitExactness
kwiberg98ab3a42015-09-30 21:54:21 -0700453#else
minyue5f026d02015-12-16 07:36:04 -0800454#define MAYBE_TestBitExactness DISABLED_TestBitExactness
kwiberg98ab3a42015-09-30 21:54:21 -0700455#endif
minyue5f026d02015-12-16 07:36:04 -0800456TEST_F(NetEqDecodingTest, MAYBE_TestBitExactness) {
minyue49c454e2016-01-08 11:30:14 -0800457 const std::string input_rtp_file =
458 webrtc::test::ResourcePath("audio_coding/neteq_universal_new", "rtp");
henrik.lundin@webrtc.org4e4b0982014-08-11 14:48:49 +0000459
Yves Gerey665174f2018-06-19 15:03:05 +0200460 const std::string output_checksum =
Jakob Ivarssona36c5912019-06-27 10:12:02 +0200461 PlatformChecksum("998be2e5a707e636af0b6298f54bedfabe72aae1",
462 "61e238ece4cd3b67d66a0b7047e06b20607dcb79", "not used",
463 "998be2e5a707e636af0b6298f54bedfabe72aae1",
464 "4116ac2a6e75baac3194b712d6fabe28b384275e");
minyue4f906772016-04-29 11:05:14 -0700465
henrik.lundin2979f552017-05-05 05:04:16 -0700466 const std::string network_stats_checksum =
Jakob Ivarssona36c5912019-06-27 10:12:02 +0200467 PlatformChecksum("3689c9f0ab9e50cefab3e44c37c3d7aa0de82ca4",
468 "0a596217fccd8d90eff7d1666b8cc63143eeda12", "not used",
469 "3689c9f0ab9e50cefab3e44c37c3d7aa0de82ca4",
470 "3689c9f0ab9e50cefab3e44c37c3d7aa0de82ca4");
minyue4f906772016-04-29 11:05:14 -0700471
Yves Gerey665174f2018-06-19 15:03:05 +0200472 DecodeAndCompare(input_rtp_file, output_checksum, network_stats_checksum,
Niels Möllerd51b3552018-11-21 13:51:24 +0100473 FLAG_gen_ref);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000474}
475
Yves Gerey665174f2018-06-19 15:03:05 +0200476#if !defined(WEBRTC_IOS) && defined(WEBRTC_NETEQ_UNITTEST_BITEXACT) && \
minyue-webrtc516711c2017-07-27 17:45:49 +0200477 defined(WEBRTC_CODEC_OPUS)
minyue93c08b72015-12-22 09:57:41 -0800478#define MAYBE_TestOpusBitExactness TestOpusBitExactness
479#else
480#define MAYBE_TestOpusBitExactness DISABLED_TestOpusBitExactness
481#endif
minyue-webrtcadb58b82017-07-26 17:59:59 +0200482TEST_F(NetEqDecodingTest, MAYBE_TestOpusBitExactness) {
minyue93c08b72015-12-22 09:57:41 -0800483 const std::string input_rtp_file =
484 webrtc::test::ResourcePath("audio_coding/neteq_opus", "rtp");
minyue93c08b72015-12-22 09:57:41 -0800485
Yves Gereya038e712018-11-14 10:45:50 +0100486 // Checksum depends on libopus being compiled with or without SSE.
487 const std::string maybe_sse =
Jakob Ivarssona36c5912019-06-27 10:12:02 +0200488 "6b602683ca7285a98118b4824d72f4257952c18f|"
489 "eb0b68bddcac00fc85403df64f83126f8ea9bc93";
Yves Gereya038e712018-11-14 10:45:50 +0100490 const std::string output_checksum = PlatformChecksum(
Jakob Ivarssona36c5912019-06-27 10:12:02 +0200491 maybe_sse, "f95f2a220c9ca5d60b81c4653d46e0de2bee159f",
492 "6f288a03d34958f62496f18fa85655593eef4dbe", maybe_sse, maybe_sse);
minyue4f906772016-04-29 11:05:14 -0700493
henrik.lundin2979f552017-05-05 05:04:16 -0700494 const std::string network_stats_checksum =
Jakob Ivarssona36c5912019-06-27 10:12:02 +0200495 PlatformChecksum("0b3d34baffaf651812ffaf06ea1b5ce45ea1c47a",
496 "a71dce66c7bea85ba22d4e29a5298f606f810444",
497 "7c64e1e915bace7c4bf583484efd64eaf234552f",
498 "0b3d34baffaf651812ffaf06ea1b5ce45ea1c47a",
499 "0b3d34baffaf651812ffaf06ea1b5ce45ea1c47a");
minyue4f906772016-04-29 11:05:14 -0700500
Yves Gerey665174f2018-06-19 15:03:05 +0200501 DecodeAndCompare(input_rtp_file, output_checksum, network_stats_checksum,
Niels Möllerd51b3552018-11-21 13:51:24 +0100502 FLAG_gen_ref);
minyue93c08b72015-12-22 09:57:41 -0800503}
504
Yves Gerey665174f2018-06-19 15:03:05 +0200505#if !defined(WEBRTC_IOS) && defined(WEBRTC_NETEQ_UNITTEST_BITEXACT) && \
Henrik Lundine9619f82017-11-27 14:05:27 +0100506 defined(WEBRTC_CODEC_OPUS)
507#define MAYBE_TestOpusDtxBitExactness TestOpusDtxBitExactness
508#else
509#define MAYBE_TestOpusDtxBitExactness DISABLED_TestOpusDtxBitExactness
510#endif
Henrik Lundin4f2a4a12018-01-26 17:32:56 +0100511TEST_F(NetEqDecodingTest, MAYBE_TestOpusDtxBitExactness) {
Henrik Lundine9619f82017-11-27 14:05:27 +0100512 const std::string input_rtp_file =
513 webrtc::test::ResourcePath("audio_coding/neteq_opus_dtx", "rtp");
514
Yves Gereya038e712018-11-14 10:45:50 +0100515 const std::string maybe_sse =
516 "713af6c92881f5aab1285765ee6680da9d1c06ce|"
517 "2ac10c4e79aeedd0df2863b079da5848b40f00b5";
518 const std::string output_checksum = PlatformChecksum(
519 maybe_sse, "3ec991b96872123f1554c03c543ca5d518431e46",
520 "da9f9a2d94e0c2d67342fad4965d7b91cda50b25", maybe_sse, maybe_sse);
Henrik Lundine9619f82017-11-27 14:05:27 +0100521
522 const std::string network_stats_checksum =
523 "bab58dc587d956f326056d7340c96eb9d2d3cc21";
524
Henrik Lundine9619f82017-11-27 14:05:27 +0100525 DecodeAndCompare(input_rtp_file, output_checksum, network_stats_checksum,
Niels Möllerd51b3552018-11-21 13:51:24 +0100526 FLAG_gen_ref);
Henrik Lundine9619f82017-11-27 14:05:27 +0100527}
528
henrik.lundin@webrtc.org7cbc4f92014-10-07 06:37:39 +0000529// Use fax mode to avoid time-scaling. This is to simplify the testing of
530// packet waiting times in the packet buffer.
531class NetEqDecodingTestFaxMode : public NetEqDecodingTest {
532 protected:
533 NetEqDecodingTestFaxMode() : NetEqDecodingTest() {
Henrik Lundin7687ad52018-07-02 10:14:46 +0200534 config_.for_test_no_time_stretching = true;
henrik.lundin@webrtc.org7cbc4f92014-10-07 06:37:39 +0000535 }
Gustaf Ullbergb0a02072017-10-02 12:00:34 +0200536 void TestJitterBufferDelay(bool apply_packet_loss);
henrik.lundin@webrtc.org7cbc4f92014-10-07 06:37:39 +0000537};
538
539TEST_F(NetEqDecodingTestFaxMode, TestFrameWaitingTimeStatistics) {
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000540 // Insert 30 dummy packets at once. Each packet contains 10 ms 16 kHz audio.
541 size_t num_frames = 30;
pkasting@chromium.org4591fbd2014-11-20 22:28:14 +0000542 const size_t kSamples = 10 * 16;
543 const size_t kPayloadBytes = kSamples * 2;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000544 for (size_t i = 0; i < num_frames; ++i) {
kwibergee2bac22015-11-11 10:34:00 -0800545 const uint8_t payload[kPayloadBytes] = {0};
henrik.lundin246ef3e2017-04-24 09:14:32 -0700546 RTPHeader rtp_info;
Mirko Bonadeia8110272017-10-18 14:22:50 +0200547 rtp_info.sequenceNumber = rtc::checked_cast<uint16_t>(i);
548 rtp_info.timestamp = rtc::checked_cast<uint32_t>(i * kSamples);
henrik.lundin246ef3e2017-04-24 09:14:32 -0700549 rtp_info.ssrc = 0x1234; // Just an arbitrary SSRC.
550 rtp_info.payloadType = 94; // PCM16b WB codec.
551 rtp_info.markerBit = 0;
552 ASSERT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000553 }
554 // Pull out all data.
555 for (size_t i = 0; i < num_frames; ++i) {
henrik.lundin7a926812016-05-12 13:51:28 -0700556 bool muted;
557 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -0800558 ASSERT_EQ(kBlockSize16kHz, out_frame_.samples_per_channel_);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000559 }
560
Henrik Lundin1bb8cf82015-08-25 13:08:04 +0200561 NetEqNetworkStatistics stats;
562 EXPECT_EQ(0, neteq_->NetworkStatistics(&stats));
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000563 // Since all frames are dumped into NetEQ at once, but pulled out with 10 ms
564 // spacing (per definition), we expect the delay to increase with 10 ms for
Henrik Lundin1bb8cf82015-08-25 13:08:04 +0200565 // each packet. Thus, we are calculating the statistics for a series from 10
566 // to 300, in steps of 10 ms.
567 EXPECT_EQ(155, stats.mean_waiting_time_ms);
568 EXPECT_EQ(155, stats.median_waiting_time_ms);
569 EXPECT_EQ(10, stats.min_waiting_time_ms);
570 EXPECT_EQ(300, stats.max_waiting_time_ms);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000571
572 // Check statistics again and make sure it's been reset.
Henrik Lundin1bb8cf82015-08-25 13:08:04 +0200573 EXPECT_EQ(0, neteq_->NetworkStatistics(&stats));
574 EXPECT_EQ(-1, stats.mean_waiting_time_ms);
575 EXPECT_EQ(-1, stats.median_waiting_time_ms);
576 EXPECT_EQ(-1, stats.min_waiting_time_ms);
577 EXPECT_EQ(-1, stats.max_waiting_time_ms);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000578}
579
henrik.lundin@webrtc.orgb4e80e02014-05-15 07:14:00 +0000580TEST_F(NetEqDecodingTest, TestAverageInterArrivalTimeNegative) {
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000581 const int kNumFrames = 3000; // Needed for convergence.
582 int frame_index = 0;
pkasting@chromium.org4591fbd2014-11-20 22:28:14 +0000583 const size_t kSamples = 10 * 16;
584 const size_t kPayloadBytes = kSamples * 2;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000585 while (frame_index < kNumFrames) {
586 // Insert one packet each time, except every 10th time where we insert two
587 // packets at once. This will create a negative clock-drift of approx. 10%.
588 int num_packets = (frame_index % 10 == 0 ? 2 : 1);
589 for (int n = 0; n < num_packets; ++n) {
590 uint8_t payload[kPayloadBytes] = {0};
henrik.lundin246ef3e2017-04-24 09:14:32 -0700591 RTPHeader rtp_info;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000592 PopulateRtpInfo(frame_index, frame_index * kSamples, &rtp_info);
henrik.lundin246ef3e2017-04-24 09:14:32 -0700593 ASSERT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000594 ++frame_index;
595 }
596
597 // Pull out data once.
henrik.lundin7a926812016-05-12 13:51:28 -0700598 bool muted;
599 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -0800600 ASSERT_EQ(kBlockSize16kHz, out_frame_.samples_per_channel_);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000601 }
602
603 NetEqNetworkStatistics network_stats;
604 ASSERT_EQ(0, neteq_->NetworkStatistics(&network_stats));
henrik.lundin0d838572016-10-13 03:35:55 -0700605 EXPECT_EQ(-103192, network_stats.clockdrift_ppm);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000606}
607
henrik.lundin@webrtc.orgb4e80e02014-05-15 07:14:00 +0000608TEST_F(NetEqDecodingTest, TestAverageInterArrivalTimePositive) {
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000609 const int kNumFrames = 5000; // Needed for convergence.
610 int frame_index = 0;
pkasting@chromium.org4591fbd2014-11-20 22:28:14 +0000611 const size_t kSamples = 10 * 16;
612 const size_t kPayloadBytes = kSamples * 2;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000613 for (int i = 0; i < kNumFrames; ++i) {
614 // Insert one packet each time, except every 10th time where we don't insert
615 // any packet. This will create a positive clock-drift of approx. 11%.
616 int num_packets = (i % 10 == 9 ? 0 : 1);
617 for (int n = 0; n < num_packets; ++n) {
618 uint8_t payload[kPayloadBytes] = {0};
henrik.lundin246ef3e2017-04-24 09:14:32 -0700619 RTPHeader rtp_info;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000620 PopulateRtpInfo(frame_index, frame_index * kSamples, &rtp_info);
henrik.lundin246ef3e2017-04-24 09:14:32 -0700621 ASSERT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000622 ++frame_index;
623 }
624
625 // Pull out data once.
henrik.lundin7a926812016-05-12 13:51:28 -0700626 bool muted;
627 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -0800628 ASSERT_EQ(kBlockSize16kHz, out_frame_.samples_per_channel_);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000629 }
630
631 NetEqNetworkStatistics network_stats;
632 ASSERT_EQ(0, neteq_->NetworkStatistics(&network_stats));
henrik.lundin0d838572016-10-13 03:35:55 -0700633 EXPECT_EQ(110953, network_stats.clockdrift_ppm);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000634}
635
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000636void NetEqDecodingTest::LongCngWithClockDrift(double drift_factor,
637 double network_freeze_ms,
638 bool pull_audio_during_freeze,
639 int delay_tolerance_ms,
640 int max_time_to_speech_ms) {
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000641 uint16_t seq_no = 0;
642 uint32_t timestamp = 0;
643 const int kFrameSizeMs = 30;
pkasting@chromium.org4591fbd2014-11-20 22:28:14 +0000644 const size_t kSamples = kFrameSizeMs * 16;
645 const size_t kPayloadBytes = kSamples * 2;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000646 double next_input_time_ms = 0.0;
647 double t_ms;
henrik.lundin7a926812016-05-12 13:51:28 -0700648 bool muted;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000649
650 // Insert speech for 5 seconds.
651 const int kSpeechDurationMs = 5000;
652 for (t_ms = 0; t_ms < kSpeechDurationMs; t_ms += 10) {
653 // Each turn in this for loop is 10 ms.
654 while (next_input_time_ms <= t_ms) {
655 // Insert one 30 ms speech frame.
656 uint8_t payload[kPayloadBytes] = {0};
henrik.lundin246ef3e2017-04-24 09:14:32 -0700657 RTPHeader rtp_info;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000658 PopulateRtpInfo(seq_no, timestamp, &rtp_info);
henrik.lundin246ef3e2017-04-24 09:14:32 -0700659 ASSERT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000660 ++seq_no;
661 timestamp += kSamples;
henrik.lundin@webrtc.orgfcfc6a92014-02-13 11:42:28 +0000662 next_input_time_ms += static_cast<double>(kFrameSizeMs) * drift_factor;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000663 }
664 // Pull out data once.
henrik.lundin7a926812016-05-12 13:51:28 -0700665 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -0800666 ASSERT_EQ(kBlockSize16kHz, out_frame_.samples_per_channel_);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000667 }
668
henrik.lundin55480f52016-03-08 02:37:57 -0800669 EXPECT_EQ(AudioFrame::kNormalSpeech, out_frame_.speech_type_);
Danil Chapovalovb6021232018-06-19 13:26:36 +0200670 absl::optional<uint32_t> playout_timestamp = neteq_->GetPlayoutTimestamp();
henrik.lundin0d96ab72016-04-06 12:28:26 -0700671 ASSERT_TRUE(playout_timestamp);
672 int32_t delay_before = timestamp - *playout_timestamp;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000673
674 // Insert CNG for 1 minute (= 60000 ms).
675 const int kCngPeriodMs = 100;
676 const int kCngPeriodSamples = kCngPeriodMs * 16; // Period in 16 kHz samples.
677 const int kCngDurationMs = 60000;
678 for (; t_ms < kSpeechDurationMs + kCngDurationMs; t_ms += 10) {
679 // Each turn in this for loop is 10 ms.
680 while (next_input_time_ms <= t_ms) {
681 // Insert one CNG frame each 100 ms.
682 uint8_t payload[kPayloadBytes];
pkasting@chromium.org4591fbd2014-11-20 22:28:14 +0000683 size_t payload_len;
henrik.lundin246ef3e2017-04-24 09:14:32 -0700684 RTPHeader rtp_info;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000685 PopulateCng(seq_no, timestamp, &rtp_info, payload, &payload_len);
kwibergee2bac22015-11-11 10:34:00 -0800686 ASSERT_EQ(0, neteq_->InsertPacket(
henrik.lundin246ef3e2017-04-24 09:14:32 -0700687 rtp_info,
kwibergee2bac22015-11-11 10:34:00 -0800688 rtc::ArrayView<const uint8_t>(payload, payload_len), 0));
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000689 ++seq_no;
690 timestamp += kCngPeriodSamples;
henrik.lundin@webrtc.orgfcfc6a92014-02-13 11:42:28 +0000691 next_input_time_ms += static_cast<double>(kCngPeriodMs) * drift_factor;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000692 }
693 // Pull out data once.
henrik.lundin7a926812016-05-12 13:51:28 -0700694 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -0800695 ASSERT_EQ(kBlockSize16kHz, out_frame_.samples_per_channel_);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000696 }
697
henrik.lundin55480f52016-03-08 02:37:57 -0800698 EXPECT_EQ(AudioFrame::kCNG, out_frame_.speech_type_);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000699
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000700 if (network_freeze_ms > 0) {
701 // First keep pulling audio for |network_freeze_ms| without inserting
702 // any data, then insert CNG data corresponding to |network_freeze_ms|
703 // without pulling any output audio.
704 const double loop_end_time = t_ms + network_freeze_ms;
705 for (; t_ms < loop_end_time; t_ms += 10) {
706 // Pull out data once.
henrik.lundin7a926812016-05-12 13:51:28 -0700707 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -0800708 ASSERT_EQ(kBlockSize16kHz, out_frame_.samples_per_channel_);
henrik.lundin55480f52016-03-08 02:37:57 -0800709 EXPECT_EQ(AudioFrame::kCNG, out_frame_.speech_type_);
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000710 }
711 bool pull_once = pull_audio_during_freeze;
712 // If |pull_once| is true, GetAudio will be called once half-way through
713 // the network recovery period.
714 double pull_time_ms = (t_ms + next_input_time_ms) / 2;
715 while (next_input_time_ms <= t_ms) {
716 if (pull_once && next_input_time_ms >= pull_time_ms) {
717 pull_once = false;
718 // Pull out data once.
henrik.lundin7a926812016-05-12 13:51:28 -0700719 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -0800720 ASSERT_EQ(kBlockSize16kHz, out_frame_.samples_per_channel_);
henrik.lundin55480f52016-03-08 02:37:57 -0800721 EXPECT_EQ(AudioFrame::kCNG, out_frame_.speech_type_);
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000722 t_ms += 10;
723 }
724 // Insert one CNG frame each 100 ms.
725 uint8_t payload[kPayloadBytes];
pkasting@chromium.org4591fbd2014-11-20 22:28:14 +0000726 size_t payload_len;
henrik.lundin246ef3e2017-04-24 09:14:32 -0700727 RTPHeader rtp_info;
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000728 PopulateCng(seq_no, timestamp, &rtp_info, payload, &payload_len);
kwibergee2bac22015-11-11 10:34:00 -0800729 ASSERT_EQ(0, neteq_->InsertPacket(
henrik.lundin246ef3e2017-04-24 09:14:32 -0700730 rtp_info,
kwibergee2bac22015-11-11 10:34:00 -0800731 rtc::ArrayView<const uint8_t>(payload, payload_len), 0));
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000732 ++seq_no;
733 timestamp += kCngPeriodSamples;
734 next_input_time_ms += kCngPeriodMs * drift_factor;
735 }
736 }
737
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000738 // Insert speech again until output type is speech.
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000739 double speech_restart_time_ms = t_ms;
henrik.lundin55480f52016-03-08 02:37:57 -0800740 while (out_frame_.speech_type_ != AudioFrame::kNormalSpeech) {
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000741 // Each turn in this for loop is 10 ms.
742 while (next_input_time_ms <= t_ms) {
743 // Insert one 30 ms speech frame.
744 uint8_t payload[kPayloadBytes] = {0};
henrik.lundin246ef3e2017-04-24 09:14:32 -0700745 RTPHeader rtp_info;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000746 PopulateRtpInfo(seq_no, timestamp, &rtp_info);
henrik.lundin246ef3e2017-04-24 09:14:32 -0700747 ASSERT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000748 ++seq_no;
749 timestamp += kSamples;
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000750 next_input_time_ms += kFrameSizeMs * drift_factor;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000751 }
752 // Pull out data once.
henrik.lundin7a926812016-05-12 13:51:28 -0700753 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -0800754 ASSERT_EQ(kBlockSize16kHz, out_frame_.samples_per_channel_);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000755 // Increase clock.
756 t_ms += 10;
757 }
758
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000759 // Check that the speech starts again within reasonable time.
760 double time_until_speech_returns_ms = t_ms - speech_restart_time_ms;
761 EXPECT_LT(time_until_speech_returns_ms, max_time_to_speech_ms);
henrik.lundin114c1b32017-04-26 07:47:32 -0700762 playout_timestamp = neteq_->GetPlayoutTimestamp();
henrik.lundin0d96ab72016-04-06 12:28:26 -0700763 ASSERT_TRUE(playout_timestamp);
764 int32_t delay_after = timestamp - *playout_timestamp;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000765 // Compare delay before and after, and make sure it differs less than 20 ms.
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000766 EXPECT_LE(delay_after, delay_before + delay_tolerance_ms * 16);
767 EXPECT_GE(delay_after, delay_before - delay_tolerance_ms * 16);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000768}
769
henrik.lundin@webrtc.orgb4e80e02014-05-15 07:14:00 +0000770TEST_F(NetEqDecodingTest, LongCngWithNegativeClockDrift) {
henrik.lundin@webrtc.orgfcfc6a92014-02-13 11:42:28 +0000771 // Apply a clock drift of -25 ms / s (sender faster than receiver).
772 const double kDriftFactor = 1000.0 / (1000.0 + 25.0);
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000773 const double kNetworkFreezeTimeMs = 0.0;
774 const bool kGetAudioDuringFreezeRecovery = false;
775 const int kDelayToleranceMs = 20;
776 const int kMaxTimeToSpeechMs = 100;
Yves Gerey665174f2018-06-19 15:03:05 +0200777 LongCngWithClockDrift(kDriftFactor, kNetworkFreezeTimeMs,
778 kGetAudioDuringFreezeRecovery, kDelayToleranceMs,
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000779 kMaxTimeToSpeechMs);
henrik.lundin@webrtc.orgfcfc6a92014-02-13 11:42:28 +0000780}
781
henrik.lundin@webrtc.orgb4e80e02014-05-15 07:14:00 +0000782TEST_F(NetEqDecodingTest, LongCngWithPositiveClockDrift) {
henrik.lundin@webrtc.orgfcfc6a92014-02-13 11:42:28 +0000783 // Apply a clock drift of +25 ms / s (sender slower than receiver).
784 const double kDriftFactor = 1000.0 / (1000.0 - 25.0);
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000785 const double kNetworkFreezeTimeMs = 0.0;
786 const bool kGetAudioDuringFreezeRecovery = false;
787 const int kDelayToleranceMs = 20;
788 const int kMaxTimeToSpeechMs = 100;
Yves Gerey665174f2018-06-19 15:03:05 +0200789 LongCngWithClockDrift(kDriftFactor, kNetworkFreezeTimeMs,
790 kGetAudioDuringFreezeRecovery, kDelayToleranceMs,
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000791 kMaxTimeToSpeechMs);
792}
793
henrik.lundin@webrtc.orgb4e80e02014-05-15 07:14:00 +0000794TEST_F(NetEqDecodingTest, LongCngWithNegativeClockDriftNetworkFreeze) {
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000795 // Apply a clock drift of -25 ms / s (sender faster than receiver).
796 const double kDriftFactor = 1000.0 / (1000.0 + 25.0);
797 const double kNetworkFreezeTimeMs = 5000.0;
798 const bool kGetAudioDuringFreezeRecovery = false;
Jakob Ivarssona36c5912019-06-27 10:12:02 +0200799 const int kDelayToleranceMs = 60;
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000800 const int kMaxTimeToSpeechMs = 200;
Yves Gerey665174f2018-06-19 15:03:05 +0200801 LongCngWithClockDrift(kDriftFactor, kNetworkFreezeTimeMs,
802 kGetAudioDuringFreezeRecovery, kDelayToleranceMs,
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000803 kMaxTimeToSpeechMs);
804}
805
henrik.lundin@webrtc.orgb4e80e02014-05-15 07:14:00 +0000806TEST_F(NetEqDecodingTest, LongCngWithPositiveClockDriftNetworkFreeze) {
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000807 // Apply a clock drift of +25 ms / s (sender slower than receiver).
808 const double kDriftFactor = 1000.0 / (1000.0 - 25.0);
809 const double kNetworkFreezeTimeMs = 5000.0;
810 const bool kGetAudioDuringFreezeRecovery = false;
811 const int kDelayToleranceMs = 20;
812 const int kMaxTimeToSpeechMs = 100;
Yves Gerey665174f2018-06-19 15:03:05 +0200813 LongCngWithClockDrift(kDriftFactor, kNetworkFreezeTimeMs,
814 kGetAudioDuringFreezeRecovery, kDelayToleranceMs,
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000815 kMaxTimeToSpeechMs);
816}
817
henrik.lundin@webrtc.orgb4e80e02014-05-15 07:14:00 +0000818TEST_F(NetEqDecodingTest, LongCngWithPositiveClockDriftNetworkFreezeExtraPull) {
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000819 // Apply a clock drift of +25 ms / s (sender slower than receiver).
820 const double kDriftFactor = 1000.0 / (1000.0 - 25.0);
821 const double kNetworkFreezeTimeMs = 5000.0;
822 const bool kGetAudioDuringFreezeRecovery = true;
823 const int kDelayToleranceMs = 20;
824 const int kMaxTimeToSpeechMs = 100;
Yves Gerey665174f2018-06-19 15:03:05 +0200825 LongCngWithClockDrift(kDriftFactor, kNetworkFreezeTimeMs,
826 kGetAudioDuringFreezeRecovery, kDelayToleranceMs,
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000827 kMaxTimeToSpeechMs);
828}
829
henrik.lundin@webrtc.orgb4e80e02014-05-15 07:14:00 +0000830TEST_F(NetEqDecodingTest, LongCngWithoutClockDrift) {
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000831 const double kDriftFactor = 1.0; // No drift.
832 const double kNetworkFreezeTimeMs = 0.0;
833 const bool kGetAudioDuringFreezeRecovery = false;
834 const int kDelayToleranceMs = 10;
835 const int kMaxTimeToSpeechMs = 50;
Yves Gerey665174f2018-06-19 15:03:05 +0200836 LongCngWithClockDrift(kDriftFactor, kNetworkFreezeTimeMs,
837 kGetAudioDuringFreezeRecovery, kDelayToleranceMs,
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000838 kMaxTimeToSpeechMs);
henrik.lundin@webrtc.orgfcfc6a92014-02-13 11:42:28 +0000839}
840
henrik.lundin@webrtc.orgb4e80e02014-05-15 07:14:00 +0000841TEST_F(NetEqDecodingTest, UnknownPayloadType) {
pkasting@chromium.org4591fbd2014-11-20 22:28:14 +0000842 const size_t kPayloadBytes = 100;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000843 uint8_t payload[kPayloadBytes] = {0};
henrik.lundin246ef3e2017-04-24 09:14:32 -0700844 RTPHeader rtp_info;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000845 PopulateRtpInfo(0, 0, &rtp_info);
henrik.lundin246ef3e2017-04-24 09:14:32 -0700846 rtp_info.payloadType = 1; // Not registered as a decoder.
847 EXPECT_EQ(NetEq::kFail, neteq_->InsertPacket(rtp_info, payload, 0));
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000848}
849
Peter Boströme2976c82016-01-04 22:44:05 +0100850#if defined(WEBRTC_CODEC_ISAC) || defined(WEBRTC_CODEC_ISACFX)
ivoc72c08ed2016-01-20 07:26:24 -0800851#define MAYBE_DecoderError DecoderError
852#else
853#define MAYBE_DecoderError DISABLED_DecoderError
854#endif
855
Peter Boströme2976c82016-01-04 22:44:05 +0100856TEST_F(NetEqDecodingTest, MAYBE_DecoderError) {
pkasting@chromium.org4591fbd2014-11-20 22:28:14 +0000857 const size_t kPayloadBytes = 100;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000858 uint8_t payload[kPayloadBytes] = {0};
henrik.lundin246ef3e2017-04-24 09:14:32 -0700859 RTPHeader rtp_info;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000860 PopulateRtpInfo(0, 0, &rtp_info);
henrik.lundin246ef3e2017-04-24 09:14:32 -0700861 rtp_info.payloadType = 103; // iSAC, but the payload is invalid.
862 EXPECT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000863 // Set all of |out_data_| to 1, and verify that it was set to 0 by the call
864 // to GetAudio.
yujo36b1a5f2017-06-12 12:45:32 -0700865 int16_t* out_frame_data = out_frame_.mutable_data();
henrik.lundin6d8e0112016-03-04 10:34:21 -0800866 for (size_t i = 0; i < AudioFrame::kMaxDataSizeSamples; ++i) {
yujo36b1a5f2017-06-12 12:45:32 -0700867 out_frame_data[i] = 1;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000868 }
henrik.lundin7a926812016-05-12 13:51:28 -0700869 bool muted;
870 EXPECT_EQ(NetEq::kFail, neteq_->GetAudio(&out_frame_, &muted));
871 ASSERT_FALSE(muted);
ivoc72c08ed2016-01-20 07:26:24 -0800872
yujo36b1a5f2017-06-12 12:45:32 -0700873 // Verify that the first 160 samples are set to 0.
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000874 static const int kExpectedOutputLength = 160; // 10 ms at 16 kHz sample rate.
yujo36b1a5f2017-06-12 12:45:32 -0700875 const int16_t* const_out_frame_data = out_frame_.data();
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000876 for (int i = 0; i < kExpectedOutputLength; ++i) {
Jonas Olsson366a50c2018-09-06 13:41:30 +0200877 rtc::StringBuilder ss;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000878 ss << "i = " << i;
879 SCOPED_TRACE(ss.str()); // Print out the parameter values on failure.
yujo36b1a5f2017-06-12 12:45:32 -0700880 EXPECT_EQ(0, const_out_frame_data[i]);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000881 }
882}
883
henrik.lundin@webrtc.orgb4e80e02014-05-15 07:14:00 +0000884TEST_F(NetEqDecodingTest, GetAudioBeforeInsertPacket) {
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000885 // Set all of |out_data_| to 1, and verify that it was set to 0 by the call
886 // to GetAudio.
yujo36b1a5f2017-06-12 12:45:32 -0700887 int16_t* out_frame_data = out_frame_.mutable_data();
henrik.lundin6d8e0112016-03-04 10:34:21 -0800888 for (size_t i = 0; i < AudioFrame::kMaxDataSizeSamples; ++i) {
yujo36b1a5f2017-06-12 12:45:32 -0700889 out_frame_data[i] = 1;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000890 }
henrik.lundin7a926812016-05-12 13:51:28 -0700891 bool muted;
892 EXPECT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
893 ASSERT_FALSE(muted);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000894 // Verify that the first block of samples is set to 0.
895 static const int kExpectedOutputLength =
896 kInitSampleRateHz / 100; // 10 ms at initial sample rate.
yujo36b1a5f2017-06-12 12:45:32 -0700897 const int16_t* const_out_frame_data = out_frame_.data();
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000898 for (int i = 0; i < kExpectedOutputLength; ++i) {
Jonas Olsson366a50c2018-09-06 13:41:30 +0200899 rtc::StringBuilder ss;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000900 ss << "i = " << i;
901 SCOPED_TRACE(ss.str()); // Print out the parameter values on failure.
yujo36b1a5f2017-06-12 12:45:32 -0700902 EXPECT_EQ(0, const_out_frame_data[i]);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000903 }
henrik.lundind89814b2015-11-23 06:49:25 -0800904 // Verify that the sample rate did not change from the initial configuration.
905 EXPECT_EQ(config_.sample_rate_hz, neteq_->last_output_sample_rate_hz());
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000906}
turaj@webrtc.orgff43c852013-09-25 00:07:27 +0000907
henrik.lundin@webrtc.org9b8102c2014-08-21 08:27:44 +0000908class NetEqBgnTest : public NetEqDecodingTest {
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +0000909 protected:
henrik.lundin@webrtc.org9b8102c2014-08-21 08:27:44 +0000910 void CheckBgn(int sampling_rate_hz) {
Peter Kastingdce40cf2015-08-24 14:52:23 -0700911 size_t expected_samples_per_channel = 0;
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +0000912 uint8_t payload_type = 0xFF; // Invalid.
913 if (sampling_rate_hz == 8000) {
914 expected_samples_per_channel = kBlockSize8kHz;
915 payload_type = 93; // PCM 16, 8 kHz.
916 } else if (sampling_rate_hz == 16000) {
917 expected_samples_per_channel = kBlockSize16kHz;
918 payload_type = 94; // PCM 16, 16 kHZ.
919 } else if (sampling_rate_hz == 32000) {
920 expected_samples_per_channel = kBlockSize32kHz;
921 payload_type = 95; // PCM 16, 32 kHz.
922 } else {
923 ASSERT_TRUE(false); // Unsupported test case.
924 }
turaj@webrtc.orgff43c852013-09-25 00:07:27 +0000925
henrik.lundin6d8e0112016-03-04 10:34:21 -0800926 AudioFrame output;
henrik.lundin@webrtc.org9b8102c2014-08-21 08:27:44 +0000927 test::AudioLoop input;
928 // We are using the same 32 kHz input file for all tests, regardless of
929 // |sampling_rate_hz|. The output may sound weird, but the test is still
930 // valid.
931 ASSERT_TRUE(input.Init(
932 webrtc::test::ResourcePath("audio_coding/testfile32kHz", "pcm"),
933 10 * sampling_rate_hz, // Max 10 seconds loop length.
Peter Kastingdce40cf2015-08-24 14:52:23 -0700934 expected_samples_per_channel));
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +0000935
936 // Payload of 10 ms of PCM16 32 kHz.
937 uint8_t payload[kBlockSize32kHz * sizeof(int16_t)];
henrik.lundin246ef3e2017-04-24 09:14:32 -0700938 RTPHeader rtp_info;
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +0000939 PopulateRtpInfo(0, 0, &rtp_info);
henrik.lundin246ef3e2017-04-24 09:14:32 -0700940 rtp_info.payloadType = payload_type;
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +0000941
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +0000942 uint32_t receive_timestamp = 0;
henrik.lundin7a926812016-05-12 13:51:28 -0700943 bool muted;
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +0000944 for (int n = 0; n < 10; ++n) { // Insert few packets and get audio.
kwiberg288886b2015-11-06 01:21:35 -0800945 auto block = input.GetNextBlock();
946 ASSERT_EQ(expected_samples_per_channel, block.size());
947 size_t enc_len_bytes =
948 WebRtcPcm16b_Encode(block.data(), block.size(), payload);
henrik.lundin@webrtc.org9b8102c2014-08-21 08:27:44 +0000949 ASSERT_EQ(enc_len_bytes, expected_samples_per_channel * 2);
950
Henrik Lundin70c09bd2017-04-24 15:56:56 +0200951 ASSERT_EQ(0, neteq_->InsertPacket(
henrik.lundin246ef3e2017-04-24 09:14:32 -0700952 rtp_info,
Henrik Lundin70c09bd2017-04-24 15:56:56 +0200953 rtc::ArrayView<const uint8_t>(payload, enc_len_bytes),
954 receive_timestamp));
henrik.lundin6d8e0112016-03-04 10:34:21 -0800955 output.Reset();
henrik.lundin7a926812016-05-12 13:51:28 -0700956 ASSERT_EQ(0, neteq_->GetAudio(&output, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -0800957 ASSERT_EQ(1u, output.num_channels_);
958 ASSERT_EQ(expected_samples_per_channel, output.samples_per_channel_);
henrik.lundin55480f52016-03-08 02:37:57 -0800959 ASSERT_EQ(AudioFrame::kNormalSpeech, output.speech_type_);
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +0000960
961 // Next packet.
Yves Gerey665174f2018-06-19 15:03:05 +0200962 rtp_info.timestamp +=
963 rtc::checked_cast<uint32_t>(expected_samples_per_channel);
henrik.lundin246ef3e2017-04-24 09:14:32 -0700964 rtp_info.sequenceNumber++;
Yves Gerey665174f2018-06-19 15:03:05 +0200965 receive_timestamp +=
966 rtc::checked_cast<uint32_t>(expected_samples_per_channel);
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +0000967 }
968
henrik.lundin6d8e0112016-03-04 10:34:21 -0800969 output.Reset();
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +0000970
971 // Get audio without inserting packets, expecting PLC and PLC-to-CNG. Pull
972 // one frame without checking speech-type. This is the first frame pulled
973 // without inserting any packet, and might not be labeled as PLC.
henrik.lundin7a926812016-05-12 13:51:28 -0700974 ASSERT_EQ(0, neteq_->GetAudio(&output, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -0800975 ASSERT_EQ(1u, output.num_channels_);
976 ASSERT_EQ(expected_samples_per_channel, output.samples_per_channel_);
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +0000977
978 // To be able to test the fading of background noise we need at lease to
979 // pull 611 frames.
980 const int kFadingThreshold = 611;
981
982 // Test several CNG-to-PLC packet for the expected behavior. The number 20
983 // is arbitrary, but sufficiently large to test enough number of frames.
984 const int kNumPlcToCngTestFrames = 20;
985 bool plc_to_cng = false;
986 for (int n = 0; n < kFadingThreshold + kNumPlcToCngTestFrames; ++n) {
henrik.lundin6d8e0112016-03-04 10:34:21 -0800987 output.Reset();
yujo36b1a5f2017-06-12 12:45:32 -0700988 // Set to non-zero.
989 memset(output.mutable_data(), 1, AudioFrame::kMaxDataSizeBytes);
henrik.lundin7a926812016-05-12 13:51:28 -0700990 ASSERT_EQ(0, neteq_->GetAudio(&output, &muted));
991 ASSERT_FALSE(muted);
henrik.lundin6d8e0112016-03-04 10:34:21 -0800992 ASSERT_EQ(1u, output.num_channels_);
993 ASSERT_EQ(expected_samples_per_channel, output.samples_per_channel_);
henrik.lundin55480f52016-03-08 02:37:57 -0800994 if (output.speech_type_ == AudioFrame::kPLCCNG) {
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +0000995 plc_to_cng = true;
996 double sum_squared = 0;
yujo36b1a5f2017-06-12 12:45:32 -0700997 const int16_t* output_data = output.data();
henrik.lundin6d8e0112016-03-04 10:34:21 -0800998 for (size_t k = 0;
999 k < output.num_channels_ * output.samples_per_channel_; ++k)
yujo36b1a5f2017-06-12 12:45:32 -07001000 sum_squared += output_data[k] * output_data[k];
Henrik Lundin67190172018-04-20 15:34:48 +02001001 EXPECT_EQ(0, sum_squared);
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +00001002 } else {
henrik.lundin55480f52016-03-08 02:37:57 -08001003 EXPECT_EQ(AudioFrame::kPLC, output.speech_type_);
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +00001004 }
1005 }
1006 EXPECT_TRUE(plc_to_cng); // Just to be sure that PLC-to-CNG has occurred.
1007 }
1008};
1009
Henrik Lundin67190172018-04-20 15:34:48 +02001010TEST_F(NetEqBgnTest, RunTest) {
henrik.lundin@webrtc.org9b8102c2014-08-21 08:27:44 +00001011 CheckBgn(8000);
1012 CheckBgn(16000);
1013 CheckBgn(32000);
1014}
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +00001015
turaj@webrtc.org78b41a02013-11-22 20:27:07 +00001016void NetEqDecodingTest::WrapTest(uint16_t start_seq_no,
1017 uint32_t start_timestamp,
1018 const std::set<uint16_t>& drop_seq_numbers,
1019 bool expect_seq_no_wrap,
1020 bool expect_timestamp_wrap) {
1021 uint16_t seq_no = start_seq_no;
1022 uint32_t timestamp = start_timestamp;
1023 const int kBlocksPerFrame = 3; // Number of 10 ms blocks per frame.
1024 const int kFrameSizeMs = kBlocksPerFrame * kTimeStepMs;
1025 const int kSamples = kBlockSize16kHz * kBlocksPerFrame;
pkasting@chromium.org4591fbd2014-11-20 22:28:14 +00001026 const size_t kPayloadBytes = kSamples * sizeof(int16_t);
turaj@webrtc.org78b41a02013-11-22 20:27:07 +00001027 double next_input_time_ms = 0.0;
turaj@webrtc.org78b41a02013-11-22 20:27:07 +00001028 uint32_t receive_timestamp = 0;
1029
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001030 // Insert speech for 2 seconds.
turaj@webrtc.org78b41a02013-11-22 20:27:07 +00001031 const int kSpeechDurationMs = 2000;
1032 int packets_inserted = 0;
1033 uint16_t last_seq_no;
1034 uint32_t last_timestamp;
1035 bool timestamp_wrapped = false;
1036 bool seq_no_wrapped = false;
1037 for (double t_ms = 0; t_ms < kSpeechDurationMs; t_ms += 10) {
1038 // Each turn in this for loop is 10 ms.
1039 while (next_input_time_ms <= t_ms) {
1040 // Insert one 30 ms speech frame.
1041 uint8_t payload[kPayloadBytes] = {0};
henrik.lundin246ef3e2017-04-24 09:14:32 -07001042 RTPHeader rtp_info;
turaj@webrtc.org78b41a02013-11-22 20:27:07 +00001043 PopulateRtpInfo(seq_no, timestamp, &rtp_info);
1044 if (drop_seq_numbers.find(seq_no) == drop_seq_numbers.end()) {
1045 // This sequence number was not in the set to drop. Insert it.
henrik.lundin246ef3e2017-04-24 09:14:32 -07001046 ASSERT_EQ(0,
1047 neteq_->InsertPacket(rtp_info, payload, receive_timestamp));
turaj@webrtc.org78b41a02013-11-22 20:27:07 +00001048 ++packets_inserted;
1049 }
1050 NetEqNetworkStatistics network_stats;
1051 ASSERT_EQ(0, neteq_->NetworkStatistics(&network_stats));
1052
1053 // Due to internal NetEq logic, preferred buffer-size is about 4 times the
1054 // packet size for first few packets. Therefore we refrain from checking
1055 // the criteria.
1056 if (packets_inserted > 4) {
1057 // Expect preferred and actual buffer size to be no more than 2 frames.
1058 EXPECT_LE(network_stats.preferred_buffer_size_ms, kFrameSizeMs * 2);
Yves Gerey665174f2018-06-19 15:03:05 +02001059 EXPECT_LE(network_stats.current_buffer_size_ms,
1060 kFrameSizeMs * 2 + algorithmic_delay_ms_);
turaj@webrtc.org78b41a02013-11-22 20:27:07 +00001061 }
1062 last_seq_no = seq_no;
1063 last_timestamp = timestamp;
1064
1065 ++seq_no;
1066 timestamp += kSamples;
1067 receive_timestamp += kSamples;
1068 next_input_time_ms += static_cast<double>(kFrameSizeMs);
1069
1070 seq_no_wrapped |= seq_no < last_seq_no;
1071 timestamp_wrapped |= timestamp < last_timestamp;
1072 }
1073 // Pull out data once.
henrik.lundin6d8e0112016-03-04 10:34:21 -08001074 AudioFrame output;
henrik.lundin7a926812016-05-12 13:51:28 -07001075 bool muted;
1076 ASSERT_EQ(0, neteq_->GetAudio(&output, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -08001077 ASSERT_EQ(kBlockSize16kHz, output.samples_per_channel_);
1078 ASSERT_EQ(1u, output.num_channels_);
turaj@webrtc.org78b41a02013-11-22 20:27:07 +00001079
1080 // Expect delay (in samples) to be less than 2 packets.
Danil Chapovalovb6021232018-06-19 13:26:36 +02001081 absl::optional<uint32_t> playout_timestamp = neteq_->GetPlayoutTimestamp();
henrik.lundin0d96ab72016-04-06 12:28:26 -07001082 ASSERT_TRUE(playout_timestamp);
1083 EXPECT_LE(timestamp - *playout_timestamp,
turaj@webrtc.org78b41a02013-11-22 20:27:07 +00001084 static_cast<uint32_t>(kSamples * 2));
turaj@webrtc.org78b41a02013-11-22 20:27:07 +00001085 }
1086 // Make sure we have actually tested wrap-around.
1087 ASSERT_EQ(expect_seq_no_wrap, seq_no_wrapped);
1088 ASSERT_EQ(expect_timestamp_wrap, timestamp_wrapped);
1089}
1090
1091TEST_F(NetEqDecodingTest, SequenceNumberWrap) {
1092 // Start with a sequence number that will soon wrap.
1093 std::set<uint16_t> drop_seq_numbers; // Don't drop any packets.
1094 WrapTest(0xFFFF - 10, 0, drop_seq_numbers, true, false);
1095}
1096
1097TEST_F(NetEqDecodingTest, SequenceNumberWrapAndDrop) {
1098 // Start with a sequence number that will soon wrap.
1099 std::set<uint16_t> drop_seq_numbers;
1100 drop_seq_numbers.insert(0xFFFF);
1101 drop_seq_numbers.insert(0x0);
1102 WrapTest(0xFFFF - 10, 0, drop_seq_numbers, true, false);
1103}
1104
1105TEST_F(NetEqDecodingTest, TimestampWrap) {
1106 // Start with a timestamp that will soon wrap.
1107 std::set<uint16_t> drop_seq_numbers;
1108 WrapTest(0, 0xFFFFFFFF - 3000, drop_seq_numbers, false, true);
1109}
1110
1111TEST_F(NetEqDecodingTest, TimestampAndSequenceNumberWrap) {
1112 // Start with a timestamp and a sequence number that will wrap at the same
1113 // time.
1114 std::set<uint16_t> drop_seq_numbers;
1115 WrapTest(0xFFFF - 10, 0xFFFFFFFF - 5000, drop_seq_numbers, true, true);
1116}
1117
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001118void NetEqDecodingTest::DuplicateCng() {
1119 uint16_t seq_no = 0;
1120 uint32_t timestamp = 0;
1121 const int kFrameSizeMs = 10;
1122 const int kSampleRateKhz = 16;
1123 const int kSamples = kFrameSizeMs * kSampleRateKhz;
pkasting@chromium.org4591fbd2014-11-20 22:28:14 +00001124 const size_t kPayloadBytes = kSamples * 2;
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001125
Yves Gerey665174f2018-06-19 15:03:05 +02001126 const int algorithmic_delay_samples =
1127 std::max(algorithmic_delay_ms_ * kSampleRateKhz, 5 * kSampleRateKhz / 8);
henrik.lundin@webrtc.orgc93437e2014-12-01 11:42:42 +00001128 // Insert three speech packets. Three are needed to get the frame length
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001129 // correct.
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001130 uint8_t payload[kPayloadBytes] = {0};
henrik.lundin246ef3e2017-04-24 09:14:32 -07001131 RTPHeader rtp_info;
henrik.lundin7a926812016-05-12 13:51:28 -07001132 bool muted;
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001133 for (int i = 0; i < 3; ++i) {
1134 PopulateRtpInfo(seq_no, timestamp, &rtp_info);
henrik.lundin246ef3e2017-04-24 09:14:32 -07001135 ASSERT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001136 ++seq_no;
1137 timestamp += kSamples;
1138
1139 // Pull audio once.
henrik.lundin7a926812016-05-12 13:51:28 -07001140 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -08001141 ASSERT_EQ(kBlockSize16kHz, out_frame_.samples_per_channel_);
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001142 }
1143 // Verify speech output.
henrik.lundin55480f52016-03-08 02:37:57 -08001144 EXPECT_EQ(AudioFrame::kNormalSpeech, out_frame_.speech_type_);
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001145
1146 // Insert same CNG packet twice.
1147 const int kCngPeriodMs = 100;
1148 const int kCngPeriodSamples = kCngPeriodMs * kSampleRateKhz;
pkasting@chromium.org4591fbd2014-11-20 22:28:14 +00001149 size_t payload_len;
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001150 PopulateCng(seq_no, timestamp, &rtp_info, payload, &payload_len);
1151 // This is the first time this CNG packet is inserted.
henrik.lundin246ef3e2017-04-24 09:14:32 -07001152 ASSERT_EQ(
1153 0, neteq_->InsertPacket(
1154 rtp_info, rtc::ArrayView<const uint8_t>(payload, payload_len), 0));
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001155
1156 // Pull audio once and make sure CNG is played.
henrik.lundin7a926812016-05-12 13:51:28 -07001157 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -08001158 ASSERT_EQ(kBlockSize16kHz, out_frame_.samples_per_channel_);
henrik.lundin55480f52016-03-08 02:37:57 -08001159 EXPECT_EQ(AudioFrame::kCNG, out_frame_.speech_type_);
henrik.lundin114c1b32017-04-26 07:47:32 -07001160 EXPECT_FALSE(
1161 neteq_->GetPlayoutTimestamp()); // Returns empty value during CNG.
henrik.lundin0d96ab72016-04-06 12:28:26 -07001162 EXPECT_EQ(timestamp - algorithmic_delay_samples,
1163 out_frame_.timestamp_ + out_frame_.samples_per_channel_);
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001164
1165 // Insert the same CNG packet again. Note that at this point it is old, since
1166 // we have already decoded the first copy of it.
henrik.lundin246ef3e2017-04-24 09:14:32 -07001167 ASSERT_EQ(
1168 0, neteq_->InsertPacket(
1169 rtp_info, rtc::ArrayView<const uint8_t>(payload, payload_len), 0));
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001170
1171 // Pull audio until we have played |kCngPeriodMs| of CNG. Start at 10 ms since
1172 // we have already pulled out CNG once.
1173 for (int cng_time_ms = 10; cng_time_ms < kCngPeriodMs; cng_time_ms += 10) {
henrik.lundin7a926812016-05-12 13:51:28 -07001174 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -08001175 ASSERT_EQ(kBlockSize16kHz, out_frame_.samples_per_channel_);
henrik.lundin55480f52016-03-08 02:37:57 -08001176 EXPECT_EQ(AudioFrame::kCNG, out_frame_.speech_type_);
henrik.lundin114c1b32017-04-26 07:47:32 -07001177 EXPECT_FALSE(
1178 neteq_->GetPlayoutTimestamp()); // Returns empty value during CNG.
turaj@webrtc.org8d1cdaa2014-04-11 18:47:55 +00001179 EXPECT_EQ(timestamp - algorithmic_delay_samples,
henrik.lundin0d96ab72016-04-06 12:28:26 -07001180 out_frame_.timestamp_ + out_frame_.samples_per_channel_);
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001181 }
1182
1183 // Insert speech again.
1184 ++seq_no;
1185 timestamp += kCngPeriodSamples;
1186 PopulateRtpInfo(seq_no, timestamp, &rtp_info);
henrik.lundin246ef3e2017-04-24 09:14:32 -07001187 ASSERT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001188
1189 // Pull audio once and verify that the output is speech again.
henrik.lundin7a926812016-05-12 13:51:28 -07001190 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -08001191 ASSERT_EQ(kBlockSize16kHz, out_frame_.samples_per_channel_);
henrik.lundin55480f52016-03-08 02:37:57 -08001192 EXPECT_EQ(AudioFrame::kNormalSpeech, out_frame_.speech_type_);
Danil Chapovalovb6021232018-06-19 13:26:36 +02001193 absl::optional<uint32_t> playout_timestamp = neteq_->GetPlayoutTimestamp();
henrik.lundin0d96ab72016-04-06 12:28:26 -07001194 ASSERT_TRUE(playout_timestamp);
turaj@webrtc.org8d1cdaa2014-04-11 18:47:55 +00001195 EXPECT_EQ(timestamp + kSamples - algorithmic_delay_samples,
henrik.lundin0d96ab72016-04-06 12:28:26 -07001196 *playout_timestamp);
wu@webrtc.org94454b72014-06-05 20:34:08 +00001197}
1198
Yves Gerey665174f2018-06-19 15:03:05 +02001199TEST_F(NetEqDecodingTest, DiscardDuplicateCng) {
1200 DuplicateCng();
1201}
henrik.lundin@webrtc.orgc93437e2014-12-01 11:42:42 +00001202
1203TEST_F(NetEqDecodingTest, CngFirst) {
1204 uint16_t seq_no = 0;
1205 uint32_t timestamp = 0;
1206 const int kFrameSizeMs = 10;
1207 const int kSampleRateKhz = 16;
1208 const int kSamples = kFrameSizeMs * kSampleRateKhz;
1209 const int kPayloadBytes = kSamples * 2;
1210 const int kCngPeriodMs = 100;
1211 const int kCngPeriodSamples = kCngPeriodMs * kSampleRateKhz;
1212 size_t payload_len;
1213
1214 uint8_t payload[kPayloadBytes] = {0};
henrik.lundin246ef3e2017-04-24 09:14:32 -07001215 RTPHeader rtp_info;
henrik.lundin@webrtc.orgc93437e2014-12-01 11:42:42 +00001216
1217 PopulateCng(seq_no, timestamp, &rtp_info, payload, &payload_len);
henrik.lundin246ef3e2017-04-24 09:14:32 -07001218 ASSERT_EQ(
1219 NetEq::kOK,
1220 neteq_->InsertPacket(
1221 rtp_info, rtc::ArrayView<const uint8_t>(payload, payload_len), 0));
henrik.lundin@webrtc.orgc93437e2014-12-01 11:42:42 +00001222 ++seq_no;
1223 timestamp += kCngPeriodSamples;
1224
1225 // Pull audio once and make sure CNG is played.
henrik.lundin7a926812016-05-12 13:51:28 -07001226 bool muted;
1227 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -08001228 ASSERT_EQ(kBlockSize16kHz, out_frame_.samples_per_channel_);
henrik.lundin55480f52016-03-08 02:37:57 -08001229 EXPECT_EQ(AudioFrame::kCNG, out_frame_.speech_type_);
henrik.lundin@webrtc.orgc93437e2014-12-01 11:42:42 +00001230
1231 // Insert some speech packets.
henrik.lundin549d80b2016-08-25 00:44:24 -07001232 const uint32_t first_speech_timestamp = timestamp;
1233 int timeout_counter = 0;
1234 do {
1235 ASSERT_LT(timeout_counter++, 20) << "Test timed out";
henrik.lundin@webrtc.orgc93437e2014-12-01 11:42:42 +00001236 PopulateRtpInfo(seq_no, timestamp, &rtp_info);
henrik.lundin246ef3e2017-04-24 09:14:32 -07001237 ASSERT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
henrik.lundin@webrtc.orgc93437e2014-12-01 11:42:42 +00001238 ++seq_no;
1239 timestamp += kSamples;
1240
1241 // Pull audio once.
henrik.lundin7a926812016-05-12 13:51:28 -07001242 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -08001243 ASSERT_EQ(kBlockSize16kHz, out_frame_.samples_per_channel_);
henrik.lundin549d80b2016-08-25 00:44:24 -07001244 } while (!IsNewerTimestamp(out_frame_.timestamp_, first_speech_timestamp));
henrik.lundin@webrtc.orgc93437e2014-12-01 11:42:42 +00001245 // Verify speech output.
henrik.lundin55480f52016-03-08 02:37:57 -08001246 EXPECT_EQ(AudioFrame::kNormalSpeech, out_frame_.speech_type_);
henrik.lundin@webrtc.orgc93437e2014-12-01 11:42:42 +00001247}
henrik.lundin7a926812016-05-12 13:51:28 -07001248
1249class NetEqDecodingTestWithMutedState : public NetEqDecodingTest {
1250 public:
1251 NetEqDecodingTestWithMutedState() : NetEqDecodingTest() {
1252 config_.enable_muted_state = true;
1253 }
1254
1255 protected:
1256 static constexpr size_t kSamples = 10 * 16;
1257 static constexpr size_t kPayloadBytes = kSamples * 2;
1258
1259 void InsertPacket(uint32_t rtp_timestamp) {
1260 uint8_t payload[kPayloadBytes] = {0};
henrik.lundin246ef3e2017-04-24 09:14:32 -07001261 RTPHeader rtp_info;
henrik.lundin7a926812016-05-12 13:51:28 -07001262 PopulateRtpInfo(0, rtp_timestamp, &rtp_info);
henrik.lundin246ef3e2017-04-24 09:14:32 -07001263 EXPECT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
henrik.lundin7a926812016-05-12 13:51:28 -07001264 }
1265
henrik.lundin42feb512016-09-20 06:51:40 -07001266 void InsertCngPacket(uint32_t rtp_timestamp) {
1267 uint8_t payload[kPayloadBytes] = {0};
henrik.lundin246ef3e2017-04-24 09:14:32 -07001268 RTPHeader rtp_info;
henrik.lundin42feb512016-09-20 06:51:40 -07001269 size_t payload_len;
1270 PopulateCng(0, rtp_timestamp, &rtp_info, payload, &payload_len);
henrik.lundin246ef3e2017-04-24 09:14:32 -07001271 EXPECT_EQ(
1272 NetEq::kOK,
1273 neteq_->InsertPacket(
1274 rtp_info, rtc::ArrayView<const uint8_t>(payload, payload_len), 0));
henrik.lundin42feb512016-09-20 06:51:40 -07001275 }
1276
henrik.lundin7a926812016-05-12 13:51:28 -07001277 bool GetAudioReturnMuted() {
1278 bool muted;
1279 EXPECT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
1280 return muted;
1281 }
1282
1283 void GetAudioUntilMuted() {
1284 while (!GetAudioReturnMuted()) {
1285 ASSERT_LT(counter_++, 1000) << "Test timed out";
1286 }
1287 }
1288
1289 void GetAudioUntilNormal() {
1290 bool muted = false;
1291 while (out_frame_.speech_type_ != AudioFrame::kNormalSpeech) {
1292 EXPECT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
1293 ASSERT_LT(counter_++, 1000) << "Test timed out";
1294 }
1295 EXPECT_FALSE(muted);
1296 }
1297
1298 int counter_ = 0;
1299};
1300
1301// Verifies that NetEq goes in and out of muted state as expected.
1302TEST_F(NetEqDecodingTestWithMutedState, MutedState) {
1303 // Insert one speech packet.
1304 InsertPacket(0);
1305 // Pull out audio once and expect it not to be muted.
1306 EXPECT_FALSE(GetAudioReturnMuted());
1307 // Pull data until faded out.
1308 GetAudioUntilMuted();
henrik.lundina4491072017-07-06 05:23:53 -07001309 EXPECT_TRUE(out_frame_.muted());
henrik.lundin7a926812016-05-12 13:51:28 -07001310
1311 // Verify that output audio is not written during muted mode. Other parameters
1312 // should be correct, though.
1313 AudioFrame new_frame;
yujo36b1a5f2017-06-12 12:45:32 -07001314 int16_t* frame_data = new_frame.mutable_data();
1315 for (size_t i = 0; i < AudioFrame::kMaxDataSizeSamples; i++) {
1316 frame_data[i] = 17;
henrik.lundin7a926812016-05-12 13:51:28 -07001317 }
1318 bool muted;
1319 EXPECT_EQ(0, neteq_->GetAudio(&new_frame, &muted));
1320 EXPECT_TRUE(muted);
henrik.lundina4491072017-07-06 05:23:53 -07001321 EXPECT_TRUE(out_frame_.muted());
yujo36b1a5f2017-06-12 12:45:32 -07001322 for (size_t i = 0; i < AudioFrame::kMaxDataSizeSamples; i++) {
1323 EXPECT_EQ(17, frame_data[i]);
henrik.lundin7a926812016-05-12 13:51:28 -07001324 }
1325 EXPECT_EQ(out_frame_.timestamp_ + out_frame_.samples_per_channel_,
1326 new_frame.timestamp_);
1327 EXPECT_EQ(out_frame_.samples_per_channel_, new_frame.samples_per_channel_);
1328 EXPECT_EQ(out_frame_.sample_rate_hz_, new_frame.sample_rate_hz_);
1329 EXPECT_EQ(out_frame_.num_channels_, new_frame.num_channels_);
1330 EXPECT_EQ(out_frame_.speech_type_, new_frame.speech_type_);
1331 EXPECT_EQ(out_frame_.vad_activity_, new_frame.vad_activity_);
1332
1333 // Insert new data. Timestamp is corrected for the time elapsed since the last
1334 // packet. Verify that normal operation resumes.
1335 InsertPacket(kSamples * counter_);
1336 GetAudioUntilNormal();
henrik.lundina4491072017-07-06 05:23:53 -07001337 EXPECT_FALSE(out_frame_.muted());
henrik.lundin612c25e2016-05-25 08:21:04 -07001338
1339 NetEqNetworkStatistics stats;
1340 EXPECT_EQ(0, neteq_->NetworkStatistics(&stats));
1341 // NetEqNetworkStatistics::expand_rate tells the fraction of samples that were
1342 // concealment samples, in Q14 (16384 = 100%) .The vast majority should be
1343 // concealment samples in this test.
1344 EXPECT_GT(stats.expand_rate, 14000);
1345 // And, it should be greater than the speech_expand_rate.
1346 EXPECT_GT(stats.expand_rate, stats.speech_expand_rate);
henrik.lundin7a926812016-05-12 13:51:28 -07001347}
1348
1349// Verifies that NetEq goes out of muted state when given a delayed packet.
1350TEST_F(NetEqDecodingTestWithMutedState, MutedStateDelayedPacket) {
1351 // Insert one speech packet.
1352 InsertPacket(0);
1353 // Pull out audio once and expect it not to be muted.
1354 EXPECT_FALSE(GetAudioReturnMuted());
1355 // Pull data until faded out.
1356 GetAudioUntilMuted();
1357 // Insert new data. Timestamp is only corrected for the half of the time
1358 // elapsed since the last packet. That is, the new packet is delayed. Verify
1359 // that normal operation resumes.
1360 InsertPacket(kSamples * counter_ / 2);
1361 GetAudioUntilNormal();
1362}
1363
1364// Verifies that NetEq goes out of muted state when given a future packet.
1365TEST_F(NetEqDecodingTestWithMutedState, MutedStateFuturePacket) {
1366 // Insert one speech packet.
1367 InsertPacket(0);
1368 // Pull out audio once and expect it not to be muted.
1369 EXPECT_FALSE(GetAudioReturnMuted());
1370 // Pull data until faded out.
1371 GetAudioUntilMuted();
1372 // Insert new data. Timestamp is over-corrected for the time elapsed since the
1373 // last packet. That is, the new packet is too early. Verify that normal
1374 // operation resumes.
1375 InsertPacket(kSamples * counter_ * 2);
1376 GetAudioUntilNormal();
1377}
1378
1379// Verifies that NetEq goes out of muted state when given an old packet.
1380TEST_F(NetEqDecodingTestWithMutedState, MutedStateOldPacket) {
1381 // Insert one speech packet.
1382 InsertPacket(0);
1383 // Pull out audio once and expect it not to be muted.
1384 EXPECT_FALSE(GetAudioReturnMuted());
1385 // Pull data until faded out.
1386 GetAudioUntilMuted();
1387
1388 EXPECT_NE(AudioFrame::kNormalSpeech, out_frame_.speech_type_);
1389 // Insert packet which is older than the first packet.
1390 InsertPacket(kSamples * (counter_ - 1000));
1391 EXPECT_FALSE(GetAudioReturnMuted());
1392 EXPECT_EQ(AudioFrame::kNormalSpeech, out_frame_.speech_type_);
1393}
1394
henrik.lundin42feb512016-09-20 06:51:40 -07001395// Verifies that NetEq doesn't enter muted state when CNG mode is active and the
1396// packet stream is suspended for a long time.
1397TEST_F(NetEqDecodingTestWithMutedState, DoNotMuteExtendedCngWithoutPackets) {
1398 // Insert one CNG packet.
1399 InsertCngPacket(0);
1400
1401 // Pull 10 seconds of audio (10 ms audio generated per lap).
1402 for (int i = 0; i < 1000; ++i) {
1403 bool muted;
1404 EXPECT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
1405 ASSERT_FALSE(muted);
1406 }
1407 EXPECT_EQ(AudioFrame::kCNG, out_frame_.speech_type_);
1408}
1409
1410// Verifies that NetEq goes back to normal after a long CNG period with the
1411// packet stream suspended.
1412TEST_F(NetEqDecodingTestWithMutedState, RecoverAfterExtendedCngWithoutPackets) {
1413 // Insert one CNG packet.
1414 InsertCngPacket(0);
1415
1416 // Pull 10 seconds of audio (10 ms audio generated per lap).
1417 for (int i = 0; i < 1000; ++i) {
1418 bool muted;
1419 EXPECT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
1420 }
1421
1422 // Insert new data. Timestamp is corrected for the time elapsed since the last
1423 // packet. Verify that normal operation resumes.
1424 InsertPacket(kSamples * counter_);
1425 GetAudioUntilNormal();
1426}
1427
henrik.lundin7a926812016-05-12 13:51:28 -07001428class NetEqDecodingTestTwoInstances : public NetEqDecodingTest {
1429 public:
1430 NetEqDecodingTestTwoInstances() : NetEqDecodingTest() {}
1431
1432 void SetUp() override {
1433 NetEqDecodingTest::SetUp();
1434 config2_ = config_;
1435 }
1436
1437 void CreateSecondInstance() {
Ivo Creusen24192c22019-07-12 17:00:25 +02001438 neteq2_.reset(NetEq::Create(config2_, CreateBuiltinAudioDecoderFactory()));
henrik.lundin7a926812016-05-12 13:51:28 -07001439 ASSERT_TRUE(neteq2_);
1440 LoadDecoders(neteq2_.get());
1441 }
1442
1443 protected:
1444 std::unique_ptr<NetEq> neteq2_;
1445 NetEq::Config config2_;
1446};
1447
1448namespace {
1449::testing::AssertionResult AudioFramesEqualExceptData(const AudioFrame& a,
1450 const AudioFrame& b) {
1451 if (a.timestamp_ != b.timestamp_)
1452 return ::testing::AssertionFailure() << "timestamp_ diff (" << a.timestamp_
1453 << " != " << b.timestamp_ << ")";
1454 if (a.sample_rate_hz_ != b.sample_rate_hz_)
Yves Gerey665174f2018-06-19 15:03:05 +02001455 return ::testing::AssertionFailure()
1456 << "sample_rate_hz_ diff (" << a.sample_rate_hz_
1457 << " != " << b.sample_rate_hz_ << ")";
henrik.lundin7a926812016-05-12 13:51:28 -07001458 if (a.samples_per_channel_ != b.samples_per_channel_)
1459 return ::testing::AssertionFailure()
1460 << "samples_per_channel_ diff (" << a.samples_per_channel_
1461 << " != " << b.samples_per_channel_ << ")";
1462 if (a.num_channels_ != b.num_channels_)
Yves Gerey665174f2018-06-19 15:03:05 +02001463 return ::testing::AssertionFailure()
1464 << "num_channels_ diff (" << a.num_channels_
1465 << " != " << b.num_channels_ << ")";
henrik.lundin7a926812016-05-12 13:51:28 -07001466 if (a.speech_type_ != b.speech_type_)
Yves Gerey665174f2018-06-19 15:03:05 +02001467 return ::testing::AssertionFailure()
1468 << "speech_type_ diff (" << a.speech_type_
1469 << " != " << b.speech_type_ << ")";
henrik.lundin7a926812016-05-12 13:51:28 -07001470 if (a.vad_activity_ != b.vad_activity_)
Yves Gerey665174f2018-06-19 15:03:05 +02001471 return ::testing::AssertionFailure()
1472 << "vad_activity_ diff (" << a.vad_activity_
1473 << " != " << b.vad_activity_ << ")";
henrik.lundin7a926812016-05-12 13:51:28 -07001474 return ::testing::AssertionSuccess();
1475}
1476
1477::testing::AssertionResult AudioFramesEqual(const AudioFrame& a,
1478 const AudioFrame& b) {
1479 ::testing::AssertionResult res = AudioFramesEqualExceptData(a, b);
1480 if (!res)
1481 return res;
Yves Gerey665174f2018-06-19 15:03:05 +02001482 if (memcmp(a.data(), b.data(),
1483 a.samples_per_channel_ * a.num_channels_ * sizeof(*a.data())) !=
1484 0) {
henrik.lundin7a926812016-05-12 13:51:28 -07001485 return ::testing::AssertionFailure() << "data_ diff";
1486 }
1487 return ::testing::AssertionSuccess();
1488}
1489
1490} // namespace
1491
1492TEST_F(NetEqDecodingTestTwoInstances, CompareMutedStateOnOff) {
1493 ASSERT_FALSE(config_.enable_muted_state);
1494 config2_.enable_muted_state = true;
1495 CreateSecondInstance();
1496
1497 // Insert one speech packet into both NetEqs.
1498 const size_t kSamples = 10 * 16;
1499 const size_t kPayloadBytes = kSamples * 2;
1500 uint8_t payload[kPayloadBytes] = {0};
henrik.lundin246ef3e2017-04-24 09:14:32 -07001501 RTPHeader rtp_info;
henrik.lundin7a926812016-05-12 13:51:28 -07001502 PopulateRtpInfo(0, 0, &rtp_info);
henrik.lundin246ef3e2017-04-24 09:14:32 -07001503 EXPECT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
1504 EXPECT_EQ(0, neteq2_->InsertPacket(rtp_info, payload, 0));
henrik.lundin7a926812016-05-12 13:51:28 -07001505
1506 AudioFrame out_frame1, out_frame2;
1507 bool muted;
1508 for (int i = 0; i < 1000; ++i) {
Jonas Olsson366a50c2018-09-06 13:41:30 +02001509 rtc::StringBuilder ss;
henrik.lundin7a926812016-05-12 13:51:28 -07001510 ss << "i = " << i;
1511 SCOPED_TRACE(ss.str()); // Print out the loop iterator on failure.
1512 EXPECT_EQ(0, neteq_->GetAudio(&out_frame1, &muted));
1513 EXPECT_FALSE(muted);
1514 EXPECT_EQ(0, neteq2_->GetAudio(&out_frame2, &muted));
1515 if (muted) {
1516 EXPECT_TRUE(AudioFramesEqualExceptData(out_frame1, out_frame2));
1517 } else {
1518 EXPECT_TRUE(AudioFramesEqual(out_frame1, out_frame2));
1519 }
1520 }
1521 EXPECT_TRUE(muted);
1522
1523 // Insert new data. Timestamp is corrected for the time elapsed since the last
1524 // packet.
1525 PopulateRtpInfo(0, kSamples * 1000, &rtp_info);
henrik.lundin246ef3e2017-04-24 09:14:32 -07001526 EXPECT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
1527 EXPECT_EQ(0, neteq2_->InsertPacket(rtp_info, payload, 0));
henrik.lundin7a926812016-05-12 13:51:28 -07001528
1529 int counter = 0;
1530 while (out_frame1.speech_type_ != AudioFrame::kNormalSpeech) {
1531 ASSERT_LT(counter++, 1000) << "Test timed out";
Jonas Olsson366a50c2018-09-06 13:41:30 +02001532 rtc::StringBuilder ss;
henrik.lundin7a926812016-05-12 13:51:28 -07001533 ss << "counter = " << counter;
1534 SCOPED_TRACE(ss.str()); // Print out the loop iterator on failure.
1535 EXPECT_EQ(0, neteq_->GetAudio(&out_frame1, &muted));
1536 EXPECT_FALSE(muted);
1537 EXPECT_EQ(0, neteq2_->GetAudio(&out_frame2, &muted));
1538 if (muted) {
1539 EXPECT_TRUE(AudioFramesEqualExceptData(out_frame1, out_frame2));
1540 } else {
1541 EXPECT_TRUE(AudioFramesEqual(out_frame1, out_frame2));
1542 }
1543 }
1544 EXPECT_FALSE(muted);
1545}
1546
henrik.lundin114c1b32017-04-26 07:47:32 -07001547TEST_F(NetEqDecodingTest, LastDecodedTimestampsEmpty) {
1548 EXPECT_TRUE(neteq_->LastDecodedTimestamps().empty());
1549
1550 // Pull out data once.
1551 AudioFrame output;
1552 bool muted;
1553 ASSERT_EQ(0, neteq_->GetAudio(&output, &muted));
1554
1555 EXPECT_TRUE(neteq_->LastDecodedTimestamps().empty());
1556}
1557
1558TEST_F(NetEqDecodingTest, LastDecodedTimestampsOneDecoded) {
1559 // Insert one packet with PCM16b WB data (this is what PopulateRtpInfo does by
1560 // default). Make the length 10 ms.
1561 constexpr size_t kPayloadSamples = 16 * 10;
1562 constexpr size_t kPayloadBytes = 2 * kPayloadSamples;
1563 uint8_t payload[kPayloadBytes] = {0};
1564
1565 RTPHeader rtp_info;
1566 constexpr uint32_t kRtpTimestamp = 0x1234;
1567 PopulateRtpInfo(0, kRtpTimestamp, &rtp_info);
1568 EXPECT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
1569
1570 // Pull out data once.
1571 AudioFrame output;
1572 bool muted;
1573 ASSERT_EQ(0, neteq_->GetAudio(&output, &muted));
1574
1575 EXPECT_EQ(std::vector<uint32_t>({kRtpTimestamp}),
1576 neteq_->LastDecodedTimestamps());
1577
1578 // Nothing decoded on the second call.
1579 ASSERT_EQ(0, neteq_->GetAudio(&output, &muted));
1580 EXPECT_TRUE(neteq_->LastDecodedTimestamps().empty());
1581}
1582
1583TEST_F(NetEqDecodingTest, LastDecodedTimestampsTwoDecoded) {
1584 // Insert two packets with PCM16b WB data (this is what PopulateRtpInfo does
1585 // by default). Make the length 5 ms so that NetEq must decode them both in
1586 // the same GetAudio call.
1587 constexpr size_t kPayloadSamples = 16 * 5;
1588 constexpr size_t kPayloadBytes = 2 * kPayloadSamples;
1589 uint8_t payload[kPayloadBytes] = {0};
1590
1591 RTPHeader rtp_info;
1592 constexpr uint32_t kRtpTimestamp1 = 0x1234;
1593 PopulateRtpInfo(0, kRtpTimestamp1, &rtp_info);
1594 EXPECT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
1595 constexpr uint32_t kRtpTimestamp2 = kRtpTimestamp1 + kPayloadSamples;
1596 PopulateRtpInfo(1, kRtpTimestamp2, &rtp_info);
1597 EXPECT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
1598
1599 // Pull out data once.
1600 AudioFrame output;
1601 bool muted;
1602 ASSERT_EQ(0, neteq_->GetAudio(&output, &muted));
1603
1604 EXPECT_EQ(std::vector<uint32_t>({kRtpTimestamp1, kRtpTimestamp2}),
1605 neteq_->LastDecodedTimestamps());
1606}
1607
Gustaf Ullberg9a2e9062017-09-18 09:28:20 +02001608TEST_F(NetEqDecodingTest, TestConcealmentEvents) {
1609 const int kNumConcealmentEvents = 19;
1610 const size_t kSamples = 10 * 16;
1611 const size_t kPayloadBytes = kSamples * 2;
1612 int seq_no = 0;
1613 RTPHeader rtp_info;
1614 rtp_info.ssrc = 0x1234; // Just an arbitrary SSRC.
1615 rtp_info.payloadType = 94; // PCM16b WB codec.
1616 rtp_info.markerBit = 0;
1617 const uint8_t payload[kPayloadBytes] = {0};
1618 bool muted;
1619
1620 for (int i = 0; i < kNumConcealmentEvents; i++) {
1621 // Insert some packets of 10 ms size.
1622 for (int j = 0; j < 10; j++) {
1623 rtp_info.sequenceNumber = seq_no++;
1624 rtp_info.timestamp = rtp_info.sequenceNumber * kSamples;
1625 neteq_->InsertPacket(rtp_info, payload, 0);
1626 neteq_->GetAudio(&out_frame_, &muted);
1627 }
1628
1629 // Lose a number of packets.
1630 int num_lost = 1 + i;
1631 for (int j = 0; j < num_lost; j++) {
1632 seq_no++;
1633 neteq_->GetAudio(&out_frame_, &muted);
1634 }
1635 }
1636
1637 // Check number of concealment events.
1638 NetEqLifetimeStatistics stats = neteq_->GetLifetimeStatistics();
1639 EXPECT_EQ(kNumConcealmentEvents, static_cast<int>(stats.concealment_events));
1640}
1641
Gustaf Ullbergb0a02072017-10-02 12:00:34 +02001642// Test that the jitter buffer delay stat is computed correctly.
1643void NetEqDecodingTestFaxMode::TestJitterBufferDelay(bool apply_packet_loss) {
1644 const int kNumPackets = 10;
1645 const int kDelayInNumPackets = 2;
1646 const int kPacketLenMs = 10; // All packets are of 10 ms size.
1647 const size_t kSamples = kPacketLenMs * 16;
1648 const size_t kPayloadBytes = kSamples * 2;
1649 RTPHeader rtp_info;
1650 rtp_info.ssrc = 0x1234; // Just an arbitrary SSRC.
1651 rtp_info.payloadType = 94; // PCM16b WB codec.
1652 rtp_info.markerBit = 0;
1653 const uint8_t payload[kPayloadBytes] = {0};
1654 bool muted;
1655 int packets_sent = 0;
1656 int packets_received = 0;
1657 int expected_delay = 0;
Chen Xing0acffb52019-01-15 15:46:29 +01001658 uint64_t expected_emitted_count = 0;
Gustaf Ullbergb0a02072017-10-02 12:00:34 +02001659 while (packets_received < kNumPackets) {
1660 // Insert packet.
1661 if (packets_sent < kNumPackets) {
1662 rtp_info.sequenceNumber = packets_sent++;
1663 rtp_info.timestamp = rtp_info.sequenceNumber * kSamples;
1664 neteq_->InsertPacket(rtp_info, payload, 0);
1665 }
1666
1667 // Get packet.
1668 if (packets_sent > kDelayInNumPackets) {
1669 neteq_->GetAudio(&out_frame_, &muted);
1670 packets_received++;
1671
1672 // The delay reported by the jitter buffer never exceeds
1673 // the number of samples previously fetched with GetAudio
1674 // (hence the min()).
1675 int packets_delay = std::min(packets_received, kDelayInNumPackets + 1);
1676
1677 // The increase of the expected delay is the product of
1678 // the current delay of the jitter buffer in ms * the
1679 // number of samples that are sent for play out.
1680 int current_delay_ms = packets_delay * kPacketLenMs;
1681 expected_delay += current_delay_ms * kSamples;
Chen Xing0acffb52019-01-15 15:46:29 +01001682 expected_emitted_count += kSamples;
Gustaf Ullbergb0a02072017-10-02 12:00:34 +02001683 }
1684 }
1685
1686 if (apply_packet_loss) {
1687 // Extra call to GetAudio to cause concealment.
1688 neteq_->GetAudio(&out_frame_, &muted);
1689 }
1690
1691 // Check jitter buffer delay.
1692 NetEqLifetimeStatistics stats = neteq_->GetLifetimeStatistics();
1693 EXPECT_EQ(expected_delay, static_cast<int>(stats.jitter_buffer_delay_ms));
Chen Xing0acffb52019-01-15 15:46:29 +01001694 EXPECT_EQ(expected_emitted_count, stats.jitter_buffer_emitted_count);
Gustaf Ullbergb0a02072017-10-02 12:00:34 +02001695}
1696
1697TEST_F(NetEqDecodingTestFaxMode, TestJitterBufferDelayWithoutLoss) {
1698 TestJitterBufferDelay(false);
1699}
1700
1701TEST_F(NetEqDecodingTestFaxMode, TestJitterBufferDelayWithLoss) {
1702 TestJitterBufferDelay(true);
1703}
1704
Jakob Ivarsson26c59ff2019-02-28 09:55:49 +01001705TEST_F(NetEqDecodingTestFaxMode, TestJitterBufferDelayWithAcceleration) {
1706 const int kPacketLenMs = 10; // All packets are of 10 ms size.
1707 const size_t kSamples = kPacketLenMs * 16;
1708 const size_t kPayloadBytes = kSamples * 2;
1709 RTPHeader rtp_info;
1710 rtp_info.ssrc = 0x1234; // Just an arbitrary SSRC.
1711 rtp_info.payloadType = 94; // PCM16b WB codec.
1712 rtp_info.markerBit = 0;
1713 const uint8_t payload[kPayloadBytes] = {0};
1714
1715 neteq_->InsertPacket(rtp_info, payload, 0);
1716
1717 bool muted;
1718 neteq_->GetAudio(&out_frame_, &muted);
1719
1720 rtp_info.sequenceNumber += 1;
1721 rtp_info.timestamp += kSamples;
1722 neteq_->InsertPacket(rtp_info, payload, 0);
1723 rtp_info.sequenceNumber += 1;
1724 rtp_info.timestamp += kSamples;
1725 neteq_->InsertPacket(rtp_info, payload, 0);
1726
1727 // We have two packets in the buffer and kAccelerate operation will
1728 // extract 20 ms of data.
1729 neteq_->GetAudio(&out_frame_, &muted, Operations::kAccelerate);
1730
1731 // Check jitter buffer delay.
1732 NetEqLifetimeStatistics stats = neteq_->GetLifetimeStatistics();
1733 EXPECT_EQ(10 * kSamples * 3, stats.jitter_buffer_delay_ms);
1734 EXPECT_EQ(kSamples * 3, stats.jitter_buffer_emitted_count);
1735}
1736
Henrik Lundin7687ad52018-07-02 10:14:46 +02001737namespace test {
Henrik Lundin7687ad52018-07-02 10:14:46 +02001738TEST(NetEqNoTimeStretchingMode, RunTest) {
1739 NetEq::Config config;
1740 config.for_test_no_time_stretching = true;
1741 auto codecs = NetEqTest::StandardDecoderMap();
Henrik Lundin7687ad52018-07-02 10:14:46 +02001742 NetEqPacketSourceInput::RtpHeaderExtensionMap rtp_ext_map = {
1743 {1, kRtpExtensionAudioLevel},
1744 {3, kRtpExtensionAbsoluteSendTime},
1745 {5, kRtpExtensionTransportSequenceNumber},
1746 {7, kRtpExtensionVideoContentType},
1747 {8, kRtpExtensionVideoTiming}};
1748 std::unique_ptr<NetEqInput> input(new NetEqRtpDumpInput(
1749 webrtc::test::ResourcePath("audio_coding/neteq_universal_new", "rtp"),
Bjorn Terelius5350d1c2018-10-11 16:51:23 +02001750 rtp_ext_map, absl::nullopt /*No SSRC filter*/));
Henrik Lundin7687ad52018-07-02 10:14:46 +02001751 std::unique_ptr<TimeLimitedNetEqInput> input_time_limit(
1752 new TimeLimitedNetEqInput(std::move(input), 20000));
1753 std::unique_ptr<AudioSink> output(new VoidAudioSink);
1754 NetEqTest::Callbacks callbacks;
Niels Möllerbd6dee82019-01-02 09:39:47 +01001755 NetEqTest test(config, CreateBuiltinAudioDecoderFactory(), codecs, nullptr,
1756 std::move(input_time_limit), std::move(output), callbacks);
Henrik Lundin7687ad52018-07-02 10:14:46 +02001757 test.Run();
1758 const auto stats = test.SimulationStats();
1759 EXPECT_EQ(0, stats.accelerate_rate);
1760 EXPECT_EQ(0, stats.preemptive_rate);
1761}
Henrik Lundin7687ad52018-07-02 10:14:46 +02001762
1763} // namespace test
henrik.lundin@webrtc.orge7ce4372014-01-09 14:01:55 +00001764} // namespace webrtc