blob: 4ed7a6b0882aecf20f9482f2cb7d15e5cc2e4d5b [file] [log] [blame]
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +00001/*
2 * Copyright (c) 2011 The WebRTC project authors. All Rights Reserved.
3 *
4 * Use of this source code is governed by a BSD-style license
5 * that can be found in the LICENSE file in the root of the source
6 * tree. An additional intellectual property rights grant can be found
7 * in the file PATENTS. All contributing project authors may
8 * be found in the AUTHORS file in the root of the source tree.
9 */
10
Mirko Bonadei92ea95e2017-09-15 06:47:31 +020011#include "modules/audio_coding/neteq/include/neteq.h"
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +000012
pbos@webrtc.org3ecc1622014-03-07 15:23:34 +000013#include <math.h>
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +000014#include <stdlib.h>
15#include <string.h> // memset
16
turaj@webrtc.org8d1cdaa2014-04-11 18:47:55 +000017#include <algorithm>
kwiberg2d0c3322016-02-14 09:28:33 -080018#include <memory>
turaj@webrtc.org78b41a02013-11-22 20:27:07 +000019#include <set>
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +000020#include <string>
21#include <vector>
22
Fredrik Solenbergbbf21a32018-04-12 22:44:09 +020023#include "api/audio/audio_frame.h"
Mirko Bonadei92ea95e2017-09-15 06:47:31 +020024#include "api/audio_codecs/builtin_audio_decoder_factory.h"
Mirko Bonadei71207422017-09-15 13:58:09 +020025#include "common_types.h" // NOLINT(build/include)
Mirko Bonadei92ea95e2017-09-15 06:47:31 +020026#include "modules/audio_coding/codecs/pcm16b/pcm16b.h"
27#include "modules/audio_coding/neteq/tools/audio_loop.h"
28#include "modules/audio_coding/neteq/tools/rtp_file_source.h"
Mirko Bonadei92ea95e2017-09-15 06:47:31 +020029#include "rtc_base/ignore_wundef.h"
Joachim Bauch4e909192017-12-19 22:27:51 +010030#include "rtc_base/messagedigest.h"
Karl Wiberge40468b2017-11-22 10:42:26 +010031#include "rtc_base/numerics/safe_conversions.h"
Mirko Bonadei92ea95e2017-09-15 06:47:31 +020032#include "rtc_base/protobuf_utils.h"
Mirko Bonadei92ea95e2017-09-15 06:47:31 +020033#include "rtc_base/stringencode.h"
Henrik Lundine9619f82017-11-27 14:05:27 +010034#include "test/field_trial.h"
Mirko Bonadei92ea95e2017-09-15 06:47:31 +020035#include "test/gtest.h"
36#include "test/testsupport/fileutils.h"
Mirko Bonadei71207422017-09-15 13:58:09 +020037#include "typedefs.h" // NOLINT(build/include)
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +000038
Mirko Bonadei81ca3bf2018-01-09 09:40:39 +010039// This must come after test/gtest.h
40#include "rtc_base/flags.h" // NOLINT(build/include)
41
minyue5f026d02015-12-16 07:36:04 -080042#ifdef WEBRTC_NETEQ_UNITTEST_BITEXACT
kwiberg77eab702016-09-28 17:42:01 -070043RTC_PUSH_IGNORING_WUNDEF()
minyue5f026d02015-12-16 07:36:04 -080044#ifdef WEBRTC_ANDROID_PLATFORM_BUILD
45#include "external/webrtc/webrtc/modules/audio_coding/neteq/neteq_unittest.pb.h"
46#else
Mirko Bonadei92ea95e2017-09-15 06:47:31 +020047#include "modules/audio_coding/neteq/neteq_unittest.pb.h"
minyue5f026d02015-12-16 07:36:04 -080048#endif
kwiberg77eab702016-09-28 17:42:01 -070049RTC_POP_IGNORING_WUNDEF()
minyue5f026d02015-12-16 07:36:04 -080050#endif
51
turaj@webrtc.orga6101d72013-10-01 22:01:09 +000052DEFINE_bool(gen_ref, false, "Generate reference files.");
53
kwiberg5adaf732016-10-04 09:33:27 -070054namespace webrtc {
55
minyue5f026d02015-12-16 07:36:04 -080056namespace {
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +000057
minyue4f906772016-04-29 11:05:14 -070058const std::string& PlatformChecksum(const std::string& checksum_general,
Henrik Lundin8cd750d2017-10-12 13:07:11 +020059 const std::string& checksum_android_32,
60 const std::string& checksum_android_64,
minyue4f906772016-04-29 11:05:14 -070061 const std::string& checksum_win_32,
62 const std::string& checksum_win_64) {
kwiberg77eab702016-09-28 17:42:01 -070063#if defined(WEBRTC_ANDROID)
Yves Gerey665174f2018-06-19 15:03:05 +020064#ifdef WEBRTC_ARCH_64_BITS
65 return checksum_android_64;
66#else
67 return checksum_android_32;
68#endif // WEBRTC_ARCH_64_BITS
kwiberg77eab702016-09-28 17:42:01 -070069#elif defined(WEBRTC_WIN)
Yves Gerey665174f2018-06-19 15:03:05 +020070#ifdef WEBRTC_ARCH_64_BITS
71 return checksum_win_64;
72#else
73 return checksum_win_32;
74#endif // WEBRTC_ARCH_64_BITS
minyue4f906772016-04-29 11:05:14 -070075#else
76 return checksum_general;
77#endif // WEBRTC_WIN
78}
79
minyue5f026d02015-12-16 07:36:04 -080080#ifdef WEBRTC_NETEQ_UNITTEST_BITEXACT
81void Convert(const webrtc::NetEqNetworkStatistics& stats_raw,
82 webrtc::neteq_unittest::NetEqNetworkStatistics* stats) {
83 stats->set_current_buffer_size_ms(stats_raw.current_buffer_size_ms);
84 stats->set_preferred_buffer_size_ms(stats_raw.preferred_buffer_size_ms);
85 stats->set_jitter_peaks_found(stats_raw.jitter_peaks_found);
86 stats->set_packet_loss_rate(stats_raw.packet_loss_rate);
minyue5f026d02015-12-16 07:36:04 -080087 stats->set_expand_rate(stats_raw.expand_rate);
88 stats->set_speech_expand_rate(stats_raw.speech_expand_rate);
89 stats->set_preemptive_rate(stats_raw.preemptive_rate);
90 stats->set_accelerate_rate(stats_raw.accelerate_rate);
91 stats->set_secondary_decoded_rate(stats_raw.secondary_decoded_rate);
minyue-webrtc0c3ca752017-08-23 15:59:38 +020092 stats->set_secondary_discarded_rate(stats_raw.secondary_discarded_rate);
minyue5f026d02015-12-16 07:36:04 -080093 stats->set_clockdrift_ppm(stats_raw.clockdrift_ppm);
94 stats->set_added_zero_samples(stats_raw.added_zero_samples);
95 stats->set_mean_waiting_time_ms(stats_raw.mean_waiting_time_ms);
96 stats->set_median_waiting_time_ms(stats_raw.median_waiting_time_ms);
97 stats->set_min_waiting_time_ms(stats_raw.min_waiting_time_ms);
98 stats->set_max_waiting_time_ms(stats_raw.max_waiting_time_ms);
99}
100
101void Convert(const webrtc::RtcpStatistics& stats_raw,
102 webrtc::neteq_unittest::RtcpStatistics* stats) {
103 stats->set_fraction_lost(stats_raw.fraction_lost);
srte186d9c32017-08-04 05:03:53 -0700104 stats->set_cumulative_lost(stats_raw.packets_lost);
minyue5f026d02015-12-16 07:36:04 -0800105 stats->set_extended_max_sequence_number(
srte186d9c32017-08-04 05:03:53 -0700106 stats_raw.extended_highest_sequence_number);
minyue5f026d02015-12-16 07:36:04 -0800107 stats->set_jitter(stats_raw.jitter);
108}
109
Yves Gerey665174f2018-06-19 15:03:05 +0200110void AddMessage(FILE* file,
111 rtc::MessageDigest* digest,
minyue4f906772016-04-29 11:05:14 -0700112 const std::string& message) {
minyue5f026d02015-12-16 07:36:04 -0800113 int32_t size = message.length();
minyue4f906772016-04-29 11:05:14 -0700114 if (file)
115 ASSERT_EQ(1u, fwrite(&size, sizeof(size), 1, file));
116 digest->Update(&size, sizeof(size));
117
118 if (file)
119 ASSERT_EQ(static_cast<size_t>(size),
120 fwrite(message.data(), sizeof(char), size, file));
121 digest->Update(message.data(), sizeof(char) * size);
minyue5f026d02015-12-16 07:36:04 -0800122}
123
minyue5f026d02015-12-16 07:36:04 -0800124#endif // WEBRTC_NETEQ_UNITTEST_BITEXACT
125
henrik.lundin7a926812016-05-12 13:51:28 -0700126void LoadDecoders(webrtc::NetEq* neteq) {
kwiberg5adaf732016-10-04 09:33:27 -0700127 ASSERT_EQ(true,
128 neteq->RegisterPayloadType(0, SdpAudioFormat("pcmu", 8000, 1)));
129 // Use non-SdpAudioFormat argument when registering PCMa, so that we get test
130 // coverage for that as well.
henrik.lundin7a926812016-05-12 13:51:28 -0700131 ASSERT_EQ(0, neteq->RegisterPayloadType(webrtc::NetEqDecoder::kDecoderPCMa,
132 "pcma", 8));
133#ifdef WEBRTC_CODEC_ILBC
kwiberg5adaf732016-10-04 09:33:27 -0700134 ASSERT_EQ(true,
135 neteq->RegisterPayloadType(102, SdpAudioFormat("ilbc", 8000, 1)));
henrik.lundin7a926812016-05-12 13:51:28 -0700136#endif
137#if defined(WEBRTC_CODEC_ISAC) || defined(WEBRTC_CODEC_ISACFX)
kwiberg5adaf732016-10-04 09:33:27 -0700138 ASSERT_EQ(true,
139 neteq->RegisterPayloadType(103, SdpAudioFormat("isac", 16000, 1)));
henrik.lundin7a926812016-05-12 13:51:28 -0700140#endif
141#ifdef WEBRTC_CODEC_ISAC
kwiberg5adaf732016-10-04 09:33:27 -0700142 ASSERT_EQ(true,
143 neteq->RegisterPayloadType(104, SdpAudioFormat("isac", 32000, 1)));
henrik.lundin7a926812016-05-12 13:51:28 -0700144#endif
145#ifdef WEBRTC_CODEC_OPUS
kwiberg5adaf732016-10-04 09:33:27 -0700146 ASSERT_EQ(true,
147 neteq->RegisterPayloadType(
148 111, SdpAudioFormat("opus", 48000, 2, {{"stereo", "0"}})));
henrik.lundin7a926812016-05-12 13:51:28 -0700149#endif
kwiberg5adaf732016-10-04 09:33:27 -0700150 ASSERT_EQ(true,
151 neteq->RegisterPayloadType(93, SdpAudioFormat("L16", 8000, 1)));
152 ASSERT_EQ(true,
153 neteq->RegisterPayloadType(94, SdpAudioFormat("L16", 16000, 1)));
154 ASSERT_EQ(true,
155 neteq->RegisterPayloadType(95, SdpAudioFormat("L16", 32000, 1)));
156 ASSERT_EQ(true,
157 neteq->RegisterPayloadType(13, SdpAudioFormat("cn", 8000, 1)));
158 ASSERT_EQ(true,
159 neteq->RegisterPayloadType(98, SdpAudioFormat("cn", 16000, 1)));
henrik.lundin7a926812016-05-12 13:51:28 -0700160}
minyue5f026d02015-12-16 07:36:04 -0800161} // namespace
162
minyue4f906772016-04-29 11:05:14 -0700163class ResultSink {
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000164 public:
minyue4f906772016-04-29 11:05:14 -0700165 explicit ResultSink(const std::string& output_file);
166 ~ResultSink();
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000167
Yves Gerey665174f2018-06-19 15:03:05 +0200168 template <typename T>
169 void AddResult(const T* test_results, size_t length);
minyue4f906772016-04-29 11:05:14 -0700170
171 void AddResult(const NetEqNetworkStatistics& stats);
172 void AddResult(const RtcpStatistics& stats);
173
174 void VerifyChecksum(const std::string& ref_check_sum);
175
176 private:
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000177 FILE* output_fp_;
minyue4f906772016-04-29 11:05:14 -0700178 std::unique_ptr<rtc::MessageDigest> digest_;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000179};
180
Joachim Bauch4e909192017-12-19 22:27:51 +0100181ResultSink::ResultSink(const std::string& output_file)
minyue4f906772016-04-29 11:05:14 -0700182 : output_fp_(nullptr),
Joachim Bauch4e909192017-12-19 22:27:51 +0100183 digest_(rtc::MessageDigestFactory::Create(rtc::DIGEST_SHA_1)) {
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000184 if (!output_file.empty()) {
185 output_fp_ = fopen(output_file.c_str(), "wb");
186 EXPECT_TRUE(output_fp_ != NULL);
187 }
188}
189
minyue4f906772016-04-29 11:05:14 -0700190ResultSink::~ResultSink() {
191 if (output_fp_)
192 fclose(output_fp_);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000193}
194
Yves Gerey665174f2018-06-19 15:03:05 +0200195template <typename T>
yujo36b1a5f2017-06-12 12:45:32 -0700196void ResultSink::AddResult(const T* test_results, size_t length) {
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000197 if (output_fp_) {
yujo36b1a5f2017-06-12 12:45:32 -0700198 ASSERT_EQ(length, fwrite(test_results, sizeof(T), length, output_fp_));
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000199 }
yujo36b1a5f2017-06-12 12:45:32 -0700200 digest_->Update(test_results, sizeof(T) * length);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000201}
202
minyue4f906772016-04-29 11:05:14 -0700203void ResultSink::AddResult(const NetEqNetworkStatistics& stats_raw) {
minyue5f026d02015-12-16 07:36:04 -0800204#ifdef WEBRTC_NETEQ_UNITTEST_BITEXACT
minyue5f026d02015-12-16 07:36:04 -0800205 neteq_unittest::NetEqNetworkStatistics stats;
206 Convert(stats_raw, &stats);
207
mbonadei7c2c8432017-04-07 00:59:12 -0700208 ProtoString stats_string;
minyue5f026d02015-12-16 07:36:04 -0800209 ASSERT_TRUE(stats.SerializeToString(&stats_string));
minyue4f906772016-04-29 11:05:14 -0700210 AddMessage(output_fp_, digest_.get(), stats_string);
minyue5f026d02015-12-16 07:36:04 -0800211#else
212 FAIL() << "Writing to reference file requires Proto Buffer.";
213#endif // WEBRTC_NETEQ_UNITTEST_BITEXACT
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000214}
215
minyue4f906772016-04-29 11:05:14 -0700216void ResultSink::AddResult(const RtcpStatistics& stats_raw) {
minyue5f026d02015-12-16 07:36:04 -0800217#ifdef WEBRTC_NETEQ_UNITTEST_BITEXACT
minyue5f026d02015-12-16 07:36:04 -0800218 neteq_unittest::RtcpStatistics stats;
219 Convert(stats_raw, &stats);
220
mbonadei7c2c8432017-04-07 00:59:12 -0700221 ProtoString stats_string;
minyue5f026d02015-12-16 07:36:04 -0800222 ASSERT_TRUE(stats.SerializeToString(&stats_string));
minyue4f906772016-04-29 11:05:14 -0700223 AddMessage(output_fp_, digest_.get(), stats_string);
minyue5f026d02015-12-16 07:36:04 -0800224#else
225 FAIL() << "Writing to reference file requires Proto Buffer.";
226#endif // WEBRTC_NETEQ_UNITTEST_BITEXACT
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000227}
228
minyue4f906772016-04-29 11:05:14 -0700229void ResultSink::VerifyChecksum(const std::string& checksum) {
230 std::vector<char> buffer;
231 buffer.resize(digest_->Size());
232 digest_->Finish(&buffer[0], buffer.size());
233 const std::string result = rtc::hex_encode(&buffer[0], digest_->Size());
234 EXPECT_EQ(checksum, result);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000235}
236
237class NetEqDecodingTest : public ::testing::Test {
238 protected:
239 // NetEQ must be polled for data once every 10 ms. Thus, neither of the
240 // constants below can be changed.
241 static const int kTimeStepMs = 10;
Peter Kastingdce40cf2015-08-24 14:52:23 -0700242 static const size_t kBlockSize8kHz = kTimeStepMs * 8;
243 static const size_t kBlockSize16kHz = kTimeStepMs * 16;
244 static const size_t kBlockSize32kHz = kTimeStepMs * 32;
minyue93c08b72015-12-22 09:57:41 -0800245 static const size_t kBlockSize48kHz = kTimeStepMs * 48;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000246 static const int kInitSampleRateHz = 8000;
247
248 NetEqDecodingTest();
249 virtual void SetUp();
250 virtual void TearDown();
251 void SelectDecoders(NetEqDecoder* used_codec);
Yves Gerey665174f2018-06-19 15:03:05 +0200252 void OpenInputFile(const std::string& rtp_file);
henrik.lundin6d8e0112016-03-04 10:34:21 -0800253 void Process();
minyue5f026d02015-12-16 07:36:04 -0800254
henrik.lundin@webrtc.org4e4b0982014-08-11 14:48:49 +0000255 void DecodeAndCompare(const std::string& rtp_file,
minyue4f906772016-04-29 11:05:14 -0700256 const std::string& output_checksum,
257 const std::string& network_stats_checksum,
258 const std::string& rtcp_stats_checksum,
259 bool gen_ref);
minyue5f026d02015-12-16 07:36:04 -0800260
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000261 static void PopulateRtpInfo(int frame_index,
262 int timestamp,
henrik.lundin246ef3e2017-04-24 09:14:32 -0700263 RTPHeader* rtp_info);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000264 static void PopulateCng(int frame_index,
265 int timestamp,
henrik.lundin246ef3e2017-04-24 09:14:32 -0700266 RTPHeader* rtp_info,
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000267 uint8_t* payload,
pkasting@chromium.org4591fbd2014-11-20 22:28:14 +0000268 size_t* payload_len);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000269
Yves Gerey665174f2018-06-19 15:03:05 +0200270 void WrapTest(uint16_t start_seq_no,
271 uint32_t start_timestamp,
turaj@webrtc.org78b41a02013-11-22 20:27:07 +0000272 const std::set<uint16_t>& drop_seq_numbers,
Yves Gerey665174f2018-06-19 15:03:05 +0200273 bool expect_seq_no_wrap,
274 bool expect_timestamp_wrap);
turaj@webrtc.org78b41a02013-11-22 20:27:07 +0000275
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000276 void LongCngWithClockDrift(double drift_factor,
277 double network_freeze_ms,
278 bool pull_audio_during_freeze,
279 int delay_tolerance_ms,
280 int max_time_to_speech_ms);
281
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +0000282 void DuplicateCng();
henrik.lundin@webrtc.orgfcfc6a92014-02-13 11:42:28 +0000283
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000284 NetEq* neteq_;
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +0000285 NetEq::Config config_;
kwiberg2d0c3322016-02-14 09:28:33 -0800286 std::unique_ptr<test::RtpFileSource> rtp_source_;
287 std::unique_ptr<test::Packet> packet_;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000288 unsigned int sim_clock_;
henrik.lundin6d8e0112016-03-04 10:34:21 -0800289 AudioFrame out_frame_;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000290 int output_sample_rate_;
turaj@webrtc.org8d1cdaa2014-04-11 18:47:55 +0000291 int algorithmic_delay_ms_;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000292};
293
294// Allocating the static const so that it can be passed by reference.
295const int NetEqDecodingTest::kTimeStepMs;
Peter Kastingdce40cf2015-08-24 14:52:23 -0700296const size_t NetEqDecodingTest::kBlockSize8kHz;
297const size_t NetEqDecodingTest::kBlockSize16kHz;
298const size_t NetEqDecodingTest::kBlockSize32kHz;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000299const int NetEqDecodingTest::kInitSampleRateHz;
300
301NetEqDecodingTest::NetEqDecodingTest()
302 : neteq_(NULL),
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +0000303 config_(),
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000304 sim_clock_(0),
turaj@webrtc.org8d1cdaa2014-04-11 18:47:55 +0000305 output_sample_rate_(kInitSampleRateHz),
306 algorithmic_delay_ms_(0) {
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +0000307 config_.sample_rate_hz = kInitSampleRateHz;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000308}
309
310void NetEqDecodingTest::SetUp() {
ossue3525782016-05-25 07:37:43 -0700311 neteq_ = NetEq::Create(config_, CreateBuiltinAudioDecoderFactory());
turaj@webrtc.org8d1cdaa2014-04-11 18:47:55 +0000312 NetEqNetworkStatistics stat;
313 ASSERT_EQ(0, neteq_->NetworkStatistics(&stat));
314 algorithmic_delay_ms_ = stat.current_buffer_size_ms;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000315 ASSERT_TRUE(neteq_);
henrik.lundin7a926812016-05-12 13:51:28 -0700316 LoadDecoders(neteq_);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000317}
318
319void NetEqDecodingTest::TearDown() {
320 delete neteq_;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000321}
322
Yves Gerey665174f2018-06-19 15:03:05 +0200323void NetEqDecodingTest::OpenInputFile(const std::string& rtp_file) {
henrik.lundin@webrtc.org966a7082014-11-17 09:08:38 +0000324 rtp_source_.reset(test::RtpFileSource::Create(rtp_file));
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000325}
326
henrik.lundin6d8e0112016-03-04 10:34:21 -0800327void NetEqDecodingTest::Process() {
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000328 // Check if time to receive.
henrik.lundin@webrtc.org966a7082014-11-17 09:08:38 +0000329 while (packet_ && sim_clock_ >= packet_->time_ms()) {
330 if (packet_->payload_length_bytes() > 0) {
ivoc72c08ed2016-01-20 07:26:24 -0800331#ifndef WEBRTC_CODEC_ISAC
332 // Ignore payload type 104 (iSAC-swb) if ISAC is not supported.
henrik.lundin246ef3e2017-04-24 09:14:32 -0700333 if (packet_->header().payloadType != 104)
ivoc72c08ed2016-01-20 07:26:24 -0800334#endif
Henrik Lundin70c09bd2017-04-24 15:56:56 +0200335 ASSERT_EQ(0,
336 neteq_->InsertPacket(
henrik.lundin246ef3e2017-04-24 09:14:32 -0700337 packet_->header(),
Henrik Lundin70c09bd2017-04-24 15:56:56 +0200338 rtc::ArrayView<const uint8_t>(
339 packet_->payload(), packet_->payload_length_bytes()),
340 static_cast<uint32_t>(packet_->time_ms() *
341 (output_sample_rate_ / 1000))));
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000342 }
343 // Get next packet.
henrik.lundin46ba49c2016-05-24 22:50:47 -0700344 packet_ = rtp_source_->NextPacket();
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000345 }
346
henrik.lundin@webrtc.orge1d468c2013-01-30 07:37:20 +0000347 // Get audio from NetEq.
henrik.lundin7a926812016-05-12 13:51:28 -0700348 bool muted;
349 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
350 ASSERT_FALSE(muted);
henrik.lundin6d8e0112016-03-04 10:34:21 -0800351 ASSERT_TRUE((out_frame_.samples_per_channel_ == kBlockSize8kHz) ||
352 (out_frame_.samples_per_channel_ == kBlockSize16kHz) ||
353 (out_frame_.samples_per_channel_ == kBlockSize32kHz) ||
354 (out_frame_.samples_per_channel_ == kBlockSize48kHz));
355 output_sample_rate_ = out_frame_.sample_rate_hz_;
henrik.lundind89814b2015-11-23 06:49:25 -0800356 EXPECT_EQ(output_sample_rate_, neteq_->last_output_sample_rate_hz());
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000357
358 // Increase time.
359 sim_clock_ += kTimeStepMs;
360}
361
minyue4f906772016-04-29 11:05:14 -0700362void NetEqDecodingTest::DecodeAndCompare(
363 const std::string& rtp_file,
364 const std::string& output_checksum,
365 const std::string& network_stats_checksum,
366 const std::string& rtcp_stats_checksum,
367 bool gen_ref) {
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000368 OpenInputFile(rtp_file);
369
minyue4f906772016-04-29 11:05:14 -0700370 std::string ref_out_file =
371 gen_ref ? webrtc::test::OutputPath() + "neteq_universal_ref.pcm" : "";
372 ResultSink output(ref_out_file);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000373
minyue4f906772016-04-29 11:05:14 -0700374 std::string stat_out_file =
375 gen_ref ? webrtc::test::OutputPath() + "neteq_network_stats.dat" : "";
376 ResultSink network_stats(stat_out_file);
henrik.lundin@webrtc.org4e4b0982014-08-11 14:48:49 +0000377
minyue4f906772016-04-29 11:05:14 -0700378 std::string rtcp_out_file =
379 gen_ref ? webrtc::test::OutputPath() + "neteq_rtcp_stats.dat" : "";
380 ResultSink rtcp_stats(rtcp_out_file);
henrik.lundin@webrtc.org4e4b0982014-08-11 14:48:49 +0000381
henrik.lundin46ba49c2016-05-24 22:50:47 -0700382 packet_ = rtp_source_->NextPacket();
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000383 int i = 0;
Henrik Lundinac0a5032017-09-25 12:22:46 +0200384 uint64_t last_concealed_samples = 0;
385 uint64_t last_total_samples_received = 0;
henrik.lundin@webrtc.org966a7082014-11-17 09:08:38 +0000386 while (packet_) {
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000387 std::ostringstream ss;
388 ss << "Lap number " << i++ << " in DecodeAndCompare while loop";
389 SCOPED_TRACE(ss.str()); // Print out the parameter values on failure.
henrik.lundin6d8e0112016-03-04 10:34:21 -0800390 ASSERT_NO_FATAL_FAILURE(Process());
Yves Gerey665174f2018-06-19 15:03:05 +0200391 ASSERT_NO_FATAL_FAILURE(
392 output.AddResult(out_frame_.data(), out_frame_.samples_per_channel_));
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000393
394 // Query the network statistics API once per second
395 if (sim_clock_ % 1000 == 0) {
396 // Process NetworkStatistics.
minyue4f906772016-04-29 11:05:14 -0700397 NetEqNetworkStatistics current_network_stats;
398 ASSERT_EQ(0, neteq_->NetworkStatistics(&current_network_stats));
399 ASSERT_NO_FATAL_FAILURE(network_stats.AddResult(current_network_stats));
400
henrik.lundin9c3efd02015-08-27 13:12:22 -0700401 // Compare with CurrentDelay, which should be identical.
minyue4f906772016-04-29 11:05:14 -0700402 EXPECT_EQ(current_network_stats.current_buffer_size_ms,
403 neteq_->CurrentDelayMs());
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000404
Henrik Lundinac0a5032017-09-25 12:22:46 +0200405 // Verify that liftime stats and network stats report similar loss
406 // concealment rates.
407 auto lifetime_stats = neteq_->GetLifetimeStatistics();
408 const uint64_t delta_concealed_samples =
409 lifetime_stats.concealed_samples - last_concealed_samples;
410 last_concealed_samples = lifetime_stats.concealed_samples;
411 const uint64_t delta_total_samples_received =
412 lifetime_stats.total_samples_received - last_total_samples_received;
413 last_total_samples_received = lifetime_stats.total_samples_received;
414 // The tolerance is 1% but expressed in Q14.
415 EXPECT_NEAR(
416 (delta_concealed_samples << 14) / delta_total_samples_received,
417 current_network_stats.expand_rate, (2 << 14) / 100.0);
418
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000419 // Process RTCPstat.
minyue4f906772016-04-29 11:05:14 -0700420 RtcpStatistics current_rtcp_stats;
421 neteq_->GetRtcpStatistics(&current_rtcp_stats);
422 ASSERT_NO_FATAL_FAILURE(rtcp_stats.AddResult(current_rtcp_stats));
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000423 }
424 }
minyue4f906772016-04-29 11:05:14 -0700425
426 SCOPED_TRACE("Check output audio.");
427 output.VerifyChecksum(output_checksum);
428 SCOPED_TRACE("Check network stats.");
429 network_stats.VerifyChecksum(network_stats_checksum);
430 SCOPED_TRACE("Check rtcp stats.");
431 rtcp_stats.VerifyChecksum(rtcp_stats_checksum);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000432}
433
434void NetEqDecodingTest::PopulateRtpInfo(int frame_index,
435 int timestamp,
henrik.lundin246ef3e2017-04-24 09:14:32 -0700436 RTPHeader* rtp_info) {
437 rtp_info->sequenceNumber = frame_index;
438 rtp_info->timestamp = timestamp;
439 rtp_info->ssrc = 0x1234; // Just an arbitrary SSRC.
440 rtp_info->payloadType = 94; // PCM16b WB codec.
441 rtp_info->markerBit = 0;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000442}
443
444void NetEqDecodingTest::PopulateCng(int frame_index,
445 int timestamp,
henrik.lundin246ef3e2017-04-24 09:14:32 -0700446 RTPHeader* rtp_info,
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000447 uint8_t* payload,
pkasting@chromium.org4591fbd2014-11-20 22:28:14 +0000448 size_t* payload_len) {
henrik.lundin246ef3e2017-04-24 09:14:32 -0700449 rtp_info->sequenceNumber = frame_index;
450 rtp_info->timestamp = timestamp;
451 rtp_info->ssrc = 0x1234; // Just an arbitrary SSRC.
452 rtp_info->payloadType = 98; // WB CNG.
453 rtp_info->markerBit = 0;
Yves Gerey665174f2018-06-19 15:03:05 +0200454 payload[0] = 64; // Noise level -64 dBov, quite arbitrarily chosen.
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000455 *payload_len = 1; // Only noise level, no spectral parameters.
456}
457
ivoc72c08ed2016-01-20 07:26:24 -0800458#if !defined(WEBRTC_IOS) && defined(WEBRTC_NETEQ_UNITTEST_BITEXACT) && \
459 (defined(WEBRTC_CODEC_ISAC) || defined(WEBRTC_CODEC_ISACFX)) && \
Karl Wibergeb254b42017-11-01 15:08:12 +0100460 defined(WEBRTC_CODEC_ILBC) && !defined(WEBRTC_ARCH_ARM64)
minyue5f026d02015-12-16 07:36:04 -0800461#define MAYBE_TestBitExactness TestBitExactness
kwiberg98ab3a42015-09-30 21:54:21 -0700462#else
minyue5f026d02015-12-16 07:36:04 -0800463#define MAYBE_TestBitExactness DISABLED_TestBitExactness
kwiberg98ab3a42015-09-30 21:54:21 -0700464#endif
minyue5f026d02015-12-16 07:36:04 -0800465TEST_F(NetEqDecodingTest, MAYBE_TestBitExactness) {
minyue49c454e2016-01-08 11:30:14 -0800466 const std::string input_rtp_file =
467 webrtc::test::ResourcePath("audio_coding/neteq_universal_new", "rtp");
henrik.lundin@webrtc.org4e4b0982014-08-11 14:48:49 +0000468
Yves Gerey665174f2018-06-19 15:03:05 +0200469 const std::string output_checksum =
470 PlatformChecksum("0c6dc227f781c81a229970f8fceda1a012498cba",
471 "15c4a2202877a414515e218bdb7992f0ad53e5af", "not used",
472 "0c6dc227f781c81a229970f8fceda1a012498cba",
473 "25fc4c863caa499aa447a5b8d059f5452cbcc500");
minyue4f906772016-04-29 11:05:14 -0700474
henrik.lundin2979f552017-05-05 05:04:16 -0700475 const std::string network_stats_checksum =
Henrik Lundin6dc82e82018-05-22 10:40:23 +0200476 PlatformChecksum("4b2370f5c794741d2a46be5c7935c66ef3fb53e9",
Yves Gerey665174f2018-06-19 15:03:05 +0200477 "e339cb2adf5ab3dfc21cb7205d670a34751e8336", "not used",
Henrik Lundin6dc82e82018-05-22 10:40:23 +0200478 "4b2370f5c794741d2a46be5c7935c66ef3fb53e9",
479 "4b2370f5c794741d2a46be5c7935c66ef3fb53e9");
minyue4f906772016-04-29 11:05:14 -0700480
Yves Gerey665174f2018-06-19 15:03:05 +0200481 const std::string rtcp_stats_checksum =
482 PlatformChecksum("b8880bf9fed2487efbddcb8d94b9937a29ae521d",
483 "f3f7b3d3e71d7e635240b5373b57df6a7e4ce9d4", "not used",
484 "b8880bf9fed2487efbddcb8d94b9937a29ae521d",
485 "b8880bf9fed2487efbddcb8d94b9937a29ae521d");
minyue4f906772016-04-29 11:05:14 -0700486
Yves Gerey665174f2018-06-19 15:03:05 +0200487 DecodeAndCompare(input_rtp_file, output_checksum, network_stats_checksum,
488 rtcp_stats_checksum, FLAG_gen_ref);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000489}
490
Yves Gerey665174f2018-06-19 15:03:05 +0200491#if !defined(WEBRTC_IOS) && defined(WEBRTC_NETEQ_UNITTEST_BITEXACT) && \
minyue-webrtc516711c2017-07-27 17:45:49 +0200492 defined(WEBRTC_CODEC_OPUS)
minyue93c08b72015-12-22 09:57:41 -0800493#define MAYBE_TestOpusBitExactness TestOpusBitExactness
494#else
495#define MAYBE_TestOpusBitExactness DISABLED_TestOpusBitExactness
496#endif
minyue-webrtcadb58b82017-07-26 17:59:59 +0200497TEST_F(NetEqDecodingTest, MAYBE_TestOpusBitExactness) {
minyue93c08b72015-12-22 09:57:41 -0800498 const std::string input_rtp_file =
499 webrtc::test::ResourcePath("audio_coding/neteq_opus", "rtp");
minyue93c08b72015-12-22 09:57:41 -0800500
Yves Gerey665174f2018-06-19 15:03:05 +0200501 const std::string output_checksum =
502 PlatformChecksum("14a63b3c7b925c82296be4bafc71bec85f2915c2",
503 "b7b7ed802b0e18ee416973bf3b9ae98599b0181d",
504 "5876e52dda90d5ca433c3726555b907b97c86374",
505 "14a63b3c7b925c82296be4bafc71bec85f2915c2",
506 "14a63b3c7b925c82296be4bafc71bec85f2915c2");
minyue4f906772016-04-29 11:05:14 -0700507
henrik.lundin2979f552017-05-05 05:04:16 -0700508 const std::string network_stats_checksum =
Henrik Lundin6dc82e82018-05-22 10:40:23 +0200509 PlatformChecksum("adb3272498e436d1c019cbfd71610e9510c54497",
510 "fa935a91abc7291db47428a2d7c5361b98713a92",
511 "42106aa5267300f709f63737707ef07afd9dac61",
512 "adb3272498e436d1c019cbfd71610e9510c54497",
513 "adb3272498e436d1c019cbfd71610e9510c54497");
minyue4f906772016-04-29 11:05:14 -0700514
Yves Gerey665174f2018-06-19 15:03:05 +0200515 const std::string rtcp_stats_checksum =
516 PlatformChecksum("e37c797e3de6a64dda88c9ade7a013d022a2e1e0",
517 "e37c797e3de6a64dda88c9ade7a013d022a2e1e0",
518 "e37c797e3de6a64dda88c9ade7a013d022a2e1e0",
519 "e37c797e3de6a64dda88c9ade7a013d022a2e1e0",
520 "e37c797e3de6a64dda88c9ade7a013d022a2e1e0");
minyue4f906772016-04-29 11:05:14 -0700521
Yves Gerey665174f2018-06-19 15:03:05 +0200522 DecodeAndCompare(input_rtp_file, output_checksum, network_stats_checksum,
523 rtcp_stats_checksum, FLAG_gen_ref);
minyue93c08b72015-12-22 09:57:41 -0800524}
525
Yves Gerey665174f2018-06-19 15:03:05 +0200526#if !defined(WEBRTC_IOS) && defined(WEBRTC_NETEQ_UNITTEST_BITEXACT) && \
Henrik Lundine9619f82017-11-27 14:05:27 +0100527 defined(WEBRTC_CODEC_OPUS)
528#define MAYBE_TestOpusDtxBitExactness TestOpusDtxBitExactness
529#else
530#define MAYBE_TestOpusDtxBitExactness DISABLED_TestOpusDtxBitExactness
531#endif
Henrik Lundin4f2a4a12018-01-26 17:32:56 +0100532TEST_F(NetEqDecodingTest, MAYBE_TestOpusDtxBitExactness) {
Henrik Lundine9619f82017-11-27 14:05:27 +0100533 const std::string input_rtp_file =
534 webrtc::test::ResourcePath("audio_coding/neteq_opus_dtx", "rtp");
535
536 const std::string output_checksum =
537 PlatformChecksum("713af6c92881f5aab1285765ee6680da9d1c06ce",
538 "3ec991b96872123f1554c03c543ca5d518431e46",
539 "da9f9a2d94e0c2d67342fad4965d7b91cda50b25",
540 "713af6c92881f5aab1285765ee6680da9d1c06ce",
541 "713af6c92881f5aab1285765ee6680da9d1c06ce");
542
543 const std::string network_stats_checksum =
544 "bab58dc587d956f326056d7340c96eb9d2d3cc21";
545
546 const std::string rtcp_stats_checksum =
547 "ac27a7f305efb58b39bf123dccee25dee5758e63";
548
549 DecodeAndCompare(input_rtp_file, output_checksum, network_stats_checksum,
550 rtcp_stats_checksum, FLAG_gen_ref);
551}
552
henrik.lundin@webrtc.org7cbc4f92014-10-07 06:37:39 +0000553// Use fax mode to avoid time-scaling. This is to simplify the testing of
554// packet waiting times in the packet buffer.
555class NetEqDecodingTestFaxMode : public NetEqDecodingTest {
556 protected:
557 NetEqDecodingTestFaxMode() : NetEqDecodingTest() {
558 config_.playout_mode = kPlayoutFax;
559 }
Gustaf Ullbergb0a02072017-10-02 12:00:34 +0200560 void TestJitterBufferDelay(bool apply_packet_loss);
henrik.lundin@webrtc.org7cbc4f92014-10-07 06:37:39 +0000561};
562
563TEST_F(NetEqDecodingTestFaxMode, TestFrameWaitingTimeStatistics) {
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000564 // Insert 30 dummy packets at once. Each packet contains 10 ms 16 kHz audio.
565 size_t num_frames = 30;
pkasting@chromium.org4591fbd2014-11-20 22:28:14 +0000566 const size_t kSamples = 10 * 16;
567 const size_t kPayloadBytes = kSamples * 2;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000568 for (size_t i = 0; i < num_frames; ++i) {
kwibergee2bac22015-11-11 10:34:00 -0800569 const uint8_t payload[kPayloadBytes] = {0};
henrik.lundin246ef3e2017-04-24 09:14:32 -0700570 RTPHeader rtp_info;
Mirko Bonadeia8110272017-10-18 14:22:50 +0200571 rtp_info.sequenceNumber = rtc::checked_cast<uint16_t>(i);
572 rtp_info.timestamp = rtc::checked_cast<uint32_t>(i * kSamples);
henrik.lundin246ef3e2017-04-24 09:14:32 -0700573 rtp_info.ssrc = 0x1234; // Just an arbitrary SSRC.
574 rtp_info.payloadType = 94; // PCM16b WB codec.
575 rtp_info.markerBit = 0;
576 ASSERT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000577 }
578 // Pull out all data.
579 for (size_t i = 0; i < num_frames; ++i) {
henrik.lundin7a926812016-05-12 13:51:28 -0700580 bool muted;
581 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -0800582 ASSERT_EQ(kBlockSize16kHz, out_frame_.samples_per_channel_);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000583 }
584
Henrik Lundin1bb8cf82015-08-25 13:08:04 +0200585 NetEqNetworkStatistics stats;
586 EXPECT_EQ(0, neteq_->NetworkStatistics(&stats));
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000587 // Since all frames are dumped into NetEQ at once, but pulled out with 10 ms
588 // spacing (per definition), we expect the delay to increase with 10 ms for
Henrik Lundin1bb8cf82015-08-25 13:08:04 +0200589 // each packet. Thus, we are calculating the statistics for a series from 10
590 // to 300, in steps of 10 ms.
591 EXPECT_EQ(155, stats.mean_waiting_time_ms);
592 EXPECT_EQ(155, stats.median_waiting_time_ms);
593 EXPECT_EQ(10, stats.min_waiting_time_ms);
594 EXPECT_EQ(300, stats.max_waiting_time_ms);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000595
596 // Check statistics again and make sure it's been reset.
Henrik Lundin1bb8cf82015-08-25 13:08:04 +0200597 EXPECT_EQ(0, neteq_->NetworkStatistics(&stats));
598 EXPECT_EQ(-1, stats.mean_waiting_time_ms);
599 EXPECT_EQ(-1, stats.median_waiting_time_ms);
600 EXPECT_EQ(-1, stats.min_waiting_time_ms);
601 EXPECT_EQ(-1, stats.max_waiting_time_ms);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000602}
603
henrik.lundin@webrtc.orgb4e80e02014-05-15 07:14:00 +0000604TEST_F(NetEqDecodingTest, TestAverageInterArrivalTimeNegative) {
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000605 const int kNumFrames = 3000; // Needed for convergence.
606 int frame_index = 0;
pkasting@chromium.org4591fbd2014-11-20 22:28:14 +0000607 const size_t kSamples = 10 * 16;
608 const size_t kPayloadBytes = kSamples * 2;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000609 while (frame_index < kNumFrames) {
610 // Insert one packet each time, except every 10th time where we insert two
611 // packets at once. This will create a negative clock-drift of approx. 10%.
612 int num_packets = (frame_index % 10 == 0 ? 2 : 1);
613 for (int n = 0; n < num_packets; ++n) {
614 uint8_t payload[kPayloadBytes] = {0};
henrik.lundin246ef3e2017-04-24 09:14:32 -0700615 RTPHeader rtp_info;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000616 PopulateRtpInfo(frame_index, frame_index * kSamples, &rtp_info);
henrik.lundin246ef3e2017-04-24 09:14:32 -0700617 ASSERT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000618 ++frame_index;
619 }
620
621 // Pull out data once.
henrik.lundin7a926812016-05-12 13:51:28 -0700622 bool muted;
623 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -0800624 ASSERT_EQ(kBlockSize16kHz, out_frame_.samples_per_channel_);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000625 }
626
627 NetEqNetworkStatistics network_stats;
628 ASSERT_EQ(0, neteq_->NetworkStatistics(&network_stats));
henrik.lundin0d838572016-10-13 03:35:55 -0700629 EXPECT_EQ(-103192, network_stats.clockdrift_ppm);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000630}
631
henrik.lundin@webrtc.orgb4e80e02014-05-15 07:14:00 +0000632TEST_F(NetEqDecodingTest, TestAverageInterArrivalTimePositive) {
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000633 const int kNumFrames = 5000; // Needed for convergence.
634 int frame_index = 0;
pkasting@chromium.org4591fbd2014-11-20 22:28:14 +0000635 const size_t kSamples = 10 * 16;
636 const size_t kPayloadBytes = kSamples * 2;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000637 for (int i = 0; i < kNumFrames; ++i) {
638 // Insert one packet each time, except every 10th time where we don't insert
639 // any packet. This will create a positive clock-drift of approx. 11%.
640 int num_packets = (i % 10 == 9 ? 0 : 1);
641 for (int n = 0; n < num_packets; ++n) {
642 uint8_t payload[kPayloadBytes] = {0};
henrik.lundin246ef3e2017-04-24 09:14:32 -0700643 RTPHeader rtp_info;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000644 PopulateRtpInfo(frame_index, frame_index * kSamples, &rtp_info);
henrik.lundin246ef3e2017-04-24 09:14:32 -0700645 ASSERT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000646 ++frame_index;
647 }
648
649 // Pull out data once.
henrik.lundin7a926812016-05-12 13:51:28 -0700650 bool muted;
651 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -0800652 ASSERT_EQ(kBlockSize16kHz, out_frame_.samples_per_channel_);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000653 }
654
655 NetEqNetworkStatistics network_stats;
656 ASSERT_EQ(0, neteq_->NetworkStatistics(&network_stats));
henrik.lundin0d838572016-10-13 03:35:55 -0700657 EXPECT_EQ(110953, network_stats.clockdrift_ppm);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000658}
659
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000660void NetEqDecodingTest::LongCngWithClockDrift(double drift_factor,
661 double network_freeze_ms,
662 bool pull_audio_during_freeze,
663 int delay_tolerance_ms,
664 int max_time_to_speech_ms) {
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000665 uint16_t seq_no = 0;
666 uint32_t timestamp = 0;
667 const int kFrameSizeMs = 30;
pkasting@chromium.org4591fbd2014-11-20 22:28:14 +0000668 const size_t kSamples = kFrameSizeMs * 16;
669 const size_t kPayloadBytes = kSamples * 2;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000670 double next_input_time_ms = 0.0;
671 double t_ms;
henrik.lundin7a926812016-05-12 13:51:28 -0700672 bool muted;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000673
674 // Insert speech for 5 seconds.
675 const int kSpeechDurationMs = 5000;
676 for (t_ms = 0; t_ms < kSpeechDurationMs; t_ms += 10) {
677 // Each turn in this for loop is 10 ms.
678 while (next_input_time_ms <= t_ms) {
679 // Insert one 30 ms speech frame.
680 uint8_t payload[kPayloadBytes] = {0};
henrik.lundin246ef3e2017-04-24 09:14:32 -0700681 RTPHeader rtp_info;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000682 PopulateRtpInfo(seq_no, timestamp, &rtp_info);
henrik.lundin246ef3e2017-04-24 09:14:32 -0700683 ASSERT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000684 ++seq_no;
685 timestamp += kSamples;
henrik.lundin@webrtc.orgfcfc6a92014-02-13 11:42:28 +0000686 next_input_time_ms += static_cast<double>(kFrameSizeMs) * drift_factor;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000687 }
688 // Pull out data once.
henrik.lundin7a926812016-05-12 13:51:28 -0700689 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -0800690 ASSERT_EQ(kBlockSize16kHz, out_frame_.samples_per_channel_);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000691 }
692
henrik.lundin55480f52016-03-08 02:37:57 -0800693 EXPECT_EQ(AudioFrame::kNormalSpeech, out_frame_.speech_type_);
Danil Chapovalovb6021232018-06-19 13:26:36 +0200694 absl::optional<uint32_t> playout_timestamp = neteq_->GetPlayoutTimestamp();
henrik.lundin0d96ab72016-04-06 12:28:26 -0700695 ASSERT_TRUE(playout_timestamp);
696 int32_t delay_before = timestamp - *playout_timestamp;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000697
698 // Insert CNG for 1 minute (= 60000 ms).
699 const int kCngPeriodMs = 100;
700 const int kCngPeriodSamples = kCngPeriodMs * 16; // Period in 16 kHz samples.
701 const int kCngDurationMs = 60000;
702 for (; t_ms < kSpeechDurationMs + kCngDurationMs; t_ms += 10) {
703 // Each turn in this for loop is 10 ms.
704 while (next_input_time_ms <= t_ms) {
705 // Insert one CNG frame each 100 ms.
706 uint8_t payload[kPayloadBytes];
pkasting@chromium.org4591fbd2014-11-20 22:28:14 +0000707 size_t payload_len;
henrik.lundin246ef3e2017-04-24 09:14:32 -0700708 RTPHeader rtp_info;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000709 PopulateCng(seq_no, timestamp, &rtp_info, payload, &payload_len);
kwibergee2bac22015-11-11 10:34:00 -0800710 ASSERT_EQ(0, neteq_->InsertPacket(
henrik.lundin246ef3e2017-04-24 09:14:32 -0700711 rtp_info,
kwibergee2bac22015-11-11 10:34:00 -0800712 rtc::ArrayView<const uint8_t>(payload, payload_len), 0));
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000713 ++seq_no;
714 timestamp += kCngPeriodSamples;
henrik.lundin@webrtc.orgfcfc6a92014-02-13 11:42:28 +0000715 next_input_time_ms += static_cast<double>(kCngPeriodMs) * drift_factor;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000716 }
717 // Pull out data once.
henrik.lundin7a926812016-05-12 13:51:28 -0700718 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -0800719 ASSERT_EQ(kBlockSize16kHz, out_frame_.samples_per_channel_);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000720 }
721
henrik.lundin55480f52016-03-08 02:37:57 -0800722 EXPECT_EQ(AudioFrame::kCNG, out_frame_.speech_type_);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000723
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000724 if (network_freeze_ms > 0) {
725 // First keep pulling audio for |network_freeze_ms| without inserting
726 // any data, then insert CNG data corresponding to |network_freeze_ms|
727 // without pulling any output audio.
728 const double loop_end_time = t_ms + network_freeze_ms;
729 for (; t_ms < loop_end_time; t_ms += 10) {
730 // Pull out data once.
henrik.lundin7a926812016-05-12 13:51:28 -0700731 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -0800732 ASSERT_EQ(kBlockSize16kHz, out_frame_.samples_per_channel_);
henrik.lundin55480f52016-03-08 02:37:57 -0800733 EXPECT_EQ(AudioFrame::kCNG, out_frame_.speech_type_);
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000734 }
735 bool pull_once = pull_audio_during_freeze;
736 // If |pull_once| is true, GetAudio will be called once half-way through
737 // the network recovery period.
738 double pull_time_ms = (t_ms + next_input_time_ms) / 2;
739 while (next_input_time_ms <= t_ms) {
740 if (pull_once && next_input_time_ms >= pull_time_ms) {
741 pull_once = false;
742 // Pull out data once.
henrik.lundin7a926812016-05-12 13:51:28 -0700743 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -0800744 ASSERT_EQ(kBlockSize16kHz, out_frame_.samples_per_channel_);
henrik.lundin55480f52016-03-08 02:37:57 -0800745 EXPECT_EQ(AudioFrame::kCNG, out_frame_.speech_type_);
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000746 t_ms += 10;
747 }
748 // Insert one CNG frame each 100 ms.
749 uint8_t payload[kPayloadBytes];
pkasting@chromium.org4591fbd2014-11-20 22:28:14 +0000750 size_t payload_len;
henrik.lundin246ef3e2017-04-24 09:14:32 -0700751 RTPHeader rtp_info;
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000752 PopulateCng(seq_no, timestamp, &rtp_info, payload, &payload_len);
kwibergee2bac22015-11-11 10:34:00 -0800753 ASSERT_EQ(0, neteq_->InsertPacket(
henrik.lundin246ef3e2017-04-24 09:14:32 -0700754 rtp_info,
kwibergee2bac22015-11-11 10:34:00 -0800755 rtc::ArrayView<const uint8_t>(payload, payload_len), 0));
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000756 ++seq_no;
757 timestamp += kCngPeriodSamples;
758 next_input_time_ms += kCngPeriodMs * drift_factor;
759 }
760 }
761
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000762 // Insert speech again until output type is speech.
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000763 double speech_restart_time_ms = t_ms;
henrik.lundin55480f52016-03-08 02:37:57 -0800764 while (out_frame_.speech_type_ != AudioFrame::kNormalSpeech) {
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000765 // Each turn in this for loop is 10 ms.
766 while (next_input_time_ms <= t_ms) {
767 // Insert one 30 ms speech frame.
768 uint8_t payload[kPayloadBytes] = {0};
henrik.lundin246ef3e2017-04-24 09:14:32 -0700769 RTPHeader rtp_info;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000770 PopulateRtpInfo(seq_no, timestamp, &rtp_info);
henrik.lundin246ef3e2017-04-24 09:14:32 -0700771 ASSERT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000772 ++seq_no;
773 timestamp += kSamples;
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000774 next_input_time_ms += kFrameSizeMs * drift_factor;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000775 }
776 // Pull out data once.
henrik.lundin7a926812016-05-12 13:51:28 -0700777 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -0800778 ASSERT_EQ(kBlockSize16kHz, out_frame_.samples_per_channel_);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000779 // Increase clock.
780 t_ms += 10;
781 }
782
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000783 // Check that the speech starts again within reasonable time.
784 double time_until_speech_returns_ms = t_ms - speech_restart_time_ms;
785 EXPECT_LT(time_until_speech_returns_ms, max_time_to_speech_ms);
henrik.lundin114c1b32017-04-26 07:47:32 -0700786 playout_timestamp = neteq_->GetPlayoutTimestamp();
henrik.lundin0d96ab72016-04-06 12:28:26 -0700787 ASSERT_TRUE(playout_timestamp);
788 int32_t delay_after = timestamp - *playout_timestamp;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000789 // Compare delay before and after, and make sure it differs less than 20 ms.
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000790 EXPECT_LE(delay_after, delay_before + delay_tolerance_ms * 16);
791 EXPECT_GE(delay_after, delay_before - delay_tolerance_ms * 16);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000792}
793
henrik.lundin@webrtc.orgb4e80e02014-05-15 07:14:00 +0000794TEST_F(NetEqDecodingTest, LongCngWithNegativeClockDrift) {
henrik.lundin@webrtc.orgfcfc6a92014-02-13 11:42:28 +0000795 // Apply a clock drift of -25 ms / s (sender faster than receiver).
796 const double kDriftFactor = 1000.0 / (1000.0 + 25.0);
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000797 const double kNetworkFreezeTimeMs = 0.0;
798 const bool kGetAudioDuringFreezeRecovery = false;
799 const int kDelayToleranceMs = 20;
800 const int kMaxTimeToSpeechMs = 100;
Yves Gerey665174f2018-06-19 15:03:05 +0200801 LongCngWithClockDrift(kDriftFactor, kNetworkFreezeTimeMs,
802 kGetAudioDuringFreezeRecovery, kDelayToleranceMs,
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000803 kMaxTimeToSpeechMs);
henrik.lundin@webrtc.orgfcfc6a92014-02-13 11:42:28 +0000804}
805
henrik.lundin@webrtc.orgb4e80e02014-05-15 07:14:00 +0000806TEST_F(NetEqDecodingTest, LongCngWithPositiveClockDrift) {
henrik.lundin@webrtc.orgfcfc6a92014-02-13 11:42:28 +0000807 // Apply a clock drift of +25 ms / s (sender slower than receiver).
808 const double kDriftFactor = 1000.0 / (1000.0 - 25.0);
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000809 const double kNetworkFreezeTimeMs = 0.0;
810 const bool kGetAudioDuringFreezeRecovery = false;
811 const int kDelayToleranceMs = 20;
812 const int kMaxTimeToSpeechMs = 100;
Yves Gerey665174f2018-06-19 15:03:05 +0200813 LongCngWithClockDrift(kDriftFactor, kNetworkFreezeTimeMs,
814 kGetAudioDuringFreezeRecovery, kDelayToleranceMs,
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000815 kMaxTimeToSpeechMs);
816}
817
henrik.lundin@webrtc.orgb4e80e02014-05-15 07:14:00 +0000818TEST_F(NetEqDecodingTest, LongCngWithNegativeClockDriftNetworkFreeze) {
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000819 // Apply a clock drift of -25 ms / s (sender faster than receiver).
820 const double kDriftFactor = 1000.0 / (1000.0 + 25.0);
821 const double kNetworkFreezeTimeMs = 5000.0;
822 const bool kGetAudioDuringFreezeRecovery = false;
823 const int kDelayToleranceMs = 50;
824 const int kMaxTimeToSpeechMs = 200;
Yves Gerey665174f2018-06-19 15:03:05 +0200825 LongCngWithClockDrift(kDriftFactor, kNetworkFreezeTimeMs,
826 kGetAudioDuringFreezeRecovery, kDelayToleranceMs,
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000827 kMaxTimeToSpeechMs);
828}
829
henrik.lundin@webrtc.orgb4e80e02014-05-15 07:14:00 +0000830TEST_F(NetEqDecodingTest, LongCngWithPositiveClockDriftNetworkFreeze) {
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000831 // Apply a clock drift of +25 ms / s (sender slower than receiver).
832 const double kDriftFactor = 1000.0 / (1000.0 - 25.0);
833 const double kNetworkFreezeTimeMs = 5000.0;
834 const bool kGetAudioDuringFreezeRecovery = false;
835 const int kDelayToleranceMs = 20;
836 const int kMaxTimeToSpeechMs = 100;
Yves Gerey665174f2018-06-19 15:03:05 +0200837 LongCngWithClockDrift(kDriftFactor, kNetworkFreezeTimeMs,
838 kGetAudioDuringFreezeRecovery, kDelayToleranceMs,
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000839 kMaxTimeToSpeechMs);
840}
841
henrik.lundin@webrtc.orgb4e80e02014-05-15 07:14:00 +0000842TEST_F(NetEqDecodingTest, LongCngWithPositiveClockDriftNetworkFreezeExtraPull) {
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000843 // Apply a clock drift of +25 ms / s (sender slower than receiver).
844 const double kDriftFactor = 1000.0 / (1000.0 - 25.0);
845 const double kNetworkFreezeTimeMs = 5000.0;
846 const bool kGetAudioDuringFreezeRecovery = true;
847 const int kDelayToleranceMs = 20;
848 const int kMaxTimeToSpeechMs = 100;
Yves Gerey665174f2018-06-19 15:03:05 +0200849 LongCngWithClockDrift(kDriftFactor, kNetworkFreezeTimeMs,
850 kGetAudioDuringFreezeRecovery, kDelayToleranceMs,
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000851 kMaxTimeToSpeechMs);
852}
853
henrik.lundin@webrtc.orgb4e80e02014-05-15 07:14:00 +0000854TEST_F(NetEqDecodingTest, LongCngWithoutClockDrift) {
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000855 const double kDriftFactor = 1.0; // No drift.
856 const double kNetworkFreezeTimeMs = 0.0;
857 const bool kGetAudioDuringFreezeRecovery = false;
858 const int kDelayToleranceMs = 10;
859 const int kMaxTimeToSpeechMs = 50;
Yves Gerey665174f2018-06-19 15:03:05 +0200860 LongCngWithClockDrift(kDriftFactor, kNetworkFreezeTimeMs,
861 kGetAudioDuringFreezeRecovery, kDelayToleranceMs,
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000862 kMaxTimeToSpeechMs);
henrik.lundin@webrtc.orgfcfc6a92014-02-13 11:42:28 +0000863}
864
henrik.lundin@webrtc.orgb4e80e02014-05-15 07:14:00 +0000865TEST_F(NetEqDecodingTest, UnknownPayloadType) {
pkasting@chromium.org4591fbd2014-11-20 22:28:14 +0000866 const size_t kPayloadBytes = 100;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000867 uint8_t payload[kPayloadBytes] = {0};
henrik.lundin246ef3e2017-04-24 09:14:32 -0700868 RTPHeader rtp_info;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000869 PopulateRtpInfo(0, 0, &rtp_info);
henrik.lundin246ef3e2017-04-24 09:14:32 -0700870 rtp_info.payloadType = 1; // Not registered as a decoder.
871 EXPECT_EQ(NetEq::kFail, neteq_->InsertPacket(rtp_info, payload, 0));
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000872}
873
Peter Boströme2976c82016-01-04 22:44:05 +0100874#if defined(WEBRTC_CODEC_ISAC) || defined(WEBRTC_CODEC_ISACFX)
ivoc72c08ed2016-01-20 07:26:24 -0800875#define MAYBE_DecoderError DecoderError
876#else
877#define MAYBE_DecoderError DISABLED_DecoderError
878#endif
879
Peter Boströme2976c82016-01-04 22:44:05 +0100880TEST_F(NetEqDecodingTest, MAYBE_DecoderError) {
pkasting@chromium.org4591fbd2014-11-20 22:28:14 +0000881 const size_t kPayloadBytes = 100;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000882 uint8_t payload[kPayloadBytes] = {0};
henrik.lundin246ef3e2017-04-24 09:14:32 -0700883 RTPHeader rtp_info;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000884 PopulateRtpInfo(0, 0, &rtp_info);
henrik.lundin246ef3e2017-04-24 09:14:32 -0700885 rtp_info.payloadType = 103; // iSAC, but the payload is invalid.
886 EXPECT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000887 // Set all of |out_data_| to 1, and verify that it was set to 0 by the call
888 // to GetAudio.
yujo36b1a5f2017-06-12 12:45:32 -0700889 int16_t* out_frame_data = out_frame_.mutable_data();
henrik.lundin6d8e0112016-03-04 10:34:21 -0800890 for (size_t i = 0; i < AudioFrame::kMaxDataSizeSamples; ++i) {
yujo36b1a5f2017-06-12 12:45:32 -0700891 out_frame_data[i] = 1;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000892 }
henrik.lundin7a926812016-05-12 13:51:28 -0700893 bool muted;
894 EXPECT_EQ(NetEq::kFail, neteq_->GetAudio(&out_frame_, &muted));
895 ASSERT_FALSE(muted);
ivoc72c08ed2016-01-20 07:26:24 -0800896
yujo36b1a5f2017-06-12 12:45:32 -0700897 // Verify that the first 160 samples are set to 0.
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000898 static const int kExpectedOutputLength = 160; // 10 ms at 16 kHz sample rate.
yujo36b1a5f2017-06-12 12:45:32 -0700899 const int16_t* const_out_frame_data = out_frame_.data();
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000900 for (int i = 0; i < kExpectedOutputLength; ++i) {
901 std::ostringstream ss;
902 ss << "i = " << i;
903 SCOPED_TRACE(ss.str()); // Print out the parameter values on failure.
yujo36b1a5f2017-06-12 12:45:32 -0700904 EXPECT_EQ(0, const_out_frame_data[i]);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000905 }
906}
907
henrik.lundin@webrtc.orgb4e80e02014-05-15 07:14:00 +0000908TEST_F(NetEqDecodingTest, GetAudioBeforeInsertPacket) {
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000909 // Set all of |out_data_| to 1, and verify that it was set to 0 by the call
910 // to GetAudio.
yujo36b1a5f2017-06-12 12:45:32 -0700911 int16_t* out_frame_data = out_frame_.mutable_data();
henrik.lundin6d8e0112016-03-04 10:34:21 -0800912 for (size_t i = 0; i < AudioFrame::kMaxDataSizeSamples; ++i) {
yujo36b1a5f2017-06-12 12:45:32 -0700913 out_frame_data[i] = 1;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000914 }
henrik.lundin7a926812016-05-12 13:51:28 -0700915 bool muted;
916 EXPECT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
917 ASSERT_FALSE(muted);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000918 // Verify that the first block of samples is set to 0.
919 static const int kExpectedOutputLength =
920 kInitSampleRateHz / 100; // 10 ms at initial sample rate.
yujo36b1a5f2017-06-12 12:45:32 -0700921 const int16_t* const_out_frame_data = out_frame_.data();
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000922 for (int i = 0; i < kExpectedOutputLength; ++i) {
923 std::ostringstream ss;
924 ss << "i = " << i;
925 SCOPED_TRACE(ss.str()); // Print out the parameter values on failure.
yujo36b1a5f2017-06-12 12:45:32 -0700926 EXPECT_EQ(0, const_out_frame_data[i]);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000927 }
henrik.lundind89814b2015-11-23 06:49:25 -0800928 // Verify that the sample rate did not change from the initial configuration.
929 EXPECT_EQ(config_.sample_rate_hz, neteq_->last_output_sample_rate_hz());
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000930}
turaj@webrtc.orgff43c852013-09-25 00:07:27 +0000931
henrik.lundin@webrtc.org9b8102c2014-08-21 08:27:44 +0000932class NetEqBgnTest : public NetEqDecodingTest {
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +0000933 protected:
henrik.lundin@webrtc.org9b8102c2014-08-21 08:27:44 +0000934 void CheckBgn(int sampling_rate_hz) {
Peter Kastingdce40cf2015-08-24 14:52:23 -0700935 size_t expected_samples_per_channel = 0;
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +0000936 uint8_t payload_type = 0xFF; // Invalid.
937 if (sampling_rate_hz == 8000) {
938 expected_samples_per_channel = kBlockSize8kHz;
939 payload_type = 93; // PCM 16, 8 kHz.
940 } else if (sampling_rate_hz == 16000) {
941 expected_samples_per_channel = kBlockSize16kHz;
942 payload_type = 94; // PCM 16, 16 kHZ.
943 } else if (sampling_rate_hz == 32000) {
944 expected_samples_per_channel = kBlockSize32kHz;
945 payload_type = 95; // PCM 16, 32 kHz.
946 } else {
947 ASSERT_TRUE(false); // Unsupported test case.
948 }
turaj@webrtc.orgff43c852013-09-25 00:07:27 +0000949
henrik.lundin6d8e0112016-03-04 10:34:21 -0800950 AudioFrame output;
henrik.lundin@webrtc.org9b8102c2014-08-21 08:27:44 +0000951 test::AudioLoop input;
952 // We are using the same 32 kHz input file for all tests, regardless of
953 // |sampling_rate_hz|. The output may sound weird, but the test is still
954 // valid.
955 ASSERT_TRUE(input.Init(
956 webrtc::test::ResourcePath("audio_coding/testfile32kHz", "pcm"),
957 10 * sampling_rate_hz, // Max 10 seconds loop length.
Peter Kastingdce40cf2015-08-24 14:52:23 -0700958 expected_samples_per_channel));
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +0000959
960 // Payload of 10 ms of PCM16 32 kHz.
961 uint8_t payload[kBlockSize32kHz * sizeof(int16_t)];
henrik.lundin246ef3e2017-04-24 09:14:32 -0700962 RTPHeader rtp_info;
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +0000963 PopulateRtpInfo(0, 0, &rtp_info);
henrik.lundin246ef3e2017-04-24 09:14:32 -0700964 rtp_info.payloadType = payload_type;
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +0000965
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +0000966 uint32_t receive_timestamp = 0;
henrik.lundin7a926812016-05-12 13:51:28 -0700967 bool muted;
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +0000968 for (int n = 0; n < 10; ++n) { // Insert few packets and get audio.
kwiberg288886b2015-11-06 01:21:35 -0800969 auto block = input.GetNextBlock();
970 ASSERT_EQ(expected_samples_per_channel, block.size());
971 size_t enc_len_bytes =
972 WebRtcPcm16b_Encode(block.data(), block.size(), payload);
henrik.lundin@webrtc.org9b8102c2014-08-21 08:27:44 +0000973 ASSERT_EQ(enc_len_bytes, expected_samples_per_channel * 2);
974
Henrik Lundin70c09bd2017-04-24 15:56:56 +0200975 ASSERT_EQ(0, neteq_->InsertPacket(
henrik.lundin246ef3e2017-04-24 09:14:32 -0700976 rtp_info,
Henrik Lundin70c09bd2017-04-24 15:56:56 +0200977 rtc::ArrayView<const uint8_t>(payload, enc_len_bytes),
978 receive_timestamp));
henrik.lundin6d8e0112016-03-04 10:34:21 -0800979 output.Reset();
henrik.lundin7a926812016-05-12 13:51:28 -0700980 ASSERT_EQ(0, neteq_->GetAudio(&output, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -0800981 ASSERT_EQ(1u, output.num_channels_);
982 ASSERT_EQ(expected_samples_per_channel, output.samples_per_channel_);
henrik.lundin55480f52016-03-08 02:37:57 -0800983 ASSERT_EQ(AudioFrame::kNormalSpeech, output.speech_type_);
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +0000984
985 // Next packet.
Yves Gerey665174f2018-06-19 15:03:05 +0200986 rtp_info.timestamp +=
987 rtc::checked_cast<uint32_t>(expected_samples_per_channel);
henrik.lundin246ef3e2017-04-24 09:14:32 -0700988 rtp_info.sequenceNumber++;
Yves Gerey665174f2018-06-19 15:03:05 +0200989 receive_timestamp +=
990 rtc::checked_cast<uint32_t>(expected_samples_per_channel);
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +0000991 }
992
henrik.lundin6d8e0112016-03-04 10:34:21 -0800993 output.Reset();
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +0000994
995 // Get audio without inserting packets, expecting PLC and PLC-to-CNG. Pull
996 // one frame without checking speech-type. This is the first frame pulled
997 // without inserting any packet, and might not be labeled as PLC.
henrik.lundin7a926812016-05-12 13:51:28 -0700998 ASSERT_EQ(0, neteq_->GetAudio(&output, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -0800999 ASSERT_EQ(1u, output.num_channels_);
1000 ASSERT_EQ(expected_samples_per_channel, output.samples_per_channel_);
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +00001001
1002 // To be able to test the fading of background noise we need at lease to
1003 // pull 611 frames.
1004 const int kFadingThreshold = 611;
1005
1006 // Test several CNG-to-PLC packet for the expected behavior. The number 20
1007 // is arbitrary, but sufficiently large to test enough number of frames.
1008 const int kNumPlcToCngTestFrames = 20;
1009 bool plc_to_cng = false;
1010 for (int n = 0; n < kFadingThreshold + kNumPlcToCngTestFrames; ++n) {
henrik.lundin6d8e0112016-03-04 10:34:21 -08001011 output.Reset();
yujo36b1a5f2017-06-12 12:45:32 -07001012 // Set to non-zero.
1013 memset(output.mutable_data(), 1, AudioFrame::kMaxDataSizeBytes);
henrik.lundin7a926812016-05-12 13:51:28 -07001014 ASSERT_EQ(0, neteq_->GetAudio(&output, &muted));
1015 ASSERT_FALSE(muted);
henrik.lundin6d8e0112016-03-04 10:34:21 -08001016 ASSERT_EQ(1u, output.num_channels_);
1017 ASSERT_EQ(expected_samples_per_channel, output.samples_per_channel_);
henrik.lundin55480f52016-03-08 02:37:57 -08001018 if (output.speech_type_ == AudioFrame::kPLCCNG) {
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +00001019 plc_to_cng = true;
1020 double sum_squared = 0;
yujo36b1a5f2017-06-12 12:45:32 -07001021 const int16_t* output_data = output.data();
henrik.lundin6d8e0112016-03-04 10:34:21 -08001022 for (size_t k = 0;
1023 k < output.num_channels_ * output.samples_per_channel_; ++k)
yujo36b1a5f2017-06-12 12:45:32 -07001024 sum_squared += output_data[k] * output_data[k];
Henrik Lundin67190172018-04-20 15:34:48 +02001025 EXPECT_EQ(0, sum_squared);
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +00001026 } else {
henrik.lundin55480f52016-03-08 02:37:57 -08001027 EXPECT_EQ(AudioFrame::kPLC, output.speech_type_);
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +00001028 }
1029 }
1030 EXPECT_TRUE(plc_to_cng); // Just to be sure that PLC-to-CNG has occurred.
1031 }
1032};
1033
Henrik Lundin67190172018-04-20 15:34:48 +02001034TEST_F(NetEqBgnTest, RunTest) {
henrik.lundin@webrtc.org9b8102c2014-08-21 08:27:44 +00001035 CheckBgn(8000);
1036 CheckBgn(16000);
1037 CheckBgn(32000);
1038}
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +00001039
turaj@webrtc.org78b41a02013-11-22 20:27:07 +00001040void NetEqDecodingTest::WrapTest(uint16_t start_seq_no,
1041 uint32_t start_timestamp,
1042 const std::set<uint16_t>& drop_seq_numbers,
1043 bool expect_seq_no_wrap,
1044 bool expect_timestamp_wrap) {
1045 uint16_t seq_no = start_seq_no;
1046 uint32_t timestamp = start_timestamp;
1047 const int kBlocksPerFrame = 3; // Number of 10 ms blocks per frame.
1048 const int kFrameSizeMs = kBlocksPerFrame * kTimeStepMs;
1049 const int kSamples = kBlockSize16kHz * kBlocksPerFrame;
pkasting@chromium.org4591fbd2014-11-20 22:28:14 +00001050 const size_t kPayloadBytes = kSamples * sizeof(int16_t);
turaj@webrtc.org78b41a02013-11-22 20:27:07 +00001051 double next_input_time_ms = 0.0;
turaj@webrtc.org78b41a02013-11-22 20:27:07 +00001052 uint32_t receive_timestamp = 0;
1053
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001054 // Insert speech for 2 seconds.
turaj@webrtc.org78b41a02013-11-22 20:27:07 +00001055 const int kSpeechDurationMs = 2000;
1056 int packets_inserted = 0;
1057 uint16_t last_seq_no;
1058 uint32_t last_timestamp;
1059 bool timestamp_wrapped = false;
1060 bool seq_no_wrapped = false;
1061 for (double t_ms = 0; t_ms < kSpeechDurationMs; t_ms += 10) {
1062 // Each turn in this for loop is 10 ms.
1063 while (next_input_time_ms <= t_ms) {
1064 // Insert one 30 ms speech frame.
1065 uint8_t payload[kPayloadBytes] = {0};
henrik.lundin246ef3e2017-04-24 09:14:32 -07001066 RTPHeader rtp_info;
turaj@webrtc.org78b41a02013-11-22 20:27:07 +00001067 PopulateRtpInfo(seq_no, timestamp, &rtp_info);
1068 if (drop_seq_numbers.find(seq_no) == drop_seq_numbers.end()) {
1069 // This sequence number was not in the set to drop. Insert it.
henrik.lundin246ef3e2017-04-24 09:14:32 -07001070 ASSERT_EQ(0,
1071 neteq_->InsertPacket(rtp_info, payload, receive_timestamp));
turaj@webrtc.org78b41a02013-11-22 20:27:07 +00001072 ++packets_inserted;
1073 }
1074 NetEqNetworkStatistics network_stats;
1075 ASSERT_EQ(0, neteq_->NetworkStatistics(&network_stats));
1076
1077 // Due to internal NetEq logic, preferred buffer-size is about 4 times the
1078 // packet size for first few packets. Therefore we refrain from checking
1079 // the criteria.
1080 if (packets_inserted > 4) {
1081 // Expect preferred and actual buffer size to be no more than 2 frames.
1082 EXPECT_LE(network_stats.preferred_buffer_size_ms, kFrameSizeMs * 2);
Yves Gerey665174f2018-06-19 15:03:05 +02001083 EXPECT_LE(network_stats.current_buffer_size_ms,
1084 kFrameSizeMs * 2 + algorithmic_delay_ms_);
turaj@webrtc.org78b41a02013-11-22 20:27:07 +00001085 }
1086 last_seq_no = seq_no;
1087 last_timestamp = timestamp;
1088
1089 ++seq_no;
1090 timestamp += kSamples;
1091 receive_timestamp += kSamples;
1092 next_input_time_ms += static_cast<double>(kFrameSizeMs);
1093
1094 seq_no_wrapped |= seq_no < last_seq_no;
1095 timestamp_wrapped |= timestamp < last_timestamp;
1096 }
1097 // Pull out data once.
henrik.lundin6d8e0112016-03-04 10:34:21 -08001098 AudioFrame output;
henrik.lundin7a926812016-05-12 13:51:28 -07001099 bool muted;
1100 ASSERT_EQ(0, neteq_->GetAudio(&output, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -08001101 ASSERT_EQ(kBlockSize16kHz, output.samples_per_channel_);
1102 ASSERT_EQ(1u, output.num_channels_);
turaj@webrtc.org78b41a02013-11-22 20:27:07 +00001103
1104 // Expect delay (in samples) to be less than 2 packets.
Danil Chapovalovb6021232018-06-19 13:26:36 +02001105 absl::optional<uint32_t> playout_timestamp = neteq_->GetPlayoutTimestamp();
henrik.lundin0d96ab72016-04-06 12:28:26 -07001106 ASSERT_TRUE(playout_timestamp);
1107 EXPECT_LE(timestamp - *playout_timestamp,
turaj@webrtc.org78b41a02013-11-22 20:27:07 +00001108 static_cast<uint32_t>(kSamples * 2));
turaj@webrtc.org78b41a02013-11-22 20:27:07 +00001109 }
1110 // Make sure we have actually tested wrap-around.
1111 ASSERT_EQ(expect_seq_no_wrap, seq_no_wrapped);
1112 ASSERT_EQ(expect_timestamp_wrap, timestamp_wrapped);
1113}
1114
1115TEST_F(NetEqDecodingTest, SequenceNumberWrap) {
1116 // Start with a sequence number that will soon wrap.
1117 std::set<uint16_t> drop_seq_numbers; // Don't drop any packets.
1118 WrapTest(0xFFFF - 10, 0, drop_seq_numbers, true, false);
1119}
1120
1121TEST_F(NetEqDecodingTest, SequenceNumberWrapAndDrop) {
1122 // Start with a sequence number that will soon wrap.
1123 std::set<uint16_t> drop_seq_numbers;
1124 drop_seq_numbers.insert(0xFFFF);
1125 drop_seq_numbers.insert(0x0);
1126 WrapTest(0xFFFF - 10, 0, drop_seq_numbers, true, false);
1127}
1128
1129TEST_F(NetEqDecodingTest, TimestampWrap) {
1130 // Start with a timestamp that will soon wrap.
1131 std::set<uint16_t> drop_seq_numbers;
1132 WrapTest(0, 0xFFFFFFFF - 3000, drop_seq_numbers, false, true);
1133}
1134
1135TEST_F(NetEqDecodingTest, TimestampAndSequenceNumberWrap) {
1136 // Start with a timestamp and a sequence number that will wrap at the same
1137 // time.
1138 std::set<uint16_t> drop_seq_numbers;
1139 WrapTest(0xFFFF - 10, 0xFFFFFFFF - 5000, drop_seq_numbers, true, true);
1140}
1141
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001142void NetEqDecodingTest::DuplicateCng() {
1143 uint16_t seq_no = 0;
1144 uint32_t timestamp = 0;
1145 const int kFrameSizeMs = 10;
1146 const int kSampleRateKhz = 16;
1147 const int kSamples = kFrameSizeMs * kSampleRateKhz;
pkasting@chromium.org4591fbd2014-11-20 22:28:14 +00001148 const size_t kPayloadBytes = kSamples * 2;
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001149
Yves Gerey665174f2018-06-19 15:03:05 +02001150 const int algorithmic_delay_samples =
1151 std::max(algorithmic_delay_ms_ * kSampleRateKhz, 5 * kSampleRateKhz / 8);
henrik.lundin@webrtc.orgc93437e2014-12-01 11:42:42 +00001152 // Insert three speech packets. Three are needed to get the frame length
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001153 // correct.
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001154 uint8_t payload[kPayloadBytes] = {0};
henrik.lundin246ef3e2017-04-24 09:14:32 -07001155 RTPHeader rtp_info;
henrik.lundin7a926812016-05-12 13:51:28 -07001156 bool muted;
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001157 for (int i = 0; i < 3; ++i) {
1158 PopulateRtpInfo(seq_no, timestamp, &rtp_info);
henrik.lundin246ef3e2017-04-24 09:14:32 -07001159 ASSERT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001160 ++seq_no;
1161 timestamp += kSamples;
1162
1163 // Pull audio once.
henrik.lundin7a926812016-05-12 13:51:28 -07001164 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -08001165 ASSERT_EQ(kBlockSize16kHz, out_frame_.samples_per_channel_);
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001166 }
1167 // Verify speech output.
henrik.lundin55480f52016-03-08 02:37:57 -08001168 EXPECT_EQ(AudioFrame::kNormalSpeech, out_frame_.speech_type_);
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001169
1170 // Insert same CNG packet twice.
1171 const int kCngPeriodMs = 100;
1172 const int kCngPeriodSamples = kCngPeriodMs * kSampleRateKhz;
pkasting@chromium.org4591fbd2014-11-20 22:28:14 +00001173 size_t payload_len;
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001174 PopulateCng(seq_no, timestamp, &rtp_info, payload, &payload_len);
1175 // This is the first time this CNG packet is inserted.
henrik.lundin246ef3e2017-04-24 09:14:32 -07001176 ASSERT_EQ(
1177 0, neteq_->InsertPacket(
1178 rtp_info, rtc::ArrayView<const uint8_t>(payload, payload_len), 0));
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001179
1180 // Pull audio once and make sure CNG is played.
henrik.lundin7a926812016-05-12 13:51:28 -07001181 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -08001182 ASSERT_EQ(kBlockSize16kHz, out_frame_.samples_per_channel_);
henrik.lundin55480f52016-03-08 02:37:57 -08001183 EXPECT_EQ(AudioFrame::kCNG, out_frame_.speech_type_);
henrik.lundin114c1b32017-04-26 07:47:32 -07001184 EXPECT_FALSE(
1185 neteq_->GetPlayoutTimestamp()); // Returns empty value during CNG.
henrik.lundin0d96ab72016-04-06 12:28:26 -07001186 EXPECT_EQ(timestamp - algorithmic_delay_samples,
1187 out_frame_.timestamp_ + out_frame_.samples_per_channel_);
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001188
1189 // Insert the same CNG packet again. Note that at this point it is old, since
1190 // we have already decoded the first copy of it.
henrik.lundin246ef3e2017-04-24 09:14:32 -07001191 ASSERT_EQ(
1192 0, neteq_->InsertPacket(
1193 rtp_info, rtc::ArrayView<const uint8_t>(payload, payload_len), 0));
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001194
1195 // Pull audio until we have played |kCngPeriodMs| of CNG. Start at 10 ms since
1196 // we have already pulled out CNG once.
1197 for (int cng_time_ms = 10; cng_time_ms < kCngPeriodMs; cng_time_ms += 10) {
henrik.lundin7a926812016-05-12 13:51:28 -07001198 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -08001199 ASSERT_EQ(kBlockSize16kHz, out_frame_.samples_per_channel_);
henrik.lundin55480f52016-03-08 02:37:57 -08001200 EXPECT_EQ(AudioFrame::kCNG, out_frame_.speech_type_);
henrik.lundin114c1b32017-04-26 07:47:32 -07001201 EXPECT_FALSE(
1202 neteq_->GetPlayoutTimestamp()); // Returns empty value during CNG.
turaj@webrtc.org8d1cdaa2014-04-11 18:47:55 +00001203 EXPECT_EQ(timestamp - algorithmic_delay_samples,
henrik.lundin0d96ab72016-04-06 12:28:26 -07001204 out_frame_.timestamp_ + out_frame_.samples_per_channel_);
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001205 }
1206
1207 // Insert speech again.
1208 ++seq_no;
1209 timestamp += kCngPeriodSamples;
1210 PopulateRtpInfo(seq_no, timestamp, &rtp_info);
henrik.lundin246ef3e2017-04-24 09:14:32 -07001211 ASSERT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001212
1213 // Pull audio once and verify that the output is speech again.
henrik.lundin7a926812016-05-12 13:51:28 -07001214 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -08001215 ASSERT_EQ(kBlockSize16kHz, out_frame_.samples_per_channel_);
henrik.lundin55480f52016-03-08 02:37:57 -08001216 EXPECT_EQ(AudioFrame::kNormalSpeech, out_frame_.speech_type_);
Danil Chapovalovb6021232018-06-19 13:26:36 +02001217 absl::optional<uint32_t> playout_timestamp = neteq_->GetPlayoutTimestamp();
henrik.lundin0d96ab72016-04-06 12:28:26 -07001218 ASSERT_TRUE(playout_timestamp);
turaj@webrtc.org8d1cdaa2014-04-11 18:47:55 +00001219 EXPECT_EQ(timestamp + kSamples - algorithmic_delay_samples,
henrik.lundin0d96ab72016-04-06 12:28:26 -07001220 *playout_timestamp);
wu@webrtc.org94454b72014-06-05 20:34:08 +00001221}
1222
Yves Gerey665174f2018-06-19 15:03:05 +02001223TEST_F(NetEqDecodingTest, DiscardDuplicateCng) {
1224 DuplicateCng();
1225}
henrik.lundin@webrtc.orgc93437e2014-12-01 11:42:42 +00001226
1227TEST_F(NetEqDecodingTest, CngFirst) {
1228 uint16_t seq_no = 0;
1229 uint32_t timestamp = 0;
1230 const int kFrameSizeMs = 10;
1231 const int kSampleRateKhz = 16;
1232 const int kSamples = kFrameSizeMs * kSampleRateKhz;
1233 const int kPayloadBytes = kSamples * 2;
1234 const int kCngPeriodMs = 100;
1235 const int kCngPeriodSamples = kCngPeriodMs * kSampleRateKhz;
1236 size_t payload_len;
1237
1238 uint8_t payload[kPayloadBytes] = {0};
henrik.lundin246ef3e2017-04-24 09:14:32 -07001239 RTPHeader rtp_info;
henrik.lundin@webrtc.orgc93437e2014-12-01 11:42:42 +00001240
1241 PopulateCng(seq_no, timestamp, &rtp_info, payload, &payload_len);
henrik.lundin246ef3e2017-04-24 09:14:32 -07001242 ASSERT_EQ(
1243 NetEq::kOK,
1244 neteq_->InsertPacket(
1245 rtp_info, rtc::ArrayView<const uint8_t>(payload, payload_len), 0));
henrik.lundin@webrtc.orgc93437e2014-12-01 11:42:42 +00001246 ++seq_no;
1247 timestamp += kCngPeriodSamples;
1248
1249 // Pull audio once and make sure CNG is played.
henrik.lundin7a926812016-05-12 13:51:28 -07001250 bool muted;
1251 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -08001252 ASSERT_EQ(kBlockSize16kHz, out_frame_.samples_per_channel_);
henrik.lundin55480f52016-03-08 02:37:57 -08001253 EXPECT_EQ(AudioFrame::kCNG, out_frame_.speech_type_);
henrik.lundin@webrtc.orgc93437e2014-12-01 11:42:42 +00001254
1255 // Insert some speech packets.
henrik.lundin549d80b2016-08-25 00:44:24 -07001256 const uint32_t first_speech_timestamp = timestamp;
1257 int timeout_counter = 0;
1258 do {
1259 ASSERT_LT(timeout_counter++, 20) << "Test timed out";
henrik.lundin@webrtc.orgc93437e2014-12-01 11:42:42 +00001260 PopulateRtpInfo(seq_no, timestamp, &rtp_info);
henrik.lundin246ef3e2017-04-24 09:14:32 -07001261 ASSERT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
henrik.lundin@webrtc.orgc93437e2014-12-01 11:42:42 +00001262 ++seq_no;
1263 timestamp += kSamples;
1264
1265 // Pull audio once.
henrik.lundin7a926812016-05-12 13:51:28 -07001266 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -08001267 ASSERT_EQ(kBlockSize16kHz, out_frame_.samples_per_channel_);
henrik.lundin549d80b2016-08-25 00:44:24 -07001268 } while (!IsNewerTimestamp(out_frame_.timestamp_, first_speech_timestamp));
henrik.lundin@webrtc.orgc93437e2014-12-01 11:42:42 +00001269 // Verify speech output.
henrik.lundin55480f52016-03-08 02:37:57 -08001270 EXPECT_EQ(AudioFrame::kNormalSpeech, out_frame_.speech_type_);
henrik.lundin@webrtc.orgc93437e2014-12-01 11:42:42 +00001271}
henrik.lundin7a926812016-05-12 13:51:28 -07001272
1273class NetEqDecodingTestWithMutedState : public NetEqDecodingTest {
1274 public:
1275 NetEqDecodingTestWithMutedState() : NetEqDecodingTest() {
1276 config_.enable_muted_state = true;
1277 }
1278
1279 protected:
1280 static constexpr size_t kSamples = 10 * 16;
1281 static constexpr size_t kPayloadBytes = kSamples * 2;
1282
1283 void InsertPacket(uint32_t rtp_timestamp) {
1284 uint8_t payload[kPayloadBytes] = {0};
henrik.lundin246ef3e2017-04-24 09:14:32 -07001285 RTPHeader rtp_info;
henrik.lundin7a926812016-05-12 13:51:28 -07001286 PopulateRtpInfo(0, rtp_timestamp, &rtp_info);
henrik.lundin246ef3e2017-04-24 09:14:32 -07001287 EXPECT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
henrik.lundin7a926812016-05-12 13:51:28 -07001288 }
1289
henrik.lundin42feb512016-09-20 06:51:40 -07001290 void InsertCngPacket(uint32_t rtp_timestamp) {
1291 uint8_t payload[kPayloadBytes] = {0};
henrik.lundin246ef3e2017-04-24 09:14:32 -07001292 RTPHeader rtp_info;
henrik.lundin42feb512016-09-20 06:51:40 -07001293 size_t payload_len;
1294 PopulateCng(0, rtp_timestamp, &rtp_info, payload, &payload_len);
henrik.lundin246ef3e2017-04-24 09:14:32 -07001295 EXPECT_EQ(
1296 NetEq::kOK,
1297 neteq_->InsertPacket(
1298 rtp_info, rtc::ArrayView<const uint8_t>(payload, payload_len), 0));
henrik.lundin42feb512016-09-20 06:51:40 -07001299 }
1300
henrik.lundin7a926812016-05-12 13:51:28 -07001301 bool GetAudioReturnMuted() {
1302 bool muted;
1303 EXPECT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
1304 return muted;
1305 }
1306
1307 void GetAudioUntilMuted() {
1308 while (!GetAudioReturnMuted()) {
1309 ASSERT_LT(counter_++, 1000) << "Test timed out";
1310 }
1311 }
1312
1313 void GetAudioUntilNormal() {
1314 bool muted = false;
1315 while (out_frame_.speech_type_ != AudioFrame::kNormalSpeech) {
1316 EXPECT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
1317 ASSERT_LT(counter_++, 1000) << "Test timed out";
1318 }
1319 EXPECT_FALSE(muted);
1320 }
1321
1322 int counter_ = 0;
1323};
1324
1325// Verifies that NetEq goes in and out of muted state as expected.
1326TEST_F(NetEqDecodingTestWithMutedState, MutedState) {
1327 // Insert one speech packet.
1328 InsertPacket(0);
1329 // Pull out audio once and expect it not to be muted.
1330 EXPECT_FALSE(GetAudioReturnMuted());
1331 // Pull data until faded out.
1332 GetAudioUntilMuted();
henrik.lundina4491072017-07-06 05:23:53 -07001333 EXPECT_TRUE(out_frame_.muted());
henrik.lundin7a926812016-05-12 13:51:28 -07001334
1335 // Verify that output audio is not written during muted mode. Other parameters
1336 // should be correct, though.
1337 AudioFrame new_frame;
yujo36b1a5f2017-06-12 12:45:32 -07001338 int16_t* frame_data = new_frame.mutable_data();
1339 for (size_t i = 0; i < AudioFrame::kMaxDataSizeSamples; i++) {
1340 frame_data[i] = 17;
henrik.lundin7a926812016-05-12 13:51:28 -07001341 }
1342 bool muted;
1343 EXPECT_EQ(0, neteq_->GetAudio(&new_frame, &muted));
1344 EXPECT_TRUE(muted);
henrik.lundina4491072017-07-06 05:23:53 -07001345 EXPECT_TRUE(out_frame_.muted());
yujo36b1a5f2017-06-12 12:45:32 -07001346 for (size_t i = 0; i < AudioFrame::kMaxDataSizeSamples; i++) {
1347 EXPECT_EQ(17, frame_data[i]);
henrik.lundin7a926812016-05-12 13:51:28 -07001348 }
1349 EXPECT_EQ(out_frame_.timestamp_ + out_frame_.samples_per_channel_,
1350 new_frame.timestamp_);
1351 EXPECT_EQ(out_frame_.samples_per_channel_, new_frame.samples_per_channel_);
1352 EXPECT_EQ(out_frame_.sample_rate_hz_, new_frame.sample_rate_hz_);
1353 EXPECT_EQ(out_frame_.num_channels_, new_frame.num_channels_);
1354 EXPECT_EQ(out_frame_.speech_type_, new_frame.speech_type_);
1355 EXPECT_EQ(out_frame_.vad_activity_, new_frame.vad_activity_);
1356
1357 // Insert new data. Timestamp is corrected for the time elapsed since the last
1358 // packet. Verify that normal operation resumes.
1359 InsertPacket(kSamples * counter_);
1360 GetAudioUntilNormal();
henrik.lundina4491072017-07-06 05:23:53 -07001361 EXPECT_FALSE(out_frame_.muted());
henrik.lundin612c25e2016-05-25 08:21:04 -07001362
1363 NetEqNetworkStatistics stats;
1364 EXPECT_EQ(0, neteq_->NetworkStatistics(&stats));
1365 // NetEqNetworkStatistics::expand_rate tells the fraction of samples that were
1366 // concealment samples, in Q14 (16384 = 100%) .The vast majority should be
1367 // concealment samples in this test.
1368 EXPECT_GT(stats.expand_rate, 14000);
1369 // And, it should be greater than the speech_expand_rate.
1370 EXPECT_GT(stats.expand_rate, stats.speech_expand_rate);
henrik.lundin7a926812016-05-12 13:51:28 -07001371}
1372
1373// Verifies that NetEq goes out of muted state when given a delayed packet.
1374TEST_F(NetEqDecodingTestWithMutedState, MutedStateDelayedPacket) {
1375 // Insert one speech packet.
1376 InsertPacket(0);
1377 // Pull out audio once and expect it not to be muted.
1378 EXPECT_FALSE(GetAudioReturnMuted());
1379 // Pull data until faded out.
1380 GetAudioUntilMuted();
1381 // Insert new data. Timestamp is only corrected for the half of the time
1382 // elapsed since the last packet. That is, the new packet is delayed. Verify
1383 // that normal operation resumes.
1384 InsertPacket(kSamples * counter_ / 2);
1385 GetAudioUntilNormal();
1386}
1387
1388// Verifies that NetEq goes out of muted state when given a future packet.
1389TEST_F(NetEqDecodingTestWithMutedState, MutedStateFuturePacket) {
1390 // Insert one speech packet.
1391 InsertPacket(0);
1392 // Pull out audio once and expect it not to be muted.
1393 EXPECT_FALSE(GetAudioReturnMuted());
1394 // Pull data until faded out.
1395 GetAudioUntilMuted();
1396 // Insert new data. Timestamp is over-corrected for the time elapsed since the
1397 // last packet. That is, the new packet is too early. Verify that normal
1398 // operation resumes.
1399 InsertPacket(kSamples * counter_ * 2);
1400 GetAudioUntilNormal();
1401}
1402
1403// Verifies that NetEq goes out of muted state when given an old packet.
1404TEST_F(NetEqDecodingTestWithMutedState, MutedStateOldPacket) {
1405 // Insert one speech packet.
1406 InsertPacket(0);
1407 // Pull out audio once and expect it not to be muted.
1408 EXPECT_FALSE(GetAudioReturnMuted());
1409 // Pull data until faded out.
1410 GetAudioUntilMuted();
1411
1412 EXPECT_NE(AudioFrame::kNormalSpeech, out_frame_.speech_type_);
1413 // Insert packet which is older than the first packet.
1414 InsertPacket(kSamples * (counter_ - 1000));
1415 EXPECT_FALSE(GetAudioReturnMuted());
1416 EXPECT_EQ(AudioFrame::kNormalSpeech, out_frame_.speech_type_);
1417}
1418
henrik.lundin42feb512016-09-20 06:51:40 -07001419// Verifies that NetEq doesn't enter muted state when CNG mode is active and the
1420// packet stream is suspended for a long time.
1421TEST_F(NetEqDecodingTestWithMutedState, DoNotMuteExtendedCngWithoutPackets) {
1422 // Insert one CNG packet.
1423 InsertCngPacket(0);
1424
1425 // Pull 10 seconds of audio (10 ms audio generated per lap).
1426 for (int i = 0; i < 1000; ++i) {
1427 bool muted;
1428 EXPECT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
1429 ASSERT_FALSE(muted);
1430 }
1431 EXPECT_EQ(AudioFrame::kCNG, out_frame_.speech_type_);
1432}
1433
1434// Verifies that NetEq goes back to normal after a long CNG period with the
1435// packet stream suspended.
1436TEST_F(NetEqDecodingTestWithMutedState, RecoverAfterExtendedCngWithoutPackets) {
1437 // Insert one CNG packet.
1438 InsertCngPacket(0);
1439
1440 // Pull 10 seconds of audio (10 ms audio generated per lap).
1441 for (int i = 0; i < 1000; ++i) {
1442 bool muted;
1443 EXPECT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
1444 }
1445
1446 // Insert new data. Timestamp is corrected for the time elapsed since the last
1447 // packet. Verify that normal operation resumes.
1448 InsertPacket(kSamples * counter_);
1449 GetAudioUntilNormal();
1450}
1451
henrik.lundin7a926812016-05-12 13:51:28 -07001452class NetEqDecodingTestTwoInstances : public NetEqDecodingTest {
1453 public:
1454 NetEqDecodingTestTwoInstances() : NetEqDecodingTest() {}
1455
1456 void SetUp() override {
1457 NetEqDecodingTest::SetUp();
1458 config2_ = config_;
1459 }
1460
1461 void CreateSecondInstance() {
ossue3525782016-05-25 07:37:43 -07001462 neteq2_.reset(NetEq::Create(config2_, CreateBuiltinAudioDecoderFactory()));
henrik.lundin7a926812016-05-12 13:51:28 -07001463 ASSERT_TRUE(neteq2_);
1464 LoadDecoders(neteq2_.get());
1465 }
1466
1467 protected:
1468 std::unique_ptr<NetEq> neteq2_;
1469 NetEq::Config config2_;
1470};
1471
1472namespace {
1473::testing::AssertionResult AudioFramesEqualExceptData(const AudioFrame& a,
1474 const AudioFrame& b) {
1475 if (a.timestamp_ != b.timestamp_)
1476 return ::testing::AssertionFailure() << "timestamp_ diff (" << a.timestamp_
1477 << " != " << b.timestamp_ << ")";
1478 if (a.sample_rate_hz_ != b.sample_rate_hz_)
Yves Gerey665174f2018-06-19 15:03:05 +02001479 return ::testing::AssertionFailure()
1480 << "sample_rate_hz_ diff (" << a.sample_rate_hz_
1481 << " != " << b.sample_rate_hz_ << ")";
henrik.lundin7a926812016-05-12 13:51:28 -07001482 if (a.samples_per_channel_ != b.samples_per_channel_)
1483 return ::testing::AssertionFailure()
1484 << "samples_per_channel_ diff (" << a.samples_per_channel_
1485 << " != " << b.samples_per_channel_ << ")";
1486 if (a.num_channels_ != b.num_channels_)
Yves Gerey665174f2018-06-19 15:03:05 +02001487 return ::testing::AssertionFailure()
1488 << "num_channels_ diff (" << a.num_channels_
1489 << " != " << b.num_channels_ << ")";
henrik.lundin7a926812016-05-12 13:51:28 -07001490 if (a.speech_type_ != b.speech_type_)
Yves Gerey665174f2018-06-19 15:03:05 +02001491 return ::testing::AssertionFailure()
1492 << "speech_type_ diff (" << a.speech_type_
1493 << " != " << b.speech_type_ << ")";
henrik.lundin7a926812016-05-12 13:51:28 -07001494 if (a.vad_activity_ != b.vad_activity_)
Yves Gerey665174f2018-06-19 15:03:05 +02001495 return ::testing::AssertionFailure()
1496 << "vad_activity_ diff (" << a.vad_activity_
1497 << " != " << b.vad_activity_ << ")";
henrik.lundin7a926812016-05-12 13:51:28 -07001498 return ::testing::AssertionSuccess();
1499}
1500
1501::testing::AssertionResult AudioFramesEqual(const AudioFrame& a,
1502 const AudioFrame& b) {
1503 ::testing::AssertionResult res = AudioFramesEqualExceptData(a, b);
1504 if (!res)
1505 return res;
Yves Gerey665174f2018-06-19 15:03:05 +02001506 if (memcmp(a.data(), b.data(),
1507 a.samples_per_channel_ * a.num_channels_ * sizeof(*a.data())) !=
1508 0) {
henrik.lundin7a926812016-05-12 13:51:28 -07001509 return ::testing::AssertionFailure() << "data_ diff";
1510 }
1511 return ::testing::AssertionSuccess();
1512}
1513
1514} // namespace
1515
1516TEST_F(NetEqDecodingTestTwoInstances, CompareMutedStateOnOff) {
1517 ASSERT_FALSE(config_.enable_muted_state);
1518 config2_.enable_muted_state = true;
1519 CreateSecondInstance();
1520
1521 // Insert one speech packet into both NetEqs.
1522 const size_t kSamples = 10 * 16;
1523 const size_t kPayloadBytes = kSamples * 2;
1524 uint8_t payload[kPayloadBytes] = {0};
henrik.lundin246ef3e2017-04-24 09:14:32 -07001525 RTPHeader rtp_info;
henrik.lundin7a926812016-05-12 13:51:28 -07001526 PopulateRtpInfo(0, 0, &rtp_info);
henrik.lundin246ef3e2017-04-24 09:14:32 -07001527 EXPECT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
1528 EXPECT_EQ(0, neteq2_->InsertPacket(rtp_info, payload, 0));
henrik.lundin7a926812016-05-12 13:51:28 -07001529
1530 AudioFrame out_frame1, out_frame2;
1531 bool muted;
1532 for (int i = 0; i < 1000; ++i) {
1533 std::ostringstream ss;
1534 ss << "i = " << i;
1535 SCOPED_TRACE(ss.str()); // Print out the loop iterator on failure.
1536 EXPECT_EQ(0, neteq_->GetAudio(&out_frame1, &muted));
1537 EXPECT_FALSE(muted);
1538 EXPECT_EQ(0, neteq2_->GetAudio(&out_frame2, &muted));
1539 if (muted) {
1540 EXPECT_TRUE(AudioFramesEqualExceptData(out_frame1, out_frame2));
1541 } else {
1542 EXPECT_TRUE(AudioFramesEqual(out_frame1, out_frame2));
1543 }
1544 }
1545 EXPECT_TRUE(muted);
1546
1547 // Insert new data. Timestamp is corrected for the time elapsed since the last
1548 // packet.
1549 PopulateRtpInfo(0, kSamples * 1000, &rtp_info);
henrik.lundin246ef3e2017-04-24 09:14:32 -07001550 EXPECT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
1551 EXPECT_EQ(0, neteq2_->InsertPacket(rtp_info, payload, 0));
henrik.lundin7a926812016-05-12 13:51:28 -07001552
1553 int counter = 0;
1554 while (out_frame1.speech_type_ != AudioFrame::kNormalSpeech) {
1555 ASSERT_LT(counter++, 1000) << "Test timed out";
1556 std::ostringstream ss;
1557 ss << "counter = " << counter;
1558 SCOPED_TRACE(ss.str()); // Print out the loop iterator on failure.
1559 EXPECT_EQ(0, neteq_->GetAudio(&out_frame1, &muted));
1560 EXPECT_FALSE(muted);
1561 EXPECT_EQ(0, neteq2_->GetAudio(&out_frame2, &muted));
1562 if (muted) {
1563 EXPECT_TRUE(AudioFramesEqualExceptData(out_frame1, out_frame2));
1564 } else {
1565 EXPECT_TRUE(AudioFramesEqual(out_frame1, out_frame2));
1566 }
1567 }
1568 EXPECT_FALSE(muted);
1569}
1570
henrik.lundin114c1b32017-04-26 07:47:32 -07001571TEST_F(NetEqDecodingTest, LastDecodedTimestampsEmpty) {
1572 EXPECT_TRUE(neteq_->LastDecodedTimestamps().empty());
1573
1574 // Pull out data once.
1575 AudioFrame output;
1576 bool muted;
1577 ASSERT_EQ(0, neteq_->GetAudio(&output, &muted));
1578
1579 EXPECT_TRUE(neteq_->LastDecodedTimestamps().empty());
1580}
1581
1582TEST_F(NetEqDecodingTest, LastDecodedTimestampsOneDecoded) {
1583 // Insert one packet with PCM16b WB data (this is what PopulateRtpInfo does by
1584 // default). Make the length 10 ms.
1585 constexpr size_t kPayloadSamples = 16 * 10;
1586 constexpr size_t kPayloadBytes = 2 * kPayloadSamples;
1587 uint8_t payload[kPayloadBytes] = {0};
1588
1589 RTPHeader rtp_info;
1590 constexpr uint32_t kRtpTimestamp = 0x1234;
1591 PopulateRtpInfo(0, kRtpTimestamp, &rtp_info);
1592 EXPECT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
1593
1594 // Pull out data once.
1595 AudioFrame output;
1596 bool muted;
1597 ASSERT_EQ(0, neteq_->GetAudio(&output, &muted));
1598
1599 EXPECT_EQ(std::vector<uint32_t>({kRtpTimestamp}),
1600 neteq_->LastDecodedTimestamps());
1601
1602 // Nothing decoded on the second call.
1603 ASSERT_EQ(0, neteq_->GetAudio(&output, &muted));
1604 EXPECT_TRUE(neteq_->LastDecodedTimestamps().empty());
1605}
1606
1607TEST_F(NetEqDecodingTest, LastDecodedTimestampsTwoDecoded) {
1608 // Insert two packets with PCM16b WB data (this is what PopulateRtpInfo does
1609 // by default). Make the length 5 ms so that NetEq must decode them both in
1610 // the same GetAudio call.
1611 constexpr size_t kPayloadSamples = 16 * 5;
1612 constexpr size_t kPayloadBytes = 2 * kPayloadSamples;
1613 uint8_t payload[kPayloadBytes] = {0};
1614
1615 RTPHeader rtp_info;
1616 constexpr uint32_t kRtpTimestamp1 = 0x1234;
1617 PopulateRtpInfo(0, kRtpTimestamp1, &rtp_info);
1618 EXPECT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
1619 constexpr uint32_t kRtpTimestamp2 = kRtpTimestamp1 + kPayloadSamples;
1620 PopulateRtpInfo(1, kRtpTimestamp2, &rtp_info);
1621 EXPECT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
1622
1623 // Pull out data once.
1624 AudioFrame output;
1625 bool muted;
1626 ASSERT_EQ(0, neteq_->GetAudio(&output, &muted));
1627
1628 EXPECT_EQ(std::vector<uint32_t>({kRtpTimestamp1, kRtpTimestamp2}),
1629 neteq_->LastDecodedTimestamps());
1630}
1631
Gustaf Ullberg9a2e9062017-09-18 09:28:20 +02001632TEST_F(NetEqDecodingTest, TestConcealmentEvents) {
1633 const int kNumConcealmentEvents = 19;
1634 const size_t kSamples = 10 * 16;
1635 const size_t kPayloadBytes = kSamples * 2;
1636 int seq_no = 0;
1637 RTPHeader rtp_info;
1638 rtp_info.ssrc = 0x1234; // Just an arbitrary SSRC.
1639 rtp_info.payloadType = 94; // PCM16b WB codec.
1640 rtp_info.markerBit = 0;
1641 const uint8_t payload[kPayloadBytes] = {0};
1642 bool muted;
1643
1644 for (int i = 0; i < kNumConcealmentEvents; i++) {
1645 // Insert some packets of 10 ms size.
1646 for (int j = 0; j < 10; j++) {
1647 rtp_info.sequenceNumber = seq_no++;
1648 rtp_info.timestamp = rtp_info.sequenceNumber * kSamples;
1649 neteq_->InsertPacket(rtp_info, payload, 0);
1650 neteq_->GetAudio(&out_frame_, &muted);
1651 }
1652
1653 // Lose a number of packets.
1654 int num_lost = 1 + i;
1655 for (int j = 0; j < num_lost; j++) {
1656 seq_no++;
1657 neteq_->GetAudio(&out_frame_, &muted);
1658 }
1659 }
1660
1661 // Check number of concealment events.
1662 NetEqLifetimeStatistics stats = neteq_->GetLifetimeStatistics();
1663 EXPECT_EQ(kNumConcealmentEvents, static_cast<int>(stats.concealment_events));
1664}
1665
Gustaf Ullbergb0a02072017-10-02 12:00:34 +02001666// Test that the jitter buffer delay stat is computed correctly.
1667void NetEqDecodingTestFaxMode::TestJitterBufferDelay(bool apply_packet_loss) {
1668 const int kNumPackets = 10;
1669 const int kDelayInNumPackets = 2;
1670 const int kPacketLenMs = 10; // All packets are of 10 ms size.
1671 const size_t kSamples = kPacketLenMs * 16;
1672 const size_t kPayloadBytes = kSamples * 2;
1673 RTPHeader rtp_info;
1674 rtp_info.ssrc = 0x1234; // Just an arbitrary SSRC.
1675 rtp_info.payloadType = 94; // PCM16b WB codec.
1676 rtp_info.markerBit = 0;
1677 const uint8_t payload[kPayloadBytes] = {0};
1678 bool muted;
1679 int packets_sent = 0;
1680 int packets_received = 0;
1681 int expected_delay = 0;
1682 while (packets_received < kNumPackets) {
1683 // Insert packet.
1684 if (packets_sent < kNumPackets) {
1685 rtp_info.sequenceNumber = packets_sent++;
1686 rtp_info.timestamp = rtp_info.sequenceNumber * kSamples;
1687 neteq_->InsertPacket(rtp_info, payload, 0);
1688 }
1689
1690 // Get packet.
1691 if (packets_sent > kDelayInNumPackets) {
1692 neteq_->GetAudio(&out_frame_, &muted);
1693 packets_received++;
1694
1695 // The delay reported by the jitter buffer never exceeds
1696 // the number of samples previously fetched with GetAudio
1697 // (hence the min()).
1698 int packets_delay = std::min(packets_received, kDelayInNumPackets + 1);
1699
1700 // The increase of the expected delay is the product of
1701 // the current delay of the jitter buffer in ms * the
1702 // number of samples that are sent for play out.
1703 int current_delay_ms = packets_delay * kPacketLenMs;
1704 expected_delay += current_delay_ms * kSamples;
1705 }
1706 }
1707
1708 if (apply_packet_loss) {
1709 // Extra call to GetAudio to cause concealment.
1710 neteq_->GetAudio(&out_frame_, &muted);
1711 }
1712
1713 // Check jitter buffer delay.
1714 NetEqLifetimeStatistics stats = neteq_->GetLifetimeStatistics();
1715 EXPECT_EQ(expected_delay, static_cast<int>(stats.jitter_buffer_delay_ms));
1716}
1717
1718TEST_F(NetEqDecodingTestFaxMode, TestJitterBufferDelayWithoutLoss) {
1719 TestJitterBufferDelay(false);
1720}
1721
1722TEST_F(NetEqDecodingTestFaxMode, TestJitterBufferDelayWithLoss) {
1723 TestJitterBufferDelay(true);
1724}
1725
henrik.lundin@webrtc.orge7ce4372014-01-09 14:01:55 +00001726} // namespace webrtc