blob: c1b32cb311754735c1cfc2b017104a8fca7cf52f [file] [log] [blame]
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +00001/*
2 * Copyright (c) 2011 The WebRTC project authors. All Rights Reserved.
3 *
4 * Use of this source code is governed by a BSD-style license
5 * that can be found in the LICENSE file in the root of the source
6 * tree. An additional intellectual property rights grant can be found
7 * in the file PATENTS. All contributing project authors may
8 * be found in the AUTHORS file in the root of the source tree.
9 */
10
Mirko Bonadei92ea95e2017-09-15 06:47:31 +020011#include "modules/audio_coding/neteq/include/neteq.h"
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +000012
pbos@webrtc.org3ecc1622014-03-07 15:23:34 +000013#include <math.h>
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +000014#include <stdlib.h>
15#include <string.h> // memset
16
turaj@webrtc.org8d1cdaa2014-04-11 18:47:55 +000017#include <algorithm>
kwiberg2d0c3322016-02-14 09:28:33 -080018#include <memory>
turaj@webrtc.org78b41a02013-11-22 20:27:07 +000019#include <set>
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +000020#include <string>
21#include <vector>
22
Mirko Bonadei92ea95e2017-09-15 06:47:31 +020023#include "api/audio_codecs/builtin_audio_decoder_factory.h"
Mirko Bonadei71207422017-09-15 13:58:09 +020024#include "common_types.h" // NOLINT(build/include)
Mirko Bonadei92ea95e2017-09-15 06:47:31 +020025#include "modules/audio_coding/codecs/pcm16b/pcm16b.h"
26#include "modules/audio_coding/neteq/tools/audio_loop.h"
27#include "modules/audio_coding/neteq/tools/rtp_file_source.h"
28#include "modules/include/module_common_types.h"
29#include "rtc_base/flags.h"
30#include "rtc_base/ignore_wundef.h"
Joachim Bauch4e909192017-12-19 22:27:51 +010031#include "rtc_base/messagedigest.h"
Karl Wiberge40468b2017-11-22 10:42:26 +010032#include "rtc_base/numerics/safe_conversions.h"
Mirko Bonadei92ea95e2017-09-15 06:47:31 +020033#include "rtc_base/protobuf_utils.h"
Mirko Bonadei92ea95e2017-09-15 06:47:31 +020034#include "rtc_base/stringencode.h"
Henrik Lundine9619f82017-11-27 14:05:27 +010035#include "test/field_trial.h"
Mirko Bonadei92ea95e2017-09-15 06:47:31 +020036#include "test/gtest.h"
37#include "test/testsupport/fileutils.h"
Mirko Bonadei71207422017-09-15 13:58:09 +020038#include "typedefs.h" // NOLINT(build/include)
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +000039
minyue5f026d02015-12-16 07:36:04 -080040#ifdef WEBRTC_NETEQ_UNITTEST_BITEXACT
kwiberg77eab702016-09-28 17:42:01 -070041RTC_PUSH_IGNORING_WUNDEF()
minyue5f026d02015-12-16 07:36:04 -080042#ifdef WEBRTC_ANDROID_PLATFORM_BUILD
43#include "external/webrtc/webrtc/modules/audio_coding/neteq/neteq_unittest.pb.h"
44#else
Mirko Bonadei92ea95e2017-09-15 06:47:31 +020045#include "modules/audio_coding/neteq/neteq_unittest.pb.h"
minyue5f026d02015-12-16 07:36:04 -080046#endif
kwiberg77eab702016-09-28 17:42:01 -070047RTC_POP_IGNORING_WUNDEF()
minyue5f026d02015-12-16 07:36:04 -080048#endif
49
turaj@webrtc.orga6101d72013-10-01 22:01:09 +000050DEFINE_bool(gen_ref, false, "Generate reference files.");
51
kwiberg5adaf732016-10-04 09:33:27 -070052namespace webrtc {
53
minyue5f026d02015-12-16 07:36:04 -080054namespace {
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +000055
minyue4f906772016-04-29 11:05:14 -070056const std::string& PlatformChecksum(const std::string& checksum_general,
Henrik Lundin8cd750d2017-10-12 13:07:11 +020057 const std::string& checksum_android_32,
58 const std::string& checksum_android_64,
minyue4f906772016-04-29 11:05:14 -070059 const std::string& checksum_win_32,
60 const std::string& checksum_win_64) {
kwiberg77eab702016-09-28 17:42:01 -070061#if defined(WEBRTC_ANDROID)
Henrik Lundin8cd750d2017-10-12 13:07:11 +020062 #ifdef WEBRTC_ARCH_64_BITS
63 return checksum_android_64;
64 #else
65 return checksum_android_32;
66 #endif // WEBRTC_ARCH_64_BITS
kwiberg77eab702016-09-28 17:42:01 -070067#elif defined(WEBRTC_WIN)
minyue4f906772016-04-29 11:05:14 -070068 #ifdef WEBRTC_ARCH_64_BITS
69 return checksum_win_64;
70 #else
71 return checksum_win_32;
72 #endif // WEBRTC_ARCH_64_BITS
73#else
74 return checksum_general;
75#endif // WEBRTC_WIN
76}
77
minyue5f026d02015-12-16 07:36:04 -080078#ifdef WEBRTC_NETEQ_UNITTEST_BITEXACT
79void Convert(const webrtc::NetEqNetworkStatistics& stats_raw,
80 webrtc::neteq_unittest::NetEqNetworkStatistics* stats) {
81 stats->set_current_buffer_size_ms(stats_raw.current_buffer_size_ms);
82 stats->set_preferred_buffer_size_ms(stats_raw.preferred_buffer_size_ms);
83 stats->set_jitter_peaks_found(stats_raw.jitter_peaks_found);
84 stats->set_packet_loss_rate(stats_raw.packet_loss_rate);
minyue5f026d02015-12-16 07:36:04 -080085 stats->set_expand_rate(stats_raw.expand_rate);
86 stats->set_speech_expand_rate(stats_raw.speech_expand_rate);
87 stats->set_preemptive_rate(stats_raw.preemptive_rate);
88 stats->set_accelerate_rate(stats_raw.accelerate_rate);
89 stats->set_secondary_decoded_rate(stats_raw.secondary_decoded_rate);
minyue-webrtc0c3ca752017-08-23 15:59:38 +020090 stats->set_secondary_discarded_rate(stats_raw.secondary_discarded_rate);
minyue5f026d02015-12-16 07:36:04 -080091 stats->set_clockdrift_ppm(stats_raw.clockdrift_ppm);
92 stats->set_added_zero_samples(stats_raw.added_zero_samples);
93 stats->set_mean_waiting_time_ms(stats_raw.mean_waiting_time_ms);
94 stats->set_median_waiting_time_ms(stats_raw.median_waiting_time_ms);
95 stats->set_min_waiting_time_ms(stats_raw.min_waiting_time_ms);
96 stats->set_max_waiting_time_ms(stats_raw.max_waiting_time_ms);
97}
98
99void Convert(const webrtc::RtcpStatistics& stats_raw,
100 webrtc::neteq_unittest::RtcpStatistics* stats) {
101 stats->set_fraction_lost(stats_raw.fraction_lost);
srte186d9c32017-08-04 05:03:53 -0700102 stats->set_cumulative_lost(stats_raw.packets_lost);
minyue5f026d02015-12-16 07:36:04 -0800103 stats->set_extended_max_sequence_number(
srte186d9c32017-08-04 05:03:53 -0700104 stats_raw.extended_highest_sequence_number);
minyue5f026d02015-12-16 07:36:04 -0800105 stats->set_jitter(stats_raw.jitter);
106}
107
minyue4f906772016-04-29 11:05:14 -0700108void AddMessage(FILE* file, rtc::MessageDigest* digest,
109 const std::string& message) {
minyue5f026d02015-12-16 07:36:04 -0800110 int32_t size = message.length();
minyue4f906772016-04-29 11:05:14 -0700111 if (file)
112 ASSERT_EQ(1u, fwrite(&size, sizeof(size), 1, file));
113 digest->Update(&size, sizeof(size));
114
115 if (file)
116 ASSERT_EQ(static_cast<size_t>(size),
117 fwrite(message.data(), sizeof(char), size, file));
118 digest->Update(message.data(), sizeof(char) * size);
minyue5f026d02015-12-16 07:36:04 -0800119}
120
minyue5f026d02015-12-16 07:36:04 -0800121#endif // WEBRTC_NETEQ_UNITTEST_BITEXACT
122
henrik.lundin7a926812016-05-12 13:51:28 -0700123void LoadDecoders(webrtc::NetEq* neteq) {
kwiberg5adaf732016-10-04 09:33:27 -0700124 ASSERT_EQ(true,
125 neteq->RegisterPayloadType(0, SdpAudioFormat("pcmu", 8000, 1)));
126 // Use non-SdpAudioFormat argument when registering PCMa, so that we get test
127 // coverage for that as well.
henrik.lundin7a926812016-05-12 13:51:28 -0700128 ASSERT_EQ(0, neteq->RegisterPayloadType(webrtc::NetEqDecoder::kDecoderPCMa,
129 "pcma", 8));
130#ifdef WEBRTC_CODEC_ILBC
kwiberg5adaf732016-10-04 09:33:27 -0700131 ASSERT_EQ(true,
132 neteq->RegisterPayloadType(102, SdpAudioFormat("ilbc", 8000, 1)));
henrik.lundin7a926812016-05-12 13:51:28 -0700133#endif
134#if defined(WEBRTC_CODEC_ISAC) || defined(WEBRTC_CODEC_ISACFX)
kwiberg5adaf732016-10-04 09:33:27 -0700135 ASSERT_EQ(true,
136 neteq->RegisterPayloadType(103, SdpAudioFormat("isac", 16000, 1)));
henrik.lundin7a926812016-05-12 13:51:28 -0700137#endif
138#ifdef WEBRTC_CODEC_ISAC
kwiberg5adaf732016-10-04 09:33:27 -0700139 ASSERT_EQ(true,
140 neteq->RegisterPayloadType(104, SdpAudioFormat("isac", 32000, 1)));
henrik.lundin7a926812016-05-12 13:51:28 -0700141#endif
142#ifdef WEBRTC_CODEC_OPUS
kwiberg5adaf732016-10-04 09:33:27 -0700143 ASSERT_EQ(true,
144 neteq->RegisterPayloadType(
145 111, SdpAudioFormat("opus", 48000, 2, {{"stereo", "0"}})));
henrik.lundin7a926812016-05-12 13:51:28 -0700146#endif
kwiberg5adaf732016-10-04 09:33:27 -0700147 ASSERT_EQ(true,
148 neteq->RegisterPayloadType(93, SdpAudioFormat("L16", 8000, 1)));
149 ASSERT_EQ(true,
150 neteq->RegisterPayloadType(94, SdpAudioFormat("L16", 16000, 1)));
151 ASSERT_EQ(true,
152 neteq->RegisterPayloadType(95, SdpAudioFormat("L16", 32000, 1)));
153 ASSERT_EQ(true,
154 neteq->RegisterPayloadType(13, SdpAudioFormat("cn", 8000, 1)));
155 ASSERT_EQ(true,
156 neteq->RegisterPayloadType(98, SdpAudioFormat("cn", 16000, 1)));
henrik.lundin7a926812016-05-12 13:51:28 -0700157}
minyue5f026d02015-12-16 07:36:04 -0800158} // namespace
159
minyue4f906772016-04-29 11:05:14 -0700160class ResultSink {
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000161 public:
minyue4f906772016-04-29 11:05:14 -0700162 explicit ResultSink(const std::string& output_file);
163 ~ResultSink();
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000164
yujo36b1a5f2017-06-12 12:45:32 -0700165 template<typename T> void AddResult(const T* test_results, size_t length);
minyue4f906772016-04-29 11:05:14 -0700166
167 void AddResult(const NetEqNetworkStatistics& stats);
168 void AddResult(const RtcpStatistics& stats);
169
170 void VerifyChecksum(const std::string& ref_check_sum);
171
172 private:
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000173 FILE* output_fp_;
minyue4f906772016-04-29 11:05:14 -0700174 std::unique_ptr<rtc::MessageDigest> digest_;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000175};
176
Joachim Bauch4e909192017-12-19 22:27:51 +0100177ResultSink::ResultSink(const std::string& output_file)
minyue4f906772016-04-29 11:05:14 -0700178 : output_fp_(nullptr),
Joachim Bauch4e909192017-12-19 22:27:51 +0100179 digest_(rtc::MessageDigestFactory::Create(rtc::DIGEST_SHA_1)) {
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000180 if (!output_file.empty()) {
181 output_fp_ = fopen(output_file.c_str(), "wb");
182 EXPECT_TRUE(output_fp_ != NULL);
183 }
184}
185
minyue4f906772016-04-29 11:05:14 -0700186ResultSink::~ResultSink() {
187 if (output_fp_)
188 fclose(output_fp_);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000189}
190
yujo36b1a5f2017-06-12 12:45:32 -0700191template<typename T>
192void ResultSink::AddResult(const T* test_results, size_t length) {
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000193 if (output_fp_) {
yujo36b1a5f2017-06-12 12:45:32 -0700194 ASSERT_EQ(length, fwrite(test_results, sizeof(T), length, output_fp_));
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000195 }
yujo36b1a5f2017-06-12 12:45:32 -0700196 digest_->Update(test_results, sizeof(T) * length);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000197}
198
minyue4f906772016-04-29 11:05:14 -0700199void ResultSink::AddResult(const NetEqNetworkStatistics& stats_raw) {
minyue5f026d02015-12-16 07:36:04 -0800200#ifdef WEBRTC_NETEQ_UNITTEST_BITEXACT
minyue5f026d02015-12-16 07:36:04 -0800201 neteq_unittest::NetEqNetworkStatistics stats;
202 Convert(stats_raw, &stats);
203
mbonadei7c2c8432017-04-07 00:59:12 -0700204 ProtoString stats_string;
minyue5f026d02015-12-16 07:36:04 -0800205 ASSERT_TRUE(stats.SerializeToString(&stats_string));
minyue4f906772016-04-29 11:05:14 -0700206 AddMessage(output_fp_, digest_.get(), stats_string);
minyue5f026d02015-12-16 07:36:04 -0800207#else
208 FAIL() << "Writing to reference file requires Proto Buffer.";
209#endif // WEBRTC_NETEQ_UNITTEST_BITEXACT
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000210}
211
minyue4f906772016-04-29 11:05:14 -0700212void ResultSink::AddResult(const RtcpStatistics& stats_raw) {
minyue5f026d02015-12-16 07:36:04 -0800213#ifdef WEBRTC_NETEQ_UNITTEST_BITEXACT
minyue5f026d02015-12-16 07:36:04 -0800214 neteq_unittest::RtcpStatistics stats;
215 Convert(stats_raw, &stats);
216
mbonadei7c2c8432017-04-07 00:59:12 -0700217 ProtoString stats_string;
minyue5f026d02015-12-16 07:36:04 -0800218 ASSERT_TRUE(stats.SerializeToString(&stats_string));
minyue4f906772016-04-29 11:05:14 -0700219 AddMessage(output_fp_, digest_.get(), stats_string);
minyue5f026d02015-12-16 07:36:04 -0800220#else
221 FAIL() << "Writing to reference file requires Proto Buffer.";
222#endif // WEBRTC_NETEQ_UNITTEST_BITEXACT
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000223}
224
minyue4f906772016-04-29 11:05:14 -0700225void ResultSink::VerifyChecksum(const std::string& checksum) {
226 std::vector<char> buffer;
227 buffer.resize(digest_->Size());
228 digest_->Finish(&buffer[0], buffer.size());
229 const std::string result = rtc::hex_encode(&buffer[0], digest_->Size());
230 EXPECT_EQ(checksum, result);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000231}
232
233class NetEqDecodingTest : public ::testing::Test {
234 protected:
235 // NetEQ must be polled for data once every 10 ms. Thus, neither of the
236 // constants below can be changed.
237 static const int kTimeStepMs = 10;
Peter Kastingdce40cf2015-08-24 14:52:23 -0700238 static const size_t kBlockSize8kHz = kTimeStepMs * 8;
239 static const size_t kBlockSize16kHz = kTimeStepMs * 16;
240 static const size_t kBlockSize32kHz = kTimeStepMs * 32;
minyue93c08b72015-12-22 09:57:41 -0800241 static const size_t kBlockSize48kHz = kTimeStepMs * 48;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000242 static const int kInitSampleRateHz = 8000;
243
244 NetEqDecodingTest();
245 virtual void SetUp();
246 virtual void TearDown();
247 void SelectDecoders(NetEqDecoder* used_codec);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000248 void OpenInputFile(const std::string &rtp_file);
henrik.lundin6d8e0112016-03-04 10:34:21 -0800249 void Process();
minyue5f026d02015-12-16 07:36:04 -0800250
henrik.lundin@webrtc.org4e4b0982014-08-11 14:48:49 +0000251 void DecodeAndCompare(const std::string& rtp_file,
minyue4f906772016-04-29 11:05:14 -0700252 const std::string& output_checksum,
253 const std::string& network_stats_checksum,
254 const std::string& rtcp_stats_checksum,
255 bool gen_ref);
minyue5f026d02015-12-16 07:36:04 -0800256
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000257 static void PopulateRtpInfo(int frame_index,
258 int timestamp,
henrik.lundin246ef3e2017-04-24 09:14:32 -0700259 RTPHeader* rtp_info);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000260 static void PopulateCng(int frame_index,
261 int timestamp,
henrik.lundin246ef3e2017-04-24 09:14:32 -0700262 RTPHeader* rtp_info,
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000263 uint8_t* payload,
pkasting@chromium.org4591fbd2014-11-20 22:28:14 +0000264 size_t* payload_len);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000265
turaj@webrtc.org78b41a02013-11-22 20:27:07 +0000266 void WrapTest(uint16_t start_seq_no, uint32_t start_timestamp,
267 const std::set<uint16_t>& drop_seq_numbers,
268 bool expect_seq_no_wrap, bool expect_timestamp_wrap);
269
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000270 void LongCngWithClockDrift(double drift_factor,
271 double network_freeze_ms,
272 bool pull_audio_during_freeze,
273 int delay_tolerance_ms,
274 int max_time_to_speech_ms);
275
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +0000276 void DuplicateCng();
henrik.lundin@webrtc.orgfcfc6a92014-02-13 11:42:28 +0000277
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000278 NetEq* neteq_;
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +0000279 NetEq::Config config_;
kwiberg2d0c3322016-02-14 09:28:33 -0800280 std::unique_ptr<test::RtpFileSource> rtp_source_;
281 std::unique_ptr<test::Packet> packet_;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000282 unsigned int sim_clock_;
henrik.lundin6d8e0112016-03-04 10:34:21 -0800283 AudioFrame out_frame_;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000284 int output_sample_rate_;
turaj@webrtc.org8d1cdaa2014-04-11 18:47:55 +0000285 int algorithmic_delay_ms_;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000286};
287
288// Allocating the static const so that it can be passed by reference.
289const int NetEqDecodingTest::kTimeStepMs;
Peter Kastingdce40cf2015-08-24 14:52:23 -0700290const size_t NetEqDecodingTest::kBlockSize8kHz;
291const size_t NetEqDecodingTest::kBlockSize16kHz;
292const size_t NetEqDecodingTest::kBlockSize32kHz;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000293const int NetEqDecodingTest::kInitSampleRateHz;
294
295NetEqDecodingTest::NetEqDecodingTest()
296 : neteq_(NULL),
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +0000297 config_(),
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000298 sim_clock_(0),
turaj@webrtc.org8d1cdaa2014-04-11 18:47:55 +0000299 output_sample_rate_(kInitSampleRateHz),
300 algorithmic_delay_ms_(0) {
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +0000301 config_.sample_rate_hz = kInitSampleRateHz;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000302}
303
304void NetEqDecodingTest::SetUp() {
ossue3525782016-05-25 07:37:43 -0700305 neteq_ = NetEq::Create(config_, CreateBuiltinAudioDecoderFactory());
turaj@webrtc.org8d1cdaa2014-04-11 18:47:55 +0000306 NetEqNetworkStatistics stat;
307 ASSERT_EQ(0, neteq_->NetworkStatistics(&stat));
308 algorithmic_delay_ms_ = stat.current_buffer_size_ms;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000309 ASSERT_TRUE(neteq_);
henrik.lundin7a926812016-05-12 13:51:28 -0700310 LoadDecoders(neteq_);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000311}
312
313void NetEqDecodingTest::TearDown() {
314 delete neteq_;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000315}
316
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000317void NetEqDecodingTest::OpenInputFile(const std::string &rtp_file) {
henrik.lundin@webrtc.org966a7082014-11-17 09:08:38 +0000318 rtp_source_.reset(test::RtpFileSource::Create(rtp_file));
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000319}
320
henrik.lundin6d8e0112016-03-04 10:34:21 -0800321void NetEqDecodingTest::Process() {
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000322 // Check if time to receive.
henrik.lundin@webrtc.org966a7082014-11-17 09:08:38 +0000323 while (packet_ && sim_clock_ >= packet_->time_ms()) {
324 if (packet_->payload_length_bytes() > 0) {
ivoc72c08ed2016-01-20 07:26:24 -0800325#ifndef WEBRTC_CODEC_ISAC
326 // Ignore payload type 104 (iSAC-swb) if ISAC is not supported.
henrik.lundin246ef3e2017-04-24 09:14:32 -0700327 if (packet_->header().payloadType != 104)
ivoc72c08ed2016-01-20 07:26:24 -0800328#endif
Henrik Lundin70c09bd2017-04-24 15:56:56 +0200329 ASSERT_EQ(0,
330 neteq_->InsertPacket(
henrik.lundin246ef3e2017-04-24 09:14:32 -0700331 packet_->header(),
Henrik Lundin70c09bd2017-04-24 15:56:56 +0200332 rtc::ArrayView<const uint8_t>(
333 packet_->payload(), packet_->payload_length_bytes()),
334 static_cast<uint32_t>(packet_->time_ms() *
335 (output_sample_rate_ / 1000))));
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000336 }
337 // Get next packet.
henrik.lundin46ba49c2016-05-24 22:50:47 -0700338 packet_ = rtp_source_->NextPacket();
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000339 }
340
henrik.lundin@webrtc.orge1d468c2013-01-30 07:37:20 +0000341 // Get audio from NetEq.
henrik.lundin7a926812016-05-12 13:51:28 -0700342 bool muted;
343 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
344 ASSERT_FALSE(muted);
henrik.lundin6d8e0112016-03-04 10:34:21 -0800345 ASSERT_TRUE((out_frame_.samples_per_channel_ == kBlockSize8kHz) ||
346 (out_frame_.samples_per_channel_ == kBlockSize16kHz) ||
347 (out_frame_.samples_per_channel_ == kBlockSize32kHz) ||
348 (out_frame_.samples_per_channel_ == kBlockSize48kHz));
349 output_sample_rate_ = out_frame_.sample_rate_hz_;
henrik.lundind89814b2015-11-23 06:49:25 -0800350 EXPECT_EQ(output_sample_rate_, neteq_->last_output_sample_rate_hz());
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000351
352 // Increase time.
353 sim_clock_ += kTimeStepMs;
354}
355
minyue4f906772016-04-29 11:05:14 -0700356void NetEqDecodingTest::DecodeAndCompare(
357 const std::string& rtp_file,
358 const std::string& output_checksum,
359 const std::string& network_stats_checksum,
360 const std::string& rtcp_stats_checksum,
361 bool gen_ref) {
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000362 OpenInputFile(rtp_file);
363
minyue4f906772016-04-29 11:05:14 -0700364 std::string ref_out_file =
365 gen_ref ? webrtc::test::OutputPath() + "neteq_universal_ref.pcm" : "";
366 ResultSink output(ref_out_file);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000367
minyue4f906772016-04-29 11:05:14 -0700368 std::string stat_out_file =
369 gen_ref ? webrtc::test::OutputPath() + "neteq_network_stats.dat" : "";
370 ResultSink network_stats(stat_out_file);
henrik.lundin@webrtc.org4e4b0982014-08-11 14:48:49 +0000371
minyue4f906772016-04-29 11:05:14 -0700372 std::string rtcp_out_file =
373 gen_ref ? webrtc::test::OutputPath() + "neteq_rtcp_stats.dat" : "";
374 ResultSink rtcp_stats(rtcp_out_file);
henrik.lundin@webrtc.org4e4b0982014-08-11 14:48:49 +0000375
henrik.lundin46ba49c2016-05-24 22:50:47 -0700376 packet_ = rtp_source_->NextPacket();
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000377 int i = 0;
Henrik Lundinac0a5032017-09-25 12:22:46 +0200378 uint64_t last_concealed_samples = 0;
379 uint64_t last_total_samples_received = 0;
henrik.lundin@webrtc.org966a7082014-11-17 09:08:38 +0000380 while (packet_) {
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000381 std::ostringstream ss;
382 ss << "Lap number " << i++ << " in DecodeAndCompare while loop";
383 SCOPED_TRACE(ss.str()); // Print out the parameter values on failure.
henrik.lundin6d8e0112016-03-04 10:34:21 -0800384 ASSERT_NO_FATAL_FAILURE(Process());
minyue4f906772016-04-29 11:05:14 -0700385 ASSERT_NO_FATAL_FAILURE(output.AddResult(
yujo36b1a5f2017-06-12 12:45:32 -0700386 out_frame_.data(), out_frame_.samples_per_channel_));
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000387
388 // Query the network statistics API once per second
389 if (sim_clock_ % 1000 == 0) {
390 // Process NetworkStatistics.
minyue4f906772016-04-29 11:05:14 -0700391 NetEqNetworkStatistics current_network_stats;
392 ASSERT_EQ(0, neteq_->NetworkStatistics(&current_network_stats));
393 ASSERT_NO_FATAL_FAILURE(network_stats.AddResult(current_network_stats));
394
henrik.lundin9c3efd02015-08-27 13:12:22 -0700395 // Compare with CurrentDelay, which should be identical.
minyue4f906772016-04-29 11:05:14 -0700396 EXPECT_EQ(current_network_stats.current_buffer_size_ms,
397 neteq_->CurrentDelayMs());
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000398
Henrik Lundinac0a5032017-09-25 12:22:46 +0200399 // Verify that liftime stats and network stats report similar loss
400 // concealment rates.
401 auto lifetime_stats = neteq_->GetLifetimeStatistics();
402 const uint64_t delta_concealed_samples =
403 lifetime_stats.concealed_samples - last_concealed_samples;
404 last_concealed_samples = lifetime_stats.concealed_samples;
405 const uint64_t delta_total_samples_received =
406 lifetime_stats.total_samples_received - last_total_samples_received;
407 last_total_samples_received = lifetime_stats.total_samples_received;
408 // The tolerance is 1% but expressed in Q14.
409 EXPECT_NEAR(
410 (delta_concealed_samples << 14) / delta_total_samples_received,
411 current_network_stats.expand_rate, (2 << 14) / 100.0);
412
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000413 // Process RTCPstat.
minyue4f906772016-04-29 11:05:14 -0700414 RtcpStatistics current_rtcp_stats;
415 neteq_->GetRtcpStatistics(&current_rtcp_stats);
416 ASSERT_NO_FATAL_FAILURE(rtcp_stats.AddResult(current_rtcp_stats));
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000417 }
418 }
minyue4f906772016-04-29 11:05:14 -0700419
420 SCOPED_TRACE("Check output audio.");
421 output.VerifyChecksum(output_checksum);
422 SCOPED_TRACE("Check network stats.");
423 network_stats.VerifyChecksum(network_stats_checksum);
424 SCOPED_TRACE("Check rtcp stats.");
425 rtcp_stats.VerifyChecksum(rtcp_stats_checksum);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000426}
427
428void NetEqDecodingTest::PopulateRtpInfo(int frame_index,
429 int timestamp,
henrik.lundin246ef3e2017-04-24 09:14:32 -0700430 RTPHeader* rtp_info) {
431 rtp_info->sequenceNumber = frame_index;
432 rtp_info->timestamp = timestamp;
433 rtp_info->ssrc = 0x1234; // Just an arbitrary SSRC.
434 rtp_info->payloadType = 94; // PCM16b WB codec.
435 rtp_info->markerBit = 0;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000436}
437
438void NetEqDecodingTest::PopulateCng(int frame_index,
439 int timestamp,
henrik.lundin246ef3e2017-04-24 09:14:32 -0700440 RTPHeader* rtp_info,
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000441 uint8_t* payload,
pkasting@chromium.org4591fbd2014-11-20 22:28:14 +0000442 size_t* payload_len) {
henrik.lundin246ef3e2017-04-24 09:14:32 -0700443 rtp_info->sequenceNumber = frame_index;
444 rtp_info->timestamp = timestamp;
445 rtp_info->ssrc = 0x1234; // Just an arbitrary SSRC.
446 rtp_info->payloadType = 98; // WB CNG.
447 rtp_info->markerBit = 0;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000448 payload[0] = 64; // Noise level -64 dBov, quite arbitrarily chosen.
449 *payload_len = 1; // Only noise level, no spectral parameters.
450}
451
ivoc72c08ed2016-01-20 07:26:24 -0800452#if !defined(WEBRTC_IOS) && defined(WEBRTC_NETEQ_UNITTEST_BITEXACT) && \
453 (defined(WEBRTC_CODEC_ISAC) || defined(WEBRTC_CODEC_ISACFX)) && \
Karl Wibergeb254b42017-11-01 15:08:12 +0100454 defined(WEBRTC_CODEC_ILBC) && !defined(WEBRTC_ARCH_ARM64)
minyue5f026d02015-12-16 07:36:04 -0800455#define MAYBE_TestBitExactness TestBitExactness
kwiberg98ab3a42015-09-30 21:54:21 -0700456#else
minyue5f026d02015-12-16 07:36:04 -0800457#define MAYBE_TestBitExactness DISABLED_TestBitExactness
kwiberg98ab3a42015-09-30 21:54:21 -0700458#endif
minyue5f026d02015-12-16 07:36:04 -0800459TEST_F(NetEqDecodingTest, MAYBE_TestBitExactness) {
minyue49c454e2016-01-08 11:30:14 -0800460 const std::string input_rtp_file =
461 webrtc::test::ResourcePath("audio_coding/neteq_universal_new", "rtp");
henrik.lundin@webrtc.org4e4b0982014-08-11 14:48:49 +0000462
minyue4f906772016-04-29 11:05:14 -0700463 const std::string output_checksum = PlatformChecksum(
soren9f2c18e2017-04-10 02:22:46 -0700464 "09fa7646e2ad032a0b156177b95f09012430f81f",
465 "1c64eb8b55ce8878676c6a1e6ddd78f48de0668b",
Henrik Lundin8cd750d2017-10-12 13:07:11 +0200466 "not used",
soren9f2c18e2017-04-10 02:22:46 -0700467 "09fa7646e2ad032a0b156177b95f09012430f81f",
468 "759fef89a5de52bd17e733dc255c671ce86be909");
minyue4f906772016-04-29 11:05:14 -0700469
henrik.lundin2979f552017-05-05 05:04:16 -0700470 const std::string network_stats_checksum =
minyue-webrtc0c3ca752017-08-23 15:59:38 +0200471 PlatformChecksum("5b4262ca328e5f066af5d34f3380521583dd20de",
472 "80235b6d727281203acb63b98f9a9e85d95f7ec0",
Henrik Lundin8cd750d2017-10-12 13:07:11 +0200473 "not used",
minyue-webrtc0c3ca752017-08-23 15:59:38 +0200474 "5b4262ca328e5f066af5d34f3380521583dd20de",
475 "5b4262ca328e5f066af5d34f3380521583dd20de");
minyue4f906772016-04-29 11:05:14 -0700476
477 const std::string rtcp_stats_checksum = PlatformChecksum(
478 "b8880bf9fed2487efbddcb8d94b9937a29ae521d",
479 "f3f7b3d3e71d7e635240b5373b57df6a7e4ce9d4",
Henrik Lundin8cd750d2017-10-12 13:07:11 +0200480 "not used",
minyue4f906772016-04-29 11:05:14 -0700481 "b8880bf9fed2487efbddcb8d94b9937a29ae521d",
482 "b8880bf9fed2487efbddcb8d94b9937a29ae521d");
483
484 DecodeAndCompare(input_rtp_file,
485 output_checksum,
486 network_stats_checksum,
487 rtcp_stats_checksum,
oprypin9b2f20c2017-08-29 05:51:57 -0700488 FLAG_gen_ref);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000489}
490
Henrik Lundin8cd750d2017-10-12 13:07:11 +0200491#if !defined(WEBRTC_IOS) && \
minyue93c08b72015-12-22 09:57:41 -0800492 defined(WEBRTC_NETEQ_UNITTEST_BITEXACT) && \
minyue-webrtc516711c2017-07-27 17:45:49 +0200493 defined(WEBRTC_CODEC_OPUS)
minyue93c08b72015-12-22 09:57:41 -0800494#define MAYBE_TestOpusBitExactness TestOpusBitExactness
495#else
496#define MAYBE_TestOpusBitExactness DISABLED_TestOpusBitExactness
497#endif
minyue-webrtcadb58b82017-07-26 17:59:59 +0200498TEST_F(NetEqDecodingTest, MAYBE_TestOpusBitExactness) {
minyue93c08b72015-12-22 09:57:41 -0800499 const std::string input_rtp_file =
500 webrtc::test::ResourcePath("audio_coding/neteq_opus", "rtp");
minyue93c08b72015-12-22 09:57:41 -0800501
minyue4f906772016-04-29 11:05:14 -0700502 const std::string output_checksum = PlatformChecksum(
Ivo Creusenfd7c0a52017-10-20 12:35:04 +0200503 "7ea28d7edf9395f4ac8e8d8dd3a9e5c620b1bf48",
504 "5b1e691ab1c4465c742d6d944bc71e3b1c0e4c0e",
505 "b096114dd8c233eaf2b0ce9802ac95af13933772",
506 "7ea28d7edf9395f4ac8e8d8dd3a9e5c620b1bf48",
507 "7ea28d7edf9395f4ac8e8d8dd3a9e5c620b1bf48");
minyue4f906772016-04-29 11:05:14 -0700508
henrik.lundin2979f552017-05-05 05:04:16 -0700509 const std::string network_stats_checksum =
Ivo Creusenfd7c0a52017-10-20 12:35:04 +0200510 PlatformChecksum("9e72233c78baf685e500dd6c94212b30a4c5f27d",
511 "9a37270e4242fbd31e80bb47dc5e7ab82cf2d557",
512 "4f1e9734bc80a290faaf9d611efcb8d7802dbc4f",
513 "9e72233c78baf685e500dd6c94212b30a4c5f27d",
514 "9e72233c78baf685e500dd6c94212b30a4c5f27d");
minyue4f906772016-04-29 11:05:14 -0700515
516 const std::string rtcp_stats_checksum = PlatformChecksum(
517 "e37c797e3de6a64dda88c9ade7a013d022a2e1e0",
518 "e37c797e3de6a64dda88c9ade7a013d022a2e1e0",
519 "e37c797e3de6a64dda88c9ade7a013d022a2e1e0",
Henrik Lundin8cd750d2017-10-12 13:07:11 +0200520 "e37c797e3de6a64dda88c9ade7a013d022a2e1e0",
minyue4f906772016-04-29 11:05:14 -0700521 "e37c797e3de6a64dda88c9ade7a013d022a2e1e0");
522
523 DecodeAndCompare(input_rtp_file,
524 output_checksum,
525 network_stats_checksum,
526 rtcp_stats_checksum,
oprypin9b2f20c2017-08-29 05:51:57 -0700527 FLAG_gen_ref);
minyue93c08b72015-12-22 09:57:41 -0800528}
529
Henrik Lundine9619f82017-11-27 14:05:27 +0100530// This test fixture is identical to NetEqDecodingTest, except that it enables
531// the WebRTC-NetEqOpusDtxDelayFix field trial.
532// TODO(bugs.webrtc.org/8488): When the field trial is over and the feature is
533// default enabled, remove this fixture class and let the
534// TestOpusDtxBitExactness test build directly on NetEqDecodingTest.
535class NetEqDecodingTestWithOpusDtxFieldTrial : public NetEqDecodingTest {
536 public:
537 NetEqDecodingTestWithOpusDtxFieldTrial()
538 : override_field_trials_("WebRTC-NetEqOpusDtxDelayFix/Enabled/") {}
539
540 private:
541 test::ScopedFieldTrials override_field_trials_;
542};
543
544#if !defined(WEBRTC_IOS) && \
545 defined(WEBRTC_NETEQ_UNITTEST_BITEXACT) && \
546 defined(WEBRTC_CODEC_OPUS)
547#define MAYBE_TestOpusDtxBitExactness TestOpusDtxBitExactness
548#else
549#define MAYBE_TestOpusDtxBitExactness DISABLED_TestOpusDtxBitExactness
550#endif
551TEST_F(NetEqDecodingTestWithOpusDtxFieldTrial, MAYBE_TestOpusDtxBitExactness) {
552 const std::string input_rtp_file =
553 webrtc::test::ResourcePath("audio_coding/neteq_opus_dtx", "rtp");
554
555 const std::string output_checksum =
556 PlatformChecksum("713af6c92881f5aab1285765ee6680da9d1c06ce",
557 "3ec991b96872123f1554c03c543ca5d518431e46",
558 "da9f9a2d94e0c2d67342fad4965d7b91cda50b25",
559 "713af6c92881f5aab1285765ee6680da9d1c06ce",
560 "713af6c92881f5aab1285765ee6680da9d1c06ce");
561
562 const std::string network_stats_checksum =
563 "bab58dc587d956f326056d7340c96eb9d2d3cc21";
564
565 const std::string rtcp_stats_checksum =
566 "ac27a7f305efb58b39bf123dccee25dee5758e63";
567
568 DecodeAndCompare(input_rtp_file, output_checksum, network_stats_checksum,
569 rtcp_stats_checksum, FLAG_gen_ref);
570}
571
henrik.lundin@webrtc.org7cbc4f92014-10-07 06:37:39 +0000572// Use fax mode to avoid time-scaling. This is to simplify the testing of
573// packet waiting times in the packet buffer.
574class NetEqDecodingTestFaxMode : public NetEqDecodingTest {
575 protected:
576 NetEqDecodingTestFaxMode() : NetEqDecodingTest() {
577 config_.playout_mode = kPlayoutFax;
578 }
Gustaf Ullbergb0a02072017-10-02 12:00:34 +0200579 void TestJitterBufferDelay(bool apply_packet_loss);
henrik.lundin@webrtc.org7cbc4f92014-10-07 06:37:39 +0000580};
581
582TEST_F(NetEqDecodingTestFaxMode, TestFrameWaitingTimeStatistics) {
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000583 // Insert 30 dummy packets at once. Each packet contains 10 ms 16 kHz audio.
584 size_t num_frames = 30;
pkasting@chromium.org4591fbd2014-11-20 22:28:14 +0000585 const size_t kSamples = 10 * 16;
586 const size_t kPayloadBytes = kSamples * 2;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000587 for (size_t i = 0; i < num_frames; ++i) {
kwibergee2bac22015-11-11 10:34:00 -0800588 const uint8_t payload[kPayloadBytes] = {0};
henrik.lundin246ef3e2017-04-24 09:14:32 -0700589 RTPHeader rtp_info;
Mirko Bonadeia8110272017-10-18 14:22:50 +0200590 rtp_info.sequenceNumber = rtc::checked_cast<uint16_t>(i);
591 rtp_info.timestamp = rtc::checked_cast<uint32_t>(i * kSamples);
henrik.lundin246ef3e2017-04-24 09:14:32 -0700592 rtp_info.ssrc = 0x1234; // Just an arbitrary SSRC.
593 rtp_info.payloadType = 94; // PCM16b WB codec.
594 rtp_info.markerBit = 0;
595 ASSERT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000596 }
597 // Pull out all data.
598 for (size_t i = 0; i < num_frames; ++i) {
henrik.lundin7a926812016-05-12 13:51:28 -0700599 bool muted;
600 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -0800601 ASSERT_EQ(kBlockSize16kHz, out_frame_.samples_per_channel_);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000602 }
603
Henrik Lundin1bb8cf82015-08-25 13:08:04 +0200604 NetEqNetworkStatistics stats;
605 EXPECT_EQ(0, neteq_->NetworkStatistics(&stats));
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000606 // Since all frames are dumped into NetEQ at once, but pulled out with 10 ms
607 // spacing (per definition), we expect the delay to increase with 10 ms for
Henrik Lundin1bb8cf82015-08-25 13:08:04 +0200608 // each packet. Thus, we are calculating the statistics for a series from 10
609 // to 300, in steps of 10 ms.
610 EXPECT_EQ(155, stats.mean_waiting_time_ms);
611 EXPECT_EQ(155, stats.median_waiting_time_ms);
612 EXPECT_EQ(10, stats.min_waiting_time_ms);
613 EXPECT_EQ(300, stats.max_waiting_time_ms);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000614
615 // Check statistics again and make sure it's been reset.
Henrik Lundin1bb8cf82015-08-25 13:08:04 +0200616 EXPECT_EQ(0, neteq_->NetworkStatistics(&stats));
617 EXPECT_EQ(-1, stats.mean_waiting_time_ms);
618 EXPECT_EQ(-1, stats.median_waiting_time_ms);
619 EXPECT_EQ(-1, stats.min_waiting_time_ms);
620 EXPECT_EQ(-1, stats.max_waiting_time_ms);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000621}
622
henrik.lundin@webrtc.orgb4e80e02014-05-15 07:14:00 +0000623TEST_F(NetEqDecodingTest, TestAverageInterArrivalTimeNegative) {
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000624 const int kNumFrames = 3000; // Needed for convergence.
625 int frame_index = 0;
pkasting@chromium.org4591fbd2014-11-20 22:28:14 +0000626 const size_t kSamples = 10 * 16;
627 const size_t kPayloadBytes = kSamples * 2;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000628 while (frame_index < kNumFrames) {
629 // Insert one packet each time, except every 10th time where we insert two
630 // packets at once. This will create a negative clock-drift of approx. 10%.
631 int num_packets = (frame_index % 10 == 0 ? 2 : 1);
632 for (int n = 0; n < num_packets; ++n) {
633 uint8_t payload[kPayloadBytes] = {0};
henrik.lundin246ef3e2017-04-24 09:14:32 -0700634 RTPHeader rtp_info;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000635 PopulateRtpInfo(frame_index, frame_index * kSamples, &rtp_info);
henrik.lundin246ef3e2017-04-24 09:14:32 -0700636 ASSERT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000637 ++frame_index;
638 }
639
640 // Pull out data once.
henrik.lundin7a926812016-05-12 13:51:28 -0700641 bool muted;
642 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -0800643 ASSERT_EQ(kBlockSize16kHz, out_frame_.samples_per_channel_);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000644 }
645
646 NetEqNetworkStatistics network_stats;
647 ASSERT_EQ(0, neteq_->NetworkStatistics(&network_stats));
henrik.lundin0d838572016-10-13 03:35:55 -0700648 EXPECT_EQ(-103192, network_stats.clockdrift_ppm);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000649}
650
henrik.lundin@webrtc.orgb4e80e02014-05-15 07:14:00 +0000651TEST_F(NetEqDecodingTest, TestAverageInterArrivalTimePositive) {
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000652 const int kNumFrames = 5000; // Needed for convergence.
653 int frame_index = 0;
pkasting@chromium.org4591fbd2014-11-20 22:28:14 +0000654 const size_t kSamples = 10 * 16;
655 const size_t kPayloadBytes = kSamples * 2;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000656 for (int i = 0; i < kNumFrames; ++i) {
657 // Insert one packet each time, except every 10th time where we don't insert
658 // any packet. This will create a positive clock-drift of approx. 11%.
659 int num_packets = (i % 10 == 9 ? 0 : 1);
660 for (int n = 0; n < num_packets; ++n) {
661 uint8_t payload[kPayloadBytes] = {0};
henrik.lundin246ef3e2017-04-24 09:14:32 -0700662 RTPHeader rtp_info;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000663 PopulateRtpInfo(frame_index, frame_index * kSamples, &rtp_info);
henrik.lundin246ef3e2017-04-24 09:14:32 -0700664 ASSERT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000665 ++frame_index;
666 }
667
668 // Pull out data once.
henrik.lundin7a926812016-05-12 13:51:28 -0700669 bool muted;
670 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -0800671 ASSERT_EQ(kBlockSize16kHz, out_frame_.samples_per_channel_);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000672 }
673
674 NetEqNetworkStatistics network_stats;
675 ASSERT_EQ(0, neteq_->NetworkStatistics(&network_stats));
henrik.lundin0d838572016-10-13 03:35:55 -0700676 EXPECT_EQ(110953, network_stats.clockdrift_ppm);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000677}
678
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000679void NetEqDecodingTest::LongCngWithClockDrift(double drift_factor,
680 double network_freeze_ms,
681 bool pull_audio_during_freeze,
682 int delay_tolerance_ms,
683 int max_time_to_speech_ms) {
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000684 uint16_t seq_no = 0;
685 uint32_t timestamp = 0;
686 const int kFrameSizeMs = 30;
pkasting@chromium.org4591fbd2014-11-20 22:28:14 +0000687 const size_t kSamples = kFrameSizeMs * 16;
688 const size_t kPayloadBytes = kSamples * 2;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000689 double next_input_time_ms = 0.0;
690 double t_ms;
henrik.lundin7a926812016-05-12 13:51:28 -0700691 bool muted;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000692
693 // Insert speech for 5 seconds.
694 const int kSpeechDurationMs = 5000;
695 for (t_ms = 0; t_ms < kSpeechDurationMs; t_ms += 10) {
696 // Each turn in this for loop is 10 ms.
697 while (next_input_time_ms <= t_ms) {
698 // Insert one 30 ms speech frame.
699 uint8_t payload[kPayloadBytes] = {0};
henrik.lundin246ef3e2017-04-24 09:14:32 -0700700 RTPHeader rtp_info;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000701 PopulateRtpInfo(seq_no, timestamp, &rtp_info);
henrik.lundin246ef3e2017-04-24 09:14:32 -0700702 ASSERT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000703 ++seq_no;
704 timestamp += kSamples;
henrik.lundin@webrtc.orgfcfc6a92014-02-13 11:42:28 +0000705 next_input_time_ms += static_cast<double>(kFrameSizeMs) * drift_factor;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000706 }
707 // Pull out data once.
henrik.lundin7a926812016-05-12 13:51:28 -0700708 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -0800709 ASSERT_EQ(kBlockSize16kHz, out_frame_.samples_per_channel_);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000710 }
711
henrik.lundin55480f52016-03-08 02:37:57 -0800712 EXPECT_EQ(AudioFrame::kNormalSpeech, out_frame_.speech_type_);
henrik.lundin114c1b32017-04-26 07:47:32 -0700713 rtc::Optional<uint32_t> playout_timestamp = neteq_->GetPlayoutTimestamp();
henrik.lundin0d96ab72016-04-06 12:28:26 -0700714 ASSERT_TRUE(playout_timestamp);
715 int32_t delay_before = timestamp - *playout_timestamp;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000716
717 // Insert CNG for 1 minute (= 60000 ms).
718 const int kCngPeriodMs = 100;
719 const int kCngPeriodSamples = kCngPeriodMs * 16; // Period in 16 kHz samples.
720 const int kCngDurationMs = 60000;
721 for (; t_ms < kSpeechDurationMs + kCngDurationMs; t_ms += 10) {
722 // Each turn in this for loop is 10 ms.
723 while (next_input_time_ms <= t_ms) {
724 // Insert one CNG frame each 100 ms.
725 uint8_t payload[kPayloadBytes];
pkasting@chromium.org4591fbd2014-11-20 22:28:14 +0000726 size_t payload_len;
henrik.lundin246ef3e2017-04-24 09:14:32 -0700727 RTPHeader rtp_info;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000728 PopulateCng(seq_no, timestamp, &rtp_info, payload, &payload_len);
kwibergee2bac22015-11-11 10:34:00 -0800729 ASSERT_EQ(0, neteq_->InsertPacket(
henrik.lundin246ef3e2017-04-24 09:14:32 -0700730 rtp_info,
kwibergee2bac22015-11-11 10:34:00 -0800731 rtc::ArrayView<const uint8_t>(payload, payload_len), 0));
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000732 ++seq_no;
733 timestamp += kCngPeriodSamples;
henrik.lundin@webrtc.orgfcfc6a92014-02-13 11:42:28 +0000734 next_input_time_ms += static_cast<double>(kCngPeriodMs) * drift_factor;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000735 }
736 // Pull out data once.
henrik.lundin7a926812016-05-12 13:51:28 -0700737 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -0800738 ASSERT_EQ(kBlockSize16kHz, out_frame_.samples_per_channel_);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000739 }
740
henrik.lundin55480f52016-03-08 02:37:57 -0800741 EXPECT_EQ(AudioFrame::kCNG, out_frame_.speech_type_);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000742
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000743 if (network_freeze_ms > 0) {
744 // First keep pulling audio for |network_freeze_ms| without inserting
745 // any data, then insert CNG data corresponding to |network_freeze_ms|
746 // without pulling any output audio.
747 const double loop_end_time = t_ms + network_freeze_ms;
748 for (; t_ms < loop_end_time; t_ms += 10) {
749 // Pull out data once.
henrik.lundin7a926812016-05-12 13:51:28 -0700750 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -0800751 ASSERT_EQ(kBlockSize16kHz, out_frame_.samples_per_channel_);
henrik.lundin55480f52016-03-08 02:37:57 -0800752 EXPECT_EQ(AudioFrame::kCNG, out_frame_.speech_type_);
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000753 }
754 bool pull_once = pull_audio_during_freeze;
755 // If |pull_once| is true, GetAudio will be called once half-way through
756 // the network recovery period.
757 double pull_time_ms = (t_ms + next_input_time_ms) / 2;
758 while (next_input_time_ms <= t_ms) {
759 if (pull_once && next_input_time_ms >= pull_time_ms) {
760 pull_once = false;
761 // Pull out data once.
henrik.lundin7a926812016-05-12 13:51:28 -0700762 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -0800763 ASSERT_EQ(kBlockSize16kHz, out_frame_.samples_per_channel_);
henrik.lundin55480f52016-03-08 02:37:57 -0800764 EXPECT_EQ(AudioFrame::kCNG, out_frame_.speech_type_);
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000765 t_ms += 10;
766 }
767 // Insert one CNG frame each 100 ms.
768 uint8_t payload[kPayloadBytes];
pkasting@chromium.org4591fbd2014-11-20 22:28:14 +0000769 size_t payload_len;
henrik.lundin246ef3e2017-04-24 09:14:32 -0700770 RTPHeader rtp_info;
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000771 PopulateCng(seq_no, timestamp, &rtp_info, payload, &payload_len);
kwibergee2bac22015-11-11 10:34:00 -0800772 ASSERT_EQ(0, neteq_->InsertPacket(
henrik.lundin246ef3e2017-04-24 09:14:32 -0700773 rtp_info,
kwibergee2bac22015-11-11 10:34:00 -0800774 rtc::ArrayView<const uint8_t>(payload, payload_len), 0));
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000775 ++seq_no;
776 timestamp += kCngPeriodSamples;
777 next_input_time_ms += kCngPeriodMs * drift_factor;
778 }
779 }
780
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000781 // Insert speech again until output type is speech.
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000782 double speech_restart_time_ms = t_ms;
henrik.lundin55480f52016-03-08 02:37:57 -0800783 while (out_frame_.speech_type_ != AudioFrame::kNormalSpeech) {
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000784 // Each turn in this for loop is 10 ms.
785 while (next_input_time_ms <= t_ms) {
786 // Insert one 30 ms speech frame.
787 uint8_t payload[kPayloadBytes] = {0};
henrik.lundin246ef3e2017-04-24 09:14:32 -0700788 RTPHeader rtp_info;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000789 PopulateRtpInfo(seq_no, timestamp, &rtp_info);
henrik.lundin246ef3e2017-04-24 09:14:32 -0700790 ASSERT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000791 ++seq_no;
792 timestamp += kSamples;
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000793 next_input_time_ms += kFrameSizeMs * drift_factor;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000794 }
795 // Pull out data once.
henrik.lundin7a926812016-05-12 13:51:28 -0700796 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -0800797 ASSERT_EQ(kBlockSize16kHz, out_frame_.samples_per_channel_);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000798 // Increase clock.
799 t_ms += 10;
800 }
801
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000802 // Check that the speech starts again within reasonable time.
803 double time_until_speech_returns_ms = t_ms - speech_restart_time_ms;
804 EXPECT_LT(time_until_speech_returns_ms, max_time_to_speech_ms);
henrik.lundin114c1b32017-04-26 07:47:32 -0700805 playout_timestamp = neteq_->GetPlayoutTimestamp();
henrik.lundin0d96ab72016-04-06 12:28:26 -0700806 ASSERT_TRUE(playout_timestamp);
807 int32_t delay_after = timestamp - *playout_timestamp;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000808 // Compare delay before and after, and make sure it differs less than 20 ms.
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000809 EXPECT_LE(delay_after, delay_before + delay_tolerance_ms * 16);
810 EXPECT_GE(delay_after, delay_before - delay_tolerance_ms * 16);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000811}
812
henrik.lundin@webrtc.orgb4e80e02014-05-15 07:14:00 +0000813TEST_F(NetEqDecodingTest, LongCngWithNegativeClockDrift) {
henrik.lundin@webrtc.orgfcfc6a92014-02-13 11:42:28 +0000814 // Apply a clock drift of -25 ms / s (sender faster than receiver).
815 const double kDriftFactor = 1000.0 / (1000.0 + 25.0);
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000816 const double kNetworkFreezeTimeMs = 0.0;
817 const bool kGetAudioDuringFreezeRecovery = false;
818 const int kDelayToleranceMs = 20;
819 const int kMaxTimeToSpeechMs = 100;
820 LongCngWithClockDrift(kDriftFactor,
821 kNetworkFreezeTimeMs,
822 kGetAudioDuringFreezeRecovery,
823 kDelayToleranceMs,
824 kMaxTimeToSpeechMs);
henrik.lundin@webrtc.orgfcfc6a92014-02-13 11:42:28 +0000825}
826
henrik.lundin@webrtc.orgb4e80e02014-05-15 07:14:00 +0000827TEST_F(NetEqDecodingTest, LongCngWithPositiveClockDrift) {
henrik.lundin@webrtc.orgfcfc6a92014-02-13 11:42:28 +0000828 // Apply a clock drift of +25 ms / s (sender slower than receiver).
829 const double kDriftFactor = 1000.0 / (1000.0 - 25.0);
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000830 const double kNetworkFreezeTimeMs = 0.0;
831 const bool kGetAudioDuringFreezeRecovery = false;
832 const int kDelayToleranceMs = 20;
833 const int kMaxTimeToSpeechMs = 100;
834 LongCngWithClockDrift(kDriftFactor,
835 kNetworkFreezeTimeMs,
836 kGetAudioDuringFreezeRecovery,
837 kDelayToleranceMs,
838 kMaxTimeToSpeechMs);
839}
840
henrik.lundin@webrtc.orgb4e80e02014-05-15 07:14:00 +0000841TEST_F(NetEqDecodingTest, LongCngWithNegativeClockDriftNetworkFreeze) {
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000842 // Apply a clock drift of -25 ms / s (sender faster than receiver).
843 const double kDriftFactor = 1000.0 / (1000.0 + 25.0);
844 const double kNetworkFreezeTimeMs = 5000.0;
845 const bool kGetAudioDuringFreezeRecovery = false;
846 const int kDelayToleranceMs = 50;
847 const int kMaxTimeToSpeechMs = 200;
848 LongCngWithClockDrift(kDriftFactor,
849 kNetworkFreezeTimeMs,
850 kGetAudioDuringFreezeRecovery,
851 kDelayToleranceMs,
852 kMaxTimeToSpeechMs);
853}
854
henrik.lundin@webrtc.orgb4e80e02014-05-15 07:14:00 +0000855TEST_F(NetEqDecodingTest, LongCngWithPositiveClockDriftNetworkFreeze) {
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000856 // Apply a clock drift of +25 ms / s (sender slower than receiver).
857 const double kDriftFactor = 1000.0 / (1000.0 - 25.0);
858 const double kNetworkFreezeTimeMs = 5000.0;
859 const bool kGetAudioDuringFreezeRecovery = false;
860 const int kDelayToleranceMs = 20;
861 const int kMaxTimeToSpeechMs = 100;
862 LongCngWithClockDrift(kDriftFactor,
863 kNetworkFreezeTimeMs,
864 kGetAudioDuringFreezeRecovery,
865 kDelayToleranceMs,
866 kMaxTimeToSpeechMs);
867}
868
henrik.lundin@webrtc.orgb4e80e02014-05-15 07:14:00 +0000869TEST_F(NetEqDecodingTest, LongCngWithPositiveClockDriftNetworkFreezeExtraPull) {
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000870 // Apply a clock drift of +25 ms / s (sender slower than receiver).
871 const double kDriftFactor = 1000.0 / (1000.0 - 25.0);
872 const double kNetworkFreezeTimeMs = 5000.0;
873 const bool kGetAudioDuringFreezeRecovery = true;
874 const int kDelayToleranceMs = 20;
875 const int kMaxTimeToSpeechMs = 100;
876 LongCngWithClockDrift(kDriftFactor,
877 kNetworkFreezeTimeMs,
878 kGetAudioDuringFreezeRecovery,
879 kDelayToleranceMs,
880 kMaxTimeToSpeechMs);
881}
882
henrik.lundin@webrtc.orgb4e80e02014-05-15 07:14:00 +0000883TEST_F(NetEqDecodingTest, LongCngWithoutClockDrift) {
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000884 const double kDriftFactor = 1.0; // No drift.
885 const double kNetworkFreezeTimeMs = 0.0;
886 const bool kGetAudioDuringFreezeRecovery = false;
887 const int kDelayToleranceMs = 10;
888 const int kMaxTimeToSpeechMs = 50;
889 LongCngWithClockDrift(kDriftFactor,
890 kNetworkFreezeTimeMs,
891 kGetAudioDuringFreezeRecovery,
892 kDelayToleranceMs,
893 kMaxTimeToSpeechMs);
henrik.lundin@webrtc.orgfcfc6a92014-02-13 11:42:28 +0000894}
895
henrik.lundin@webrtc.orgb4e80e02014-05-15 07:14:00 +0000896TEST_F(NetEqDecodingTest, UnknownPayloadType) {
pkasting@chromium.org4591fbd2014-11-20 22:28:14 +0000897 const size_t kPayloadBytes = 100;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000898 uint8_t payload[kPayloadBytes] = {0};
henrik.lundin246ef3e2017-04-24 09:14:32 -0700899 RTPHeader rtp_info;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000900 PopulateRtpInfo(0, 0, &rtp_info);
henrik.lundin246ef3e2017-04-24 09:14:32 -0700901 rtp_info.payloadType = 1; // Not registered as a decoder.
902 EXPECT_EQ(NetEq::kFail, neteq_->InsertPacket(rtp_info, payload, 0));
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000903}
904
Peter Boströme2976c82016-01-04 22:44:05 +0100905#if defined(WEBRTC_CODEC_ISAC) || defined(WEBRTC_CODEC_ISACFX)
ivoc72c08ed2016-01-20 07:26:24 -0800906#define MAYBE_DecoderError DecoderError
907#else
908#define MAYBE_DecoderError DISABLED_DecoderError
909#endif
910
Peter Boströme2976c82016-01-04 22:44:05 +0100911TEST_F(NetEqDecodingTest, MAYBE_DecoderError) {
pkasting@chromium.org4591fbd2014-11-20 22:28:14 +0000912 const size_t kPayloadBytes = 100;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000913 uint8_t payload[kPayloadBytes] = {0};
henrik.lundin246ef3e2017-04-24 09:14:32 -0700914 RTPHeader rtp_info;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000915 PopulateRtpInfo(0, 0, &rtp_info);
henrik.lundin246ef3e2017-04-24 09:14:32 -0700916 rtp_info.payloadType = 103; // iSAC, but the payload is invalid.
917 EXPECT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000918 // Set all of |out_data_| to 1, and verify that it was set to 0 by the call
919 // to GetAudio.
yujo36b1a5f2017-06-12 12:45:32 -0700920 int16_t* out_frame_data = out_frame_.mutable_data();
henrik.lundin6d8e0112016-03-04 10:34:21 -0800921 for (size_t i = 0; i < AudioFrame::kMaxDataSizeSamples; ++i) {
yujo36b1a5f2017-06-12 12:45:32 -0700922 out_frame_data[i] = 1;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000923 }
henrik.lundin7a926812016-05-12 13:51:28 -0700924 bool muted;
925 EXPECT_EQ(NetEq::kFail, neteq_->GetAudio(&out_frame_, &muted));
926 ASSERT_FALSE(muted);
ivoc72c08ed2016-01-20 07:26:24 -0800927
yujo36b1a5f2017-06-12 12:45:32 -0700928 // Verify that the first 160 samples are set to 0.
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000929 static const int kExpectedOutputLength = 160; // 10 ms at 16 kHz sample rate.
yujo36b1a5f2017-06-12 12:45:32 -0700930 const int16_t* const_out_frame_data = out_frame_.data();
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000931 for (int i = 0; i < kExpectedOutputLength; ++i) {
932 std::ostringstream ss;
933 ss << "i = " << i;
934 SCOPED_TRACE(ss.str()); // Print out the parameter values on failure.
yujo36b1a5f2017-06-12 12:45:32 -0700935 EXPECT_EQ(0, const_out_frame_data[i]);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000936 }
937}
938
henrik.lundin@webrtc.orgb4e80e02014-05-15 07:14:00 +0000939TEST_F(NetEqDecodingTest, GetAudioBeforeInsertPacket) {
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000940 // Set all of |out_data_| to 1, and verify that it was set to 0 by the call
941 // to GetAudio.
yujo36b1a5f2017-06-12 12:45:32 -0700942 int16_t* out_frame_data = out_frame_.mutable_data();
henrik.lundin6d8e0112016-03-04 10:34:21 -0800943 for (size_t i = 0; i < AudioFrame::kMaxDataSizeSamples; ++i) {
yujo36b1a5f2017-06-12 12:45:32 -0700944 out_frame_data[i] = 1;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000945 }
henrik.lundin7a926812016-05-12 13:51:28 -0700946 bool muted;
947 EXPECT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
948 ASSERT_FALSE(muted);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000949 // Verify that the first block of samples is set to 0.
950 static const int kExpectedOutputLength =
951 kInitSampleRateHz / 100; // 10 ms at initial sample rate.
yujo36b1a5f2017-06-12 12:45:32 -0700952 const int16_t* const_out_frame_data = out_frame_.data();
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000953 for (int i = 0; i < kExpectedOutputLength; ++i) {
954 std::ostringstream ss;
955 ss << "i = " << i;
956 SCOPED_TRACE(ss.str()); // Print out the parameter values on failure.
yujo36b1a5f2017-06-12 12:45:32 -0700957 EXPECT_EQ(0, const_out_frame_data[i]);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000958 }
henrik.lundind89814b2015-11-23 06:49:25 -0800959 // Verify that the sample rate did not change from the initial configuration.
960 EXPECT_EQ(config_.sample_rate_hz, neteq_->last_output_sample_rate_hz());
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000961}
turaj@webrtc.orgff43c852013-09-25 00:07:27 +0000962
henrik.lundin@webrtc.org9b8102c2014-08-21 08:27:44 +0000963class NetEqBgnTest : public NetEqDecodingTest {
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +0000964 protected:
henrik.lundin@webrtc.org9b8102c2014-08-21 08:27:44 +0000965 virtual void TestCondition(double sum_squared_noise,
966 bool should_be_faded) = 0;
turaj@webrtc.orgff43c852013-09-25 00:07:27 +0000967
henrik.lundin@webrtc.org9b8102c2014-08-21 08:27:44 +0000968 void CheckBgn(int sampling_rate_hz) {
Peter Kastingdce40cf2015-08-24 14:52:23 -0700969 size_t expected_samples_per_channel = 0;
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +0000970 uint8_t payload_type = 0xFF; // Invalid.
971 if (sampling_rate_hz == 8000) {
972 expected_samples_per_channel = kBlockSize8kHz;
973 payload_type = 93; // PCM 16, 8 kHz.
974 } else if (sampling_rate_hz == 16000) {
975 expected_samples_per_channel = kBlockSize16kHz;
976 payload_type = 94; // PCM 16, 16 kHZ.
977 } else if (sampling_rate_hz == 32000) {
978 expected_samples_per_channel = kBlockSize32kHz;
979 payload_type = 95; // PCM 16, 32 kHz.
980 } else {
981 ASSERT_TRUE(false); // Unsupported test case.
982 }
turaj@webrtc.orgff43c852013-09-25 00:07:27 +0000983
henrik.lundin6d8e0112016-03-04 10:34:21 -0800984 AudioFrame output;
henrik.lundin@webrtc.org9b8102c2014-08-21 08:27:44 +0000985 test::AudioLoop input;
986 // We are using the same 32 kHz input file for all tests, regardless of
987 // |sampling_rate_hz|. The output may sound weird, but the test is still
988 // valid.
989 ASSERT_TRUE(input.Init(
990 webrtc::test::ResourcePath("audio_coding/testfile32kHz", "pcm"),
991 10 * sampling_rate_hz, // Max 10 seconds loop length.
Peter Kastingdce40cf2015-08-24 14:52:23 -0700992 expected_samples_per_channel));
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +0000993
994 // Payload of 10 ms of PCM16 32 kHz.
995 uint8_t payload[kBlockSize32kHz * sizeof(int16_t)];
henrik.lundin246ef3e2017-04-24 09:14:32 -0700996 RTPHeader rtp_info;
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +0000997 PopulateRtpInfo(0, 0, &rtp_info);
henrik.lundin246ef3e2017-04-24 09:14:32 -0700998 rtp_info.payloadType = payload_type;
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +0000999
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +00001000 uint32_t receive_timestamp = 0;
henrik.lundin7a926812016-05-12 13:51:28 -07001001 bool muted;
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +00001002 for (int n = 0; n < 10; ++n) { // Insert few packets and get audio.
kwiberg288886b2015-11-06 01:21:35 -08001003 auto block = input.GetNextBlock();
1004 ASSERT_EQ(expected_samples_per_channel, block.size());
1005 size_t enc_len_bytes =
1006 WebRtcPcm16b_Encode(block.data(), block.size(), payload);
henrik.lundin@webrtc.org9b8102c2014-08-21 08:27:44 +00001007 ASSERT_EQ(enc_len_bytes, expected_samples_per_channel * 2);
1008
Henrik Lundin70c09bd2017-04-24 15:56:56 +02001009 ASSERT_EQ(0, neteq_->InsertPacket(
henrik.lundin246ef3e2017-04-24 09:14:32 -07001010 rtp_info,
Henrik Lundin70c09bd2017-04-24 15:56:56 +02001011 rtc::ArrayView<const uint8_t>(payload, enc_len_bytes),
1012 receive_timestamp));
henrik.lundin6d8e0112016-03-04 10:34:21 -08001013 output.Reset();
henrik.lundin7a926812016-05-12 13:51:28 -07001014 ASSERT_EQ(0, neteq_->GetAudio(&output, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -08001015 ASSERT_EQ(1u, output.num_channels_);
1016 ASSERT_EQ(expected_samples_per_channel, output.samples_per_channel_);
henrik.lundin55480f52016-03-08 02:37:57 -08001017 ASSERT_EQ(AudioFrame::kNormalSpeech, output.speech_type_);
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +00001018
1019 // Next packet.
Mirko Bonadeia8110272017-10-18 14:22:50 +02001020 rtp_info.timestamp += rtc::checked_cast<uint32_t>(
1021 expected_samples_per_channel);
henrik.lundin246ef3e2017-04-24 09:14:32 -07001022 rtp_info.sequenceNumber++;
Mirko Bonadeia8110272017-10-18 14:22:50 +02001023 receive_timestamp += rtc::checked_cast<uint32_t>(
1024 expected_samples_per_channel);
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +00001025 }
1026
henrik.lundin6d8e0112016-03-04 10:34:21 -08001027 output.Reset();
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +00001028
1029 // Get audio without inserting packets, expecting PLC and PLC-to-CNG. Pull
1030 // one frame without checking speech-type. This is the first frame pulled
1031 // without inserting any packet, and might not be labeled as PLC.
henrik.lundin7a926812016-05-12 13:51:28 -07001032 ASSERT_EQ(0, neteq_->GetAudio(&output, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -08001033 ASSERT_EQ(1u, output.num_channels_);
1034 ASSERT_EQ(expected_samples_per_channel, output.samples_per_channel_);
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +00001035
1036 // To be able to test the fading of background noise we need at lease to
1037 // pull 611 frames.
1038 const int kFadingThreshold = 611;
1039
1040 // Test several CNG-to-PLC packet for the expected behavior. The number 20
1041 // is arbitrary, but sufficiently large to test enough number of frames.
1042 const int kNumPlcToCngTestFrames = 20;
1043 bool plc_to_cng = false;
1044 for (int n = 0; n < kFadingThreshold + kNumPlcToCngTestFrames; ++n) {
henrik.lundin6d8e0112016-03-04 10:34:21 -08001045 output.Reset();
yujo36b1a5f2017-06-12 12:45:32 -07001046 // Set to non-zero.
1047 memset(output.mutable_data(), 1, AudioFrame::kMaxDataSizeBytes);
henrik.lundin7a926812016-05-12 13:51:28 -07001048 ASSERT_EQ(0, neteq_->GetAudio(&output, &muted));
1049 ASSERT_FALSE(muted);
henrik.lundin6d8e0112016-03-04 10:34:21 -08001050 ASSERT_EQ(1u, output.num_channels_);
1051 ASSERT_EQ(expected_samples_per_channel, output.samples_per_channel_);
henrik.lundin55480f52016-03-08 02:37:57 -08001052 if (output.speech_type_ == AudioFrame::kPLCCNG) {
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +00001053 plc_to_cng = true;
1054 double sum_squared = 0;
yujo36b1a5f2017-06-12 12:45:32 -07001055 const int16_t* output_data = output.data();
henrik.lundin6d8e0112016-03-04 10:34:21 -08001056 for (size_t k = 0;
1057 k < output.num_channels_ * output.samples_per_channel_; ++k)
yujo36b1a5f2017-06-12 12:45:32 -07001058 sum_squared += output_data[k] * output_data[k];
henrik.lundin@webrtc.org9b8102c2014-08-21 08:27:44 +00001059 TestCondition(sum_squared, n > kFadingThreshold);
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +00001060 } else {
henrik.lundin55480f52016-03-08 02:37:57 -08001061 EXPECT_EQ(AudioFrame::kPLC, output.speech_type_);
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +00001062 }
1063 }
1064 EXPECT_TRUE(plc_to_cng); // Just to be sure that PLC-to-CNG has occurred.
1065 }
1066};
1067
henrik.lundin@webrtc.org9b8102c2014-08-21 08:27:44 +00001068class NetEqBgnTestOn : public NetEqBgnTest {
1069 protected:
1070 NetEqBgnTestOn() : NetEqBgnTest() {
1071 config_.background_noise_mode = NetEq::kBgnOn;
1072 }
1073
1074 void TestCondition(double sum_squared_noise, bool /*should_be_faded*/) {
1075 EXPECT_NE(0, sum_squared_noise);
1076 }
1077};
1078
1079class NetEqBgnTestOff : public NetEqBgnTest {
1080 protected:
1081 NetEqBgnTestOff() : NetEqBgnTest() {
1082 config_.background_noise_mode = NetEq::kBgnOff;
1083 }
1084
1085 void TestCondition(double sum_squared_noise, bool /*should_be_faded*/) {
1086 EXPECT_EQ(0, sum_squared_noise);
1087 }
1088};
1089
1090class NetEqBgnTestFade : public NetEqBgnTest {
1091 protected:
1092 NetEqBgnTestFade() : NetEqBgnTest() {
1093 config_.background_noise_mode = NetEq::kBgnFade;
1094 }
1095
1096 void TestCondition(double sum_squared_noise, bool should_be_faded) {
1097 if (should_be_faded)
1098 EXPECT_EQ(0, sum_squared_noise);
1099 }
1100};
1101
henrika1d34fe92015-06-16 10:04:20 +02001102TEST_F(NetEqBgnTestOn, RunTest) {
henrik.lundin@webrtc.org9b8102c2014-08-21 08:27:44 +00001103 CheckBgn(8000);
1104 CheckBgn(16000);
1105 CheckBgn(32000);
turaj@webrtc.orgff43c852013-09-25 00:07:27 +00001106}
turaj@webrtc.org7b75ac62013-09-26 00:27:56 +00001107
henrika1d34fe92015-06-16 10:04:20 +02001108TEST_F(NetEqBgnTestOff, RunTest) {
henrik.lundin@webrtc.org9b8102c2014-08-21 08:27:44 +00001109 CheckBgn(8000);
1110 CheckBgn(16000);
1111 CheckBgn(32000);
1112}
1113
henrika1d34fe92015-06-16 10:04:20 +02001114TEST_F(NetEqBgnTestFade, RunTest) {
henrik.lundin@webrtc.org9b8102c2014-08-21 08:27:44 +00001115 CheckBgn(8000);
1116 CheckBgn(16000);
1117 CheckBgn(32000);
1118}
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +00001119
turaj@webrtc.org78b41a02013-11-22 20:27:07 +00001120void NetEqDecodingTest::WrapTest(uint16_t start_seq_no,
1121 uint32_t start_timestamp,
1122 const std::set<uint16_t>& drop_seq_numbers,
1123 bool expect_seq_no_wrap,
1124 bool expect_timestamp_wrap) {
1125 uint16_t seq_no = start_seq_no;
1126 uint32_t timestamp = start_timestamp;
1127 const int kBlocksPerFrame = 3; // Number of 10 ms blocks per frame.
1128 const int kFrameSizeMs = kBlocksPerFrame * kTimeStepMs;
1129 const int kSamples = kBlockSize16kHz * kBlocksPerFrame;
pkasting@chromium.org4591fbd2014-11-20 22:28:14 +00001130 const size_t kPayloadBytes = kSamples * sizeof(int16_t);
turaj@webrtc.org78b41a02013-11-22 20:27:07 +00001131 double next_input_time_ms = 0.0;
turaj@webrtc.org78b41a02013-11-22 20:27:07 +00001132 uint32_t receive_timestamp = 0;
1133
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001134 // Insert speech for 2 seconds.
turaj@webrtc.org78b41a02013-11-22 20:27:07 +00001135 const int kSpeechDurationMs = 2000;
1136 int packets_inserted = 0;
1137 uint16_t last_seq_no;
1138 uint32_t last_timestamp;
1139 bool timestamp_wrapped = false;
1140 bool seq_no_wrapped = false;
1141 for (double t_ms = 0; t_ms < kSpeechDurationMs; t_ms += 10) {
1142 // Each turn in this for loop is 10 ms.
1143 while (next_input_time_ms <= t_ms) {
1144 // Insert one 30 ms speech frame.
1145 uint8_t payload[kPayloadBytes] = {0};
henrik.lundin246ef3e2017-04-24 09:14:32 -07001146 RTPHeader rtp_info;
turaj@webrtc.org78b41a02013-11-22 20:27:07 +00001147 PopulateRtpInfo(seq_no, timestamp, &rtp_info);
1148 if (drop_seq_numbers.find(seq_no) == drop_seq_numbers.end()) {
1149 // This sequence number was not in the set to drop. Insert it.
henrik.lundin246ef3e2017-04-24 09:14:32 -07001150 ASSERT_EQ(0,
1151 neteq_->InsertPacket(rtp_info, payload, receive_timestamp));
turaj@webrtc.org78b41a02013-11-22 20:27:07 +00001152 ++packets_inserted;
1153 }
1154 NetEqNetworkStatistics network_stats;
1155 ASSERT_EQ(0, neteq_->NetworkStatistics(&network_stats));
1156
1157 // Due to internal NetEq logic, preferred buffer-size is about 4 times the
1158 // packet size for first few packets. Therefore we refrain from checking
1159 // the criteria.
1160 if (packets_inserted > 4) {
1161 // Expect preferred and actual buffer size to be no more than 2 frames.
1162 EXPECT_LE(network_stats.preferred_buffer_size_ms, kFrameSizeMs * 2);
turaj@webrtc.org8d1cdaa2014-04-11 18:47:55 +00001163 EXPECT_LE(network_stats.current_buffer_size_ms, kFrameSizeMs * 2 +
1164 algorithmic_delay_ms_);
turaj@webrtc.org78b41a02013-11-22 20:27:07 +00001165 }
1166 last_seq_no = seq_no;
1167 last_timestamp = timestamp;
1168
1169 ++seq_no;
1170 timestamp += kSamples;
1171 receive_timestamp += kSamples;
1172 next_input_time_ms += static_cast<double>(kFrameSizeMs);
1173
1174 seq_no_wrapped |= seq_no < last_seq_no;
1175 timestamp_wrapped |= timestamp < last_timestamp;
1176 }
1177 // Pull out data once.
henrik.lundin6d8e0112016-03-04 10:34:21 -08001178 AudioFrame output;
henrik.lundin7a926812016-05-12 13:51:28 -07001179 bool muted;
1180 ASSERT_EQ(0, neteq_->GetAudio(&output, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -08001181 ASSERT_EQ(kBlockSize16kHz, output.samples_per_channel_);
1182 ASSERT_EQ(1u, output.num_channels_);
turaj@webrtc.org78b41a02013-11-22 20:27:07 +00001183
1184 // Expect delay (in samples) to be less than 2 packets.
henrik.lundin114c1b32017-04-26 07:47:32 -07001185 rtc::Optional<uint32_t> playout_timestamp = neteq_->GetPlayoutTimestamp();
henrik.lundin0d96ab72016-04-06 12:28:26 -07001186 ASSERT_TRUE(playout_timestamp);
1187 EXPECT_LE(timestamp - *playout_timestamp,
turaj@webrtc.org78b41a02013-11-22 20:27:07 +00001188 static_cast<uint32_t>(kSamples * 2));
turaj@webrtc.org78b41a02013-11-22 20:27:07 +00001189 }
1190 // Make sure we have actually tested wrap-around.
1191 ASSERT_EQ(expect_seq_no_wrap, seq_no_wrapped);
1192 ASSERT_EQ(expect_timestamp_wrap, timestamp_wrapped);
1193}
1194
1195TEST_F(NetEqDecodingTest, SequenceNumberWrap) {
1196 // Start with a sequence number that will soon wrap.
1197 std::set<uint16_t> drop_seq_numbers; // Don't drop any packets.
1198 WrapTest(0xFFFF - 10, 0, drop_seq_numbers, true, false);
1199}
1200
1201TEST_F(NetEqDecodingTest, SequenceNumberWrapAndDrop) {
1202 // Start with a sequence number that will soon wrap.
1203 std::set<uint16_t> drop_seq_numbers;
1204 drop_seq_numbers.insert(0xFFFF);
1205 drop_seq_numbers.insert(0x0);
1206 WrapTest(0xFFFF - 10, 0, drop_seq_numbers, true, false);
1207}
1208
1209TEST_F(NetEqDecodingTest, TimestampWrap) {
1210 // Start with a timestamp that will soon wrap.
1211 std::set<uint16_t> drop_seq_numbers;
1212 WrapTest(0, 0xFFFFFFFF - 3000, drop_seq_numbers, false, true);
1213}
1214
1215TEST_F(NetEqDecodingTest, TimestampAndSequenceNumberWrap) {
1216 // Start with a timestamp and a sequence number that will wrap at the same
1217 // time.
1218 std::set<uint16_t> drop_seq_numbers;
1219 WrapTest(0xFFFF - 10, 0xFFFFFFFF - 5000, drop_seq_numbers, true, true);
1220}
1221
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001222void NetEqDecodingTest::DuplicateCng() {
1223 uint16_t seq_no = 0;
1224 uint32_t timestamp = 0;
1225 const int kFrameSizeMs = 10;
1226 const int kSampleRateKhz = 16;
1227 const int kSamples = kFrameSizeMs * kSampleRateKhz;
pkasting@chromium.org4591fbd2014-11-20 22:28:14 +00001228 const size_t kPayloadBytes = kSamples * 2;
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001229
turaj@webrtc.org8d1cdaa2014-04-11 18:47:55 +00001230 const int algorithmic_delay_samples = std::max(
1231 algorithmic_delay_ms_ * kSampleRateKhz, 5 * kSampleRateKhz / 8);
henrik.lundin@webrtc.orgc93437e2014-12-01 11:42:42 +00001232 // Insert three speech packets. Three are needed to get the frame length
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001233 // correct.
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001234 uint8_t payload[kPayloadBytes] = {0};
henrik.lundin246ef3e2017-04-24 09:14:32 -07001235 RTPHeader rtp_info;
henrik.lundin7a926812016-05-12 13:51:28 -07001236 bool muted;
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001237 for (int i = 0; i < 3; ++i) {
1238 PopulateRtpInfo(seq_no, timestamp, &rtp_info);
henrik.lundin246ef3e2017-04-24 09:14:32 -07001239 ASSERT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001240 ++seq_no;
1241 timestamp += kSamples;
1242
1243 // Pull audio once.
henrik.lundin7a926812016-05-12 13:51:28 -07001244 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -08001245 ASSERT_EQ(kBlockSize16kHz, out_frame_.samples_per_channel_);
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001246 }
1247 // Verify speech output.
henrik.lundin55480f52016-03-08 02:37:57 -08001248 EXPECT_EQ(AudioFrame::kNormalSpeech, out_frame_.speech_type_);
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001249
1250 // Insert same CNG packet twice.
1251 const int kCngPeriodMs = 100;
1252 const int kCngPeriodSamples = kCngPeriodMs * kSampleRateKhz;
pkasting@chromium.org4591fbd2014-11-20 22:28:14 +00001253 size_t payload_len;
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001254 PopulateCng(seq_no, timestamp, &rtp_info, payload, &payload_len);
1255 // This is the first time this CNG packet is inserted.
henrik.lundin246ef3e2017-04-24 09:14:32 -07001256 ASSERT_EQ(
1257 0, neteq_->InsertPacket(
1258 rtp_info, rtc::ArrayView<const uint8_t>(payload, payload_len), 0));
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001259
1260 // Pull audio once and make sure CNG is played.
henrik.lundin7a926812016-05-12 13:51:28 -07001261 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -08001262 ASSERT_EQ(kBlockSize16kHz, out_frame_.samples_per_channel_);
henrik.lundin55480f52016-03-08 02:37:57 -08001263 EXPECT_EQ(AudioFrame::kCNG, out_frame_.speech_type_);
henrik.lundin114c1b32017-04-26 07:47:32 -07001264 EXPECT_FALSE(
1265 neteq_->GetPlayoutTimestamp()); // Returns empty value during CNG.
henrik.lundin0d96ab72016-04-06 12:28:26 -07001266 EXPECT_EQ(timestamp - algorithmic_delay_samples,
1267 out_frame_.timestamp_ + out_frame_.samples_per_channel_);
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001268
1269 // Insert the same CNG packet again. Note that at this point it is old, since
1270 // we have already decoded the first copy of it.
henrik.lundin246ef3e2017-04-24 09:14:32 -07001271 ASSERT_EQ(
1272 0, neteq_->InsertPacket(
1273 rtp_info, rtc::ArrayView<const uint8_t>(payload, payload_len), 0));
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001274
1275 // Pull audio until we have played |kCngPeriodMs| of CNG. Start at 10 ms since
1276 // we have already pulled out CNG once.
1277 for (int cng_time_ms = 10; cng_time_ms < kCngPeriodMs; cng_time_ms += 10) {
henrik.lundin7a926812016-05-12 13:51:28 -07001278 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -08001279 ASSERT_EQ(kBlockSize16kHz, out_frame_.samples_per_channel_);
henrik.lundin55480f52016-03-08 02:37:57 -08001280 EXPECT_EQ(AudioFrame::kCNG, out_frame_.speech_type_);
henrik.lundin114c1b32017-04-26 07:47:32 -07001281 EXPECT_FALSE(
1282 neteq_->GetPlayoutTimestamp()); // Returns empty value during CNG.
turaj@webrtc.org8d1cdaa2014-04-11 18:47:55 +00001283 EXPECT_EQ(timestamp - algorithmic_delay_samples,
henrik.lundin0d96ab72016-04-06 12:28:26 -07001284 out_frame_.timestamp_ + out_frame_.samples_per_channel_);
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001285 }
1286
1287 // Insert speech again.
1288 ++seq_no;
1289 timestamp += kCngPeriodSamples;
1290 PopulateRtpInfo(seq_no, timestamp, &rtp_info);
henrik.lundin246ef3e2017-04-24 09:14:32 -07001291 ASSERT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001292
1293 // Pull audio once and verify that the output is speech again.
henrik.lundin7a926812016-05-12 13:51:28 -07001294 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -08001295 ASSERT_EQ(kBlockSize16kHz, out_frame_.samples_per_channel_);
henrik.lundin55480f52016-03-08 02:37:57 -08001296 EXPECT_EQ(AudioFrame::kNormalSpeech, out_frame_.speech_type_);
henrik.lundin114c1b32017-04-26 07:47:32 -07001297 rtc::Optional<uint32_t> playout_timestamp = neteq_->GetPlayoutTimestamp();
henrik.lundin0d96ab72016-04-06 12:28:26 -07001298 ASSERT_TRUE(playout_timestamp);
turaj@webrtc.org8d1cdaa2014-04-11 18:47:55 +00001299 EXPECT_EQ(timestamp + kSamples - algorithmic_delay_samples,
henrik.lundin0d96ab72016-04-06 12:28:26 -07001300 *playout_timestamp);
wu@webrtc.org94454b72014-06-05 20:34:08 +00001301}
1302
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001303TEST_F(NetEqDecodingTest, DiscardDuplicateCng) { DuplicateCng(); }
henrik.lundin@webrtc.orgc93437e2014-12-01 11:42:42 +00001304
1305TEST_F(NetEqDecodingTest, CngFirst) {
1306 uint16_t seq_no = 0;
1307 uint32_t timestamp = 0;
1308 const int kFrameSizeMs = 10;
1309 const int kSampleRateKhz = 16;
1310 const int kSamples = kFrameSizeMs * kSampleRateKhz;
1311 const int kPayloadBytes = kSamples * 2;
1312 const int kCngPeriodMs = 100;
1313 const int kCngPeriodSamples = kCngPeriodMs * kSampleRateKhz;
1314 size_t payload_len;
1315
1316 uint8_t payload[kPayloadBytes] = {0};
henrik.lundin246ef3e2017-04-24 09:14:32 -07001317 RTPHeader rtp_info;
henrik.lundin@webrtc.orgc93437e2014-12-01 11:42:42 +00001318
1319 PopulateCng(seq_no, timestamp, &rtp_info, payload, &payload_len);
henrik.lundin246ef3e2017-04-24 09:14:32 -07001320 ASSERT_EQ(
1321 NetEq::kOK,
1322 neteq_->InsertPacket(
1323 rtp_info, rtc::ArrayView<const uint8_t>(payload, payload_len), 0));
henrik.lundin@webrtc.orgc93437e2014-12-01 11:42:42 +00001324 ++seq_no;
1325 timestamp += kCngPeriodSamples;
1326
1327 // Pull audio once and make sure CNG is played.
henrik.lundin7a926812016-05-12 13:51:28 -07001328 bool muted;
1329 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -08001330 ASSERT_EQ(kBlockSize16kHz, out_frame_.samples_per_channel_);
henrik.lundin55480f52016-03-08 02:37:57 -08001331 EXPECT_EQ(AudioFrame::kCNG, out_frame_.speech_type_);
henrik.lundin@webrtc.orgc93437e2014-12-01 11:42:42 +00001332
1333 // Insert some speech packets.
henrik.lundin549d80b2016-08-25 00:44:24 -07001334 const uint32_t first_speech_timestamp = timestamp;
1335 int timeout_counter = 0;
1336 do {
1337 ASSERT_LT(timeout_counter++, 20) << "Test timed out";
henrik.lundin@webrtc.orgc93437e2014-12-01 11:42:42 +00001338 PopulateRtpInfo(seq_no, timestamp, &rtp_info);
henrik.lundin246ef3e2017-04-24 09:14:32 -07001339 ASSERT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
henrik.lundin@webrtc.orgc93437e2014-12-01 11:42:42 +00001340 ++seq_no;
1341 timestamp += kSamples;
1342
1343 // Pull audio once.
henrik.lundin7a926812016-05-12 13:51:28 -07001344 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -08001345 ASSERT_EQ(kBlockSize16kHz, out_frame_.samples_per_channel_);
henrik.lundin549d80b2016-08-25 00:44:24 -07001346 } while (!IsNewerTimestamp(out_frame_.timestamp_, first_speech_timestamp));
henrik.lundin@webrtc.orgc93437e2014-12-01 11:42:42 +00001347 // Verify speech output.
henrik.lundin55480f52016-03-08 02:37:57 -08001348 EXPECT_EQ(AudioFrame::kNormalSpeech, out_frame_.speech_type_);
henrik.lundin@webrtc.orgc93437e2014-12-01 11:42:42 +00001349}
henrik.lundin7a926812016-05-12 13:51:28 -07001350
1351class NetEqDecodingTestWithMutedState : public NetEqDecodingTest {
1352 public:
1353 NetEqDecodingTestWithMutedState() : NetEqDecodingTest() {
1354 config_.enable_muted_state = true;
1355 }
1356
1357 protected:
1358 static constexpr size_t kSamples = 10 * 16;
1359 static constexpr size_t kPayloadBytes = kSamples * 2;
1360
1361 void InsertPacket(uint32_t rtp_timestamp) {
1362 uint8_t payload[kPayloadBytes] = {0};
henrik.lundin246ef3e2017-04-24 09:14:32 -07001363 RTPHeader rtp_info;
henrik.lundin7a926812016-05-12 13:51:28 -07001364 PopulateRtpInfo(0, rtp_timestamp, &rtp_info);
henrik.lundin246ef3e2017-04-24 09:14:32 -07001365 EXPECT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
henrik.lundin7a926812016-05-12 13:51:28 -07001366 }
1367
henrik.lundin42feb512016-09-20 06:51:40 -07001368 void InsertCngPacket(uint32_t rtp_timestamp) {
1369 uint8_t payload[kPayloadBytes] = {0};
henrik.lundin246ef3e2017-04-24 09:14:32 -07001370 RTPHeader rtp_info;
henrik.lundin42feb512016-09-20 06:51:40 -07001371 size_t payload_len;
1372 PopulateCng(0, rtp_timestamp, &rtp_info, payload, &payload_len);
henrik.lundin246ef3e2017-04-24 09:14:32 -07001373 EXPECT_EQ(
1374 NetEq::kOK,
1375 neteq_->InsertPacket(
1376 rtp_info, rtc::ArrayView<const uint8_t>(payload, payload_len), 0));
henrik.lundin42feb512016-09-20 06:51:40 -07001377 }
1378
henrik.lundin7a926812016-05-12 13:51:28 -07001379 bool GetAudioReturnMuted() {
1380 bool muted;
1381 EXPECT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
1382 return muted;
1383 }
1384
1385 void GetAudioUntilMuted() {
1386 while (!GetAudioReturnMuted()) {
1387 ASSERT_LT(counter_++, 1000) << "Test timed out";
1388 }
1389 }
1390
1391 void GetAudioUntilNormal() {
1392 bool muted = false;
1393 while (out_frame_.speech_type_ != AudioFrame::kNormalSpeech) {
1394 EXPECT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
1395 ASSERT_LT(counter_++, 1000) << "Test timed out";
1396 }
1397 EXPECT_FALSE(muted);
1398 }
1399
1400 int counter_ = 0;
1401};
1402
1403// Verifies that NetEq goes in and out of muted state as expected.
1404TEST_F(NetEqDecodingTestWithMutedState, MutedState) {
1405 // Insert one speech packet.
1406 InsertPacket(0);
1407 // Pull out audio once and expect it not to be muted.
1408 EXPECT_FALSE(GetAudioReturnMuted());
1409 // Pull data until faded out.
1410 GetAudioUntilMuted();
henrik.lundina4491072017-07-06 05:23:53 -07001411 EXPECT_TRUE(out_frame_.muted());
henrik.lundin7a926812016-05-12 13:51:28 -07001412
1413 // Verify that output audio is not written during muted mode. Other parameters
1414 // should be correct, though.
1415 AudioFrame new_frame;
yujo36b1a5f2017-06-12 12:45:32 -07001416 int16_t* frame_data = new_frame.mutable_data();
1417 for (size_t i = 0; i < AudioFrame::kMaxDataSizeSamples; i++) {
1418 frame_data[i] = 17;
henrik.lundin7a926812016-05-12 13:51:28 -07001419 }
1420 bool muted;
1421 EXPECT_EQ(0, neteq_->GetAudio(&new_frame, &muted));
1422 EXPECT_TRUE(muted);
henrik.lundina4491072017-07-06 05:23:53 -07001423 EXPECT_TRUE(out_frame_.muted());
yujo36b1a5f2017-06-12 12:45:32 -07001424 for (size_t i = 0; i < AudioFrame::kMaxDataSizeSamples; i++) {
1425 EXPECT_EQ(17, frame_data[i]);
henrik.lundin7a926812016-05-12 13:51:28 -07001426 }
1427 EXPECT_EQ(out_frame_.timestamp_ + out_frame_.samples_per_channel_,
1428 new_frame.timestamp_);
1429 EXPECT_EQ(out_frame_.samples_per_channel_, new_frame.samples_per_channel_);
1430 EXPECT_EQ(out_frame_.sample_rate_hz_, new_frame.sample_rate_hz_);
1431 EXPECT_EQ(out_frame_.num_channels_, new_frame.num_channels_);
1432 EXPECT_EQ(out_frame_.speech_type_, new_frame.speech_type_);
1433 EXPECT_EQ(out_frame_.vad_activity_, new_frame.vad_activity_);
1434
1435 // Insert new data. Timestamp is corrected for the time elapsed since the last
1436 // packet. Verify that normal operation resumes.
1437 InsertPacket(kSamples * counter_);
1438 GetAudioUntilNormal();
henrik.lundina4491072017-07-06 05:23:53 -07001439 EXPECT_FALSE(out_frame_.muted());
henrik.lundin612c25e2016-05-25 08:21:04 -07001440
1441 NetEqNetworkStatistics stats;
1442 EXPECT_EQ(0, neteq_->NetworkStatistics(&stats));
1443 // NetEqNetworkStatistics::expand_rate tells the fraction of samples that were
1444 // concealment samples, in Q14 (16384 = 100%) .The vast majority should be
1445 // concealment samples in this test.
1446 EXPECT_GT(stats.expand_rate, 14000);
1447 // And, it should be greater than the speech_expand_rate.
1448 EXPECT_GT(stats.expand_rate, stats.speech_expand_rate);
henrik.lundin7a926812016-05-12 13:51:28 -07001449}
1450
1451// Verifies that NetEq goes out of muted state when given a delayed packet.
1452TEST_F(NetEqDecodingTestWithMutedState, MutedStateDelayedPacket) {
1453 // Insert one speech packet.
1454 InsertPacket(0);
1455 // Pull out audio once and expect it not to be muted.
1456 EXPECT_FALSE(GetAudioReturnMuted());
1457 // Pull data until faded out.
1458 GetAudioUntilMuted();
1459 // Insert new data. Timestamp is only corrected for the half of the time
1460 // elapsed since the last packet. That is, the new packet is delayed. Verify
1461 // that normal operation resumes.
1462 InsertPacket(kSamples * counter_ / 2);
1463 GetAudioUntilNormal();
1464}
1465
1466// Verifies that NetEq goes out of muted state when given a future packet.
1467TEST_F(NetEqDecodingTestWithMutedState, MutedStateFuturePacket) {
1468 // Insert one speech packet.
1469 InsertPacket(0);
1470 // Pull out audio once and expect it not to be muted.
1471 EXPECT_FALSE(GetAudioReturnMuted());
1472 // Pull data until faded out.
1473 GetAudioUntilMuted();
1474 // Insert new data. Timestamp is over-corrected for the time elapsed since the
1475 // last packet. That is, the new packet is too early. Verify that normal
1476 // operation resumes.
1477 InsertPacket(kSamples * counter_ * 2);
1478 GetAudioUntilNormal();
1479}
1480
1481// Verifies that NetEq goes out of muted state when given an old packet.
1482TEST_F(NetEqDecodingTestWithMutedState, MutedStateOldPacket) {
1483 // Insert one speech packet.
1484 InsertPacket(0);
1485 // Pull out audio once and expect it not to be muted.
1486 EXPECT_FALSE(GetAudioReturnMuted());
1487 // Pull data until faded out.
1488 GetAudioUntilMuted();
1489
1490 EXPECT_NE(AudioFrame::kNormalSpeech, out_frame_.speech_type_);
1491 // Insert packet which is older than the first packet.
1492 InsertPacket(kSamples * (counter_ - 1000));
1493 EXPECT_FALSE(GetAudioReturnMuted());
1494 EXPECT_EQ(AudioFrame::kNormalSpeech, out_frame_.speech_type_);
1495}
1496
henrik.lundin42feb512016-09-20 06:51:40 -07001497// Verifies that NetEq doesn't enter muted state when CNG mode is active and the
1498// packet stream is suspended for a long time.
1499TEST_F(NetEqDecodingTestWithMutedState, DoNotMuteExtendedCngWithoutPackets) {
1500 // Insert one CNG packet.
1501 InsertCngPacket(0);
1502
1503 // Pull 10 seconds of audio (10 ms audio generated per lap).
1504 for (int i = 0; i < 1000; ++i) {
1505 bool muted;
1506 EXPECT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
1507 ASSERT_FALSE(muted);
1508 }
1509 EXPECT_EQ(AudioFrame::kCNG, out_frame_.speech_type_);
1510}
1511
1512// Verifies that NetEq goes back to normal after a long CNG period with the
1513// packet stream suspended.
1514TEST_F(NetEqDecodingTestWithMutedState, RecoverAfterExtendedCngWithoutPackets) {
1515 // Insert one CNG packet.
1516 InsertCngPacket(0);
1517
1518 // Pull 10 seconds of audio (10 ms audio generated per lap).
1519 for (int i = 0; i < 1000; ++i) {
1520 bool muted;
1521 EXPECT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
1522 }
1523
1524 // Insert new data. Timestamp is corrected for the time elapsed since the last
1525 // packet. Verify that normal operation resumes.
1526 InsertPacket(kSamples * counter_);
1527 GetAudioUntilNormal();
1528}
1529
henrik.lundin7a926812016-05-12 13:51:28 -07001530class NetEqDecodingTestTwoInstances : public NetEqDecodingTest {
1531 public:
1532 NetEqDecodingTestTwoInstances() : NetEqDecodingTest() {}
1533
1534 void SetUp() override {
1535 NetEqDecodingTest::SetUp();
1536 config2_ = config_;
1537 }
1538
1539 void CreateSecondInstance() {
ossue3525782016-05-25 07:37:43 -07001540 neteq2_.reset(NetEq::Create(config2_, CreateBuiltinAudioDecoderFactory()));
henrik.lundin7a926812016-05-12 13:51:28 -07001541 ASSERT_TRUE(neteq2_);
1542 LoadDecoders(neteq2_.get());
1543 }
1544
1545 protected:
1546 std::unique_ptr<NetEq> neteq2_;
1547 NetEq::Config config2_;
1548};
1549
1550namespace {
1551::testing::AssertionResult AudioFramesEqualExceptData(const AudioFrame& a,
1552 const AudioFrame& b) {
1553 if (a.timestamp_ != b.timestamp_)
1554 return ::testing::AssertionFailure() << "timestamp_ diff (" << a.timestamp_
1555 << " != " << b.timestamp_ << ")";
1556 if (a.sample_rate_hz_ != b.sample_rate_hz_)
1557 return ::testing::AssertionFailure() << "sample_rate_hz_ diff ("
1558 << a.sample_rate_hz_
1559 << " != " << b.sample_rate_hz_ << ")";
1560 if (a.samples_per_channel_ != b.samples_per_channel_)
1561 return ::testing::AssertionFailure()
1562 << "samples_per_channel_ diff (" << a.samples_per_channel_
1563 << " != " << b.samples_per_channel_ << ")";
1564 if (a.num_channels_ != b.num_channels_)
1565 return ::testing::AssertionFailure() << "num_channels_ diff ("
1566 << a.num_channels_
1567 << " != " << b.num_channels_ << ")";
1568 if (a.speech_type_ != b.speech_type_)
1569 return ::testing::AssertionFailure() << "speech_type_ diff ("
1570 << a.speech_type_
1571 << " != " << b.speech_type_ << ")";
1572 if (a.vad_activity_ != b.vad_activity_)
1573 return ::testing::AssertionFailure() << "vad_activity_ diff ("
1574 << a.vad_activity_
1575 << " != " << b.vad_activity_ << ")";
1576 return ::testing::AssertionSuccess();
1577}
1578
1579::testing::AssertionResult AudioFramesEqual(const AudioFrame& a,
1580 const AudioFrame& b) {
1581 ::testing::AssertionResult res = AudioFramesEqualExceptData(a, b);
1582 if (!res)
1583 return res;
1584 if (memcmp(
yujo36b1a5f2017-06-12 12:45:32 -07001585 a.data(), b.data(),
1586 a.samples_per_channel_ * a.num_channels_ * sizeof(*a.data())) != 0) {
henrik.lundin7a926812016-05-12 13:51:28 -07001587 return ::testing::AssertionFailure() << "data_ diff";
1588 }
1589 return ::testing::AssertionSuccess();
1590}
1591
1592} // namespace
1593
1594TEST_F(NetEqDecodingTestTwoInstances, CompareMutedStateOnOff) {
1595 ASSERT_FALSE(config_.enable_muted_state);
1596 config2_.enable_muted_state = true;
1597 CreateSecondInstance();
1598
1599 // Insert one speech packet into both NetEqs.
1600 const size_t kSamples = 10 * 16;
1601 const size_t kPayloadBytes = kSamples * 2;
1602 uint8_t payload[kPayloadBytes] = {0};
henrik.lundin246ef3e2017-04-24 09:14:32 -07001603 RTPHeader rtp_info;
henrik.lundin7a926812016-05-12 13:51:28 -07001604 PopulateRtpInfo(0, 0, &rtp_info);
henrik.lundin246ef3e2017-04-24 09:14:32 -07001605 EXPECT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
1606 EXPECT_EQ(0, neteq2_->InsertPacket(rtp_info, payload, 0));
henrik.lundin7a926812016-05-12 13:51:28 -07001607
1608 AudioFrame out_frame1, out_frame2;
1609 bool muted;
1610 for (int i = 0; i < 1000; ++i) {
1611 std::ostringstream ss;
1612 ss << "i = " << i;
1613 SCOPED_TRACE(ss.str()); // Print out the loop iterator on failure.
1614 EXPECT_EQ(0, neteq_->GetAudio(&out_frame1, &muted));
1615 EXPECT_FALSE(muted);
1616 EXPECT_EQ(0, neteq2_->GetAudio(&out_frame2, &muted));
1617 if (muted) {
1618 EXPECT_TRUE(AudioFramesEqualExceptData(out_frame1, out_frame2));
1619 } else {
1620 EXPECT_TRUE(AudioFramesEqual(out_frame1, out_frame2));
1621 }
1622 }
1623 EXPECT_TRUE(muted);
1624
1625 // Insert new data. Timestamp is corrected for the time elapsed since the last
1626 // packet.
1627 PopulateRtpInfo(0, kSamples * 1000, &rtp_info);
henrik.lundin246ef3e2017-04-24 09:14:32 -07001628 EXPECT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
1629 EXPECT_EQ(0, neteq2_->InsertPacket(rtp_info, payload, 0));
henrik.lundin7a926812016-05-12 13:51:28 -07001630
1631 int counter = 0;
1632 while (out_frame1.speech_type_ != AudioFrame::kNormalSpeech) {
1633 ASSERT_LT(counter++, 1000) << "Test timed out";
1634 std::ostringstream ss;
1635 ss << "counter = " << counter;
1636 SCOPED_TRACE(ss.str()); // Print out the loop iterator on failure.
1637 EXPECT_EQ(0, neteq_->GetAudio(&out_frame1, &muted));
1638 EXPECT_FALSE(muted);
1639 EXPECT_EQ(0, neteq2_->GetAudio(&out_frame2, &muted));
1640 if (muted) {
1641 EXPECT_TRUE(AudioFramesEqualExceptData(out_frame1, out_frame2));
1642 } else {
1643 EXPECT_TRUE(AudioFramesEqual(out_frame1, out_frame2));
1644 }
1645 }
1646 EXPECT_FALSE(muted);
1647}
1648
henrik.lundin114c1b32017-04-26 07:47:32 -07001649TEST_F(NetEqDecodingTest, LastDecodedTimestampsEmpty) {
1650 EXPECT_TRUE(neteq_->LastDecodedTimestamps().empty());
1651
1652 // Pull out data once.
1653 AudioFrame output;
1654 bool muted;
1655 ASSERT_EQ(0, neteq_->GetAudio(&output, &muted));
1656
1657 EXPECT_TRUE(neteq_->LastDecodedTimestamps().empty());
1658}
1659
1660TEST_F(NetEqDecodingTest, LastDecodedTimestampsOneDecoded) {
1661 // Insert one packet with PCM16b WB data (this is what PopulateRtpInfo does by
1662 // default). Make the length 10 ms.
1663 constexpr size_t kPayloadSamples = 16 * 10;
1664 constexpr size_t kPayloadBytes = 2 * kPayloadSamples;
1665 uint8_t payload[kPayloadBytes] = {0};
1666
1667 RTPHeader rtp_info;
1668 constexpr uint32_t kRtpTimestamp = 0x1234;
1669 PopulateRtpInfo(0, kRtpTimestamp, &rtp_info);
1670 EXPECT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
1671
1672 // Pull out data once.
1673 AudioFrame output;
1674 bool muted;
1675 ASSERT_EQ(0, neteq_->GetAudio(&output, &muted));
1676
1677 EXPECT_EQ(std::vector<uint32_t>({kRtpTimestamp}),
1678 neteq_->LastDecodedTimestamps());
1679
1680 // Nothing decoded on the second call.
1681 ASSERT_EQ(0, neteq_->GetAudio(&output, &muted));
1682 EXPECT_TRUE(neteq_->LastDecodedTimestamps().empty());
1683}
1684
1685TEST_F(NetEqDecodingTest, LastDecodedTimestampsTwoDecoded) {
1686 // Insert two packets with PCM16b WB data (this is what PopulateRtpInfo does
1687 // by default). Make the length 5 ms so that NetEq must decode them both in
1688 // the same GetAudio call.
1689 constexpr size_t kPayloadSamples = 16 * 5;
1690 constexpr size_t kPayloadBytes = 2 * kPayloadSamples;
1691 uint8_t payload[kPayloadBytes] = {0};
1692
1693 RTPHeader rtp_info;
1694 constexpr uint32_t kRtpTimestamp1 = 0x1234;
1695 PopulateRtpInfo(0, kRtpTimestamp1, &rtp_info);
1696 EXPECT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
1697 constexpr uint32_t kRtpTimestamp2 = kRtpTimestamp1 + kPayloadSamples;
1698 PopulateRtpInfo(1, kRtpTimestamp2, &rtp_info);
1699 EXPECT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
1700
1701 // Pull out data once.
1702 AudioFrame output;
1703 bool muted;
1704 ASSERT_EQ(0, neteq_->GetAudio(&output, &muted));
1705
1706 EXPECT_EQ(std::vector<uint32_t>({kRtpTimestamp1, kRtpTimestamp2}),
1707 neteq_->LastDecodedTimestamps());
1708}
1709
Gustaf Ullberg9a2e9062017-09-18 09:28:20 +02001710TEST_F(NetEqDecodingTest, TestConcealmentEvents) {
1711 const int kNumConcealmentEvents = 19;
1712 const size_t kSamples = 10 * 16;
1713 const size_t kPayloadBytes = kSamples * 2;
1714 int seq_no = 0;
1715 RTPHeader rtp_info;
1716 rtp_info.ssrc = 0x1234; // Just an arbitrary SSRC.
1717 rtp_info.payloadType = 94; // PCM16b WB codec.
1718 rtp_info.markerBit = 0;
1719 const uint8_t payload[kPayloadBytes] = {0};
1720 bool muted;
1721
1722 for (int i = 0; i < kNumConcealmentEvents; i++) {
1723 // Insert some packets of 10 ms size.
1724 for (int j = 0; j < 10; j++) {
1725 rtp_info.sequenceNumber = seq_no++;
1726 rtp_info.timestamp = rtp_info.sequenceNumber * kSamples;
1727 neteq_->InsertPacket(rtp_info, payload, 0);
1728 neteq_->GetAudio(&out_frame_, &muted);
1729 }
1730
1731 // Lose a number of packets.
1732 int num_lost = 1 + i;
1733 for (int j = 0; j < num_lost; j++) {
1734 seq_no++;
1735 neteq_->GetAudio(&out_frame_, &muted);
1736 }
1737 }
1738
1739 // Check number of concealment events.
1740 NetEqLifetimeStatistics stats = neteq_->GetLifetimeStatistics();
1741 EXPECT_EQ(kNumConcealmentEvents, static_cast<int>(stats.concealment_events));
1742}
1743
Gustaf Ullbergb0a02072017-10-02 12:00:34 +02001744// Test that the jitter buffer delay stat is computed correctly.
1745void NetEqDecodingTestFaxMode::TestJitterBufferDelay(bool apply_packet_loss) {
1746 const int kNumPackets = 10;
1747 const int kDelayInNumPackets = 2;
1748 const int kPacketLenMs = 10; // All packets are of 10 ms size.
1749 const size_t kSamples = kPacketLenMs * 16;
1750 const size_t kPayloadBytes = kSamples * 2;
1751 RTPHeader rtp_info;
1752 rtp_info.ssrc = 0x1234; // Just an arbitrary SSRC.
1753 rtp_info.payloadType = 94; // PCM16b WB codec.
1754 rtp_info.markerBit = 0;
1755 const uint8_t payload[kPayloadBytes] = {0};
1756 bool muted;
1757 int packets_sent = 0;
1758 int packets_received = 0;
1759 int expected_delay = 0;
1760 while (packets_received < kNumPackets) {
1761 // Insert packet.
1762 if (packets_sent < kNumPackets) {
1763 rtp_info.sequenceNumber = packets_sent++;
1764 rtp_info.timestamp = rtp_info.sequenceNumber * kSamples;
1765 neteq_->InsertPacket(rtp_info, payload, 0);
1766 }
1767
1768 // Get packet.
1769 if (packets_sent > kDelayInNumPackets) {
1770 neteq_->GetAudio(&out_frame_, &muted);
1771 packets_received++;
1772
1773 // The delay reported by the jitter buffer never exceeds
1774 // the number of samples previously fetched with GetAudio
1775 // (hence the min()).
1776 int packets_delay = std::min(packets_received, kDelayInNumPackets + 1);
1777
1778 // The increase of the expected delay is the product of
1779 // the current delay of the jitter buffer in ms * the
1780 // number of samples that are sent for play out.
1781 int current_delay_ms = packets_delay * kPacketLenMs;
1782 expected_delay += current_delay_ms * kSamples;
1783 }
1784 }
1785
1786 if (apply_packet_loss) {
1787 // Extra call to GetAudio to cause concealment.
1788 neteq_->GetAudio(&out_frame_, &muted);
1789 }
1790
1791 // Check jitter buffer delay.
1792 NetEqLifetimeStatistics stats = neteq_->GetLifetimeStatistics();
1793 EXPECT_EQ(expected_delay, static_cast<int>(stats.jitter_buffer_delay_ms));
1794}
1795
1796TEST_F(NetEqDecodingTestFaxMode, TestJitterBufferDelayWithoutLoss) {
1797 TestJitterBufferDelay(false);
1798}
1799
1800TEST_F(NetEqDecodingTestFaxMode, TestJitterBufferDelayWithLoss) {
1801 TestJitterBufferDelay(true);
1802}
1803
henrik.lundin@webrtc.orge7ce4372014-01-09 14:01:55 +00001804} // namespace webrtc