blob: 5b9221793f65399a0d8965cbdf5a09689de983ca [file] [log] [blame]
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +00001/*
2 * Copyright (c) 2011 The WebRTC project authors. All Rights Reserved.
3 *
4 * Use of this source code is governed by a BSD-style license
5 * that can be found in the LICENSE file in the root of the source
6 * tree. An additional intellectual property rights grant can be found
7 * in the file PATENTS. All contributing project authors may
8 * be found in the AUTHORS file in the root of the source tree.
9 */
10
Mirko Bonadei92ea95e2017-09-15 06:47:31 +020011#include "modules/audio_coding/neteq/include/neteq.h"
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +000012
pbos@webrtc.org3ecc1622014-03-07 15:23:34 +000013#include <math.h>
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +000014#include <stdlib.h>
15#include <string.h> // memset
16
turaj@webrtc.org8d1cdaa2014-04-11 18:47:55 +000017#include <algorithm>
kwiberg2d0c3322016-02-14 09:28:33 -080018#include <memory>
turaj@webrtc.org78b41a02013-11-22 20:27:07 +000019#include <set>
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +000020#include <string>
21#include <vector>
22
Mirko Bonadei92ea95e2017-09-15 06:47:31 +020023#include "api/audio_codecs/builtin_audio_decoder_factory.h"
Mirko Bonadei71207422017-09-15 13:58:09 +020024#include "common_types.h" // NOLINT(build/include)
Mirko Bonadei92ea95e2017-09-15 06:47:31 +020025#include "modules/audio_coding/codecs/pcm16b/pcm16b.h"
26#include "modules/audio_coding/neteq/tools/audio_loop.h"
27#include "modules/audio_coding/neteq/tools/rtp_file_source.h"
28#include "modules/include/module_common_types.h"
29#include "rtc_base/flags.h"
30#include "rtc_base/ignore_wundef.h"
31#include "rtc_base/protobuf_utils.h"
32#include "rtc_base/sha1digest.h"
33#include "rtc_base/stringencode.h"
34#include "test/gtest.h"
35#include "test/testsupport/fileutils.h"
Mirko Bonadei71207422017-09-15 13:58:09 +020036#include "typedefs.h" // NOLINT(build/include)
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +000037
minyue5f026d02015-12-16 07:36:04 -080038#ifdef WEBRTC_NETEQ_UNITTEST_BITEXACT
kwiberg77eab702016-09-28 17:42:01 -070039RTC_PUSH_IGNORING_WUNDEF()
minyue5f026d02015-12-16 07:36:04 -080040#ifdef WEBRTC_ANDROID_PLATFORM_BUILD
41#include "external/webrtc/webrtc/modules/audio_coding/neteq/neteq_unittest.pb.h"
42#else
Mirko Bonadei92ea95e2017-09-15 06:47:31 +020043#include "modules/audio_coding/neteq/neteq_unittest.pb.h"
minyue5f026d02015-12-16 07:36:04 -080044#endif
kwiberg77eab702016-09-28 17:42:01 -070045RTC_POP_IGNORING_WUNDEF()
minyue5f026d02015-12-16 07:36:04 -080046#endif
47
turaj@webrtc.orga6101d72013-10-01 22:01:09 +000048DEFINE_bool(gen_ref, false, "Generate reference files.");
49
kwiberg5adaf732016-10-04 09:33:27 -070050namespace webrtc {
51
minyue5f026d02015-12-16 07:36:04 -080052namespace {
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +000053
minyue4f906772016-04-29 11:05:14 -070054const std::string& PlatformChecksum(const std::string& checksum_general,
55 const std::string& checksum_android,
56 const std::string& checksum_win_32,
57 const std::string& checksum_win_64) {
kwiberg77eab702016-09-28 17:42:01 -070058#if defined(WEBRTC_ANDROID)
minyue4f906772016-04-29 11:05:14 -070059 return checksum_android;
kwiberg77eab702016-09-28 17:42:01 -070060#elif defined(WEBRTC_WIN)
minyue4f906772016-04-29 11:05:14 -070061 #ifdef WEBRTC_ARCH_64_BITS
62 return checksum_win_64;
63 #else
64 return checksum_win_32;
65 #endif // WEBRTC_ARCH_64_BITS
66#else
67 return checksum_general;
68#endif // WEBRTC_WIN
69}
70
minyue5f026d02015-12-16 07:36:04 -080071#ifdef WEBRTC_NETEQ_UNITTEST_BITEXACT
72void Convert(const webrtc::NetEqNetworkStatistics& stats_raw,
73 webrtc::neteq_unittest::NetEqNetworkStatistics* stats) {
74 stats->set_current_buffer_size_ms(stats_raw.current_buffer_size_ms);
75 stats->set_preferred_buffer_size_ms(stats_raw.preferred_buffer_size_ms);
76 stats->set_jitter_peaks_found(stats_raw.jitter_peaks_found);
77 stats->set_packet_loss_rate(stats_raw.packet_loss_rate);
minyue5f026d02015-12-16 07:36:04 -080078 stats->set_expand_rate(stats_raw.expand_rate);
79 stats->set_speech_expand_rate(stats_raw.speech_expand_rate);
80 stats->set_preemptive_rate(stats_raw.preemptive_rate);
81 stats->set_accelerate_rate(stats_raw.accelerate_rate);
82 stats->set_secondary_decoded_rate(stats_raw.secondary_decoded_rate);
minyue-webrtc0c3ca752017-08-23 15:59:38 +020083 stats->set_secondary_discarded_rate(stats_raw.secondary_discarded_rate);
minyue5f026d02015-12-16 07:36:04 -080084 stats->set_clockdrift_ppm(stats_raw.clockdrift_ppm);
85 stats->set_added_zero_samples(stats_raw.added_zero_samples);
86 stats->set_mean_waiting_time_ms(stats_raw.mean_waiting_time_ms);
87 stats->set_median_waiting_time_ms(stats_raw.median_waiting_time_ms);
88 stats->set_min_waiting_time_ms(stats_raw.min_waiting_time_ms);
89 stats->set_max_waiting_time_ms(stats_raw.max_waiting_time_ms);
90}
91
92void Convert(const webrtc::RtcpStatistics& stats_raw,
93 webrtc::neteq_unittest::RtcpStatistics* stats) {
94 stats->set_fraction_lost(stats_raw.fraction_lost);
srte186d9c32017-08-04 05:03:53 -070095 stats->set_cumulative_lost(stats_raw.packets_lost);
minyue5f026d02015-12-16 07:36:04 -080096 stats->set_extended_max_sequence_number(
srte186d9c32017-08-04 05:03:53 -070097 stats_raw.extended_highest_sequence_number);
minyue5f026d02015-12-16 07:36:04 -080098 stats->set_jitter(stats_raw.jitter);
99}
100
minyue4f906772016-04-29 11:05:14 -0700101void AddMessage(FILE* file, rtc::MessageDigest* digest,
102 const std::string& message) {
minyue5f026d02015-12-16 07:36:04 -0800103 int32_t size = message.length();
minyue4f906772016-04-29 11:05:14 -0700104 if (file)
105 ASSERT_EQ(1u, fwrite(&size, sizeof(size), 1, file));
106 digest->Update(&size, sizeof(size));
107
108 if (file)
109 ASSERT_EQ(static_cast<size_t>(size),
110 fwrite(message.data(), sizeof(char), size, file));
111 digest->Update(message.data(), sizeof(char) * size);
minyue5f026d02015-12-16 07:36:04 -0800112}
113
minyue5f026d02015-12-16 07:36:04 -0800114#endif // WEBRTC_NETEQ_UNITTEST_BITEXACT
115
henrik.lundin7a926812016-05-12 13:51:28 -0700116void LoadDecoders(webrtc::NetEq* neteq) {
kwiberg5adaf732016-10-04 09:33:27 -0700117 ASSERT_EQ(true,
118 neteq->RegisterPayloadType(0, SdpAudioFormat("pcmu", 8000, 1)));
119 // Use non-SdpAudioFormat argument when registering PCMa, so that we get test
120 // coverage for that as well.
henrik.lundin7a926812016-05-12 13:51:28 -0700121 ASSERT_EQ(0, neteq->RegisterPayloadType(webrtc::NetEqDecoder::kDecoderPCMa,
122 "pcma", 8));
123#ifdef WEBRTC_CODEC_ILBC
kwiberg5adaf732016-10-04 09:33:27 -0700124 ASSERT_EQ(true,
125 neteq->RegisterPayloadType(102, SdpAudioFormat("ilbc", 8000, 1)));
henrik.lundin7a926812016-05-12 13:51:28 -0700126#endif
127#if defined(WEBRTC_CODEC_ISAC) || defined(WEBRTC_CODEC_ISACFX)
kwiberg5adaf732016-10-04 09:33:27 -0700128 ASSERT_EQ(true,
129 neteq->RegisterPayloadType(103, SdpAudioFormat("isac", 16000, 1)));
henrik.lundin7a926812016-05-12 13:51:28 -0700130#endif
131#ifdef WEBRTC_CODEC_ISAC
kwiberg5adaf732016-10-04 09:33:27 -0700132 ASSERT_EQ(true,
133 neteq->RegisterPayloadType(104, SdpAudioFormat("isac", 32000, 1)));
henrik.lundin7a926812016-05-12 13:51:28 -0700134#endif
135#ifdef WEBRTC_CODEC_OPUS
kwiberg5adaf732016-10-04 09:33:27 -0700136 ASSERT_EQ(true,
137 neteq->RegisterPayloadType(
138 111, SdpAudioFormat("opus", 48000, 2, {{"stereo", "0"}})));
henrik.lundin7a926812016-05-12 13:51:28 -0700139#endif
kwiberg5adaf732016-10-04 09:33:27 -0700140 ASSERT_EQ(true,
141 neteq->RegisterPayloadType(93, SdpAudioFormat("L16", 8000, 1)));
142 ASSERT_EQ(true,
143 neteq->RegisterPayloadType(94, SdpAudioFormat("L16", 16000, 1)));
144 ASSERT_EQ(true,
145 neteq->RegisterPayloadType(95, SdpAudioFormat("L16", 32000, 1)));
146 ASSERT_EQ(true,
147 neteq->RegisterPayloadType(13, SdpAudioFormat("cn", 8000, 1)));
148 ASSERT_EQ(true,
149 neteq->RegisterPayloadType(98, SdpAudioFormat("cn", 16000, 1)));
henrik.lundin7a926812016-05-12 13:51:28 -0700150}
minyue5f026d02015-12-16 07:36:04 -0800151} // namespace
152
minyue4f906772016-04-29 11:05:14 -0700153class ResultSink {
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000154 public:
minyue4f906772016-04-29 11:05:14 -0700155 explicit ResultSink(const std::string& output_file);
156 ~ResultSink();
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000157
yujo36b1a5f2017-06-12 12:45:32 -0700158 template<typename T> void AddResult(const T* test_results, size_t length);
minyue4f906772016-04-29 11:05:14 -0700159
160 void AddResult(const NetEqNetworkStatistics& stats);
161 void AddResult(const RtcpStatistics& stats);
162
163 void VerifyChecksum(const std::string& ref_check_sum);
164
165 private:
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000166 FILE* output_fp_;
minyue4f906772016-04-29 11:05:14 -0700167 std::unique_ptr<rtc::MessageDigest> digest_;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000168};
169
minyue4f906772016-04-29 11:05:14 -0700170ResultSink::ResultSink(const std::string &output_file)
171 : output_fp_(nullptr),
172 digest_(new rtc::Sha1Digest()) {
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000173 if (!output_file.empty()) {
174 output_fp_ = fopen(output_file.c_str(), "wb");
175 EXPECT_TRUE(output_fp_ != NULL);
176 }
177}
178
minyue4f906772016-04-29 11:05:14 -0700179ResultSink::~ResultSink() {
180 if (output_fp_)
181 fclose(output_fp_);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000182}
183
yujo36b1a5f2017-06-12 12:45:32 -0700184template<typename T>
185void ResultSink::AddResult(const T* test_results, size_t length) {
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000186 if (output_fp_) {
yujo36b1a5f2017-06-12 12:45:32 -0700187 ASSERT_EQ(length, fwrite(test_results, sizeof(T), length, output_fp_));
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000188 }
yujo36b1a5f2017-06-12 12:45:32 -0700189 digest_->Update(test_results, sizeof(T) * length);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000190}
191
minyue4f906772016-04-29 11:05:14 -0700192void ResultSink::AddResult(const NetEqNetworkStatistics& stats_raw) {
minyue5f026d02015-12-16 07:36:04 -0800193#ifdef WEBRTC_NETEQ_UNITTEST_BITEXACT
minyue5f026d02015-12-16 07:36:04 -0800194 neteq_unittest::NetEqNetworkStatistics stats;
195 Convert(stats_raw, &stats);
196
mbonadei7c2c8432017-04-07 00:59:12 -0700197 ProtoString stats_string;
minyue5f026d02015-12-16 07:36:04 -0800198 ASSERT_TRUE(stats.SerializeToString(&stats_string));
minyue4f906772016-04-29 11:05:14 -0700199 AddMessage(output_fp_, digest_.get(), stats_string);
minyue5f026d02015-12-16 07:36:04 -0800200#else
201 FAIL() << "Writing to reference file requires Proto Buffer.";
202#endif // WEBRTC_NETEQ_UNITTEST_BITEXACT
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000203}
204
minyue4f906772016-04-29 11:05:14 -0700205void ResultSink::AddResult(const RtcpStatistics& stats_raw) {
minyue5f026d02015-12-16 07:36:04 -0800206#ifdef WEBRTC_NETEQ_UNITTEST_BITEXACT
minyue5f026d02015-12-16 07:36:04 -0800207 neteq_unittest::RtcpStatistics stats;
208 Convert(stats_raw, &stats);
209
mbonadei7c2c8432017-04-07 00:59:12 -0700210 ProtoString stats_string;
minyue5f026d02015-12-16 07:36:04 -0800211 ASSERT_TRUE(stats.SerializeToString(&stats_string));
minyue4f906772016-04-29 11:05:14 -0700212 AddMessage(output_fp_, digest_.get(), stats_string);
minyue5f026d02015-12-16 07:36:04 -0800213#else
214 FAIL() << "Writing to reference file requires Proto Buffer.";
215#endif // WEBRTC_NETEQ_UNITTEST_BITEXACT
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000216}
217
minyue4f906772016-04-29 11:05:14 -0700218void ResultSink::VerifyChecksum(const std::string& checksum) {
219 std::vector<char> buffer;
220 buffer.resize(digest_->Size());
221 digest_->Finish(&buffer[0], buffer.size());
222 const std::string result = rtc::hex_encode(&buffer[0], digest_->Size());
223 EXPECT_EQ(checksum, result);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000224}
225
226class NetEqDecodingTest : public ::testing::Test {
227 protected:
228 // NetEQ must be polled for data once every 10 ms. Thus, neither of the
229 // constants below can be changed.
230 static const int kTimeStepMs = 10;
Peter Kastingdce40cf2015-08-24 14:52:23 -0700231 static const size_t kBlockSize8kHz = kTimeStepMs * 8;
232 static const size_t kBlockSize16kHz = kTimeStepMs * 16;
233 static const size_t kBlockSize32kHz = kTimeStepMs * 32;
minyue93c08b72015-12-22 09:57:41 -0800234 static const size_t kBlockSize48kHz = kTimeStepMs * 48;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000235 static const int kInitSampleRateHz = 8000;
236
237 NetEqDecodingTest();
238 virtual void SetUp();
239 virtual void TearDown();
240 void SelectDecoders(NetEqDecoder* used_codec);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000241 void OpenInputFile(const std::string &rtp_file);
henrik.lundin6d8e0112016-03-04 10:34:21 -0800242 void Process();
minyue5f026d02015-12-16 07:36:04 -0800243
henrik.lundin@webrtc.org4e4b0982014-08-11 14:48:49 +0000244 void DecodeAndCompare(const std::string& rtp_file,
minyue4f906772016-04-29 11:05:14 -0700245 const std::string& output_checksum,
246 const std::string& network_stats_checksum,
247 const std::string& rtcp_stats_checksum,
248 bool gen_ref);
minyue5f026d02015-12-16 07:36:04 -0800249
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000250 static void PopulateRtpInfo(int frame_index,
251 int timestamp,
henrik.lundin246ef3e2017-04-24 09:14:32 -0700252 RTPHeader* rtp_info);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000253 static void PopulateCng(int frame_index,
254 int timestamp,
henrik.lundin246ef3e2017-04-24 09:14:32 -0700255 RTPHeader* rtp_info,
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000256 uint8_t* payload,
pkasting@chromium.org4591fbd2014-11-20 22:28:14 +0000257 size_t* payload_len);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000258
turaj@webrtc.org78b41a02013-11-22 20:27:07 +0000259 void WrapTest(uint16_t start_seq_no, uint32_t start_timestamp,
260 const std::set<uint16_t>& drop_seq_numbers,
261 bool expect_seq_no_wrap, bool expect_timestamp_wrap);
262
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000263 void LongCngWithClockDrift(double drift_factor,
264 double network_freeze_ms,
265 bool pull_audio_during_freeze,
266 int delay_tolerance_ms,
267 int max_time_to_speech_ms);
268
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +0000269 void DuplicateCng();
henrik.lundin@webrtc.orgfcfc6a92014-02-13 11:42:28 +0000270
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000271 NetEq* neteq_;
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +0000272 NetEq::Config config_;
kwiberg2d0c3322016-02-14 09:28:33 -0800273 std::unique_ptr<test::RtpFileSource> rtp_source_;
274 std::unique_ptr<test::Packet> packet_;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000275 unsigned int sim_clock_;
henrik.lundin6d8e0112016-03-04 10:34:21 -0800276 AudioFrame out_frame_;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000277 int output_sample_rate_;
turaj@webrtc.org8d1cdaa2014-04-11 18:47:55 +0000278 int algorithmic_delay_ms_;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000279};
280
281// Allocating the static const so that it can be passed by reference.
282const int NetEqDecodingTest::kTimeStepMs;
Peter Kastingdce40cf2015-08-24 14:52:23 -0700283const size_t NetEqDecodingTest::kBlockSize8kHz;
284const size_t NetEqDecodingTest::kBlockSize16kHz;
285const size_t NetEqDecodingTest::kBlockSize32kHz;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000286const int NetEqDecodingTest::kInitSampleRateHz;
287
288NetEqDecodingTest::NetEqDecodingTest()
289 : neteq_(NULL),
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +0000290 config_(),
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000291 sim_clock_(0),
turaj@webrtc.org8d1cdaa2014-04-11 18:47:55 +0000292 output_sample_rate_(kInitSampleRateHz),
293 algorithmic_delay_ms_(0) {
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +0000294 config_.sample_rate_hz = kInitSampleRateHz;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000295}
296
297void NetEqDecodingTest::SetUp() {
ossue3525782016-05-25 07:37:43 -0700298 neteq_ = NetEq::Create(config_, CreateBuiltinAudioDecoderFactory());
turaj@webrtc.org8d1cdaa2014-04-11 18:47:55 +0000299 NetEqNetworkStatistics stat;
300 ASSERT_EQ(0, neteq_->NetworkStatistics(&stat));
301 algorithmic_delay_ms_ = stat.current_buffer_size_ms;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000302 ASSERT_TRUE(neteq_);
henrik.lundin7a926812016-05-12 13:51:28 -0700303 LoadDecoders(neteq_);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000304}
305
306void NetEqDecodingTest::TearDown() {
307 delete neteq_;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000308}
309
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000310void NetEqDecodingTest::OpenInputFile(const std::string &rtp_file) {
henrik.lundin@webrtc.org966a7082014-11-17 09:08:38 +0000311 rtp_source_.reset(test::RtpFileSource::Create(rtp_file));
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000312}
313
henrik.lundin6d8e0112016-03-04 10:34:21 -0800314void NetEqDecodingTest::Process() {
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000315 // Check if time to receive.
henrik.lundin@webrtc.org966a7082014-11-17 09:08:38 +0000316 while (packet_ && sim_clock_ >= packet_->time_ms()) {
317 if (packet_->payload_length_bytes() > 0) {
ivoc72c08ed2016-01-20 07:26:24 -0800318#ifndef WEBRTC_CODEC_ISAC
319 // Ignore payload type 104 (iSAC-swb) if ISAC is not supported.
henrik.lundin246ef3e2017-04-24 09:14:32 -0700320 if (packet_->header().payloadType != 104)
ivoc72c08ed2016-01-20 07:26:24 -0800321#endif
Henrik Lundin70c09bd2017-04-24 15:56:56 +0200322 ASSERT_EQ(0,
323 neteq_->InsertPacket(
henrik.lundin246ef3e2017-04-24 09:14:32 -0700324 packet_->header(),
Henrik Lundin70c09bd2017-04-24 15:56:56 +0200325 rtc::ArrayView<const uint8_t>(
326 packet_->payload(), packet_->payload_length_bytes()),
327 static_cast<uint32_t>(packet_->time_ms() *
328 (output_sample_rate_ / 1000))));
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000329 }
330 // Get next packet.
henrik.lundin46ba49c2016-05-24 22:50:47 -0700331 packet_ = rtp_source_->NextPacket();
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000332 }
333
henrik.lundin@webrtc.orge1d468c2013-01-30 07:37:20 +0000334 // Get audio from NetEq.
henrik.lundin7a926812016-05-12 13:51:28 -0700335 bool muted;
336 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
337 ASSERT_FALSE(muted);
henrik.lundin6d8e0112016-03-04 10:34:21 -0800338 ASSERT_TRUE((out_frame_.samples_per_channel_ == kBlockSize8kHz) ||
339 (out_frame_.samples_per_channel_ == kBlockSize16kHz) ||
340 (out_frame_.samples_per_channel_ == kBlockSize32kHz) ||
341 (out_frame_.samples_per_channel_ == kBlockSize48kHz));
342 output_sample_rate_ = out_frame_.sample_rate_hz_;
henrik.lundind89814b2015-11-23 06:49:25 -0800343 EXPECT_EQ(output_sample_rate_, neteq_->last_output_sample_rate_hz());
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000344
345 // Increase time.
346 sim_clock_ += kTimeStepMs;
347}
348
minyue4f906772016-04-29 11:05:14 -0700349void NetEqDecodingTest::DecodeAndCompare(
350 const std::string& rtp_file,
351 const std::string& output_checksum,
352 const std::string& network_stats_checksum,
353 const std::string& rtcp_stats_checksum,
354 bool gen_ref) {
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000355 OpenInputFile(rtp_file);
356
minyue4f906772016-04-29 11:05:14 -0700357 std::string ref_out_file =
358 gen_ref ? webrtc::test::OutputPath() + "neteq_universal_ref.pcm" : "";
359 ResultSink output(ref_out_file);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000360
minyue4f906772016-04-29 11:05:14 -0700361 std::string stat_out_file =
362 gen_ref ? webrtc::test::OutputPath() + "neteq_network_stats.dat" : "";
363 ResultSink network_stats(stat_out_file);
henrik.lundin@webrtc.org4e4b0982014-08-11 14:48:49 +0000364
minyue4f906772016-04-29 11:05:14 -0700365 std::string rtcp_out_file =
366 gen_ref ? webrtc::test::OutputPath() + "neteq_rtcp_stats.dat" : "";
367 ResultSink rtcp_stats(rtcp_out_file);
henrik.lundin@webrtc.org4e4b0982014-08-11 14:48:49 +0000368
henrik.lundin46ba49c2016-05-24 22:50:47 -0700369 packet_ = rtp_source_->NextPacket();
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000370 int i = 0;
Henrik Lundinac0a5032017-09-25 12:22:46 +0200371 uint64_t last_concealed_samples = 0;
372 uint64_t last_total_samples_received = 0;
henrik.lundin@webrtc.org966a7082014-11-17 09:08:38 +0000373 while (packet_) {
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000374 std::ostringstream ss;
375 ss << "Lap number " << i++ << " in DecodeAndCompare while loop";
376 SCOPED_TRACE(ss.str()); // Print out the parameter values on failure.
henrik.lundin6d8e0112016-03-04 10:34:21 -0800377 ASSERT_NO_FATAL_FAILURE(Process());
minyue4f906772016-04-29 11:05:14 -0700378 ASSERT_NO_FATAL_FAILURE(output.AddResult(
yujo36b1a5f2017-06-12 12:45:32 -0700379 out_frame_.data(), out_frame_.samples_per_channel_));
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000380
381 // Query the network statistics API once per second
382 if (sim_clock_ % 1000 == 0) {
383 // Process NetworkStatistics.
minyue4f906772016-04-29 11:05:14 -0700384 NetEqNetworkStatistics current_network_stats;
385 ASSERT_EQ(0, neteq_->NetworkStatistics(&current_network_stats));
386 ASSERT_NO_FATAL_FAILURE(network_stats.AddResult(current_network_stats));
387
henrik.lundin9c3efd02015-08-27 13:12:22 -0700388 // Compare with CurrentDelay, which should be identical.
minyue4f906772016-04-29 11:05:14 -0700389 EXPECT_EQ(current_network_stats.current_buffer_size_ms,
390 neteq_->CurrentDelayMs());
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000391
Henrik Lundinac0a5032017-09-25 12:22:46 +0200392 // Verify that liftime stats and network stats report similar loss
393 // concealment rates.
394 auto lifetime_stats = neteq_->GetLifetimeStatistics();
395 const uint64_t delta_concealed_samples =
396 lifetime_stats.concealed_samples - last_concealed_samples;
397 last_concealed_samples = lifetime_stats.concealed_samples;
398 const uint64_t delta_total_samples_received =
399 lifetime_stats.total_samples_received - last_total_samples_received;
400 last_total_samples_received = lifetime_stats.total_samples_received;
401 // The tolerance is 1% but expressed in Q14.
402 EXPECT_NEAR(
403 (delta_concealed_samples << 14) / delta_total_samples_received,
404 current_network_stats.expand_rate, (2 << 14) / 100.0);
405
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000406 // Process RTCPstat.
minyue4f906772016-04-29 11:05:14 -0700407 RtcpStatistics current_rtcp_stats;
408 neteq_->GetRtcpStatistics(&current_rtcp_stats);
409 ASSERT_NO_FATAL_FAILURE(rtcp_stats.AddResult(current_rtcp_stats));
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000410 }
411 }
minyue4f906772016-04-29 11:05:14 -0700412
413 SCOPED_TRACE("Check output audio.");
414 output.VerifyChecksum(output_checksum);
415 SCOPED_TRACE("Check network stats.");
416 network_stats.VerifyChecksum(network_stats_checksum);
417 SCOPED_TRACE("Check rtcp stats.");
418 rtcp_stats.VerifyChecksum(rtcp_stats_checksum);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000419}
420
421void NetEqDecodingTest::PopulateRtpInfo(int frame_index,
422 int timestamp,
henrik.lundin246ef3e2017-04-24 09:14:32 -0700423 RTPHeader* rtp_info) {
424 rtp_info->sequenceNumber = frame_index;
425 rtp_info->timestamp = timestamp;
426 rtp_info->ssrc = 0x1234; // Just an arbitrary SSRC.
427 rtp_info->payloadType = 94; // PCM16b WB codec.
428 rtp_info->markerBit = 0;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000429}
430
431void NetEqDecodingTest::PopulateCng(int frame_index,
432 int timestamp,
henrik.lundin246ef3e2017-04-24 09:14:32 -0700433 RTPHeader* rtp_info,
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000434 uint8_t* payload,
pkasting@chromium.org4591fbd2014-11-20 22:28:14 +0000435 size_t* payload_len) {
henrik.lundin246ef3e2017-04-24 09:14:32 -0700436 rtp_info->sequenceNumber = frame_index;
437 rtp_info->timestamp = timestamp;
438 rtp_info->ssrc = 0x1234; // Just an arbitrary SSRC.
439 rtp_info->payloadType = 98; // WB CNG.
440 rtp_info->markerBit = 0;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000441 payload[0] = 64; // Noise level -64 dBov, quite arbitrarily chosen.
442 *payload_len = 1; // Only noise level, no spectral parameters.
443}
444
ivoc72c08ed2016-01-20 07:26:24 -0800445#if !defined(WEBRTC_IOS) && defined(WEBRTC_NETEQ_UNITTEST_BITEXACT) && \
446 (defined(WEBRTC_CODEC_ISAC) || defined(WEBRTC_CODEC_ISACFX)) && \
447 defined(WEBRTC_CODEC_ILBC) && defined(WEBRTC_CODEC_G722) && \
pbosc7a65692016-05-06 12:50:04 -0700448 !defined(WEBRTC_ARCH_ARM64)
minyue5f026d02015-12-16 07:36:04 -0800449#define MAYBE_TestBitExactness TestBitExactness
kwiberg98ab3a42015-09-30 21:54:21 -0700450#else
minyue5f026d02015-12-16 07:36:04 -0800451#define MAYBE_TestBitExactness DISABLED_TestBitExactness
kwiberg98ab3a42015-09-30 21:54:21 -0700452#endif
minyue5f026d02015-12-16 07:36:04 -0800453TEST_F(NetEqDecodingTest, MAYBE_TestBitExactness) {
minyue49c454e2016-01-08 11:30:14 -0800454 const std::string input_rtp_file =
455 webrtc::test::ResourcePath("audio_coding/neteq_universal_new", "rtp");
henrik.lundin@webrtc.org4e4b0982014-08-11 14:48:49 +0000456
minyue4f906772016-04-29 11:05:14 -0700457 const std::string output_checksum = PlatformChecksum(
soren9f2c18e2017-04-10 02:22:46 -0700458 "09fa7646e2ad032a0b156177b95f09012430f81f",
459 "1c64eb8b55ce8878676c6a1e6ddd78f48de0668b",
460 "09fa7646e2ad032a0b156177b95f09012430f81f",
461 "759fef89a5de52bd17e733dc255c671ce86be909");
minyue4f906772016-04-29 11:05:14 -0700462
henrik.lundin2979f552017-05-05 05:04:16 -0700463 const std::string network_stats_checksum =
minyue-webrtc0c3ca752017-08-23 15:59:38 +0200464 PlatformChecksum("5b4262ca328e5f066af5d34f3380521583dd20de",
465 "80235b6d727281203acb63b98f9a9e85d95f7ec0",
466 "5b4262ca328e5f066af5d34f3380521583dd20de",
467 "5b4262ca328e5f066af5d34f3380521583dd20de");
minyue4f906772016-04-29 11:05:14 -0700468
469 const std::string rtcp_stats_checksum = PlatformChecksum(
470 "b8880bf9fed2487efbddcb8d94b9937a29ae521d",
471 "f3f7b3d3e71d7e635240b5373b57df6a7e4ce9d4",
472 "b8880bf9fed2487efbddcb8d94b9937a29ae521d",
473 "b8880bf9fed2487efbddcb8d94b9937a29ae521d");
474
475 DecodeAndCompare(input_rtp_file,
476 output_checksum,
477 network_stats_checksum,
478 rtcp_stats_checksum,
oprypin9b2f20c2017-08-29 05:51:57 -0700479 FLAG_gen_ref);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000480}
481
minyue93c08b72015-12-22 09:57:41 -0800482#if !defined(WEBRTC_IOS) && !defined(WEBRTC_ANDROID) && \
483 defined(WEBRTC_NETEQ_UNITTEST_BITEXACT) && \
minyue-webrtc516711c2017-07-27 17:45:49 +0200484 defined(WEBRTC_CODEC_OPUS)
minyue93c08b72015-12-22 09:57:41 -0800485#define MAYBE_TestOpusBitExactness TestOpusBitExactness
486#else
487#define MAYBE_TestOpusBitExactness DISABLED_TestOpusBitExactness
488#endif
minyue-webrtcadb58b82017-07-26 17:59:59 +0200489TEST_F(NetEqDecodingTest, MAYBE_TestOpusBitExactness) {
minyue93c08b72015-12-22 09:57:41 -0800490 const std::string input_rtp_file =
491 webrtc::test::ResourcePath("audio_coding/neteq_opus", "rtp");
minyue93c08b72015-12-22 09:57:41 -0800492
minyue4f906772016-04-29 11:05:14 -0700493 const std::string output_checksum = PlatformChecksum(
minyue-webrtcadb58b82017-07-26 17:59:59 +0200494 "721e1e0c6effe4b2401536a4eef11512c9fb709c",
495 "721e1e0c6effe4b2401536a4eef11512c9fb709c",
496 "721e1e0c6effe4b2401536a4eef11512c9fb709c",
497 "721e1e0c6effe4b2401536a4eef11512c9fb709c");
minyue4f906772016-04-29 11:05:14 -0700498
henrik.lundin2979f552017-05-05 05:04:16 -0700499 const std::string network_stats_checksum =
minyue-webrtc0c3ca752017-08-23 15:59:38 +0200500 PlatformChecksum("4e749c46e2611877120ac7a20cbbe555cfbd70ea",
501 "4e749c46e2611877120ac7a20cbbe555cfbd70ea",
502 "4e749c46e2611877120ac7a20cbbe555cfbd70ea",
503 "4e749c46e2611877120ac7a20cbbe555cfbd70ea");
minyue4f906772016-04-29 11:05:14 -0700504
505 const std::string rtcp_stats_checksum = PlatformChecksum(
506 "e37c797e3de6a64dda88c9ade7a013d022a2e1e0",
507 "e37c797e3de6a64dda88c9ade7a013d022a2e1e0",
508 "e37c797e3de6a64dda88c9ade7a013d022a2e1e0",
509 "e37c797e3de6a64dda88c9ade7a013d022a2e1e0");
510
511 DecodeAndCompare(input_rtp_file,
512 output_checksum,
513 network_stats_checksum,
514 rtcp_stats_checksum,
oprypin9b2f20c2017-08-29 05:51:57 -0700515 FLAG_gen_ref);
minyue93c08b72015-12-22 09:57:41 -0800516}
517
henrik.lundin@webrtc.org7cbc4f92014-10-07 06:37:39 +0000518// Use fax mode to avoid time-scaling. This is to simplify the testing of
519// packet waiting times in the packet buffer.
520class NetEqDecodingTestFaxMode : public NetEqDecodingTest {
521 protected:
522 NetEqDecodingTestFaxMode() : NetEqDecodingTest() {
523 config_.playout_mode = kPlayoutFax;
524 }
525};
526
527TEST_F(NetEqDecodingTestFaxMode, TestFrameWaitingTimeStatistics) {
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000528 // Insert 30 dummy packets at once. Each packet contains 10 ms 16 kHz audio.
529 size_t num_frames = 30;
pkasting@chromium.org4591fbd2014-11-20 22:28:14 +0000530 const size_t kSamples = 10 * 16;
531 const size_t kPayloadBytes = kSamples * 2;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000532 for (size_t i = 0; i < num_frames; ++i) {
kwibergee2bac22015-11-11 10:34:00 -0800533 const uint8_t payload[kPayloadBytes] = {0};
henrik.lundin246ef3e2017-04-24 09:14:32 -0700534 RTPHeader rtp_info;
535 rtp_info.sequenceNumber = i;
536 rtp_info.timestamp = i * kSamples;
537 rtp_info.ssrc = 0x1234; // Just an arbitrary SSRC.
538 rtp_info.payloadType = 94; // PCM16b WB codec.
539 rtp_info.markerBit = 0;
540 ASSERT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000541 }
542 // Pull out all data.
543 for (size_t i = 0; i < num_frames; ++i) {
henrik.lundin7a926812016-05-12 13:51:28 -0700544 bool muted;
545 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -0800546 ASSERT_EQ(kBlockSize16kHz, out_frame_.samples_per_channel_);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000547 }
548
Henrik Lundin1bb8cf82015-08-25 13:08:04 +0200549 NetEqNetworkStatistics stats;
550 EXPECT_EQ(0, neteq_->NetworkStatistics(&stats));
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000551 // Since all frames are dumped into NetEQ at once, but pulled out with 10 ms
552 // spacing (per definition), we expect the delay to increase with 10 ms for
Henrik Lundin1bb8cf82015-08-25 13:08:04 +0200553 // each packet. Thus, we are calculating the statistics for a series from 10
554 // to 300, in steps of 10 ms.
555 EXPECT_EQ(155, stats.mean_waiting_time_ms);
556 EXPECT_EQ(155, stats.median_waiting_time_ms);
557 EXPECT_EQ(10, stats.min_waiting_time_ms);
558 EXPECT_EQ(300, stats.max_waiting_time_ms);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000559
560 // Check statistics again and make sure it's been reset.
Henrik Lundin1bb8cf82015-08-25 13:08:04 +0200561 EXPECT_EQ(0, neteq_->NetworkStatistics(&stats));
562 EXPECT_EQ(-1, stats.mean_waiting_time_ms);
563 EXPECT_EQ(-1, stats.median_waiting_time_ms);
564 EXPECT_EQ(-1, stats.min_waiting_time_ms);
565 EXPECT_EQ(-1, stats.max_waiting_time_ms);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000566}
567
henrik.lundin@webrtc.orgb4e80e02014-05-15 07:14:00 +0000568TEST_F(NetEqDecodingTest, TestAverageInterArrivalTimeNegative) {
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000569 const int kNumFrames = 3000; // Needed for convergence.
570 int frame_index = 0;
pkasting@chromium.org4591fbd2014-11-20 22:28:14 +0000571 const size_t kSamples = 10 * 16;
572 const size_t kPayloadBytes = kSamples * 2;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000573 while (frame_index < kNumFrames) {
574 // Insert one packet each time, except every 10th time where we insert two
575 // packets at once. This will create a negative clock-drift of approx. 10%.
576 int num_packets = (frame_index % 10 == 0 ? 2 : 1);
577 for (int n = 0; n < num_packets; ++n) {
578 uint8_t payload[kPayloadBytes] = {0};
henrik.lundin246ef3e2017-04-24 09:14:32 -0700579 RTPHeader rtp_info;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000580 PopulateRtpInfo(frame_index, frame_index * kSamples, &rtp_info);
henrik.lundin246ef3e2017-04-24 09:14:32 -0700581 ASSERT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000582 ++frame_index;
583 }
584
585 // Pull out data once.
henrik.lundin7a926812016-05-12 13:51:28 -0700586 bool muted;
587 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -0800588 ASSERT_EQ(kBlockSize16kHz, out_frame_.samples_per_channel_);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000589 }
590
591 NetEqNetworkStatistics network_stats;
592 ASSERT_EQ(0, neteq_->NetworkStatistics(&network_stats));
henrik.lundin0d838572016-10-13 03:35:55 -0700593 EXPECT_EQ(-103192, network_stats.clockdrift_ppm);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000594}
595
henrik.lundin@webrtc.orgb4e80e02014-05-15 07:14:00 +0000596TEST_F(NetEqDecodingTest, TestAverageInterArrivalTimePositive) {
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000597 const int kNumFrames = 5000; // Needed for convergence.
598 int frame_index = 0;
pkasting@chromium.org4591fbd2014-11-20 22:28:14 +0000599 const size_t kSamples = 10 * 16;
600 const size_t kPayloadBytes = kSamples * 2;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000601 for (int i = 0; i < kNumFrames; ++i) {
602 // Insert one packet each time, except every 10th time where we don't insert
603 // any packet. This will create a positive clock-drift of approx. 11%.
604 int num_packets = (i % 10 == 9 ? 0 : 1);
605 for (int n = 0; n < num_packets; ++n) {
606 uint8_t payload[kPayloadBytes] = {0};
henrik.lundin246ef3e2017-04-24 09:14:32 -0700607 RTPHeader rtp_info;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000608 PopulateRtpInfo(frame_index, frame_index * kSamples, &rtp_info);
henrik.lundin246ef3e2017-04-24 09:14:32 -0700609 ASSERT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000610 ++frame_index;
611 }
612
613 // Pull out data once.
henrik.lundin7a926812016-05-12 13:51:28 -0700614 bool muted;
615 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -0800616 ASSERT_EQ(kBlockSize16kHz, out_frame_.samples_per_channel_);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000617 }
618
619 NetEqNetworkStatistics network_stats;
620 ASSERT_EQ(0, neteq_->NetworkStatistics(&network_stats));
henrik.lundin0d838572016-10-13 03:35:55 -0700621 EXPECT_EQ(110953, network_stats.clockdrift_ppm);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000622}
623
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000624void NetEqDecodingTest::LongCngWithClockDrift(double drift_factor,
625 double network_freeze_ms,
626 bool pull_audio_during_freeze,
627 int delay_tolerance_ms,
628 int max_time_to_speech_ms) {
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000629 uint16_t seq_no = 0;
630 uint32_t timestamp = 0;
631 const int kFrameSizeMs = 30;
pkasting@chromium.org4591fbd2014-11-20 22:28:14 +0000632 const size_t kSamples = kFrameSizeMs * 16;
633 const size_t kPayloadBytes = kSamples * 2;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000634 double next_input_time_ms = 0.0;
635 double t_ms;
henrik.lundin7a926812016-05-12 13:51:28 -0700636 bool muted;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000637
638 // Insert speech for 5 seconds.
639 const int kSpeechDurationMs = 5000;
640 for (t_ms = 0; t_ms < kSpeechDurationMs; t_ms += 10) {
641 // Each turn in this for loop is 10 ms.
642 while (next_input_time_ms <= t_ms) {
643 // Insert one 30 ms speech frame.
644 uint8_t payload[kPayloadBytes] = {0};
henrik.lundin246ef3e2017-04-24 09:14:32 -0700645 RTPHeader rtp_info;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000646 PopulateRtpInfo(seq_no, timestamp, &rtp_info);
henrik.lundin246ef3e2017-04-24 09:14:32 -0700647 ASSERT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000648 ++seq_no;
649 timestamp += kSamples;
henrik.lundin@webrtc.orgfcfc6a92014-02-13 11:42:28 +0000650 next_input_time_ms += static_cast<double>(kFrameSizeMs) * drift_factor;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000651 }
652 // Pull out data once.
henrik.lundin7a926812016-05-12 13:51:28 -0700653 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -0800654 ASSERT_EQ(kBlockSize16kHz, out_frame_.samples_per_channel_);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000655 }
656
henrik.lundin55480f52016-03-08 02:37:57 -0800657 EXPECT_EQ(AudioFrame::kNormalSpeech, out_frame_.speech_type_);
henrik.lundin114c1b32017-04-26 07:47:32 -0700658 rtc::Optional<uint32_t> playout_timestamp = neteq_->GetPlayoutTimestamp();
henrik.lundin0d96ab72016-04-06 12:28:26 -0700659 ASSERT_TRUE(playout_timestamp);
660 int32_t delay_before = timestamp - *playout_timestamp;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000661
662 // Insert CNG for 1 minute (= 60000 ms).
663 const int kCngPeriodMs = 100;
664 const int kCngPeriodSamples = kCngPeriodMs * 16; // Period in 16 kHz samples.
665 const int kCngDurationMs = 60000;
666 for (; t_ms < kSpeechDurationMs + kCngDurationMs; t_ms += 10) {
667 // Each turn in this for loop is 10 ms.
668 while (next_input_time_ms <= t_ms) {
669 // Insert one CNG frame each 100 ms.
670 uint8_t payload[kPayloadBytes];
pkasting@chromium.org4591fbd2014-11-20 22:28:14 +0000671 size_t payload_len;
henrik.lundin246ef3e2017-04-24 09:14:32 -0700672 RTPHeader rtp_info;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000673 PopulateCng(seq_no, timestamp, &rtp_info, payload, &payload_len);
kwibergee2bac22015-11-11 10:34:00 -0800674 ASSERT_EQ(0, neteq_->InsertPacket(
henrik.lundin246ef3e2017-04-24 09:14:32 -0700675 rtp_info,
kwibergee2bac22015-11-11 10:34:00 -0800676 rtc::ArrayView<const uint8_t>(payload, payload_len), 0));
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000677 ++seq_no;
678 timestamp += kCngPeriodSamples;
henrik.lundin@webrtc.orgfcfc6a92014-02-13 11:42:28 +0000679 next_input_time_ms += static_cast<double>(kCngPeriodMs) * drift_factor;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000680 }
681 // Pull out data once.
henrik.lundin7a926812016-05-12 13:51:28 -0700682 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -0800683 ASSERT_EQ(kBlockSize16kHz, out_frame_.samples_per_channel_);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000684 }
685
henrik.lundin55480f52016-03-08 02:37:57 -0800686 EXPECT_EQ(AudioFrame::kCNG, out_frame_.speech_type_);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000687
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000688 if (network_freeze_ms > 0) {
689 // First keep pulling audio for |network_freeze_ms| without inserting
690 // any data, then insert CNG data corresponding to |network_freeze_ms|
691 // without pulling any output audio.
692 const double loop_end_time = t_ms + network_freeze_ms;
693 for (; t_ms < loop_end_time; t_ms += 10) {
694 // Pull out data once.
henrik.lundin7a926812016-05-12 13:51:28 -0700695 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -0800696 ASSERT_EQ(kBlockSize16kHz, out_frame_.samples_per_channel_);
henrik.lundin55480f52016-03-08 02:37:57 -0800697 EXPECT_EQ(AudioFrame::kCNG, out_frame_.speech_type_);
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000698 }
699 bool pull_once = pull_audio_during_freeze;
700 // If |pull_once| is true, GetAudio will be called once half-way through
701 // the network recovery period.
702 double pull_time_ms = (t_ms + next_input_time_ms) / 2;
703 while (next_input_time_ms <= t_ms) {
704 if (pull_once && next_input_time_ms >= pull_time_ms) {
705 pull_once = false;
706 // Pull out data once.
henrik.lundin7a926812016-05-12 13:51:28 -0700707 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -0800708 ASSERT_EQ(kBlockSize16kHz, out_frame_.samples_per_channel_);
henrik.lundin55480f52016-03-08 02:37:57 -0800709 EXPECT_EQ(AudioFrame::kCNG, out_frame_.speech_type_);
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000710 t_ms += 10;
711 }
712 // Insert one CNG frame each 100 ms.
713 uint8_t payload[kPayloadBytes];
pkasting@chromium.org4591fbd2014-11-20 22:28:14 +0000714 size_t payload_len;
henrik.lundin246ef3e2017-04-24 09:14:32 -0700715 RTPHeader rtp_info;
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000716 PopulateCng(seq_no, timestamp, &rtp_info, payload, &payload_len);
kwibergee2bac22015-11-11 10:34:00 -0800717 ASSERT_EQ(0, neteq_->InsertPacket(
henrik.lundin246ef3e2017-04-24 09:14:32 -0700718 rtp_info,
kwibergee2bac22015-11-11 10:34:00 -0800719 rtc::ArrayView<const uint8_t>(payload, payload_len), 0));
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000720 ++seq_no;
721 timestamp += kCngPeriodSamples;
722 next_input_time_ms += kCngPeriodMs * drift_factor;
723 }
724 }
725
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000726 // Insert speech again until output type is speech.
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000727 double speech_restart_time_ms = t_ms;
henrik.lundin55480f52016-03-08 02:37:57 -0800728 while (out_frame_.speech_type_ != AudioFrame::kNormalSpeech) {
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000729 // Each turn in this for loop is 10 ms.
730 while (next_input_time_ms <= t_ms) {
731 // Insert one 30 ms speech frame.
732 uint8_t payload[kPayloadBytes] = {0};
henrik.lundin246ef3e2017-04-24 09:14:32 -0700733 RTPHeader rtp_info;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000734 PopulateRtpInfo(seq_no, timestamp, &rtp_info);
henrik.lundin246ef3e2017-04-24 09:14:32 -0700735 ASSERT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000736 ++seq_no;
737 timestamp += kSamples;
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000738 next_input_time_ms += kFrameSizeMs * drift_factor;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000739 }
740 // Pull out data once.
henrik.lundin7a926812016-05-12 13:51:28 -0700741 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -0800742 ASSERT_EQ(kBlockSize16kHz, out_frame_.samples_per_channel_);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000743 // Increase clock.
744 t_ms += 10;
745 }
746
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000747 // Check that the speech starts again within reasonable time.
748 double time_until_speech_returns_ms = t_ms - speech_restart_time_ms;
749 EXPECT_LT(time_until_speech_returns_ms, max_time_to_speech_ms);
henrik.lundin114c1b32017-04-26 07:47:32 -0700750 playout_timestamp = neteq_->GetPlayoutTimestamp();
henrik.lundin0d96ab72016-04-06 12:28:26 -0700751 ASSERT_TRUE(playout_timestamp);
752 int32_t delay_after = timestamp - *playout_timestamp;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000753 // Compare delay before and after, and make sure it differs less than 20 ms.
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000754 EXPECT_LE(delay_after, delay_before + delay_tolerance_ms * 16);
755 EXPECT_GE(delay_after, delay_before - delay_tolerance_ms * 16);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000756}
757
henrik.lundin@webrtc.orgb4e80e02014-05-15 07:14:00 +0000758TEST_F(NetEqDecodingTest, LongCngWithNegativeClockDrift) {
henrik.lundin@webrtc.orgfcfc6a92014-02-13 11:42:28 +0000759 // Apply a clock drift of -25 ms / s (sender faster than receiver).
760 const double kDriftFactor = 1000.0 / (1000.0 + 25.0);
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000761 const double kNetworkFreezeTimeMs = 0.0;
762 const bool kGetAudioDuringFreezeRecovery = false;
763 const int kDelayToleranceMs = 20;
764 const int kMaxTimeToSpeechMs = 100;
765 LongCngWithClockDrift(kDriftFactor,
766 kNetworkFreezeTimeMs,
767 kGetAudioDuringFreezeRecovery,
768 kDelayToleranceMs,
769 kMaxTimeToSpeechMs);
henrik.lundin@webrtc.orgfcfc6a92014-02-13 11:42:28 +0000770}
771
henrik.lundin@webrtc.orgb4e80e02014-05-15 07:14:00 +0000772TEST_F(NetEqDecodingTest, LongCngWithPositiveClockDrift) {
henrik.lundin@webrtc.orgfcfc6a92014-02-13 11:42:28 +0000773 // Apply a clock drift of +25 ms / s (sender slower than receiver).
774 const double kDriftFactor = 1000.0 / (1000.0 - 25.0);
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000775 const double kNetworkFreezeTimeMs = 0.0;
776 const bool kGetAudioDuringFreezeRecovery = false;
777 const int kDelayToleranceMs = 20;
778 const int kMaxTimeToSpeechMs = 100;
779 LongCngWithClockDrift(kDriftFactor,
780 kNetworkFreezeTimeMs,
781 kGetAudioDuringFreezeRecovery,
782 kDelayToleranceMs,
783 kMaxTimeToSpeechMs);
784}
785
henrik.lundin@webrtc.orgb4e80e02014-05-15 07:14:00 +0000786TEST_F(NetEqDecodingTest, LongCngWithNegativeClockDriftNetworkFreeze) {
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000787 // Apply a clock drift of -25 ms / s (sender faster than receiver).
788 const double kDriftFactor = 1000.0 / (1000.0 + 25.0);
789 const double kNetworkFreezeTimeMs = 5000.0;
790 const bool kGetAudioDuringFreezeRecovery = false;
791 const int kDelayToleranceMs = 50;
792 const int kMaxTimeToSpeechMs = 200;
793 LongCngWithClockDrift(kDriftFactor,
794 kNetworkFreezeTimeMs,
795 kGetAudioDuringFreezeRecovery,
796 kDelayToleranceMs,
797 kMaxTimeToSpeechMs);
798}
799
henrik.lundin@webrtc.orgb4e80e02014-05-15 07:14:00 +0000800TEST_F(NetEqDecodingTest, LongCngWithPositiveClockDriftNetworkFreeze) {
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000801 // Apply a clock drift of +25 ms / s (sender slower than receiver).
802 const double kDriftFactor = 1000.0 / (1000.0 - 25.0);
803 const double kNetworkFreezeTimeMs = 5000.0;
804 const bool kGetAudioDuringFreezeRecovery = false;
805 const int kDelayToleranceMs = 20;
806 const int kMaxTimeToSpeechMs = 100;
807 LongCngWithClockDrift(kDriftFactor,
808 kNetworkFreezeTimeMs,
809 kGetAudioDuringFreezeRecovery,
810 kDelayToleranceMs,
811 kMaxTimeToSpeechMs);
812}
813
henrik.lundin@webrtc.orgb4e80e02014-05-15 07:14:00 +0000814TEST_F(NetEqDecodingTest, LongCngWithPositiveClockDriftNetworkFreezeExtraPull) {
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000815 // Apply a clock drift of +25 ms / s (sender slower than receiver).
816 const double kDriftFactor = 1000.0 / (1000.0 - 25.0);
817 const double kNetworkFreezeTimeMs = 5000.0;
818 const bool kGetAudioDuringFreezeRecovery = true;
819 const int kDelayToleranceMs = 20;
820 const int kMaxTimeToSpeechMs = 100;
821 LongCngWithClockDrift(kDriftFactor,
822 kNetworkFreezeTimeMs,
823 kGetAudioDuringFreezeRecovery,
824 kDelayToleranceMs,
825 kMaxTimeToSpeechMs);
826}
827
henrik.lundin@webrtc.orgb4e80e02014-05-15 07:14:00 +0000828TEST_F(NetEqDecodingTest, LongCngWithoutClockDrift) {
henrik.lundin@webrtc.org24779fe2014-03-14 12:40:05 +0000829 const double kDriftFactor = 1.0; // No drift.
830 const double kNetworkFreezeTimeMs = 0.0;
831 const bool kGetAudioDuringFreezeRecovery = false;
832 const int kDelayToleranceMs = 10;
833 const int kMaxTimeToSpeechMs = 50;
834 LongCngWithClockDrift(kDriftFactor,
835 kNetworkFreezeTimeMs,
836 kGetAudioDuringFreezeRecovery,
837 kDelayToleranceMs,
838 kMaxTimeToSpeechMs);
henrik.lundin@webrtc.orgfcfc6a92014-02-13 11:42:28 +0000839}
840
henrik.lundin@webrtc.orgb4e80e02014-05-15 07:14:00 +0000841TEST_F(NetEqDecodingTest, UnknownPayloadType) {
pkasting@chromium.org4591fbd2014-11-20 22:28:14 +0000842 const size_t kPayloadBytes = 100;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000843 uint8_t payload[kPayloadBytes] = {0};
henrik.lundin246ef3e2017-04-24 09:14:32 -0700844 RTPHeader rtp_info;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000845 PopulateRtpInfo(0, 0, &rtp_info);
henrik.lundin246ef3e2017-04-24 09:14:32 -0700846 rtp_info.payloadType = 1; // Not registered as a decoder.
847 EXPECT_EQ(NetEq::kFail, neteq_->InsertPacket(rtp_info, payload, 0));
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000848}
849
Peter Boströme2976c82016-01-04 22:44:05 +0100850#if defined(WEBRTC_CODEC_ISAC) || defined(WEBRTC_CODEC_ISACFX)
ivoc72c08ed2016-01-20 07:26:24 -0800851#define MAYBE_DecoderError DecoderError
852#else
853#define MAYBE_DecoderError DISABLED_DecoderError
854#endif
855
Peter Boströme2976c82016-01-04 22:44:05 +0100856TEST_F(NetEqDecodingTest, MAYBE_DecoderError) {
pkasting@chromium.org4591fbd2014-11-20 22:28:14 +0000857 const size_t kPayloadBytes = 100;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000858 uint8_t payload[kPayloadBytes] = {0};
henrik.lundin246ef3e2017-04-24 09:14:32 -0700859 RTPHeader rtp_info;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000860 PopulateRtpInfo(0, 0, &rtp_info);
henrik.lundin246ef3e2017-04-24 09:14:32 -0700861 rtp_info.payloadType = 103; // iSAC, but the payload is invalid.
862 EXPECT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000863 // Set all of |out_data_| to 1, and verify that it was set to 0 by the call
864 // to GetAudio.
yujo36b1a5f2017-06-12 12:45:32 -0700865 int16_t* out_frame_data = out_frame_.mutable_data();
henrik.lundin6d8e0112016-03-04 10:34:21 -0800866 for (size_t i = 0; i < AudioFrame::kMaxDataSizeSamples; ++i) {
yujo36b1a5f2017-06-12 12:45:32 -0700867 out_frame_data[i] = 1;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000868 }
henrik.lundin7a926812016-05-12 13:51:28 -0700869 bool muted;
870 EXPECT_EQ(NetEq::kFail, neteq_->GetAudio(&out_frame_, &muted));
871 ASSERT_FALSE(muted);
ivoc72c08ed2016-01-20 07:26:24 -0800872
yujo36b1a5f2017-06-12 12:45:32 -0700873 // Verify that the first 160 samples are set to 0.
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000874 static const int kExpectedOutputLength = 160; // 10 ms at 16 kHz sample rate.
yujo36b1a5f2017-06-12 12:45:32 -0700875 const int16_t* const_out_frame_data = out_frame_.data();
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000876 for (int i = 0; i < kExpectedOutputLength; ++i) {
877 std::ostringstream ss;
878 ss << "i = " << i;
879 SCOPED_TRACE(ss.str()); // Print out the parameter values on failure.
yujo36b1a5f2017-06-12 12:45:32 -0700880 EXPECT_EQ(0, const_out_frame_data[i]);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000881 }
882}
883
henrik.lundin@webrtc.orgb4e80e02014-05-15 07:14:00 +0000884TEST_F(NetEqDecodingTest, GetAudioBeforeInsertPacket) {
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000885 // Set all of |out_data_| to 1, and verify that it was set to 0 by the call
886 // to GetAudio.
yujo36b1a5f2017-06-12 12:45:32 -0700887 int16_t* out_frame_data = out_frame_.mutable_data();
henrik.lundin6d8e0112016-03-04 10:34:21 -0800888 for (size_t i = 0; i < AudioFrame::kMaxDataSizeSamples; ++i) {
yujo36b1a5f2017-06-12 12:45:32 -0700889 out_frame_data[i] = 1;
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000890 }
henrik.lundin7a926812016-05-12 13:51:28 -0700891 bool muted;
892 EXPECT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
893 ASSERT_FALSE(muted);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000894 // Verify that the first block of samples is set to 0.
895 static const int kExpectedOutputLength =
896 kInitSampleRateHz / 100; // 10 ms at initial sample rate.
yujo36b1a5f2017-06-12 12:45:32 -0700897 const int16_t* const_out_frame_data = out_frame_.data();
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000898 for (int i = 0; i < kExpectedOutputLength; ++i) {
899 std::ostringstream ss;
900 ss << "i = " << i;
901 SCOPED_TRACE(ss.str()); // Print out the parameter values on failure.
yujo36b1a5f2017-06-12 12:45:32 -0700902 EXPECT_EQ(0, const_out_frame_data[i]);
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000903 }
henrik.lundind89814b2015-11-23 06:49:25 -0800904 // Verify that the sample rate did not change from the initial configuration.
905 EXPECT_EQ(config_.sample_rate_hz, neteq_->last_output_sample_rate_hz());
henrik.lundin@webrtc.orgd94659d2013-01-29 12:09:21 +0000906}
turaj@webrtc.orgff43c852013-09-25 00:07:27 +0000907
henrik.lundin@webrtc.org9b8102c2014-08-21 08:27:44 +0000908class NetEqBgnTest : public NetEqDecodingTest {
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +0000909 protected:
henrik.lundin@webrtc.org9b8102c2014-08-21 08:27:44 +0000910 virtual void TestCondition(double sum_squared_noise,
911 bool should_be_faded) = 0;
turaj@webrtc.orgff43c852013-09-25 00:07:27 +0000912
henrik.lundin@webrtc.org9b8102c2014-08-21 08:27:44 +0000913 void CheckBgn(int sampling_rate_hz) {
Peter Kastingdce40cf2015-08-24 14:52:23 -0700914 size_t expected_samples_per_channel = 0;
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +0000915 uint8_t payload_type = 0xFF; // Invalid.
916 if (sampling_rate_hz == 8000) {
917 expected_samples_per_channel = kBlockSize8kHz;
918 payload_type = 93; // PCM 16, 8 kHz.
919 } else if (sampling_rate_hz == 16000) {
920 expected_samples_per_channel = kBlockSize16kHz;
921 payload_type = 94; // PCM 16, 16 kHZ.
922 } else if (sampling_rate_hz == 32000) {
923 expected_samples_per_channel = kBlockSize32kHz;
924 payload_type = 95; // PCM 16, 32 kHz.
925 } else {
926 ASSERT_TRUE(false); // Unsupported test case.
927 }
turaj@webrtc.orgff43c852013-09-25 00:07:27 +0000928
henrik.lundin6d8e0112016-03-04 10:34:21 -0800929 AudioFrame output;
henrik.lundin@webrtc.org9b8102c2014-08-21 08:27:44 +0000930 test::AudioLoop input;
931 // We are using the same 32 kHz input file for all tests, regardless of
932 // |sampling_rate_hz|. The output may sound weird, but the test is still
933 // valid.
934 ASSERT_TRUE(input.Init(
935 webrtc::test::ResourcePath("audio_coding/testfile32kHz", "pcm"),
936 10 * sampling_rate_hz, // Max 10 seconds loop length.
Peter Kastingdce40cf2015-08-24 14:52:23 -0700937 expected_samples_per_channel));
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +0000938
939 // Payload of 10 ms of PCM16 32 kHz.
940 uint8_t payload[kBlockSize32kHz * sizeof(int16_t)];
henrik.lundin246ef3e2017-04-24 09:14:32 -0700941 RTPHeader rtp_info;
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +0000942 PopulateRtpInfo(0, 0, &rtp_info);
henrik.lundin246ef3e2017-04-24 09:14:32 -0700943 rtp_info.payloadType = payload_type;
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +0000944
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +0000945 uint32_t receive_timestamp = 0;
henrik.lundin7a926812016-05-12 13:51:28 -0700946 bool muted;
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +0000947 for (int n = 0; n < 10; ++n) { // Insert few packets and get audio.
kwiberg288886b2015-11-06 01:21:35 -0800948 auto block = input.GetNextBlock();
949 ASSERT_EQ(expected_samples_per_channel, block.size());
950 size_t enc_len_bytes =
951 WebRtcPcm16b_Encode(block.data(), block.size(), payload);
henrik.lundin@webrtc.org9b8102c2014-08-21 08:27:44 +0000952 ASSERT_EQ(enc_len_bytes, expected_samples_per_channel * 2);
953
Henrik Lundin70c09bd2017-04-24 15:56:56 +0200954 ASSERT_EQ(0, neteq_->InsertPacket(
henrik.lundin246ef3e2017-04-24 09:14:32 -0700955 rtp_info,
Henrik Lundin70c09bd2017-04-24 15:56:56 +0200956 rtc::ArrayView<const uint8_t>(payload, enc_len_bytes),
957 receive_timestamp));
henrik.lundin6d8e0112016-03-04 10:34:21 -0800958 output.Reset();
henrik.lundin7a926812016-05-12 13:51:28 -0700959 ASSERT_EQ(0, neteq_->GetAudio(&output, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -0800960 ASSERT_EQ(1u, output.num_channels_);
961 ASSERT_EQ(expected_samples_per_channel, output.samples_per_channel_);
henrik.lundin55480f52016-03-08 02:37:57 -0800962 ASSERT_EQ(AudioFrame::kNormalSpeech, output.speech_type_);
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +0000963
964 // Next packet.
henrik.lundin246ef3e2017-04-24 09:14:32 -0700965 rtp_info.timestamp += expected_samples_per_channel;
966 rtp_info.sequenceNumber++;
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +0000967 receive_timestamp += expected_samples_per_channel;
968 }
969
henrik.lundin6d8e0112016-03-04 10:34:21 -0800970 output.Reset();
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +0000971
972 // Get audio without inserting packets, expecting PLC and PLC-to-CNG. Pull
973 // one frame without checking speech-type. This is the first frame pulled
974 // without inserting any packet, and might not be labeled as PLC.
henrik.lundin7a926812016-05-12 13:51:28 -0700975 ASSERT_EQ(0, neteq_->GetAudio(&output, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -0800976 ASSERT_EQ(1u, output.num_channels_);
977 ASSERT_EQ(expected_samples_per_channel, output.samples_per_channel_);
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +0000978
979 // To be able to test the fading of background noise we need at lease to
980 // pull 611 frames.
981 const int kFadingThreshold = 611;
982
983 // Test several CNG-to-PLC packet for the expected behavior. The number 20
984 // is arbitrary, but sufficiently large to test enough number of frames.
985 const int kNumPlcToCngTestFrames = 20;
986 bool plc_to_cng = false;
987 for (int n = 0; n < kFadingThreshold + kNumPlcToCngTestFrames; ++n) {
henrik.lundin6d8e0112016-03-04 10:34:21 -0800988 output.Reset();
yujo36b1a5f2017-06-12 12:45:32 -0700989 // Set to non-zero.
990 memset(output.mutable_data(), 1, AudioFrame::kMaxDataSizeBytes);
henrik.lundin7a926812016-05-12 13:51:28 -0700991 ASSERT_EQ(0, neteq_->GetAudio(&output, &muted));
992 ASSERT_FALSE(muted);
henrik.lundin6d8e0112016-03-04 10:34:21 -0800993 ASSERT_EQ(1u, output.num_channels_);
994 ASSERT_EQ(expected_samples_per_channel, output.samples_per_channel_);
henrik.lundin55480f52016-03-08 02:37:57 -0800995 if (output.speech_type_ == AudioFrame::kPLCCNG) {
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +0000996 plc_to_cng = true;
997 double sum_squared = 0;
yujo36b1a5f2017-06-12 12:45:32 -0700998 const int16_t* output_data = output.data();
henrik.lundin6d8e0112016-03-04 10:34:21 -0800999 for (size_t k = 0;
1000 k < output.num_channels_ * output.samples_per_channel_; ++k)
yujo36b1a5f2017-06-12 12:45:32 -07001001 sum_squared += output_data[k] * output_data[k];
henrik.lundin@webrtc.org9b8102c2014-08-21 08:27:44 +00001002 TestCondition(sum_squared, n > kFadingThreshold);
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +00001003 } else {
henrik.lundin55480f52016-03-08 02:37:57 -08001004 EXPECT_EQ(AudioFrame::kPLC, output.speech_type_);
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +00001005 }
1006 }
1007 EXPECT_TRUE(plc_to_cng); // Just to be sure that PLC-to-CNG has occurred.
1008 }
1009};
1010
henrik.lundin@webrtc.org9b8102c2014-08-21 08:27:44 +00001011class NetEqBgnTestOn : public NetEqBgnTest {
1012 protected:
1013 NetEqBgnTestOn() : NetEqBgnTest() {
1014 config_.background_noise_mode = NetEq::kBgnOn;
1015 }
1016
1017 void TestCondition(double sum_squared_noise, bool /*should_be_faded*/) {
1018 EXPECT_NE(0, sum_squared_noise);
1019 }
1020};
1021
1022class NetEqBgnTestOff : public NetEqBgnTest {
1023 protected:
1024 NetEqBgnTestOff() : NetEqBgnTest() {
1025 config_.background_noise_mode = NetEq::kBgnOff;
1026 }
1027
1028 void TestCondition(double sum_squared_noise, bool /*should_be_faded*/) {
1029 EXPECT_EQ(0, sum_squared_noise);
1030 }
1031};
1032
1033class NetEqBgnTestFade : public NetEqBgnTest {
1034 protected:
1035 NetEqBgnTestFade() : NetEqBgnTest() {
1036 config_.background_noise_mode = NetEq::kBgnFade;
1037 }
1038
1039 void TestCondition(double sum_squared_noise, bool should_be_faded) {
1040 if (should_be_faded)
1041 EXPECT_EQ(0, sum_squared_noise);
1042 }
1043};
1044
henrika1d34fe92015-06-16 10:04:20 +02001045TEST_F(NetEqBgnTestOn, RunTest) {
henrik.lundin@webrtc.org9b8102c2014-08-21 08:27:44 +00001046 CheckBgn(8000);
1047 CheckBgn(16000);
1048 CheckBgn(32000);
turaj@webrtc.orgff43c852013-09-25 00:07:27 +00001049}
turaj@webrtc.org7b75ac62013-09-26 00:27:56 +00001050
henrika1d34fe92015-06-16 10:04:20 +02001051TEST_F(NetEqBgnTestOff, RunTest) {
henrik.lundin@webrtc.org9b8102c2014-08-21 08:27:44 +00001052 CheckBgn(8000);
1053 CheckBgn(16000);
1054 CheckBgn(32000);
1055}
1056
henrika1d34fe92015-06-16 10:04:20 +02001057TEST_F(NetEqBgnTestFade, RunTest) {
henrik.lundin@webrtc.org9b8102c2014-08-21 08:27:44 +00001058 CheckBgn(8000);
1059 CheckBgn(16000);
1060 CheckBgn(32000);
1061}
henrik.lundin@webrtc.orgea257842014-08-07 12:27:37 +00001062
turaj@webrtc.org78b41a02013-11-22 20:27:07 +00001063void NetEqDecodingTest::WrapTest(uint16_t start_seq_no,
1064 uint32_t start_timestamp,
1065 const std::set<uint16_t>& drop_seq_numbers,
1066 bool expect_seq_no_wrap,
1067 bool expect_timestamp_wrap) {
1068 uint16_t seq_no = start_seq_no;
1069 uint32_t timestamp = start_timestamp;
1070 const int kBlocksPerFrame = 3; // Number of 10 ms blocks per frame.
1071 const int kFrameSizeMs = kBlocksPerFrame * kTimeStepMs;
1072 const int kSamples = kBlockSize16kHz * kBlocksPerFrame;
pkasting@chromium.org4591fbd2014-11-20 22:28:14 +00001073 const size_t kPayloadBytes = kSamples * sizeof(int16_t);
turaj@webrtc.org78b41a02013-11-22 20:27:07 +00001074 double next_input_time_ms = 0.0;
turaj@webrtc.org78b41a02013-11-22 20:27:07 +00001075 uint32_t receive_timestamp = 0;
1076
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001077 // Insert speech for 2 seconds.
turaj@webrtc.org78b41a02013-11-22 20:27:07 +00001078 const int kSpeechDurationMs = 2000;
1079 int packets_inserted = 0;
1080 uint16_t last_seq_no;
1081 uint32_t last_timestamp;
1082 bool timestamp_wrapped = false;
1083 bool seq_no_wrapped = false;
1084 for (double t_ms = 0; t_ms < kSpeechDurationMs; t_ms += 10) {
1085 // Each turn in this for loop is 10 ms.
1086 while (next_input_time_ms <= t_ms) {
1087 // Insert one 30 ms speech frame.
1088 uint8_t payload[kPayloadBytes] = {0};
henrik.lundin246ef3e2017-04-24 09:14:32 -07001089 RTPHeader rtp_info;
turaj@webrtc.org78b41a02013-11-22 20:27:07 +00001090 PopulateRtpInfo(seq_no, timestamp, &rtp_info);
1091 if (drop_seq_numbers.find(seq_no) == drop_seq_numbers.end()) {
1092 // This sequence number was not in the set to drop. Insert it.
henrik.lundin246ef3e2017-04-24 09:14:32 -07001093 ASSERT_EQ(0,
1094 neteq_->InsertPacket(rtp_info, payload, receive_timestamp));
turaj@webrtc.org78b41a02013-11-22 20:27:07 +00001095 ++packets_inserted;
1096 }
1097 NetEqNetworkStatistics network_stats;
1098 ASSERT_EQ(0, neteq_->NetworkStatistics(&network_stats));
1099
1100 // Due to internal NetEq logic, preferred buffer-size is about 4 times the
1101 // packet size for first few packets. Therefore we refrain from checking
1102 // the criteria.
1103 if (packets_inserted > 4) {
1104 // Expect preferred and actual buffer size to be no more than 2 frames.
1105 EXPECT_LE(network_stats.preferred_buffer_size_ms, kFrameSizeMs * 2);
turaj@webrtc.org8d1cdaa2014-04-11 18:47:55 +00001106 EXPECT_LE(network_stats.current_buffer_size_ms, kFrameSizeMs * 2 +
1107 algorithmic_delay_ms_);
turaj@webrtc.org78b41a02013-11-22 20:27:07 +00001108 }
1109 last_seq_no = seq_no;
1110 last_timestamp = timestamp;
1111
1112 ++seq_no;
1113 timestamp += kSamples;
1114 receive_timestamp += kSamples;
1115 next_input_time_ms += static_cast<double>(kFrameSizeMs);
1116
1117 seq_no_wrapped |= seq_no < last_seq_no;
1118 timestamp_wrapped |= timestamp < last_timestamp;
1119 }
1120 // Pull out data once.
henrik.lundin6d8e0112016-03-04 10:34:21 -08001121 AudioFrame output;
henrik.lundin7a926812016-05-12 13:51:28 -07001122 bool muted;
1123 ASSERT_EQ(0, neteq_->GetAudio(&output, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -08001124 ASSERT_EQ(kBlockSize16kHz, output.samples_per_channel_);
1125 ASSERT_EQ(1u, output.num_channels_);
turaj@webrtc.org78b41a02013-11-22 20:27:07 +00001126
1127 // Expect delay (in samples) to be less than 2 packets.
henrik.lundin114c1b32017-04-26 07:47:32 -07001128 rtc::Optional<uint32_t> playout_timestamp = neteq_->GetPlayoutTimestamp();
henrik.lundin0d96ab72016-04-06 12:28:26 -07001129 ASSERT_TRUE(playout_timestamp);
1130 EXPECT_LE(timestamp - *playout_timestamp,
turaj@webrtc.org78b41a02013-11-22 20:27:07 +00001131 static_cast<uint32_t>(kSamples * 2));
turaj@webrtc.org78b41a02013-11-22 20:27:07 +00001132 }
1133 // Make sure we have actually tested wrap-around.
1134 ASSERT_EQ(expect_seq_no_wrap, seq_no_wrapped);
1135 ASSERT_EQ(expect_timestamp_wrap, timestamp_wrapped);
1136}
1137
1138TEST_F(NetEqDecodingTest, SequenceNumberWrap) {
1139 // Start with a sequence number that will soon wrap.
1140 std::set<uint16_t> drop_seq_numbers; // Don't drop any packets.
1141 WrapTest(0xFFFF - 10, 0, drop_seq_numbers, true, false);
1142}
1143
1144TEST_F(NetEqDecodingTest, SequenceNumberWrapAndDrop) {
1145 // Start with a sequence number that will soon wrap.
1146 std::set<uint16_t> drop_seq_numbers;
1147 drop_seq_numbers.insert(0xFFFF);
1148 drop_seq_numbers.insert(0x0);
1149 WrapTest(0xFFFF - 10, 0, drop_seq_numbers, true, false);
1150}
1151
1152TEST_F(NetEqDecodingTest, TimestampWrap) {
1153 // Start with a timestamp that will soon wrap.
1154 std::set<uint16_t> drop_seq_numbers;
1155 WrapTest(0, 0xFFFFFFFF - 3000, drop_seq_numbers, false, true);
1156}
1157
1158TEST_F(NetEqDecodingTest, TimestampAndSequenceNumberWrap) {
1159 // Start with a timestamp and a sequence number that will wrap at the same
1160 // time.
1161 std::set<uint16_t> drop_seq_numbers;
1162 WrapTest(0xFFFF - 10, 0xFFFFFFFF - 5000, drop_seq_numbers, true, true);
1163}
1164
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001165void NetEqDecodingTest::DuplicateCng() {
1166 uint16_t seq_no = 0;
1167 uint32_t timestamp = 0;
1168 const int kFrameSizeMs = 10;
1169 const int kSampleRateKhz = 16;
1170 const int kSamples = kFrameSizeMs * kSampleRateKhz;
pkasting@chromium.org4591fbd2014-11-20 22:28:14 +00001171 const size_t kPayloadBytes = kSamples * 2;
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001172
turaj@webrtc.org8d1cdaa2014-04-11 18:47:55 +00001173 const int algorithmic_delay_samples = std::max(
1174 algorithmic_delay_ms_ * kSampleRateKhz, 5 * kSampleRateKhz / 8);
henrik.lundin@webrtc.orgc93437e2014-12-01 11:42:42 +00001175 // Insert three speech packets. Three are needed to get the frame length
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001176 // correct.
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001177 uint8_t payload[kPayloadBytes] = {0};
henrik.lundin246ef3e2017-04-24 09:14:32 -07001178 RTPHeader rtp_info;
henrik.lundin7a926812016-05-12 13:51:28 -07001179 bool muted;
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001180 for (int i = 0; i < 3; ++i) {
1181 PopulateRtpInfo(seq_no, timestamp, &rtp_info);
henrik.lundin246ef3e2017-04-24 09:14:32 -07001182 ASSERT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001183 ++seq_no;
1184 timestamp += kSamples;
1185
1186 // Pull audio once.
henrik.lundin7a926812016-05-12 13:51:28 -07001187 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -08001188 ASSERT_EQ(kBlockSize16kHz, out_frame_.samples_per_channel_);
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001189 }
1190 // Verify speech output.
henrik.lundin55480f52016-03-08 02:37:57 -08001191 EXPECT_EQ(AudioFrame::kNormalSpeech, out_frame_.speech_type_);
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001192
1193 // Insert same CNG packet twice.
1194 const int kCngPeriodMs = 100;
1195 const int kCngPeriodSamples = kCngPeriodMs * kSampleRateKhz;
pkasting@chromium.org4591fbd2014-11-20 22:28:14 +00001196 size_t payload_len;
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001197 PopulateCng(seq_no, timestamp, &rtp_info, payload, &payload_len);
1198 // This is the first time this CNG packet is inserted.
henrik.lundin246ef3e2017-04-24 09:14:32 -07001199 ASSERT_EQ(
1200 0, neteq_->InsertPacket(
1201 rtp_info, rtc::ArrayView<const uint8_t>(payload, payload_len), 0));
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001202
1203 // Pull audio once and make sure CNG is played.
henrik.lundin7a926812016-05-12 13:51:28 -07001204 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -08001205 ASSERT_EQ(kBlockSize16kHz, out_frame_.samples_per_channel_);
henrik.lundin55480f52016-03-08 02:37:57 -08001206 EXPECT_EQ(AudioFrame::kCNG, out_frame_.speech_type_);
henrik.lundin114c1b32017-04-26 07:47:32 -07001207 EXPECT_FALSE(
1208 neteq_->GetPlayoutTimestamp()); // Returns empty value during CNG.
henrik.lundin0d96ab72016-04-06 12:28:26 -07001209 EXPECT_EQ(timestamp - algorithmic_delay_samples,
1210 out_frame_.timestamp_ + out_frame_.samples_per_channel_);
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001211
1212 // Insert the same CNG packet again. Note that at this point it is old, since
1213 // we have already decoded the first copy of it.
henrik.lundin246ef3e2017-04-24 09:14:32 -07001214 ASSERT_EQ(
1215 0, neteq_->InsertPacket(
1216 rtp_info, rtc::ArrayView<const uint8_t>(payload, payload_len), 0));
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001217
1218 // Pull audio until we have played |kCngPeriodMs| of CNG. Start at 10 ms since
1219 // we have already pulled out CNG once.
1220 for (int cng_time_ms = 10; cng_time_ms < kCngPeriodMs; cng_time_ms += 10) {
henrik.lundin7a926812016-05-12 13:51:28 -07001221 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -08001222 ASSERT_EQ(kBlockSize16kHz, out_frame_.samples_per_channel_);
henrik.lundin55480f52016-03-08 02:37:57 -08001223 EXPECT_EQ(AudioFrame::kCNG, out_frame_.speech_type_);
henrik.lundin114c1b32017-04-26 07:47:32 -07001224 EXPECT_FALSE(
1225 neteq_->GetPlayoutTimestamp()); // Returns empty value during CNG.
turaj@webrtc.org8d1cdaa2014-04-11 18:47:55 +00001226 EXPECT_EQ(timestamp - algorithmic_delay_samples,
henrik.lundin0d96ab72016-04-06 12:28:26 -07001227 out_frame_.timestamp_ + out_frame_.samples_per_channel_);
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001228 }
1229
1230 // Insert speech again.
1231 ++seq_no;
1232 timestamp += kCngPeriodSamples;
1233 PopulateRtpInfo(seq_no, timestamp, &rtp_info);
henrik.lundin246ef3e2017-04-24 09:14:32 -07001234 ASSERT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001235
1236 // Pull audio once and verify that the output is speech again.
henrik.lundin7a926812016-05-12 13:51:28 -07001237 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -08001238 ASSERT_EQ(kBlockSize16kHz, out_frame_.samples_per_channel_);
henrik.lundin55480f52016-03-08 02:37:57 -08001239 EXPECT_EQ(AudioFrame::kNormalSpeech, out_frame_.speech_type_);
henrik.lundin114c1b32017-04-26 07:47:32 -07001240 rtc::Optional<uint32_t> playout_timestamp = neteq_->GetPlayoutTimestamp();
henrik.lundin0d96ab72016-04-06 12:28:26 -07001241 ASSERT_TRUE(playout_timestamp);
turaj@webrtc.org8d1cdaa2014-04-11 18:47:55 +00001242 EXPECT_EQ(timestamp + kSamples - algorithmic_delay_samples,
henrik.lundin0d96ab72016-04-06 12:28:26 -07001243 *playout_timestamp);
wu@webrtc.org94454b72014-06-05 20:34:08 +00001244}
1245
henrik.lundin@webrtc.orgca8cb952014-03-12 10:26:52 +00001246TEST_F(NetEqDecodingTest, DiscardDuplicateCng) { DuplicateCng(); }
henrik.lundin@webrtc.orgc93437e2014-12-01 11:42:42 +00001247
1248TEST_F(NetEqDecodingTest, CngFirst) {
1249 uint16_t seq_no = 0;
1250 uint32_t timestamp = 0;
1251 const int kFrameSizeMs = 10;
1252 const int kSampleRateKhz = 16;
1253 const int kSamples = kFrameSizeMs * kSampleRateKhz;
1254 const int kPayloadBytes = kSamples * 2;
1255 const int kCngPeriodMs = 100;
1256 const int kCngPeriodSamples = kCngPeriodMs * kSampleRateKhz;
1257 size_t payload_len;
1258
1259 uint8_t payload[kPayloadBytes] = {0};
henrik.lundin246ef3e2017-04-24 09:14:32 -07001260 RTPHeader rtp_info;
henrik.lundin@webrtc.orgc93437e2014-12-01 11:42:42 +00001261
1262 PopulateCng(seq_no, timestamp, &rtp_info, payload, &payload_len);
henrik.lundin246ef3e2017-04-24 09:14:32 -07001263 ASSERT_EQ(
1264 NetEq::kOK,
1265 neteq_->InsertPacket(
1266 rtp_info, rtc::ArrayView<const uint8_t>(payload, payload_len), 0));
henrik.lundin@webrtc.orgc93437e2014-12-01 11:42:42 +00001267 ++seq_no;
1268 timestamp += kCngPeriodSamples;
1269
1270 // Pull audio once and make sure CNG is played.
henrik.lundin7a926812016-05-12 13:51:28 -07001271 bool muted;
1272 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -08001273 ASSERT_EQ(kBlockSize16kHz, out_frame_.samples_per_channel_);
henrik.lundin55480f52016-03-08 02:37:57 -08001274 EXPECT_EQ(AudioFrame::kCNG, out_frame_.speech_type_);
henrik.lundin@webrtc.orgc93437e2014-12-01 11:42:42 +00001275
1276 // Insert some speech packets.
henrik.lundin549d80b2016-08-25 00:44:24 -07001277 const uint32_t first_speech_timestamp = timestamp;
1278 int timeout_counter = 0;
1279 do {
1280 ASSERT_LT(timeout_counter++, 20) << "Test timed out";
henrik.lundin@webrtc.orgc93437e2014-12-01 11:42:42 +00001281 PopulateRtpInfo(seq_no, timestamp, &rtp_info);
henrik.lundin246ef3e2017-04-24 09:14:32 -07001282 ASSERT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
henrik.lundin@webrtc.orgc93437e2014-12-01 11:42:42 +00001283 ++seq_no;
1284 timestamp += kSamples;
1285
1286 // Pull audio once.
henrik.lundin7a926812016-05-12 13:51:28 -07001287 ASSERT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
henrik.lundin6d8e0112016-03-04 10:34:21 -08001288 ASSERT_EQ(kBlockSize16kHz, out_frame_.samples_per_channel_);
henrik.lundin549d80b2016-08-25 00:44:24 -07001289 } while (!IsNewerTimestamp(out_frame_.timestamp_, first_speech_timestamp));
henrik.lundin@webrtc.orgc93437e2014-12-01 11:42:42 +00001290 // Verify speech output.
henrik.lundin55480f52016-03-08 02:37:57 -08001291 EXPECT_EQ(AudioFrame::kNormalSpeech, out_frame_.speech_type_);
henrik.lundin@webrtc.orgc93437e2014-12-01 11:42:42 +00001292}
henrik.lundin7a926812016-05-12 13:51:28 -07001293
1294class NetEqDecodingTestWithMutedState : public NetEqDecodingTest {
1295 public:
1296 NetEqDecodingTestWithMutedState() : NetEqDecodingTest() {
1297 config_.enable_muted_state = true;
1298 }
1299
1300 protected:
1301 static constexpr size_t kSamples = 10 * 16;
1302 static constexpr size_t kPayloadBytes = kSamples * 2;
1303
1304 void InsertPacket(uint32_t rtp_timestamp) {
1305 uint8_t payload[kPayloadBytes] = {0};
henrik.lundin246ef3e2017-04-24 09:14:32 -07001306 RTPHeader rtp_info;
henrik.lundin7a926812016-05-12 13:51:28 -07001307 PopulateRtpInfo(0, rtp_timestamp, &rtp_info);
henrik.lundin246ef3e2017-04-24 09:14:32 -07001308 EXPECT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
henrik.lundin7a926812016-05-12 13:51:28 -07001309 }
1310
henrik.lundin42feb512016-09-20 06:51:40 -07001311 void InsertCngPacket(uint32_t rtp_timestamp) {
1312 uint8_t payload[kPayloadBytes] = {0};
henrik.lundin246ef3e2017-04-24 09:14:32 -07001313 RTPHeader rtp_info;
henrik.lundin42feb512016-09-20 06:51:40 -07001314 size_t payload_len;
1315 PopulateCng(0, rtp_timestamp, &rtp_info, payload, &payload_len);
henrik.lundin246ef3e2017-04-24 09:14:32 -07001316 EXPECT_EQ(
1317 NetEq::kOK,
1318 neteq_->InsertPacket(
1319 rtp_info, rtc::ArrayView<const uint8_t>(payload, payload_len), 0));
henrik.lundin42feb512016-09-20 06:51:40 -07001320 }
1321
henrik.lundin7a926812016-05-12 13:51:28 -07001322 bool GetAudioReturnMuted() {
1323 bool muted;
1324 EXPECT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
1325 return muted;
1326 }
1327
1328 void GetAudioUntilMuted() {
1329 while (!GetAudioReturnMuted()) {
1330 ASSERT_LT(counter_++, 1000) << "Test timed out";
1331 }
1332 }
1333
1334 void GetAudioUntilNormal() {
1335 bool muted = false;
1336 while (out_frame_.speech_type_ != AudioFrame::kNormalSpeech) {
1337 EXPECT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
1338 ASSERT_LT(counter_++, 1000) << "Test timed out";
1339 }
1340 EXPECT_FALSE(muted);
1341 }
1342
1343 int counter_ = 0;
1344};
1345
1346// Verifies that NetEq goes in and out of muted state as expected.
1347TEST_F(NetEqDecodingTestWithMutedState, MutedState) {
1348 // Insert one speech packet.
1349 InsertPacket(0);
1350 // Pull out audio once and expect it not to be muted.
1351 EXPECT_FALSE(GetAudioReturnMuted());
1352 // Pull data until faded out.
1353 GetAudioUntilMuted();
henrik.lundina4491072017-07-06 05:23:53 -07001354 EXPECT_TRUE(out_frame_.muted());
henrik.lundin7a926812016-05-12 13:51:28 -07001355
1356 // Verify that output audio is not written during muted mode. Other parameters
1357 // should be correct, though.
1358 AudioFrame new_frame;
yujo36b1a5f2017-06-12 12:45:32 -07001359 int16_t* frame_data = new_frame.mutable_data();
1360 for (size_t i = 0; i < AudioFrame::kMaxDataSizeSamples; i++) {
1361 frame_data[i] = 17;
henrik.lundin7a926812016-05-12 13:51:28 -07001362 }
1363 bool muted;
1364 EXPECT_EQ(0, neteq_->GetAudio(&new_frame, &muted));
1365 EXPECT_TRUE(muted);
henrik.lundina4491072017-07-06 05:23:53 -07001366 EXPECT_TRUE(out_frame_.muted());
yujo36b1a5f2017-06-12 12:45:32 -07001367 for (size_t i = 0; i < AudioFrame::kMaxDataSizeSamples; i++) {
1368 EXPECT_EQ(17, frame_data[i]);
henrik.lundin7a926812016-05-12 13:51:28 -07001369 }
1370 EXPECT_EQ(out_frame_.timestamp_ + out_frame_.samples_per_channel_,
1371 new_frame.timestamp_);
1372 EXPECT_EQ(out_frame_.samples_per_channel_, new_frame.samples_per_channel_);
1373 EXPECT_EQ(out_frame_.sample_rate_hz_, new_frame.sample_rate_hz_);
1374 EXPECT_EQ(out_frame_.num_channels_, new_frame.num_channels_);
1375 EXPECT_EQ(out_frame_.speech_type_, new_frame.speech_type_);
1376 EXPECT_EQ(out_frame_.vad_activity_, new_frame.vad_activity_);
1377
1378 // Insert new data. Timestamp is corrected for the time elapsed since the last
1379 // packet. Verify that normal operation resumes.
1380 InsertPacket(kSamples * counter_);
1381 GetAudioUntilNormal();
henrik.lundina4491072017-07-06 05:23:53 -07001382 EXPECT_FALSE(out_frame_.muted());
henrik.lundin612c25e2016-05-25 08:21:04 -07001383
1384 NetEqNetworkStatistics stats;
1385 EXPECT_EQ(0, neteq_->NetworkStatistics(&stats));
1386 // NetEqNetworkStatistics::expand_rate tells the fraction of samples that were
1387 // concealment samples, in Q14 (16384 = 100%) .The vast majority should be
1388 // concealment samples in this test.
1389 EXPECT_GT(stats.expand_rate, 14000);
1390 // And, it should be greater than the speech_expand_rate.
1391 EXPECT_GT(stats.expand_rate, stats.speech_expand_rate);
henrik.lundin7a926812016-05-12 13:51:28 -07001392}
1393
1394// Verifies that NetEq goes out of muted state when given a delayed packet.
1395TEST_F(NetEqDecodingTestWithMutedState, MutedStateDelayedPacket) {
1396 // Insert one speech packet.
1397 InsertPacket(0);
1398 // Pull out audio once and expect it not to be muted.
1399 EXPECT_FALSE(GetAudioReturnMuted());
1400 // Pull data until faded out.
1401 GetAudioUntilMuted();
1402 // Insert new data. Timestamp is only corrected for the half of the time
1403 // elapsed since the last packet. That is, the new packet is delayed. Verify
1404 // that normal operation resumes.
1405 InsertPacket(kSamples * counter_ / 2);
1406 GetAudioUntilNormal();
1407}
1408
1409// Verifies that NetEq goes out of muted state when given a future packet.
1410TEST_F(NetEqDecodingTestWithMutedState, MutedStateFuturePacket) {
1411 // Insert one speech packet.
1412 InsertPacket(0);
1413 // Pull out audio once and expect it not to be muted.
1414 EXPECT_FALSE(GetAudioReturnMuted());
1415 // Pull data until faded out.
1416 GetAudioUntilMuted();
1417 // Insert new data. Timestamp is over-corrected for the time elapsed since the
1418 // last packet. That is, the new packet is too early. Verify that normal
1419 // operation resumes.
1420 InsertPacket(kSamples * counter_ * 2);
1421 GetAudioUntilNormal();
1422}
1423
1424// Verifies that NetEq goes out of muted state when given an old packet.
1425TEST_F(NetEqDecodingTestWithMutedState, MutedStateOldPacket) {
1426 // Insert one speech packet.
1427 InsertPacket(0);
1428 // Pull out audio once and expect it not to be muted.
1429 EXPECT_FALSE(GetAudioReturnMuted());
1430 // Pull data until faded out.
1431 GetAudioUntilMuted();
1432
1433 EXPECT_NE(AudioFrame::kNormalSpeech, out_frame_.speech_type_);
1434 // Insert packet which is older than the first packet.
1435 InsertPacket(kSamples * (counter_ - 1000));
1436 EXPECT_FALSE(GetAudioReturnMuted());
1437 EXPECT_EQ(AudioFrame::kNormalSpeech, out_frame_.speech_type_);
1438}
1439
henrik.lundin42feb512016-09-20 06:51:40 -07001440// Verifies that NetEq doesn't enter muted state when CNG mode is active and the
1441// packet stream is suspended for a long time.
1442TEST_F(NetEqDecodingTestWithMutedState, DoNotMuteExtendedCngWithoutPackets) {
1443 // Insert one CNG packet.
1444 InsertCngPacket(0);
1445
1446 // Pull 10 seconds of audio (10 ms audio generated per lap).
1447 for (int i = 0; i < 1000; ++i) {
1448 bool muted;
1449 EXPECT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
1450 ASSERT_FALSE(muted);
1451 }
1452 EXPECT_EQ(AudioFrame::kCNG, out_frame_.speech_type_);
1453}
1454
1455// Verifies that NetEq goes back to normal after a long CNG period with the
1456// packet stream suspended.
1457TEST_F(NetEqDecodingTestWithMutedState, RecoverAfterExtendedCngWithoutPackets) {
1458 // Insert one CNG packet.
1459 InsertCngPacket(0);
1460
1461 // Pull 10 seconds of audio (10 ms audio generated per lap).
1462 for (int i = 0; i < 1000; ++i) {
1463 bool muted;
1464 EXPECT_EQ(0, neteq_->GetAudio(&out_frame_, &muted));
1465 }
1466
1467 // Insert new data. Timestamp is corrected for the time elapsed since the last
1468 // packet. Verify that normal operation resumes.
1469 InsertPacket(kSamples * counter_);
1470 GetAudioUntilNormal();
1471}
1472
henrik.lundin7a926812016-05-12 13:51:28 -07001473class NetEqDecodingTestTwoInstances : public NetEqDecodingTest {
1474 public:
1475 NetEqDecodingTestTwoInstances() : NetEqDecodingTest() {}
1476
1477 void SetUp() override {
1478 NetEqDecodingTest::SetUp();
1479 config2_ = config_;
1480 }
1481
1482 void CreateSecondInstance() {
ossue3525782016-05-25 07:37:43 -07001483 neteq2_.reset(NetEq::Create(config2_, CreateBuiltinAudioDecoderFactory()));
henrik.lundin7a926812016-05-12 13:51:28 -07001484 ASSERT_TRUE(neteq2_);
1485 LoadDecoders(neteq2_.get());
1486 }
1487
1488 protected:
1489 std::unique_ptr<NetEq> neteq2_;
1490 NetEq::Config config2_;
1491};
1492
1493namespace {
1494::testing::AssertionResult AudioFramesEqualExceptData(const AudioFrame& a,
1495 const AudioFrame& b) {
1496 if (a.timestamp_ != b.timestamp_)
1497 return ::testing::AssertionFailure() << "timestamp_ diff (" << a.timestamp_
1498 << " != " << b.timestamp_ << ")";
1499 if (a.sample_rate_hz_ != b.sample_rate_hz_)
1500 return ::testing::AssertionFailure() << "sample_rate_hz_ diff ("
1501 << a.sample_rate_hz_
1502 << " != " << b.sample_rate_hz_ << ")";
1503 if (a.samples_per_channel_ != b.samples_per_channel_)
1504 return ::testing::AssertionFailure()
1505 << "samples_per_channel_ diff (" << a.samples_per_channel_
1506 << " != " << b.samples_per_channel_ << ")";
1507 if (a.num_channels_ != b.num_channels_)
1508 return ::testing::AssertionFailure() << "num_channels_ diff ("
1509 << a.num_channels_
1510 << " != " << b.num_channels_ << ")";
1511 if (a.speech_type_ != b.speech_type_)
1512 return ::testing::AssertionFailure() << "speech_type_ diff ("
1513 << a.speech_type_
1514 << " != " << b.speech_type_ << ")";
1515 if (a.vad_activity_ != b.vad_activity_)
1516 return ::testing::AssertionFailure() << "vad_activity_ diff ("
1517 << a.vad_activity_
1518 << " != " << b.vad_activity_ << ")";
1519 return ::testing::AssertionSuccess();
1520}
1521
1522::testing::AssertionResult AudioFramesEqual(const AudioFrame& a,
1523 const AudioFrame& b) {
1524 ::testing::AssertionResult res = AudioFramesEqualExceptData(a, b);
1525 if (!res)
1526 return res;
1527 if (memcmp(
yujo36b1a5f2017-06-12 12:45:32 -07001528 a.data(), b.data(),
1529 a.samples_per_channel_ * a.num_channels_ * sizeof(*a.data())) != 0) {
henrik.lundin7a926812016-05-12 13:51:28 -07001530 return ::testing::AssertionFailure() << "data_ diff";
1531 }
1532 return ::testing::AssertionSuccess();
1533}
1534
1535} // namespace
1536
1537TEST_F(NetEqDecodingTestTwoInstances, CompareMutedStateOnOff) {
1538 ASSERT_FALSE(config_.enable_muted_state);
1539 config2_.enable_muted_state = true;
1540 CreateSecondInstance();
1541
1542 // Insert one speech packet into both NetEqs.
1543 const size_t kSamples = 10 * 16;
1544 const size_t kPayloadBytes = kSamples * 2;
1545 uint8_t payload[kPayloadBytes] = {0};
henrik.lundin246ef3e2017-04-24 09:14:32 -07001546 RTPHeader rtp_info;
henrik.lundin7a926812016-05-12 13:51:28 -07001547 PopulateRtpInfo(0, 0, &rtp_info);
henrik.lundin246ef3e2017-04-24 09:14:32 -07001548 EXPECT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
1549 EXPECT_EQ(0, neteq2_->InsertPacket(rtp_info, payload, 0));
henrik.lundin7a926812016-05-12 13:51:28 -07001550
1551 AudioFrame out_frame1, out_frame2;
1552 bool muted;
1553 for (int i = 0; i < 1000; ++i) {
1554 std::ostringstream ss;
1555 ss << "i = " << i;
1556 SCOPED_TRACE(ss.str()); // Print out the loop iterator on failure.
1557 EXPECT_EQ(0, neteq_->GetAudio(&out_frame1, &muted));
1558 EXPECT_FALSE(muted);
1559 EXPECT_EQ(0, neteq2_->GetAudio(&out_frame2, &muted));
1560 if (muted) {
1561 EXPECT_TRUE(AudioFramesEqualExceptData(out_frame1, out_frame2));
1562 } else {
1563 EXPECT_TRUE(AudioFramesEqual(out_frame1, out_frame2));
1564 }
1565 }
1566 EXPECT_TRUE(muted);
1567
1568 // Insert new data. Timestamp is corrected for the time elapsed since the last
1569 // packet.
1570 PopulateRtpInfo(0, kSamples * 1000, &rtp_info);
henrik.lundin246ef3e2017-04-24 09:14:32 -07001571 EXPECT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
1572 EXPECT_EQ(0, neteq2_->InsertPacket(rtp_info, payload, 0));
henrik.lundin7a926812016-05-12 13:51:28 -07001573
1574 int counter = 0;
1575 while (out_frame1.speech_type_ != AudioFrame::kNormalSpeech) {
1576 ASSERT_LT(counter++, 1000) << "Test timed out";
1577 std::ostringstream ss;
1578 ss << "counter = " << counter;
1579 SCOPED_TRACE(ss.str()); // Print out the loop iterator on failure.
1580 EXPECT_EQ(0, neteq_->GetAudio(&out_frame1, &muted));
1581 EXPECT_FALSE(muted);
1582 EXPECT_EQ(0, neteq2_->GetAudio(&out_frame2, &muted));
1583 if (muted) {
1584 EXPECT_TRUE(AudioFramesEqualExceptData(out_frame1, out_frame2));
1585 } else {
1586 EXPECT_TRUE(AudioFramesEqual(out_frame1, out_frame2));
1587 }
1588 }
1589 EXPECT_FALSE(muted);
1590}
1591
henrik.lundin114c1b32017-04-26 07:47:32 -07001592TEST_F(NetEqDecodingTest, LastDecodedTimestampsEmpty) {
1593 EXPECT_TRUE(neteq_->LastDecodedTimestamps().empty());
1594
1595 // Pull out data once.
1596 AudioFrame output;
1597 bool muted;
1598 ASSERT_EQ(0, neteq_->GetAudio(&output, &muted));
1599
1600 EXPECT_TRUE(neteq_->LastDecodedTimestamps().empty());
1601}
1602
1603TEST_F(NetEqDecodingTest, LastDecodedTimestampsOneDecoded) {
1604 // Insert one packet with PCM16b WB data (this is what PopulateRtpInfo does by
1605 // default). Make the length 10 ms.
1606 constexpr size_t kPayloadSamples = 16 * 10;
1607 constexpr size_t kPayloadBytes = 2 * kPayloadSamples;
1608 uint8_t payload[kPayloadBytes] = {0};
1609
1610 RTPHeader rtp_info;
1611 constexpr uint32_t kRtpTimestamp = 0x1234;
1612 PopulateRtpInfo(0, kRtpTimestamp, &rtp_info);
1613 EXPECT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
1614
1615 // Pull out data once.
1616 AudioFrame output;
1617 bool muted;
1618 ASSERT_EQ(0, neteq_->GetAudio(&output, &muted));
1619
1620 EXPECT_EQ(std::vector<uint32_t>({kRtpTimestamp}),
1621 neteq_->LastDecodedTimestamps());
1622
1623 // Nothing decoded on the second call.
1624 ASSERT_EQ(0, neteq_->GetAudio(&output, &muted));
1625 EXPECT_TRUE(neteq_->LastDecodedTimestamps().empty());
1626}
1627
1628TEST_F(NetEqDecodingTest, LastDecodedTimestampsTwoDecoded) {
1629 // Insert two packets with PCM16b WB data (this is what PopulateRtpInfo does
1630 // by default). Make the length 5 ms so that NetEq must decode them both in
1631 // the same GetAudio call.
1632 constexpr size_t kPayloadSamples = 16 * 5;
1633 constexpr size_t kPayloadBytes = 2 * kPayloadSamples;
1634 uint8_t payload[kPayloadBytes] = {0};
1635
1636 RTPHeader rtp_info;
1637 constexpr uint32_t kRtpTimestamp1 = 0x1234;
1638 PopulateRtpInfo(0, kRtpTimestamp1, &rtp_info);
1639 EXPECT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
1640 constexpr uint32_t kRtpTimestamp2 = kRtpTimestamp1 + kPayloadSamples;
1641 PopulateRtpInfo(1, kRtpTimestamp2, &rtp_info);
1642 EXPECT_EQ(0, neteq_->InsertPacket(rtp_info, payload, 0));
1643
1644 // Pull out data once.
1645 AudioFrame output;
1646 bool muted;
1647 ASSERT_EQ(0, neteq_->GetAudio(&output, &muted));
1648
1649 EXPECT_EQ(std::vector<uint32_t>({kRtpTimestamp1, kRtpTimestamp2}),
1650 neteq_->LastDecodedTimestamps());
1651}
1652
Gustaf Ullberg9a2e9062017-09-18 09:28:20 +02001653TEST_F(NetEqDecodingTest, TestConcealmentEvents) {
1654 const int kNumConcealmentEvents = 19;
1655 const size_t kSamples = 10 * 16;
1656 const size_t kPayloadBytes = kSamples * 2;
1657 int seq_no = 0;
1658 RTPHeader rtp_info;
1659 rtp_info.ssrc = 0x1234; // Just an arbitrary SSRC.
1660 rtp_info.payloadType = 94; // PCM16b WB codec.
1661 rtp_info.markerBit = 0;
1662 const uint8_t payload[kPayloadBytes] = {0};
1663 bool muted;
1664
1665 for (int i = 0; i < kNumConcealmentEvents; i++) {
1666 // Insert some packets of 10 ms size.
1667 for (int j = 0; j < 10; j++) {
1668 rtp_info.sequenceNumber = seq_no++;
1669 rtp_info.timestamp = rtp_info.sequenceNumber * kSamples;
1670 neteq_->InsertPacket(rtp_info, payload, 0);
1671 neteq_->GetAudio(&out_frame_, &muted);
1672 }
1673
1674 // Lose a number of packets.
1675 int num_lost = 1 + i;
1676 for (int j = 0; j < num_lost; j++) {
1677 seq_no++;
1678 neteq_->GetAudio(&out_frame_, &muted);
1679 }
1680 }
1681
1682 // Check number of concealment events.
1683 NetEqLifetimeStatistics stats = neteq_->GetLifetimeStatistics();
1684 EXPECT_EQ(kNumConcealmentEvents, static_cast<int>(stats.concealment_events));
1685}
1686
henrik.lundin@webrtc.orge7ce4372014-01-09 14:01:55 +00001687} // namespace webrtc