Reformat existing code. There should be no functional effects.
This includes changes like:
* Attempt to break lines at better positions
* Use "override" in more places, don't use "virtual" with it
* Use {} where the body is more than one line
* Make declaration and definition arg names match
* Eliminate unused code
* EXPECT_EQ(expected, actual) (but use (actual, expected) for e.g. _GT)
* Correct #include order
* Use anonymous namespaces in preference to "static" for file-scoping
* Eliminate unnecessary casts
* Update reference code in comments of ARM assembly sources to match actual current C code
* Fix indenting to be more style-guide compliant
* Use arraysize() in more places
* Use bool instead of int for "boolean" values (0/1)
* Shorten and simplify code
* Spaces around operators
* 80 column limit
* Use const more consistently
* Space goes after '*' in type name, not before
* Remove unnecessary return values
* Use "(var == const)", not "(const == var)"
* Spelling
* Prefer true, typed constants to "enum hack" constants
* Avoid "virtual" on non-overridden functions
* ASSERT(x == y) -> ASSERT_EQ(y, x)
BUG=none
R=andrew@webrtc.org, asapersson@webrtc.org, henrika@webrtc.org, juberti@webrtc.org, kjellander@webrtc.org, kwiberg@webrtc.org
Review URL: https://codereview.webrtc.org/1172163004
Cr-Commit-Position: refs/heads/master@{#9420}
diff --git a/webrtc/modules/audio_coding/codecs/g711/g711_interface.c b/webrtc/modules/audio_coding/codecs/g711/g711_interface.c
index 809a70e..b579520 100644
--- a/webrtc/modules/audio_coding/codecs/g711/g711_interface.c
+++ b/webrtc/modules/audio_coding/codecs/g711/g711_interface.c
@@ -52,13 +52,6 @@
return len;
}
-int WebRtcG711_DurationEst(const uint8_t* payload,
- int payload_length_bytes) {
- (void) payload;
- /* G.711 is one byte per sample, so we can just return the number of bytes. */
- return payload_length_bytes;
-}
-
int16_t WebRtcG711_Version(char* version, int16_t lenBytes) {
strncpy(version, "2.0.0", lenBytes);
return 0;
diff --git a/webrtc/modules/audio_coding/codecs/g711/include/g711_interface.h b/webrtc/modules/audio_coding/codecs/g711/include/g711_interface.h
index 0b798a6..5c71e98 100644
--- a/webrtc/modules/audio_coding/codecs/g711/include/g711_interface.h
+++ b/webrtc/modules/audio_coding/codecs/g711/include/g711_interface.h
@@ -111,23 +111,6 @@
int16_t* decoded,
int16_t* speechType);
-/****************************************************************************
- * WebRtcG711_DurationEst(...)
- *
- * This function estimates the duration of a G711 packet in samples.
- *
- * Input:
- * - payload : Encoded data
- * - payloadLengthBytes : Bytes in encoded vector
- *
- * Return value : The duration of the packet in samples, which is
- * just payload_length_bytes, since G.711 uses one
- * byte per sample.
- */
-
-int WebRtcG711_DurationEst(const uint8_t* payload,
- int payload_length_bytes);
-
/**********************************************************************
* WebRtcG711_Version(...)
*
diff --git a/webrtc/modules/audio_coding/codecs/g711/test/testG711.cc b/webrtc/modules/audio_coding/codecs/g711/test/testG711.cc
index 9511df7..49c671c 100644
--- a/webrtc/modules/audio_coding/codecs/g711/test/testG711.cc
+++ b/webrtc/modules/audio_coding/codecs/g711/test/testG711.cc
@@ -24,18 +24,12 @@
#define CLOCKS_PER_SEC_G711 1000
/* function for reading audio data from PCM file */
-int readframe(int16_t* data, FILE* inp, int length) {
-
- short k, rlen, status = 0;
-
- rlen = (short) fread(data, sizeof(int16_t), length, inp);
- if (rlen < length) {
- for (k = rlen; k < length; k++)
- data[k] = 0;
- status = 1;
- }
-
- return status;
+bool readframe(int16_t* data, FILE* inp, int length) {
+ short rlen = (short) fread(data, sizeof(int16_t), length, inp);
+ if (rlen >= length)
+ return false;
+ memset(data + rlen, 0, (length - rlen) * sizeof(int16_t));
+ return true;
}
int main(int argc, char* argv[]) {
@@ -43,7 +37,8 @@
FILE* inp;
FILE* outp;
FILE* bitp = NULL;
- int framecnt, endfile;
+ int framecnt;
+ bool endfile;
int16_t framelength = 80;
@@ -122,8 +117,8 @@
/* Initialize encoder and decoder */
framecnt = 0;
- endfile = 0;
- while (endfile == 0) {
+ endfile = false;
+ while (!endfile) {
framecnt++;
/* Read speech block */
endfile = readframe(shortdata, inp, framelength);
diff --git a/webrtc/modules/audio_coding/codecs/g722/g722_interface.c b/webrtc/modules/audio_coding/codecs/g722/g722_interface.c
index d06c588..25d75ee 100644
--- a/webrtc/modules/audio_coding/codecs/g722/g722_interface.c
+++ b/webrtc/modules/audio_coding/codecs/g722/g722_interface.c
@@ -52,8 +52,8 @@
{
unsigned char *codechar = (unsigned char*) encoded;
// Encode the input speech vector
- return WebRtc_g722_encode((G722EncoderState*) G722enc_inst,
- codechar, speechIn, len);
+ return WebRtc_g722_encode((G722EncoderState*) G722enc_inst, codechar,
+ speechIn, len);
}
int16_t WebRtcG722_CreateDecoder(G722DecInst **G722dec_inst)
@@ -93,8 +93,8 @@
{
// Decode the G.722 encoder stream
*speechType=G722_WEBRTC_SPEECH;
- return WebRtc_g722_decode((G722DecoderState*) G722dec_inst,
- decoded, encoded, len);
+ return WebRtc_g722_decode((G722DecoderState*) G722dec_inst, decoded,
+ encoded, len);
}
int16_t WebRtcG722_Version(char *versionStr, short len)
diff --git a/webrtc/modules/audio_coding/codecs/g722/include/g722_interface.h b/webrtc/modules/audio_coding/codecs/g722/include/g722_interface.h
index 7fe11a7..711b991 100644
--- a/webrtc/modules/audio_coding/codecs/g722/include/g722_interface.h
+++ b/webrtc/modules/audio_coding/codecs/g722/include/g722_interface.h
@@ -91,7 +91,7 @@
* Output:
* - encoded : The encoded data vector
*
- * Return value : Length (in bytes) of coded data
+ * Return value : Length (in bytes) of coded data
*/
int16_t WebRtcG722_Encode(G722EncInst* G722enc_inst,
diff --git a/webrtc/modules/audio_coding/codecs/g722/test/testG722.cc b/webrtc/modules/audio_coding/codecs/g722/test/testG722.cc
index 9b2f54c..6a6f03c 100644
--- a/webrtc/modules/audio_coding/codecs/g722/test/testG722.cc
+++ b/webrtc/modules/audio_coding/codecs/g722/test/testG722.cc
@@ -29,18 +29,13 @@
typedef struct WebRtcG722DecInst G722DecInst;
/* function for reading audio data from PCM file */
-int readframe(int16_t *data, FILE *inp, int length)
+bool readframe(int16_t *data, FILE *inp, int length)
{
- short k, rlen, status = 0;
-
- rlen = (short)fread(data, sizeof(int16_t), length, inp);
- if (rlen < length) {
- for (k = rlen; k < length; k++)
- data[k] = 0;
- status = 1;
- }
-
- return status;
+ short rlen = (short)fread(data, sizeof(int16_t), length, inp);
+ if (rlen >= length)
+ return false;
+ memset(data + rlen, 0, (length - rlen) * sizeof(int16_t));
+ return true;
}
int main(int argc, char* argv[])
@@ -48,7 +43,8 @@
char inname[60], outbit[40], outname[40];
FILE *inp, *outbitp, *outp;
- int framecnt, endfile;
+ int framecnt;
+ bool endfile;
int16_t framelength = 160;
G722EncInst *G722enc_inst;
G722DecInst *G722dec_inst;
@@ -116,8 +112,8 @@
/* Initialize encoder and decoder */
framecnt = 0;
- endfile = 0;
- while (endfile == 0) {
+ endfile = false;
+ while (!endfile) {
framecnt++;
/* Read speech block */
@@ -139,13 +135,13 @@
printf("Error in encoder/decoder\n");
} else {
/* Write coded bits to file */
- if (fwrite(streamdata, sizeof(short), stream_len/2,
- outbitp) != static_cast<size_t>(stream_len/2)) {
+ if (fwrite(streamdata, sizeof(short), stream_len / 2, outbitp) !=
+ static_cast<size_t>(stream_len / 2)) {
return -1;
}
/* Write coded speech to file */
- if (fwrite(decoded, sizeof(short), framelength,
- outp) != static_cast<size_t>(framelength)) {
+ if (fwrite(decoded, sizeof(short), framelength, outp) !=
+ static_cast<size_t>(framelength)) {
return -1;
}
}
diff --git a/webrtc/modules/audio_coding/codecs/ilbc/abs_quant_loop.c b/webrtc/modules/audio_coding/codecs/ilbc/abs_quant_loop.c
index 1a18a1d..d26fb5d 100644
--- a/webrtc/modules/audio_coding/codecs/ilbc/abs_quant_loop.c
+++ b/webrtc/modules/audio_coding/codecs/ilbc/abs_quant_loop.c
@@ -23,7 +23,7 @@
void WebRtcIlbcfix_AbsQuantLoop(int16_t *syntOutIN, int16_t *in_weightedIN,
int16_t *weightDenumIN, int16_t *quantLenIN,
int16_t *idxVecIN ) {
- int n, k1, k2;
+ int k1, k2;
int16_t index;
int32_t toQW32;
int32_t toQ32;
@@ -36,8 +36,6 @@
int16_t *quantLen = quantLenIN;
int16_t *idxVec = idxVecIN;
- n=0;
-
for(k1=0;k1<2;k1++) {
for(k2=0;k2<quantLen[k1];k2++){
@@ -81,7 +79,6 @@
*syntOut = (int16_t) (tmp16a + (int32_t)(*in_weighted) - toQW32);
- n++;
syntOut++; in_weighted++;
}
/* Update perceptual weighting filter at subframe border */
diff --git a/webrtc/modules/audio_coding/codecs/ilbc/audio_encoder_ilbc.cc b/webrtc/modules/audio_coding/codecs/ilbc/audio_encoder_ilbc.cc
index 4fea44b..66d2465 100644
--- a/webrtc/modules/audio_coding/codecs/ilbc/audio_encoder_ilbc.cc
+++ b/webrtc/modules/audio_coding/codecs/ilbc/audio_encoder_ilbc.cc
@@ -25,12 +25,9 @@
} // namespace
bool AudioEncoderIlbc::Config::IsOk() const {
- if (!(frame_size_ms == 20 || frame_size_ms == 30 || frame_size_ms == 40 ||
- frame_size_ms == 60))
- return false;
- if (kSampleRateHz / 100 * (frame_size_ms / 10) > kMaxSamplesPerPacket)
- return false;
- return true;
+ return (frame_size_ms == 20 || frame_size_ms == 30 || frame_size_ms == 40 ||
+ frame_size_ms == 60) &&
+ (kSampleRateHz / 100 * (frame_size_ms / 10)) <= kMaxSamplesPerPacket;
}
AudioEncoderIlbc::AudioEncoderIlbc(const Config& config)
diff --git a/webrtc/modules/audio_coding/codecs/ilbc/cb_mem_energy.c b/webrtc/modules/audio_coding/codecs/ilbc/cb_mem_energy.c
index f8a0933..8dfde21 100644
--- a/webrtc/modules/audio_coding/codecs/ilbc/cb_mem_energy.c
+++ b/webrtc/modules/audio_coding/codecs/ilbc/cb_mem_energy.c
@@ -35,7 +35,7 @@
int16_t *energyW16, /* (o) Energy in the CB vectors */
int16_t *energyShifts, /* (o) Shift value of the energy */
int16_t scale, /* (i) The scaling of all energy values */
- int16_t base_size /* (i) Index to where the energy values should be stored */
+ int16_t base_size /* (i) Index to where energy values should be stored */
) {
int16_t *ppi, *ppo, *pp;
int32_t energy, tmp32;
diff --git a/webrtc/modules/audio_coding/codecs/ilbc/cb_mem_energy_augmentation.c b/webrtc/modules/audio_coding/codecs/ilbc/cb_mem_energy_augmentation.c
index 7e6daf9..789d2d4 100644
--- a/webrtc/modules/audio_coding/codecs/ilbc/cb_mem_energy_augmentation.c
+++ b/webrtc/modules/audio_coding/codecs/ilbc/cb_mem_energy_augmentation.c
@@ -23,7 +23,7 @@
int16_t *interpSamples, /* (i) The interpolated samples */
int16_t *CBmem, /* (i) The CB memory */
int16_t scale, /* (i) The scaling of all energy values */
- int16_t base_size, /* (i) Index to where the energy values should be stored */
+ int16_t base_size, /* (i) Index to where energy values should be stored */
int16_t *energyW16, /* (o) Energy in the CB vectors */
int16_t *energyShifts /* (o) Shift value of the energy */
){
diff --git a/webrtc/modules/audio_coding/codecs/ilbc/cb_mem_energy_augmentation.h b/webrtc/modules/audio_coding/codecs/ilbc/cb_mem_energy_augmentation.h
index 6c181bd..9b5f85c 100644
--- a/webrtc/modules/audio_coding/codecs/ilbc/cb_mem_energy_augmentation.h
+++ b/webrtc/modules/audio_coding/codecs/ilbc/cb_mem_energy_augmentation.h
@@ -23,7 +23,7 @@
int16_t *interpSamples, /* (i) The interpolated samples */
int16_t *CBmem, /* (i) The CB memory */
int16_t scale, /* (i) The scaling of all energy values */
- int16_t base_size, /* (i) Index to where the energy values should be stored */
+ int16_t base_size, /* (i) Index to where energy values should be stored */
int16_t *energyW16, /* (o) Energy in the CB vectors */
int16_t *energyShifts /* (o) Shift value of the energy */
);
diff --git a/webrtc/modules/audio_coding/codecs/ilbc/cb_mem_energy_calc.c b/webrtc/modules/audio_coding/codecs/ilbc/cb_mem_energy_calc.c
index b1c0f8c..ec2dcaa 100644
--- a/webrtc/modules/audio_coding/codecs/ilbc/cb_mem_energy_calc.c
+++ b/webrtc/modules/audio_coding/codecs/ilbc/cb_mem_energy_calc.c
@@ -29,7 +29,7 @@
int16_t *energyW16, /* (o) Energy in the CB vectors */
int16_t *energyShifts, /* (o) Shift value of the energy */
int16_t scale, /* (i) The scaling of all energy values */
- int16_t base_size /* (i) Index to where the energy values should be stored */
+ int16_t base_size /* (i) Index to where energy values should be stored */
)
{
int16_t j,shft;
diff --git a/webrtc/modules/audio_coding/codecs/ilbc/cb_mem_energy_calc.h b/webrtc/modules/audio_coding/codecs/ilbc/cb_mem_energy_calc.h
index c7e1e54..6428269 100644
--- a/webrtc/modules/audio_coding/codecs/ilbc/cb_mem_energy_calc.h
+++ b/webrtc/modules/audio_coding/codecs/ilbc/cb_mem_energy_calc.h
@@ -27,7 +27,7 @@
int16_t *energyW16, /* (o) Energy in the CB vectors */
int16_t *energyShifts, /* (o) Shift value of the energy */
int16_t scale, /* (i) The scaling of all energy values */
- int16_t base_size /* (i) Index to where the energy values should be stored */
+ int16_t base_size /* (i) Index to where energy values should be stored */
);
#endif
diff --git a/webrtc/modules/audio_coding/codecs/ilbc/cb_search.c b/webrtc/modules/audio_coding/codecs/ilbc/cb_search.c
index 2ee9f6c..bc60149 100644
--- a/webrtc/modules/audio_coding/codecs/ilbc/cb_search.c
+++ b/webrtc/modules/audio_coding/codecs/ilbc/cb_search.c
@@ -147,7 +147,8 @@
/* Compute the CB vectors' energies for the second cb section (filtered cb) */
WebRtcIlbcfix_CbMemEnergyAugmentation(interpSamplesFilt, cbvectors,
- scale, (int16_t)(base_size+20), energyW16, energyShifts);
+ scale, (int16_t)(base_size + 20),
+ energyW16, energyShifts);
/* Compute the CB vectors' energies and store them in the vector
* energyW16. Also the corresponding shift values are stored. The
@@ -238,9 +239,12 @@
if (lTarget==SUBL) {
i=sInd;
if (sInd<20) {
- WebRtcIlbcfix_AugmentedCbCorr(target, cbvectors+lMem,
+ WebRtcIlbcfix_AugmentedCbCorr(target, cbvectors + lMem,
interpSamplesFilt, cDot,
- (int16_t)(sInd+20), (int16_t)(WEBRTC_SPL_MIN(39, (eInd+20))), scale);
+ (int16_t)(sInd + 20),
+ (int16_t)(WEBRTC_SPL_MIN(39,
+ (eInd + 20))),
+ scale);
i=20;
cDotPtr = &cDot[20 - sInd];
} else {
@@ -250,14 +254,16 @@
cb_vecPtr = cbvectors+lMem-20-i;
/* Calculate the cross correlations (main part of the filtered CB) */
- WebRtcSpl_CrossCorrelation(cDotPtr, target, cb_vecPtr, lTarget, (int16_t)(eInd-i+1), scale, -1);
+ WebRtcSpl_CrossCorrelation(cDotPtr, target, cb_vecPtr, lTarget,
+ (int16_t)(eInd - i + 1), scale, -1);
} else {
cDotPtr = cDot;
cb_vecPtr = cbvectors+lMem-lTarget-sInd;
/* Calculate the cross correlations (main part of the filtered CB) */
- WebRtcSpl_CrossCorrelation(cDotPtr, target, cb_vecPtr, lTarget, (int16_t)(eInd-sInd+1), scale, -1);
+ WebRtcSpl_CrossCorrelation(cDotPtr, target, cb_vecPtr, lTarget,
+ (int16_t)(eInd - sInd + 1), scale, -1);
}
diff --git a/webrtc/modules/audio_coding/codecs/ilbc/decode.c b/webrtc/modules/audio_coding/codecs/ilbc/decode.c
index 035460b..9918de2 100644
--- a/webrtc/modules/audio_coding/codecs/ilbc/decode.c
+++ b/webrtc/modules/audio_coding/codecs/ilbc/decode.c
@@ -103,9 +103,10 @@
WebRtcIlbcfix_DecodeResidual(iLBCdec_inst, iLBCbits_inst, decresidual, syntdenum);
/* preparing the plc for a future loss! */
- WebRtcIlbcfix_DoThePlc( PLCresidual, PLClpc, 0,
- decresidual, syntdenum + (LPC_FILTERORDER + 1)*(iLBCdec_inst->nsub - 1),
- (int16_t)(iLBCdec_inst->last_lag), iLBCdec_inst);
+ WebRtcIlbcfix_DoThePlc(
+ PLCresidual, PLClpc, 0, decresidual,
+ syntdenum + (LPC_FILTERORDER + 1) * (iLBCdec_inst->nsub - 1),
+ (int16_t)(iLBCdec_inst->last_lag), iLBCdec_inst);
/* Use the output from doThePLC */
WEBRTC_SPL_MEMCPY_W16(decresidual, PLCresidual, iLBCdec_inst->blockl);
@@ -120,8 +121,8 @@
/* packet loss conceal */
- WebRtcIlbcfix_DoThePlc( PLCresidual, PLClpc, 1,
- decresidual, syntdenum, (int16_t)(iLBCdec_inst->last_lag), iLBCdec_inst);
+ WebRtcIlbcfix_DoThePlc(PLCresidual, PLClpc, 1, decresidual, syntdenum,
+ (int16_t)(iLBCdec_inst->last_lag), iLBCdec_inst);
WEBRTC_SPL_MEMCPY_W16(decresidual, PLCresidual, iLBCdec_inst->blockl);
diff --git a/webrtc/modules/audio_coding/codecs/ilbc/enhancer_interface.c b/webrtc/modules/audio_coding/codecs/ilbc/enhancer_interface.c
index 262a564..6dca0b7 100644
--- a/webrtc/modules/audio_coding/codecs/ilbc/enhancer_interface.c
+++ b/webrtc/modules/audio_coding/codecs/ilbc/enhancer_interface.c
@@ -336,8 +336,8 @@
enh_bufPtr1,
synt,
&iLBCdec_inst->old_syntdenum[
- (iLBCdec_inst->nsub-1)*(LPC_FILTERORDER+1)],
- LPC_FILTERORDER+1, lag);
+ (iLBCdec_inst->nsub-1)*(LPC_FILTERORDER+1)],
+ LPC_FILTERORDER+1, lag);
WEBRTC_SPL_MEMCPY_W16(&synt[-LPC_FILTERORDER], &synt[lag-LPC_FILTERORDER],
LPC_FILTERORDER);
@@ -347,8 +347,8 @@
WebRtcSpl_FilterARFastQ12(
enh_bufPtr1, synt,
&iLBCdec_inst->old_syntdenum[
- (iLBCdec_inst->nsub-1)*(LPC_FILTERORDER+1)],
- LPC_FILTERORDER+1, lag);
+ (iLBCdec_inst->nsub-1)*(LPC_FILTERORDER+1)],
+ LPC_FILTERORDER+1, lag);
WEBRTC_SPL_MEMCPY_W16(iLBCdec_inst->syntMem, &synt[lag-LPC_FILTERORDER],
LPC_FILTERORDER);
diff --git a/webrtc/modules/audio_coding/codecs/ilbc/my_corr.c b/webrtc/modules/audio_coding/codecs/ilbc/my_corr.c
index ec3cf20..ab08001 100644
--- a/webrtc/modules/audio_coding/codecs/ilbc/my_corr.c
+++ b/webrtc/modules/audio_coding/codecs/ilbc/my_corr.c
@@ -23,10 +23,10 @@
*---------------------------------------------------------------*/
void WebRtcIlbcfix_MyCorr(
- int32_t *corr, /* (o) correlation of seq1 and seq2 */
- int16_t *seq1, /* (i) first sequence */
+ int32_t* corr, /* (o) correlation of seq1 and seq2 */
+ const int16_t* seq1, /* (i) first sequence */
int16_t dim1, /* (i) dimension first seq1 */
- const int16_t *seq2, /* (i) second sequence */
+ const int16_t* seq2, /* (i) second sequence */
int16_t dim2 /* (i) dimension seq2 */
){
int16_t max, scale, loops;
diff --git a/webrtc/modules/audio_coding/codecs/ilbc/my_corr.h b/webrtc/modules/audio_coding/codecs/ilbc/my_corr.h
index ee66998..a74dd1e 100644
--- a/webrtc/modules/audio_coding/codecs/ilbc/my_corr.h
+++ b/webrtc/modules/audio_coding/codecs/ilbc/my_corr.h
@@ -26,10 +26,10 @@
*---------------------------------------------------------------*/
void WebRtcIlbcfix_MyCorr(
- int32_t *corr, /* (o) correlation of seq1 and seq2 */
- int16_t *seq1, /* (i) first sequence */
+ int32_t* corr, /* (o) correlation of seq1 and seq2 */
+ const int16_t* seq1, /* (i) first sequence */
int16_t dim1, /* (i) dimension first seq1 */
- const int16_t *seq2, /* (i) second sequence */
+ const int16_t* seq2, /* (i) second sequence */
int16_t dim2 /* (i) dimension seq2 */
);
diff --git a/webrtc/modules/audio_coding/codecs/ilbc/test/iLBC_test.c b/webrtc/modules/audio_coding/codecs/ilbc/test/iLBC_test.c
index 9c42037..6ee3df4 100644
--- a/webrtc/modules/audio_coding/codecs/ilbc/test/iLBC_test.c
+++ b/webrtc/modules/audio_coding/codecs/ilbc/test/iLBC_test.c
@@ -52,6 +52,7 @@
int blockcount = 0;
int packetlosscount = 0;
int frameLen;
+ size_t len_i16s;
int16_t speechType;
IlbcEncoderInstance *Enc_Inst;
IlbcDecoderInstance *Dec_Inst;
@@ -173,9 +174,8 @@
/* write byte file */
- if (fwrite(encoded_data, sizeof(int16_t),
- ((len+1)/sizeof(int16_t)), efileid) !=
- (size_t)(((len+1)/sizeof(int16_t)))) {
+ len_i16s = (len + 1) / sizeof(int16_t);
+ if (fwrite(encoded_data, sizeof(int16_t), len_i16s, efileid) != len_i16s) {
return -1;
}
diff --git a/webrtc/modules/audio_coding/codecs/ilbc/test/iLBC_testLib.c b/webrtc/modules/audio_coding/codecs/ilbc/test/iLBC_testLib.c
index df37bec..3dcda29c 100644
--- a/webrtc/modules/audio_coding/codecs/ilbc/test/iLBC_testLib.c
+++ b/webrtc/modules/audio_coding/codecs/ilbc/test/iLBC_testLib.c
@@ -42,6 +42,7 @@
FILE *ifileid,*efileid,*ofileid, *chfileid;
short encoded_data[55], data[240], speechType;
short len, mode, pli;
+ size_t readlen;
int blockcount = 0;
IlbcEncoderInstance *Enc_Inst;
@@ -125,19 +126,16 @@
/* loop over input blocks */
#ifdef SPLIT_10MS
- while(fread(data, sizeof(short), 80, ifileid) == 80) {
+ readlen = 80;
#else
- while((short)fread(data,sizeof(short),(mode<<3),ifileid)==(mode<<3)) {
+ readlen = (size_t)(mode << 3);
#endif
+ while(fread(data, sizeof(short), readlen, ifileid) == readlen) {
blockcount++;
/* encoding */
fprintf(stderr, "--- Encoding block %i --- ",blockcount);
-#ifdef SPLIT_10MS
- len=WebRtcIlbcfix_Encode(Enc_Inst, data, 80, encoded_data);
-#else
- len=WebRtcIlbcfix_Encode(Enc_Inst, data, (short)(mode<<3), encoded_data);
-#endif
+ len=WebRtcIlbcfix_Encode(Enc_Inst, data, (short)readlen, encoded_data);
if (len < 0) {
fprintf(stderr, "Error encoding\n");
exit(0);
@@ -152,9 +150,7 @@
/* write byte file */
if(len != 0){ //len may be 0 in 10ms split case
fwrite(encoded_data,1,len,efileid);
- }
- if(len != 0){ //len may be 0 in 10ms split case
/* get channel data if provided */
if (argc==6) {
if (fread(&pli, sizeof(int16_t), 1, chfileid)) {
diff --git a/webrtc/modules/audio_coding/codecs/ilbc/xcorr_coef.c b/webrtc/modules/audio_coding/codecs/ilbc/xcorr_coef.c
index 328a5fe..53d95bf 100644
--- a/webrtc/modules/audio_coding/codecs/ilbc/xcorr_coef.c
+++ b/webrtc/modules/audio_coding/codecs/ilbc/xcorr_coef.c
@@ -57,11 +57,11 @@
if (step==1) {
max=WebRtcSpl_MaxAbsValueW16(regressor, subl + searchLen - 1);
rp_beg = regressor;
- rp_end = ®ressor[subl];
+ rp_end = regressor + subl;
} else { /* step==-1 */
- max=WebRtcSpl_MaxAbsValueW16(®ressor[-searchLen], subl + searchLen - 1);
- rp_beg = ®ressor[-1];
- rp_end = ®ressor[subl-1];
+ max = WebRtcSpl_MaxAbsValueW16(regressor - searchLen, subl + searchLen - 1);
+ rp_beg = regressor - 1;
+ rp_end = regressor + subl - 1;
}
/* Introduce a scale factor on the Energy in int32_t in
diff --git a/webrtc/modules/audio_coding/codecs/isac/fix/interface/isacfix.h b/webrtc/modules/audio_coding/codecs/isac/fix/interface/isacfix.h
index 961fd3f..421a08c 100644
--- a/webrtc/modules/audio_coding/codecs/isac/fix/interface/isacfix.h
+++ b/webrtc/modules/audio_coding/codecs/isac/fix/interface/isacfix.h
@@ -227,10 +227,10 @@
int16_t WebRtcIsacfix_UpdateBwEstimate(ISACFIX_MainStruct *ISAC_main_inst,
const uint8_t* encoded,
- int32_t packet_size,
- uint16_t rtp_seq_number,
- uint32_t send_ts,
- uint32_t arr_ts);
+ int32_t packet_size,
+ uint16_t rtp_seq_number,
+ uint32_t send_ts,
+ uint32_t arr_ts);
/****************************************************************************
* WebRtcIsacfix_Decode(...)
diff --git a/webrtc/modules/audio_coding/codecs/isac/fix/source/codec.h b/webrtc/modules/audio_coding/codecs/isac/fix/source/codec.h
index 1270cc3..55623a2 100644
--- a/webrtc/modules/audio_coding/codecs/isac/fix/source/codec.h
+++ b/webrtc/modules/audio_coding/codecs/isac/fix/source/codec.h
@@ -36,9 +36,9 @@
IsacFixDecoderInstance* ISACdec_obj,
int16_t* current_framesamples);
-int16_t WebRtcIsacfix_DecodePlcImpl(int16_t* decoded,
- IsacFixDecoderInstance* ISACdec_obj,
- int16_t* current_framesample );
+void WebRtcIsacfix_DecodePlcImpl(int16_t* decoded,
+ IsacFixDecoderInstance* ISACdec_obj,
+ int16_t* current_framesample );
int WebRtcIsacfix_EncodeImpl(int16_t* in,
IsacFixEncoderInstance* ISACenc_obj,
diff --git a/webrtc/modules/audio_coding/codecs/isac/fix/source/decode_plc.c b/webrtc/modules/audio_coding/codecs/isac/fix/source/decode_plc.c
index 1a7ff92..c3a89c3 100644
--- a/webrtc/modules/audio_coding/codecs/isac/fix/source/decode_plc.c
+++ b/webrtc/modules/audio_coding/codecs/isac/fix/source/decode_plc.c
@@ -175,7 +175,10 @@
-static void LinearResampler( int16_t *in, int16_t *out, int16_t lenIn, int16_t lenOut )
+static void LinearResampler(int16_t* in,
+ int16_t* out,
+ int16_t lenIn,
+ int16_t lenOut)
{
int32_t n = (lenIn - 1) * RESAMP_RES;
int16_t resOut, i, j, relativePos, diff; /* */
@@ -230,12 +233,11 @@
-int16_t WebRtcIsacfix_DecodePlcImpl(int16_t *signal_out16,
- IsacFixDecoderInstance *ISACdec_obj,
- int16_t *current_framesamples )
+void WebRtcIsacfix_DecodePlcImpl(int16_t *signal_out16,
+ IsacFixDecoderInstance *ISACdec_obj,
+ int16_t *current_framesamples )
{
int subframecnt;
- int16_t len = 0;
int16_t* Vector_Word16_1;
int16_t Vector_Word16_Extended_1[FRAMESAMPLES_HALF + NOISE_FILTER_LEN];
@@ -797,6 +799,4 @@
(ISACdec_obj->plcstr_obj).used = PLC_WAS_USED;
*current_framesamples = 480;
-
- return len;
}
diff --git a/webrtc/modules/audio_coding/codecs/isac/fix/source/entropy_coding.c b/webrtc/modules/audio_coding/codecs/isac/fix/source/entropy_coding.c
index 3965378..9f52c9d 100644
--- a/webrtc/modules/audio_coding/codecs/isac/fix/source/entropy_coding.c
+++ b/webrtc/modules/audio_coding/codecs/isac/fix/source/entropy_coding.c
@@ -1675,7 +1675,7 @@
int32_t meangainQ12;
int32_t CQ11, CQ10,tmp32a,tmp32b;
- int16_t shft,tmp16a,tmp16c;
+ int16_t shft;
meangainQ12=0;
for (k = 0; k < 4; k++)
@@ -1725,22 +1725,19 @@
CQ11 = WEBRTC_SPL_SHIFT_W32(CQ11,11-shft); // Scale with StepSize, Q11
for (k=0; k<PITCH_SUBFRAMES; k++) {
tmp32a = WEBRTC_SPL_MUL_16_32_RSFT11(WebRtcIsacfix_kTransform[0][k], CQ11);
- tmp16a = (int16_t)(tmp32a >> 5);
- PitchLags_Q7[k] = tmp16a;
+ PitchLags_Q7[k] = (int16_t)(tmp32a >> 5);
}
CQ10 = mean_val2Q10[index[1]];
for (k=0; k<PITCH_SUBFRAMES; k++) {
tmp32b = WebRtcIsacfix_kTransform[1][k] * (int16_t)CQ10 >> 10;
- tmp16c = (int16_t)(tmp32b >> 5);
- PitchLags_Q7[k] += tmp16c;
+ PitchLags_Q7[k] += (int16_t)(tmp32b >> 5);
}
CQ10 = mean_val4Q10[index[3]];
for (k=0; k<PITCH_SUBFRAMES; k++) {
tmp32b = WebRtcIsacfix_kTransform[3][k] * (int16_t)CQ10 >> 10;
- tmp16c = (int16_t)(tmp32b >> 5);
- PitchLags_Q7[k] += tmp16c;
+ PitchLags_Q7[k] += (int16_t)(tmp32b >> 5);
}
return 0;
@@ -1761,7 +1758,7 @@
const int16_t *mean_val2Q10,*mean_val4Q10;
const int16_t *lower_limit, *upper_limit;
const uint16_t **cdf;
- int16_t shft, tmp16a, tmp16b, tmp16c;
+ int16_t shft, tmp16b;
int32_t tmp32b;
int status = 0;
@@ -1832,22 +1829,19 @@
for (k=0; k<PITCH_SUBFRAMES; k++) {
tmp32a = WEBRTC_SPL_MUL_16_32_RSFT11(WebRtcIsacfix_kTransform[0][k], CQ11); // Q12
- tmp16a = (int16_t)(tmp32a >> 5); // Q7.
- PitchLagsQ7[k] = tmp16a;
+ PitchLagsQ7[k] = (int16_t)(tmp32a >> 5); // Q7.
}
CQ10 = mean_val2Q10[index[1]];
for (k=0; k<PITCH_SUBFRAMES; k++) {
tmp32b = WebRtcIsacfix_kTransform[1][k] * (int16_t)CQ10 >> 10;
- tmp16c = (int16_t)(tmp32b >> 5); // Q7.
- PitchLagsQ7[k] += tmp16c;
+ PitchLagsQ7[k] += (int16_t)(tmp32b >> 5); // Q7.
}
CQ10 = mean_val4Q10[index[3]];
for (k=0; k<PITCH_SUBFRAMES; k++) {
tmp32b = WebRtcIsacfix_kTransform[3][k] * (int16_t)CQ10 >> 10;
- tmp16c = (int16_t)(tmp32b >> 5); // Q7.
- PitchLagsQ7[k] += tmp16c;
+ PitchLagsQ7[k] += (int16_t)(tmp32b >> 5); // Q7.
}
/* entropy coding of quantization pitch lags */
diff --git a/webrtc/modules/audio_coding/codecs/isac/fix/source/isacfix.c b/webrtc/modules/audio_coding/codecs/isac/fix/source/isacfix.c
index 03bceec..25076d8 100644
--- a/webrtc/modules/audio_coding/codecs/isac/fix/source/isacfix.c
+++ b/webrtc/modules/audio_coding/codecs/isac/fix/source/isacfix.c
@@ -620,9 +620,9 @@
int16_t WebRtcIsacfix_UpdateBwEstimate1(ISACFIX_MainStruct *ISAC_main_inst,
const uint8_t* encoded,
- int32_t packet_size,
- uint16_t rtp_seq_number,
- uint32_t arr_ts)
+ int32_t packet_size,
+ uint16_t rtp_seq_number,
+ uint32_t arr_ts)
{
ISACFIX_SubStruct *ISAC_inst;
Bitstr_dec streamdata;
@@ -692,10 +692,10 @@
int16_t WebRtcIsacfix_UpdateBwEstimate(ISACFIX_MainStruct *ISAC_main_inst,
const uint8_t* encoded,
- int32_t packet_size,
- uint16_t rtp_seq_number,
- uint32_t send_ts,
- uint32_t arr_ts)
+ int32_t packet_size,
+ uint16_t rtp_seq_number,
+ uint32_t send_ts,
+ uint32_t arr_ts)
{
ISACFIX_SubStruct *ISAC_inst;
Bitstr_dec streamdata;
@@ -767,11 +767,11 @@
*/
-int16_t WebRtcIsacfix_Decode(ISACFIX_MainStruct *ISAC_main_inst,
+int16_t WebRtcIsacfix_Decode(ISACFIX_MainStruct* ISAC_main_inst,
const uint8_t* encoded,
- int16_t len,
- int16_t *decoded,
- int16_t *speechType)
+ int16_t len,
+ int16_t* decoded,
+ int16_t* speechType)
{
ISACFIX_SubStruct *ISAC_inst;
/* number of samples (480 or 960), output from decoder */
@@ -981,9 +981,8 @@
declen = 0;
while( noOfLostFrames > 0 )
{
- ok = WebRtcIsacfix_DecodePlcImpl( outframeWB, &ISAC_inst->ISACdec_obj, &no_of_samples );
- if(ok)
- return -1;
+ WebRtcIsacfix_DecodePlcImpl(outframeWB, &ISAC_inst->ISACdec_obj,
+ &no_of_samples);
WebRtcIsacfix_SplitAndFilter2(outframeWB, &(outframeNB[k*240]), dummy, &ISAC_inst->ISACdec_obj.decimatorstr_obj);
@@ -1029,7 +1028,7 @@
int16_t noOfLostFrames)
{
- int16_t no_of_samples, declen, k, ok;
+ int16_t no_of_samples, declen, k;
int16_t outframe16[MAX_FRAMESAMPLES];
ISACFIX_SubStruct *ISAC_inst;
@@ -1044,9 +1043,8 @@
declen = 0;
while( noOfLostFrames > 0 )
{
- ok = WebRtcIsacfix_DecodePlcImpl( &(outframe16[k*480]), &ISAC_inst->ISACdec_obj, &no_of_samples );
- if(ok)
- return -1;
+ WebRtcIsacfix_DecodePlcImpl(&(outframe16[k*480]), &ISAC_inst->ISACdec_obj,
+ &no_of_samples);
declen += no_of_samples;
noOfLostFrames--;
k++;
diff --git a/webrtc/modules/audio_coding/codecs/isac/fix/source/lattice.c b/webrtc/modules/audio_coding/codecs/isac/fix/source/lattice.c
index 3b26a98..7fcb9e3 100644
--- a/webrtc/modules/audio_coding/codecs/isac/fix/source/lattice.c
+++ b/webrtc/modules/audio_coding/codecs/isac/fix/source/lattice.c
@@ -279,7 +279,8 @@
ARfQ0vec[i] = (int16_t)WebRtcSpl_SatW32ToW16(tmp32); // Q0
}
- for (i=orderCoef;i>0;i--) //get the state of f&g for the first input, for all orders
+ // Get the state of f & g for the first input, for all orders.
+ for (i = orderCoef; i > 0; i--)
{
tmp32 = (cthQ15[i - 1] * ARfQ0vec[0] - sthQ15[i - 1] * stateGQ0[i - 1] +
16384) >> 15;
diff --git a/webrtc/modules/audio_coding/codecs/isac/fix/source/lattice_armv7.S b/webrtc/modules/audio_coding/codecs/isac/fix/source/lattice_armv7.S
index ccea467..4a0d99f 100644
--- a/webrtc/modules/audio_coding/codecs/isac/fix/source/lattice_armv7.S
+++ b/webrtc/modules/audio_coding/codecs/isac/fix/source/lattice_armv7.S
@@ -55,9 +55,11 @@
smlabb r11, r7, r5, r12 @ sth_Q15[k - 1] * tmpAR + 16384
smlabb r10, r6, r5, r12 @ cth_Q15[k - 1] * tmpAR + 16384
smulbb r7, r7, r8 @ sth_Q15[k - 1] * ar_g_Q0[k - 1]
- smlabb r11, r6, r8, r11 @ cth_Q15[k - 1]*ar_g_Q0[k - 1]+(sth_Q15[k - 1]*tmpAR+16384)
+ smlabb r11, r6, r8, r11 @ cth_Q15[k - 1] * ar_g_Q0[k - 1] +
+ @ (sth_Q15[k - 1] * tmpAR + 16384)
- sub r10, r10, r7 @ cth_Q15[k - 1]*tmpAR+16384-(sth_Q15[k - 1]*ar_g_Q0[k - 1])
+ sub r10, r10, r7 @ cth_Q15[k - 1] * tmpAR + 16384 -
+ @ (sth_Q15[k - 1] * ar_g_Q0[k - 1])
ssat r11, #16, r11, asr #15
ssat r5, #16, r10, asr #15
strh r11, [r0], #-2 @ Output: ar_g_Q0[k]
diff --git a/webrtc/modules/audio_coding/codecs/isac/fix/source/pitch_filter.c b/webrtc/modules/audio_coding/codecs/isac/fix/source/pitch_filter.c
index 1149d50..c787d6e 100644
--- a/webrtc/modules/audio_coding/codecs/isac/fix/source/pitch_filter.c
+++ b/webrtc/modules/audio_coding/codecs/isac/fix/source/pitch_filter.c
@@ -34,19 +34,6 @@
{ 271, -743, 1570, -3320, 12963, 7301, -2292, 953, -325}
};
-// Function prototype for pitch filtering.
-// TODO(Turaj): Add descriptions of input and output parameters.
-void WebRtcIsacfix_PitchFilterCore(int loopNumber,
- int16_t gain,
- int index,
- int16_t sign,
- int16_t* inputState,
- int16_t* outputBuf2,
- const int16_t* coefficient,
- int16_t* inputBuf,
- int16_t* outputBuf,
- int* index2);
-
static __inline int32_t CalcLrIntQ(int32_t fixVal,
int16_t qDomain) {
int32_t roundVal = 1 << (qDomain - 1);
diff --git a/webrtc/modules/audio_coding/codecs/isac/fix/test/isac_speed_test.cc b/webrtc/modules/audio_coding/codecs/isac/fix/test/isac_speed_test.cc
index 3089f58..8f073ad 100644
--- a/webrtc/modules/audio_coding/codecs/isac/fix/test/isac_speed_test.cc
+++ b/webrtc/modules/audio_coding/codecs/isac/fix/test/isac_speed_test.cc
@@ -83,14 +83,14 @@
return 1000.0 * clocks / CLOCKS_PER_SEC;
}
-float IsacSpeedTest::DecodeABlock(const uint8_t* bit_stream, int encoded_bytes,
+float IsacSpeedTest::DecodeABlock(const uint8_t* bit_stream,
+ int encoded_bytes,
int16_t* out_data) {
int value;
int16_t audio_type;
clock_t clocks = clock();
- value = WebRtcIsacfix_Decode(ISACFIX_main_inst_,
- bit_stream,
- encoded_bytes, out_data, &audio_type);
+ value = WebRtcIsacfix_Decode(ISACFIX_main_inst_, bit_stream, encoded_bytes,
+ out_data, &audio_type);
clocks = clock() - clocks;
EXPECT_EQ(output_length_sample_, value);
return 1000.0 * clocks / CLOCKS_PER_SEC;
diff --git a/webrtc/modules/audio_coding/codecs/isac/fix/test/test_iSACfixfloat.c b/webrtc/modules/audio_coding/codecs/isac/fix/test/test_iSACfixfloat.c
index e2a778a..fae6d6a 100644
--- a/webrtc/modules/audio_coding/codecs/isac/fix/test/test_iSACfixfloat.c
+++ b/webrtc/modules/audio_coding/codecs/isac/fix/test/test_iSACfixfloat.c
@@ -582,8 +582,7 @@
totalsmpls += declen;
totalbits += 8 * stream_len;
- kbps = ((double)FS) / ((double)cur_framesmpls) * 8.0 * stream_len /
- 1000.0; // kbits/s
+ kbps = ((double)FS) / ((double)cur_framesmpls) * 8.0 * stream_len / 1000.0;
fy = fopen("bit_rate.dat", "a");
fprintf(fy, "Frame %i = %0.14f\n", framecnt, kbps);
fclose(fy);
diff --git a/webrtc/modules/audio_coding/codecs/isac/main/source/bandwidth_estimator.c b/webrtc/modules/audio_coding/codecs/isac/main/source/bandwidth_estimator.c
index c4ceb59..ce8ceb2 100644
--- a/webrtc/modules/audio_coding/codecs/isac/main/source/bandwidth_estimator.c
+++ b/webrtc/modules/audio_coding/codecs/isac/main/source/bandwidth_estimator.c
@@ -132,12 +132,12 @@
/* Index - integer (range 0...23) indicating bottle neck & jitter as estimated by other side */
/* returns 0 if everything went fine, -1 otherwise */
int16_t WebRtcIsac_UpdateBandwidthEstimator(
- BwEstimatorstr *bwest_str,
+ BwEstimatorstr* bwest_str,
const uint16_t rtp_number,
- const int32_t frame_length,
+ const int32_t frame_length,
const uint32_t send_ts,
const uint32_t arr_ts,
- const int32_t pksize
+ const int32_t pksize
/*, const uint16_t Index*/)
{
float weight = 0.0f;
diff --git a/webrtc/modules/audio_coding/codecs/isac/main/source/bandwidth_estimator.h b/webrtc/modules/audio_coding/codecs/isac/main/source/bandwidth_estimator.h
index edabdff..8482a8c 100644
--- a/webrtc/modules/audio_coding/codecs/isac/main/source/bandwidth_estimator.h
+++ b/webrtc/modules/audio_coding/codecs/isac/main/source/bandwidth_estimator.h
@@ -90,12 +90,12 @@
/* Index - integer (range 0...23) indicating bottle neck & jitter as estimated by other side */
/* returns 0 if everything went fine, -1 otherwise */
int16_t WebRtcIsac_UpdateBandwidthEstimator(
- BwEstimatorstr* bwest_str,
+ BwEstimatorstr* bwest_str,
const uint16_t rtp_number,
- const int32_t frame_length,
+ const int32_t frame_length,
const uint32_t send_ts,
const uint32_t arr_ts,
- const int32_t pksize);
+ const int32_t pksize);
/* Update receiving estimates. Used when we only receive BWE index, no iSAC data packet. */
int16_t WebRtcIsac_UpdateUplinkBwImpl(
diff --git a/webrtc/modules/audio_coding/codecs/isac/main/source/lpc_analysis.c b/webrtc/modules/audio_coding/codecs/isac/main/source/lpc_analysis.c
index 5198ebf..4708a5c 100644
--- a/webrtc/modules/audio_coding/codecs/isac/main/source/lpc_analysis.c
+++ b/webrtc/modules/audio_coding/codecs/isac/main/source/lpc_analysis.c
@@ -78,8 +78,8 @@
double WebRtcIsac_LevDurb(double *a, double *k, double *r, int order)
{
- double sum, alpha;
- int m, m_h, i;
+ double sum, alpha;
+ int m, m_h, i;
alpha = 0; //warning -DH
a[0] = 1.0;
if (r[0] < LEVINSON_EPS) { /* if r[0] <= 0, set LPC coeff. to zero */
diff --git a/webrtc/modules/audio_coding/codecs/opus/audio_encoder_opus.cc b/webrtc/modules/audio_coding/codecs/opus/audio_encoder_opus.cc
index e69b0c8..17fa5b2 100644
--- a/webrtc/modules/audio_coding/codecs/opus/audio_encoder_opus.cc
+++ b/webrtc/modules/audio_coding/codecs/opus/audio_encoder_opus.cc
@@ -114,10 +114,10 @@
size_t AudioEncoderOpus::MaxEncodedBytes() const {
// Calculate the number of bytes we expect the encoder to produce,
// then multiply by two to give a wide margin for error.
- int frame_size_ms = num_10ms_frames_per_packet_ * 10;
size_t bytes_per_millisecond =
- static_cast<size_t>(bitrate_bps_ / (1000 * 8) + 1);
- size_t approx_encoded_bytes = frame_size_ms * bytes_per_millisecond;
+ static_cast<size_t>(bitrate_bps_ / (1000 * 8) + 1);
+ size_t approx_encoded_bytes =
+ num_10ms_frames_per_packet_ * 10 * bytes_per_millisecond;
return 2 * approx_encoded_bytes;
}
diff --git a/webrtc/modules/audio_coding/codecs/opus/opus_unittest.cc b/webrtc/modules/audio_coding/codecs/opus/opus_unittest.cc
index bb7bee9..028d2ec 100644
--- a/webrtc/modules/audio_coding/codecs/opus/opus_unittest.cc
+++ b/webrtc/modules/audio_coding/codecs/opus/opus_unittest.cc
@@ -46,7 +46,7 @@
int EncodeDecode(WebRtcOpusEncInst* encoder,
const int16_t* input_audio,
- const int input_samples,
+ int input_samples,
WebRtcOpusDecInst* decoder,
int16_t* output_audio,
int16_t* audio_type);
@@ -98,7 +98,7 @@
int OpusTest::EncodeDecode(WebRtcOpusEncInst* encoder,
const int16_t* input_audio,
- const int input_samples,
+ int input_samples,
WebRtcOpusDecInst* decoder,
int16_t* output_audio,
int16_t* audio_type) {
@@ -165,7 +165,7 @@
EXPECT_EQ(0, opus_encoder_->in_dtx_mode);
EXPECT_EQ(0, opus_decoder_->in_dtx_mode);
EXPECT_EQ(0, audio_type); // Speech.
- } else if (1 == encoded_bytes_) {
+ } else if (encoded_bytes_ == 1) {
EXPECT_EQ(1, opus_encoder_->in_dtx_mode);
EXPECT_EQ(1, opus_decoder_->in_dtx_mode);
EXPECT_EQ(2, audio_type); // Comfort noise.
diff --git a/webrtc/modules/audio_coding/main/acm2/acm_receive_test.cc b/webrtc/modules/audio_coding/main/acm2/acm_receive_test.cc
index e74ce22..dc59984 100644
--- a/webrtc/modules/audio_coding/main/acm2/acm_receive_test.cc
+++ b/webrtc/modules/audio_coding/main/acm2/acm_receive_test.cc
@@ -25,12 +25,12 @@
AcmReceiveTest::AcmReceiveTest(PacketSource* packet_source,
AudioSink* audio_sink,
int output_freq_hz,
- NumOutputChannels exptected_output_channels)
+ NumOutputChannels expected_output_channels)
: clock_(0),
packet_source_(packet_source),
audio_sink_(audio_sink),
output_freq_hz_(output_freq_hz),
- exptected_output_channels_(exptected_output_channels) {
+ expected_output_channels_(expected_output_channels) {
webrtc::AudioCoding::Config config;
config.clock = &clock_;
config.playout_frequency_hz = output_freq_hz_;
@@ -95,13 +95,13 @@
EXPECT_EQ(output_freq_hz_, output_frame.sample_rate_hz_);
const int samples_per_block = output_freq_hz_ * 10 / 1000;
EXPECT_EQ(samples_per_block, output_frame.samples_per_channel_);
- if (exptected_output_channels_ != kArbitraryChannels) {
+ if (expected_output_channels_ != kArbitraryChannels) {
if (output_frame.speech_type_ == webrtc::AudioFrame::kPLC) {
// Don't check number of channels for PLC output, since each test run
// usually starts with a short period of mono PLC before decoding the
// first packet.
} else {
- EXPECT_EQ(exptected_output_channels_, output_frame.num_channels_);
+ EXPECT_EQ(expected_output_channels_, output_frame.num_channels_);
}
}
ASSERT_TRUE(audio_sink_->WriteAudioFrame(output_frame));
diff --git a/webrtc/modules/audio_coding/main/acm2/acm_receive_test.h b/webrtc/modules/audio_coding/main/acm2/acm_receive_test.h
index 552a748..a1e0142 100644
--- a/webrtc/modules/audio_coding/main/acm2/acm_receive_test.h
+++ b/webrtc/modules/audio_coding/main/acm2/acm_receive_test.h
@@ -35,7 +35,7 @@
PacketSource* packet_source,
AudioSink* audio_sink,
int output_freq_hz,
- NumOutputChannels exptected_output_channels);
+ NumOutputChannels expected_output_channels);
virtual ~AcmReceiveTest() {}
// Registers the codecs with default parameters from ACM.
@@ -54,7 +54,7 @@
PacketSource* packet_source_;
AudioSink* audio_sink_;
const int output_freq_hz_;
- NumOutputChannels exptected_output_channels_;
+ NumOutputChannels expected_output_channels_;
DISALLOW_COPY_AND_ASSIGN(AcmReceiveTest);
};
diff --git a/webrtc/modules/audio_coding/main/acm2/audio_coding_module_impl.cc b/webrtc/modules/audio_coding/main/acm2/audio_coding_module_impl.cc
index ce98636..4e665ea 100644
--- a/webrtc/modules/audio_coding/main/acm2/audio_coding_module_impl.cc
+++ b/webrtc/modules/audio_coding/main/acm2/audio_coding_module_impl.cc
@@ -339,8 +339,8 @@
}
// If the length and frequency matches. We currently just support raw PCM.
- if ((audio_frame.sample_rate_hz_ / 100)
- != audio_frame.samples_per_channel_) {
+ if ((audio_frame.sample_rate_hz_ / 100) !=
+ audio_frame.samples_per_channel_) {
WEBRTC_TRACE(webrtc::kTraceError, webrtc::kTraceAudioCoding, id_,
"Cannot Add 10 ms audio, input frequency and length doesn't"
" match");
diff --git a/webrtc/modules/audio_coding/main/test/opus_test.cc b/webrtc/modules/audio_coding/main/test/opus_test.cc
index 09301df..a407fc5 100644
--- a/webrtc/modules/audio_coding/main/test/opus_test.cc
+++ b/webrtc/modules/audio_coding/main/test/opus_test.cc
@@ -277,12 +277,12 @@
bitstream_len_byte = WebRtcOpus_Encode(
opus_mono_encoder_, &audio[read_samples],
frame_length, kMaxBytes, bitstream);
- ASSERT_GT(bitstream_len_byte, -1);
+ ASSERT_GE(bitstream_len_byte, 0);
} else {
bitstream_len_byte = WebRtcOpus_Encode(
opus_stereo_encoder_, &audio[read_samples],
frame_length, kMaxBytes, bitstream);
- ASSERT_GT(bitstream_len_byte, -1);
+ ASSERT_GE(bitstream_len_byte, 0);
}
// Simulate packet loss by setting |packet_loss_| to "true" in
diff --git a/webrtc/modules/audio_coding/neteq/decision_logic_normal.h b/webrtc/modules/audio_coding/neteq/decision_logic_normal.h
index 0254839..047663f 100644
--- a/webrtc/modules/audio_coding/neteq/decision_logic_normal.h
+++ b/webrtc/modules/audio_coding/neteq/decision_logic_normal.h
@@ -61,7 +61,8 @@
virtual Operations FuturePacketAvailable(
const SyncBuffer& sync_buffer,
const Expand& expand,
- int decoder_frame_length, Modes prev_mode,
+ int decoder_frame_length,
+ Modes prev_mode,
uint32_t target_timestamp,
uint32_t available_timestamp,
bool play_dtmf);
diff --git a/webrtc/modules/audio_coding/neteq/expand.cc b/webrtc/modules/audio_coding/neteq/expand.cc
index d5f0f9c..bde6559 100644
--- a/webrtc/modules/audio_coding/neteq/expand.cc
+++ b/webrtc/modules/audio_coding/neteq/expand.cc
@@ -214,24 +214,24 @@
// Create combined signal by shifting in more and more of unvoiced part.
temp_shift = 8 - temp_shift; // = getbits(mix_factor_increment).
- size_t temp_lenght = (parameters.current_voice_mix_factor -
+ size_t temp_length = (parameters.current_voice_mix_factor -
parameters.voice_mix_factor) >> temp_shift;
- temp_lenght = std::min(temp_lenght, current_lag);
- DspHelper::CrossFade(voiced_vector, unvoiced_vector, temp_lenght,
+ temp_length = std::min(temp_length, current_lag);
+ DspHelper::CrossFade(voiced_vector, unvoiced_vector, temp_length,
¶meters.current_voice_mix_factor,
mix_factor_increment, temp_data);
// End of cross-fading period was reached before end of expanded signal
// path. Mix the rest with a fixed mixing factor.
- if (temp_lenght < current_lag) {
+ if (temp_length < current_lag) {
if (mix_factor_increment != 0) {
parameters.current_voice_mix_factor = parameters.voice_mix_factor;
}
int16_t temp_scale = 16384 - parameters.current_voice_mix_factor;
WebRtcSpl_ScaleAndAddVectorsWithRound(
- voiced_vector + temp_lenght, parameters.current_voice_mix_factor,
- unvoiced_vector + temp_lenght, temp_scale, 14,
- temp_data + temp_lenght, static_cast<int>(current_lag - temp_lenght));
+ voiced_vector + temp_length, parameters.current_voice_mix_factor,
+ unvoiced_vector + temp_length, temp_scale, 14,
+ temp_data + temp_length, static_cast<int>(current_lag - temp_length));
}
// Select muting slope depending on how many consecutive expands we have
@@ -428,13 +428,12 @@
// Calculate the exact best correlation in the range between
// |correlation_lag| and |distortion_lag|.
- correlation_length = distortion_lag + 10;
- correlation_length = std::min(correlation_length, fs_mult_120);
- correlation_length = std::max(correlation_length, 60 * fs_mult);
+ correlation_length =
+ std::max(std::min(distortion_lag + 10, fs_mult_120), 60 * fs_mult);
int start_index = std::min(distortion_lag, correlation_lag);
- int correlation_lags = WEBRTC_SPL_ABS_W16((distortion_lag-correlation_lag))
- + 1;
+ int correlation_lags =
+ WEBRTC_SPL_ABS_W16((distortion_lag-correlation_lag)) + 1;
assert(correlation_lags <= 99 * fs_mult + 1); // Cannot be larger.
for (size_t channel_ix = 0; channel_ix < num_channels_; ++channel_ix) {
@@ -753,8 +752,10 @@
memset(ar_filter_state, 0, sizeof(ar_filter_state));
}
-int16_t Expand::Correlation(const int16_t* input, size_t input_length,
- int16_t* output, int16_t* output_scale) const {
+void Expand::Correlation(const int16_t* input,
+ size_t input_length,
+ int16_t* output,
+ int16_t* output_scale) const {
// Set parameters depending on sample rate.
const int16_t* filter_coefficients;
int16_t num_coefficients;
@@ -818,7 +819,6 @@
norm_shift2);
// Total scale factor (right shifts) of correlation value.
*output_scale = 2 * norm_shift + kCorrelationShift + norm_shift2;
- return kNumCorrelationLags;
}
void Expand::UpdateLagIndex() {
@@ -850,7 +850,7 @@
int16_t* buffer) {
static const int kNoiseLpcOrder = BackgroundNoise::kMaxLpcOrder;
int16_t scaled_random_vector[kMaxSampleRate / 8000 * 125];
- assert(static_cast<size_t>(kMaxSampleRate / 8000 * 125) >= num_noise_samples);
+ assert(num_noise_samples <= static_cast<size_t>(kMaxSampleRate / 8000 * 125));
int16_t* noise_samples = &buffer[kNoiseLpcOrder];
if (background_noise_->initialized()) {
// Use background noise parameters.
diff --git a/webrtc/modules/audio_coding/neteq/expand.h b/webrtc/modules/audio_coding/neteq/expand.h
index 0000642..b015959 100644
--- a/webrtc/modules/audio_coding/neteq/expand.h
+++ b/webrtc/modules/audio_coding/neteq/expand.h
@@ -119,9 +119,11 @@
// Calculate the auto-correlation of |input|, with length |input_length|
// samples. The correlation is calculated from a downsampled version of
// |input|, and is written to |output|. The scale factor is written to
- // |output_scale|. Returns the length of the correlation vector.
- int16_t Correlation(const int16_t* input, size_t input_length,
- int16_t* output, int16_t* output_scale) const;
+ // |output_scale|.
+ void Correlation(const int16_t* input,
+ size_t input_length,
+ int16_t* output,
+ int16_t* output_scale) const;
void UpdateLagIndex();
diff --git a/webrtc/modules/audio_coding/neteq/merge.cc b/webrtc/modules/audio_coding/neteq/merge.cc
index 8399a78..8e686ba 100644
--- a/webrtc/modules/audio_coding/neteq/merge.cc
+++ b/webrtc/modules/audio_coding/neteq/merge.cc
@@ -312,8 +312,8 @@
int expand_period) const {
// Calculate correlation without any normalization.
const int max_corr_length = kMaxCorrelationLength;
- int stop_position_downsamp = std::min(
- max_corr_length, expand_->max_lag() / (fs_mult_ * 2) + 1);
+ int stop_position_downsamp =
+ std::min(max_corr_length, expand_->max_lag() / (fs_mult_ * 2) + 1);
int16_t correlation_shift = 0;
if (expanded_max * input_max > 26843546) {
correlation_shift = 3;
@@ -367,9 +367,9 @@
// Ensure that underrun does not occur for 10ms case => we have to get at
// least 10ms + overlap . (This should never happen thanks to the above
// modification of peak-finding starting point.)
- while ((best_correlation_index + input_length) <
- static_cast<int>(timestamps_per_call_ + expand_->overlap_length()) ||
- best_correlation_index + input_length < start_position) {
+ while (((best_correlation_index + input_length) <
+ static_cast<int>(timestamps_per_call_ + expand_->overlap_length())) ||
+ ((best_correlation_index + input_length) < start_position)) {
assert(false); // Should never happen.
best_correlation_index += expand_period; // Jump one lag ahead.
}
diff --git a/webrtc/modules/audio_coding/neteq/neteq_external_decoder_unittest.cc b/webrtc/modules/audio_coding/neteq/neteq_external_decoder_unittest.cc
index 729dbf6..bf68b3b 100644
--- a/webrtc/modules/audio_coding/neteq/neteq_external_decoder_unittest.cc
+++ b/webrtc/modules/audio_coding/neteq/neteq_external_decoder_unittest.cc
@@ -74,7 +74,7 @@
return -1;
}
payload_size_bytes_ = WebRtcPcm16b_Encode(input_, frame_size_samples_,
- encoded_);;
+ encoded_);
int next_send_time = rtp_generator_->GetRtpHeader(
kPayloadType, frame_size_samples_, &rtp_header_);
diff --git a/webrtc/modules/audio_coding/neteq/neteq_impl.cc b/webrtc/modules/audio_coding/neteq/neteq_impl.cc
index 3a3ad98..29b8d1a 100644
--- a/webrtc/modules/audio_coding/neteq/neteq_impl.cc
+++ b/webrtc/modules/audio_coding/neteq/neteq_impl.cc
@@ -315,9 +315,10 @@
int NetEqImpl::NetworkStatistics(NetEqNetworkStatistics* stats) {
CriticalSectionScoped lock(crit_sect_.get());
assert(decoder_database_.get());
- const int total_samples_in_buffers = packet_buffer_->NumSamplesInBuffer(
- decoder_database_.get(), decoder_frame_length_) +
- static_cast<int>(sync_buffer_->FutureLength());
+ const int total_samples_in_buffers =
+ packet_buffer_->NumSamplesInBuffer(decoder_database_.get(),
+ decoder_frame_length_) +
+ static_cast<int>(sync_buffer_->FutureLength());
assert(delay_manager_.get());
assert(decision_logic_.get());
stats_.GetNetworkStatistics(fs_hz_, total_samples_in_buffers,
@@ -704,8 +705,10 @@
return 0;
}
-int NetEqImpl::GetAudioInternal(size_t max_length, int16_t* output,
- int* samples_per_channel, int* num_channels) {
+int NetEqImpl::GetAudioInternal(size_t max_length,
+ int16_t* output,
+ int* samples_per_channel,
+ int* num_channels) {
PacketList packet_list;
DtmfEvent dtmf_event;
Operations operation;
diff --git a/webrtc/modules/audio_coding/neteq/packet_buffer.cc b/webrtc/modules/audio_coding/neteq/packet_buffer.cc
index 08b237f..5792b22 100644
--- a/webrtc/modules/audio_coding/neteq/packet_buffer.cc
+++ b/webrtc/modules/audio_coding/neteq/packet_buffer.cc
@@ -250,16 +250,12 @@
Packet* packet = (*it);
AudioDecoder* decoder =
decoder_database->GetDecoder(packet->header.payloadType);
- if (decoder) {
- int duration;
- if (packet->sync_packet) {
- duration = last_duration;
- } else if (packet->primary) {
- duration =
- decoder->PacketDuration(packet->payload, packet->payload_length);
- } else {
+ if (decoder && !packet->sync_packet) {
+ if (!packet->primary) {
continue;
}
+ int duration =
+ decoder->PacketDuration(packet->payload, packet->payload_length);
if (duration >= 0) {
last_duration = duration; // Save the most up-to-date (valid) duration.
}
diff --git a/webrtc/modules/audio_coding/neteq/preemptive_expand.h b/webrtc/modules/audio_coding/neteq/preemptive_expand.h
index c583a48..65da703 100644
--- a/webrtc/modules/audio_coding/neteq/preemptive_expand.h
+++ b/webrtc/modules/audio_coding/neteq/preemptive_expand.h
@@ -52,9 +52,9 @@
protected:
// Sets the parameters |best_correlation| and |peak_index| to suitable
// values when the signal contains no active speech.
- void SetParametersForPassiveSpeech(size_t len,
- int16_t* w16_bestCorr,
- int* w16_bestIndex) const override;
+ void SetParametersForPassiveSpeech(size_t input_length,
+ int16_t* best_correlation,
+ int* peak_index) const override;
// Checks the criteria for performing the time-stretching operation and,
// if possible, performs the time-stretching.
diff --git a/webrtc/modules/audio_coding/neteq/statistics_calculator.cc b/webrtc/modules/audio_coding/neteq/statistics_calculator.cc
index f637eb8..ce800dd 100644
--- a/webrtc/modules/audio_coding/neteq/statistics_calculator.cc
+++ b/webrtc/modules/audio_coding/neteq/statistics_calculator.cc
@@ -149,7 +149,7 @@
stats->speech_expand_rate =
CalculateQ14Ratio(expanded_speech_samples_,
- timestamps_since_last_report_);
+ timestamps_since_last_report_);
stats->secondary_decoded_rate =
CalculateQ14Ratio(secondary_decoded_samples_,
diff --git a/webrtc/modules/audio_coding/neteq/test/RTPencode.cc b/webrtc/modules/audio_coding/neteq/test/RTPencode.cc
index 192d374..f25a279 100644
--- a/webrtc/modules/audio_coding/neteq/test/RTPencode.cc
+++ b/webrtc/modules/audio_coding/neteq/test/RTPencode.cc
@@ -152,7 +152,7 @@
#ifdef NETEQ_ISACFIX_CODEC
#include "isacfix.h"
#ifdef CODEC_ISAC
-#error Cannot have both ISAC and ISACfix defined. Please de-select one in the beginning of RTPencode.cpp
+#error Cannot have both ISAC and ISACfix defined. Please de-select one.
#endif
#endif
#ifdef CODEC_G722
diff --git a/webrtc/modules/audio_coding/neteq/tools/constant_pcm_packet_source.cc b/webrtc/modules/audio_coding/neteq/tools/constant_pcm_packet_source.cc
index 65c4e9d..af4b8e1 100644
--- a/webrtc/modules/audio_coding/neteq/tools/constant_pcm_packet_source.cc
+++ b/webrtc/modules/audio_coding/neteq/tools/constant_pcm_packet_source.cc
@@ -32,7 +32,7 @@
timestamp_(0),
payload_ssrc_(0xABCD1234) {
int encoded_len = WebRtcPcm16b_Encode(&sample_value, 1, encoded_sample_);
- CHECK_EQ(encoded_len, 2);
+ CHECK_EQ(2, encoded_len);
}
Packet* ConstantPcmPacketSource::NextPacket() {
diff --git a/webrtc/modules/audio_coding/neteq/tools/neteq_performance_test.cc b/webrtc/modules/audio_coding/neteq/tools/neteq_performance_test.cc
index 080b99b..1c76d76 100644
--- a/webrtc/modules/audio_coding/neteq/tools/neteq_performance_test.cc
+++ b/webrtc/modules/audio_coding/neteq/tools/neteq_performance_test.cc
@@ -103,8 +103,8 @@
static const int kMaxChannels = 1;
static const int kMaxSamplesPerMs = 48000 / 1000;
static const int kOutputBlockSizeMs = 10;
- static const int kOutDataLen = kOutputBlockSizeMs * kMaxSamplesPerMs *
- kMaxChannels;
+ static const int kOutDataLen =
+ kOutputBlockSizeMs * kMaxSamplesPerMs * kMaxChannels;
int16_t out_data[kOutDataLen];
int num_channels;
int samples_per_channel;
diff --git a/webrtc/modules/audio_coding/neteq/tools/neteq_rtpplay.cc b/webrtc/modules/audio_coding/neteq/tools/neteq_rtpplay.cc
index 11dd20a..6bcd717 100644
--- a/webrtc/modules/audio_coding/neteq/tools/neteq_rtpplay.cc
+++ b/webrtc/modules/audio_coding/neteq/tools/neteq_rtpplay.cc
@@ -552,8 +552,8 @@
// Check if it is time to get output audio.
if (time_now_ms >= next_output_time_ms) {
- static const int kOutDataLen = kOutputBlockSizeMs * kMaxSamplesPerMs *
- kMaxChannels;
+ static const int kOutDataLen =
+ kOutputBlockSizeMs * kMaxSamplesPerMs * kMaxChannels;
int16_t out_data[kOutDataLen];
int num_channels;
int samples_per_channel;