996 lines
38 KiB
C++
996 lines
38 KiB
C++
/*
|
|
* Copyright (c) 2013 The WebRTC project authors. All Rights Reserved.
|
|
*
|
|
* Use of this source code is governed by a BSD-style license
|
|
* that can be found in the LICENSE file in the root of the source
|
|
* tree. An additional intellectual property rights grant can be found
|
|
* in the file PATENTS. All contributing project authors may
|
|
* be found in the AUTHORS file in the root of the source tree.
|
|
*/
|
|
|
|
#include <memory>
|
|
#include <string>
|
|
|
|
#include "modules/audio_coding/codecs/opus/opus_inst.h"
|
|
#include "modules/audio_coding/codecs/opus/opus_interface.h"
|
|
#include "modules/audio_coding/neteq/tools/audio_loop.h"
|
|
#include "rtc_base/checks.h"
|
|
#include "rtc_base/numerics/safe_conversions.h"
|
|
#include "test/gtest.h"
|
|
#include "test/testsupport/file_utils.h"
|
|
|
|
namespace webrtc {
|
|
|
|
namespace {
|
|
// Equivalent to SDP params
|
|
// {{"channel_mapping", "0,1,2,3"}, {"coupled_streams", "2"}}.
|
|
constexpr unsigned char kQuadChannelMapping[] = {0, 1, 2, 3};
|
|
constexpr int kQuadTotalStreams = 2;
|
|
constexpr int kQuadCoupledStreams = 2;
|
|
|
|
constexpr unsigned char kStereoChannelMapping[] = {0, 1};
|
|
constexpr int kStereoTotalStreams = 1;
|
|
constexpr int kStereoCoupledStreams = 1;
|
|
|
|
constexpr unsigned char kMonoChannelMapping[] = {0};
|
|
constexpr int kMonoTotalStreams = 1;
|
|
constexpr int kMonoCoupledStreams = 0;
|
|
|
|
void CreateSingleOrMultiStreamEncoder(WebRtcOpusEncInst** opus_encoder,
|
|
int channels,
|
|
int application,
|
|
bool use_multistream,
|
|
int encoder_sample_rate_hz) {
|
|
EXPECT_TRUE(channels == 1 || channels == 2 || use_multistream);
|
|
if (use_multistream) {
|
|
EXPECT_EQ(encoder_sample_rate_hz, 48000);
|
|
if (channels == 1) {
|
|
EXPECT_EQ(0, WebRtcOpus_MultistreamEncoderCreate(
|
|
opus_encoder, channels, application, kMonoTotalStreams,
|
|
kMonoCoupledStreams, kMonoChannelMapping));
|
|
} else if (channels == 2) {
|
|
EXPECT_EQ(0, WebRtcOpus_MultistreamEncoderCreate(
|
|
opus_encoder, channels, application, kStereoTotalStreams,
|
|
kStereoCoupledStreams, kStereoChannelMapping));
|
|
} else if (channels == 4) {
|
|
EXPECT_EQ(0, WebRtcOpus_MultistreamEncoderCreate(
|
|
opus_encoder, channels, application, kQuadTotalStreams,
|
|
kQuadCoupledStreams, kQuadChannelMapping));
|
|
} else {
|
|
EXPECT_TRUE(false) << channels;
|
|
}
|
|
} else {
|
|
EXPECT_EQ(0, WebRtcOpus_EncoderCreate(opus_encoder, channels, application,
|
|
encoder_sample_rate_hz));
|
|
}
|
|
}
|
|
|
|
void CreateSingleOrMultiStreamDecoder(WebRtcOpusDecInst** opus_decoder,
|
|
int channels,
|
|
bool use_multistream,
|
|
int decoder_sample_rate_hz) {
|
|
EXPECT_TRUE(channels == 1 || channels == 2 || use_multistream);
|
|
if (use_multistream) {
|
|
EXPECT_EQ(decoder_sample_rate_hz, 48000);
|
|
if (channels == 1) {
|
|
EXPECT_EQ(0, WebRtcOpus_MultistreamDecoderCreate(
|
|
opus_decoder, channels, kMonoTotalStreams,
|
|
kMonoCoupledStreams, kMonoChannelMapping));
|
|
} else if (channels == 2) {
|
|
EXPECT_EQ(0, WebRtcOpus_MultistreamDecoderCreate(
|
|
opus_decoder, channels, kStereoTotalStreams,
|
|
kStereoCoupledStreams, kStereoChannelMapping));
|
|
} else if (channels == 4) {
|
|
EXPECT_EQ(0, WebRtcOpus_MultistreamDecoderCreate(
|
|
opus_decoder, channels, kQuadTotalStreams,
|
|
kQuadCoupledStreams, kQuadChannelMapping));
|
|
} else {
|
|
EXPECT_TRUE(false) << channels;
|
|
}
|
|
} else {
|
|
EXPECT_EQ(0, WebRtcOpus_DecoderCreate(opus_decoder, channels,
|
|
decoder_sample_rate_hz));
|
|
}
|
|
}
|
|
|
|
int SamplesPerChannel(int sample_rate_hz, int duration_ms) {
|
|
const int samples_per_ms = rtc::CheckedDivExact(sample_rate_hz, 1000);
|
|
return samples_per_ms * duration_ms;
|
|
}
|
|
|
|
using test::AudioLoop;
|
|
using ::testing::Combine;
|
|
using ::testing::TestWithParam;
|
|
using ::testing::Values;
|
|
|
|
// Maximum number of bytes in output bitstream.
|
|
const size_t kMaxBytes = 2000;
|
|
|
|
class OpusTest
|
|
: public TestWithParam<::testing::tuple<size_t, int, bool, int, int>> {
|
|
protected:
|
|
OpusTest() = default;
|
|
|
|
void TestDtxEffect(bool dtx, int block_length_ms);
|
|
|
|
void TestCbrEffect(bool dtx, int block_length_ms);
|
|
|
|
// Prepare |speech_data_| for encoding, read from a hard-coded file.
|
|
// After preparation, |speech_data_.GetNextBlock()| returns a pointer to a
|
|
// block of |block_length_ms| milliseconds. The data is looped every
|
|
// |loop_length_ms| milliseconds.
|
|
void PrepareSpeechData(int block_length_ms, int loop_length_ms);
|
|
|
|
int EncodeDecode(WebRtcOpusEncInst* encoder,
|
|
rtc::ArrayView<const int16_t> input_audio,
|
|
WebRtcOpusDecInst* decoder,
|
|
int16_t* output_audio,
|
|
int16_t* audio_type);
|
|
|
|
void SetMaxPlaybackRate(WebRtcOpusEncInst* encoder,
|
|
opus_int32 expect,
|
|
int32_t set);
|
|
|
|
void CheckAudioBounded(const int16_t* audio,
|
|
size_t samples,
|
|
size_t channels,
|
|
uint16_t bound) const;
|
|
|
|
WebRtcOpusEncInst* opus_encoder_ = nullptr;
|
|
WebRtcOpusDecInst* opus_decoder_ = nullptr;
|
|
AudioLoop speech_data_;
|
|
uint8_t bitstream_[kMaxBytes];
|
|
size_t encoded_bytes_ = 0;
|
|
const size_t channels_{std::get<0>(GetParam())};
|
|
const int application_{std::get<1>(GetParam())};
|
|
const bool use_multistream_{std::get<2>(GetParam())};
|
|
const int encoder_sample_rate_hz_{std::get<3>(GetParam())};
|
|
const int decoder_sample_rate_hz_{std::get<4>(GetParam())};
|
|
};
|
|
|
|
} // namespace
|
|
|
|
// Singlestream: Try all combinations.
|
|
INSTANTIATE_TEST_SUITE_P(Singlestream,
|
|
OpusTest,
|
|
testing::Combine(testing::Values(1, 2),
|
|
testing::Values(0, 1),
|
|
testing::Values(false),
|
|
testing::Values(16000, 48000),
|
|
testing::Values(16000, 48000)));
|
|
|
|
// Multistream: Some representative cases (only 48 kHz for now).
|
|
INSTANTIATE_TEST_SUITE_P(
|
|
Multistream,
|
|
OpusTest,
|
|
testing::Values(std::make_tuple(1, 0, true, 48000, 48000),
|
|
std::make_tuple(2, 1, true, 48000, 48000),
|
|
std::make_tuple(4, 0, true, 48000, 48000),
|
|
std::make_tuple(4, 1, true, 48000, 48000)));
|
|
|
|
void OpusTest::PrepareSpeechData(int block_length_ms, int loop_length_ms) {
|
|
std::map<int, std::string> channel_to_basename = {
|
|
{1, "audio_coding/testfile32kHz"},
|
|
{2, "audio_coding/teststereo32kHz"},
|
|
{4, "audio_coding/speech_4_channels_48k_one_second"}};
|
|
std::map<int, std::string> channel_to_suffix = {
|
|
{1, "pcm"}, {2, "pcm"}, {4, "wav"}};
|
|
const std::string file_name = webrtc::test::ResourcePath(
|
|
channel_to_basename[channels_], channel_to_suffix[channels_]);
|
|
if (loop_length_ms < block_length_ms) {
|
|
loop_length_ms = block_length_ms;
|
|
}
|
|
const int sample_rate_khz =
|
|
rtc::CheckedDivExact(encoder_sample_rate_hz_, 1000);
|
|
EXPECT_TRUE(speech_data_.Init(file_name,
|
|
loop_length_ms * sample_rate_khz * channels_,
|
|
block_length_ms * sample_rate_khz * channels_));
|
|
}
|
|
|
|
void OpusTest::SetMaxPlaybackRate(WebRtcOpusEncInst* encoder,
|
|
opus_int32 expect,
|
|
int32_t set) {
|
|
opus_int32 bandwidth;
|
|
EXPECT_EQ(0, WebRtcOpus_SetMaxPlaybackRate(opus_encoder_, set));
|
|
EXPECT_EQ(0, WebRtcOpus_GetMaxPlaybackRate(opus_encoder_, &bandwidth));
|
|
EXPECT_EQ(expect, bandwidth);
|
|
}
|
|
|
|
void OpusTest::CheckAudioBounded(const int16_t* audio,
|
|
size_t samples,
|
|
size_t channels,
|
|
uint16_t bound) const {
|
|
for (size_t i = 0; i < samples; ++i) {
|
|
for (size_t c = 0; c < channels; ++c) {
|
|
ASSERT_GE(audio[i * channels + c], -bound);
|
|
ASSERT_LE(audio[i * channels + c], bound);
|
|
}
|
|
}
|
|
}
|
|
|
|
int OpusTest::EncodeDecode(WebRtcOpusEncInst* encoder,
|
|
rtc::ArrayView<const int16_t> input_audio,
|
|
WebRtcOpusDecInst* decoder,
|
|
int16_t* output_audio,
|
|
int16_t* audio_type) {
|
|
const int input_samples_per_channel =
|
|
rtc::CheckedDivExact(input_audio.size(), channels_);
|
|
int encoded_bytes_int =
|
|
WebRtcOpus_Encode(encoder, input_audio.data(), input_samples_per_channel,
|
|
kMaxBytes, bitstream_);
|
|
EXPECT_GE(encoded_bytes_int, 0);
|
|
encoded_bytes_ = static_cast<size_t>(encoded_bytes_int);
|
|
if (encoded_bytes_ != 0) {
|
|
int est_len = WebRtcOpus_DurationEst(decoder, bitstream_, encoded_bytes_);
|
|
int act_len = WebRtcOpus_Decode(decoder, bitstream_, encoded_bytes_,
|
|
output_audio, audio_type);
|
|
EXPECT_EQ(est_len, act_len);
|
|
return act_len;
|
|
} else {
|
|
int total_dtx_len = 0;
|
|
const int output_samples_per_channel = input_samples_per_channel *
|
|
decoder_sample_rate_hz_ /
|
|
encoder_sample_rate_hz_;
|
|
while (total_dtx_len < output_samples_per_channel) {
|
|
int est_len = WebRtcOpus_DurationEst(decoder, NULL, 0);
|
|
int act_len = WebRtcOpus_Decode(decoder, NULL, 0,
|
|
&output_audio[total_dtx_len * channels_],
|
|
audio_type);
|
|
EXPECT_EQ(est_len, act_len);
|
|
total_dtx_len += act_len;
|
|
}
|
|
return total_dtx_len;
|
|
}
|
|
}
|
|
|
|
// Test if encoder/decoder can enter DTX mode properly and do not enter DTX when
|
|
// they should not. This test is signal dependent.
|
|
void OpusTest::TestDtxEffect(bool dtx, int block_length_ms) {
|
|
PrepareSpeechData(block_length_ms, 2000);
|
|
const size_t input_samples =
|
|
rtc::CheckedDivExact(encoder_sample_rate_hz_, 1000) * block_length_ms;
|
|
const size_t output_samples =
|
|
rtc::CheckedDivExact(decoder_sample_rate_hz_, 1000) * block_length_ms;
|
|
|
|
// Create encoder memory.
|
|
CreateSingleOrMultiStreamEncoder(&opus_encoder_, channels_, application_,
|
|
use_multistream_, encoder_sample_rate_hz_);
|
|
CreateSingleOrMultiStreamDecoder(&opus_decoder_, channels_, use_multistream_,
|
|
decoder_sample_rate_hz_);
|
|
|
|
// Set bitrate.
|
|
EXPECT_EQ(
|
|
0, WebRtcOpus_SetBitRate(opus_encoder_, channels_ == 1 ? 32000 : 64000));
|
|
|
|
// Set input audio as silence.
|
|
std::vector<int16_t> silence(input_samples * channels_, 0);
|
|
|
|
// Setting DTX.
|
|
EXPECT_EQ(0, dtx ? WebRtcOpus_EnableDtx(opus_encoder_)
|
|
: WebRtcOpus_DisableDtx(opus_encoder_));
|
|
|
|
int16_t audio_type;
|
|
int16_t* output_data_decode = new int16_t[output_samples * channels_];
|
|
|
|
for (int i = 0; i < 100; ++i) {
|
|
EXPECT_EQ(output_samples,
|
|
static_cast<size_t>(EncodeDecode(
|
|
opus_encoder_, speech_data_.GetNextBlock(), opus_decoder_,
|
|
output_data_decode, &audio_type)));
|
|
// If not DTX, it should never enter DTX mode. If DTX, we do not care since
|
|
// whether it enters DTX depends on the signal type.
|
|
if (!dtx) {
|
|
EXPECT_GT(encoded_bytes_, 1U);
|
|
EXPECT_EQ(0, opus_encoder_->in_dtx_mode);
|
|
EXPECT_EQ(0, opus_decoder_->in_dtx_mode);
|
|
EXPECT_EQ(0, audio_type); // Speech.
|
|
}
|
|
}
|
|
|
|
// We input some silent segments. In DTX mode, the encoder will stop sending.
|
|
// However, DTX may happen after a while.
|
|
for (int i = 0; i < 30; ++i) {
|
|
EXPECT_EQ(output_samples, static_cast<size_t>(EncodeDecode(
|
|
opus_encoder_, silence, opus_decoder_,
|
|
output_data_decode, &audio_type)));
|
|
if (!dtx) {
|
|
EXPECT_GT(encoded_bytes_, 1U);
|
|
EXPECT_EQ(0, opus_encoder_->in_dtx_mode);
|
|
EXPECT_EQ(0, opus_decoder_->in_dtx_mode);
|
|
EXPECT_EQ(0, audio_type); // Speech.
|
|
} else if (encoded_bytes_ == 1) {
|
|
EXPECT_EQ(1, opus_encoder_->in_dtx_mode);
|
|
EXPECT_EQ(1, opus_decoder_->in_dtx_mode);
|
|
EXPECT_EQ(2, audio_type); // Comfort noise.
|
|
break;
|
|
}
|
|
}
|
|
|
|
// When Opus is in DTX, it wakes up in a regular basis. It sends two packets,
|
|
// one with an arbitrary size and the other of 1-byte, then stops sending for
|
|
// a certain number of frames.
|
|
|
|
// |max_dtx_frames| is the maximum number of frames Opus can stay in DTX.
|
|
// TODO(kwiberg): Why does this number depend on the encoding sample rate?
|
|
const int max_dtx_frames =
|
|
(encoder_sample_rate_hz_ == 16000 ? 800 : 400) / block_length_ms + 1;
|
|
|
|
// We run |kRunTimeMs| milliseconds of pure silence.
|
|
const int kRunTimeMs = 4500;
|
|
|
|
// We check that, after a |kCheckTimeMs| milliseconds (given that the CNG in
|
|
// Opus needs time to adapt), the absolute values of DTX decoded signal are
|
|
// bounded by |kOutputValueBound|.
|
|
const int kCheckTimeMs = 4000;
|
|
|
|
#if defined(OPUS_FIXED_POINT)
|
|
// Fixed-point Opus generates a random (comfort) noise, which has a less
|
|
// predictable value bound than its floating-point Opus. This value depends on
|
|
// input signal, and the time window for checking the output values (between
|
|
// |kCheckTimeMs| and |kRunTimeMs|).
|
|
const uint16_t kOutputValueBound = 30;
|
|
|
|
#else
|
|
const uint16_t kOutputValueBound = 2;
|
|
#endif
|
|
|
|
int time = 0;
|
|
while (time < kRunTimeMs) {
|
|
// DTX mode is maintained for maximum |max_dtx_frames| frames.
|
|
int i = 0;
|
|
for (; i < max_dtx_frames; ++i) {
|
|
time += block_length_ms;
|
|
EXPECT_EQ(output_samples, static_cast<size_t>(EncodeDecode(
|
|
opus_encoder_, silence, opus_decoder_,
|
|
output_data_decode, &audio_type)));
|
|
if (dtx) {
|
|
if (encoded_bytes_ > 1)
|
|
break;
|
|
EXPECT_EQ(0U, encoded_bytes_) // Send 0 byte.
|
|
<< "Opus should have entered DTX mode.";
|
|
EXPECT_EQ(1, opus_encoder_->in_dtx_mode);
|
|
EXPECT_EQ(1, opus_decoder_->in_dtx_mode);
|
|
EXPECT_EQ(2, audio_type); // Comfort noise.
|
|
if (time >= kCheckTimeMs) {
|
|
CheckAudioBounded(output_data_decode, output_samples, channels_,
|
|
kOutputValueBound);
|
|
}
|
|
} else {
|
|
EXPECT_GT(encoded_bytes_, 1U);
|
|
EXPECT_EQ(0, opus_encoder_->in_dtx_mode);
|
|
EXPECT_EQ(0, opus_decoder_->in_dtx_mode);
|
|
EXPECT_EQ(0, audio_type); // Speech.
|
|
}
|
|
}
|
|
|
|
if (dtx) {
|
|
// With DTX, Opus must stop transmission for some time.
|
|
EXPECT_GT(i, 1);
|
|
}
|
|
|
|
// We expect a normal payload.
|
|
EXPECT_EQ(0, opus_encoder_->in_dtx_mode);
|
|
EXPECT_EQ(0, opus_decoder_->in_dtx_mode);
|
|
EXPECT_EQ(0, audio_type); // Speech.
|
|
|
|
// Enters DTX again immediately.
|
|
time += block_length_ms;
|
|
EXPECT_EQ(output_samples, static_cast<size_t>(EncodeDecode(
|
|
opus_encoder_, silence, opus_decoder_,
|
|
output_data_decode, &audio_type)));
|
|
if (dtx) {
|
|
EXPECT_EQ(1U, encoded_bytes_); // Send 1 byte.
|
|
EXPECT_EQ(1, opus_encoder_->in_dtx_mode);
|
|
EXPECT_EQ(1, opus_decoder_->in_dtx_mode);
|
|
EXPECT_EQ(2, audio_type); // Comfort noise.
|
|
if (time >= kCheckTimeMs) {
|
|
CheckAudioBounded(output_data_decode, output_samples, channels_,
|
|
kOutputValueBound);
|
|
}
|
|
} else {
|
|
EXPECT_GT(encoded_bytes_, 1U);
|
|
EXPECT_EQ(0, opus_encoder_->in_dtx_mode);
|
|
EXPECT_EQ(0, opus_decoder_->in_dtx_mode);
|
|
EXPECT_EQ(0, audio_type); // Speech.
|
|
}
|
|
}
|
|
|
|
silence[0] = 10000;
|
|
if (dtx) {
|
|
// Verify that encoder/decoder can jump out from DTX mode.
|
|
EXPECT_EQ(output_samples, static_cast<size_t>(EncodeDecode(
|
|
opus_encoder_, silence, opus_decoder_,
|
|
output_data_decode, &audio_type)));
|
|
EXPECT_GT(encoded_bytes_, 1U);
|
|
EXPECT_EQ(0, opus_encoder_->in_dtx_mode);
|
|
EXPECT_EQ(0, opus_decoder_->in_dtx_mode);
|
|
EXPECT_EQ(0, audio_type); // Speech.
|
|
}
|
|
|
|
// Free memory.
|
|
delete[] output_data_decode;
|
|
EXPECT_EQ(0, WebRtcOpus_EncoderFree(opus_encoder_));
|
|
EXPECT_EQ(0, WebRtcOpus_DecoderFree(opus_decoder_));
|
|
}
|
|
|
|
// Test if CBR does what we expect.
|
|
void OpusTest::TestCbrEffect(bool cbr, int block_length_ms) {
|
|
PrepareSpeechData(block_length_ms, 2000);
|
|
const size_t output_samples =
|
|
rtc::CheckedDivExact(decoder_sample_rate_hz_, 1000) * block_length_ms;
|
|
|
|
int32_t max_pkt_size_diff = 0;
|
|
int32_t prev_pkt_size = 0;
|
|
|
|
// Create encoder memory.
|
|
CreateSingleOrMultiStreamEncoder(&opus_encoder_, channels_, application_,
|
|
use_multistream_, encoder_sample_rate_hz_);
|
|
CreateSingleOrMultiStreamDecoder(&opus_decoder_, channels_, use_multistream_,
|
|
decoder_sample_rate_hz_);
|
|
|
|
// Set bitrate.
|
|
EXPECT_EQ(
|
|
0, WebRtcOpus_SetBitRate(opus_encoder_, channels_ == 1 ? 32000 : 64000));
|
|
|
|
// Setting CBR.
|
|
EXPECT_EQ(0, cbr ? WebRtcOpus_EnableCbr(opus_encoder_)
|
|
: WebRtcOpus_DisableCbr(opus_encoder_));
|
|
|
|
int16_t audio_type;
|
|
std::vector<int16_t> audio_out(output_samples * channels_);
|
|
for (int i = 0; i < 100; ++i) {
|
|
EXPECT_EQ(output_samples,
|
|
static_cast<size_t>(
|
|
EncodeDecode(opus_encoder_, speech_data_.GetNextBlock(),
|
|
opus_decoder_, audio_out.data(), &audio_type)));
|
|
|
|
if (prev_pkt_size > 0) {
|
|
int32_t diff = std::abs((int32_t)encoded_bytes_ - prev_pkt_size);
|
|
max_pkt_size_diff = std::max(max_pkt_size_diff, diff);
|
|
}
|
|
prev_pkt_size = rtc::checked_cast<int32_t>(encoded_bytes_);
|
|
}
|
|
|
|
if (cbr) {
|
|
EXPECT_EQ(max_pkt_size_diff, 0);
|
|
} else {
|
|
EXPECT_GT(max_pkt_size_diff, 0);
|
|
}
|
|
|
|
// Free memory.
|
|
EXPECT_EQ(0, WebRtcOpus_EncoderFree(opus_encoder_));
|
|
EXPECT_EQ(0, WebRtcOpus_DecoderFree(opus_decoder_));
|
|
}
|
|
|
|
// Test failing Create.
|
|
TEST(OpusTest, OpusCreateFail) {
|
|
WebRtcOpusEncInst* opus_encoder;
|
|
WebRtcOpusDecInst* opus_decoder;
|
|
|
|
// Test to see that an invalid pointer is caught.
|
|
EXPECT_EQ(-1, WebRtcOpus_EncoderCreate(NULL, 1, 0, 48000));
|
|
// Invalid channel number.
|
|
EXPECT_EQ(-1, WebRtcOpus_EncoderCreate(&opus_encoder, 257, 0, 48000));
|
|
// Invalid applciation mode.
|
|
EXPECT_EQ(-1, WebRtcOpus_EncoderCreate(&opus_encoder, 1, 2, 48000));
|
|
// Invalid sample rate.
|
|
EXPECT_EQ(-1, WebRtcOpus_EncoderCreate(&opus_encoder, 1, 0, 12345));
|
|
|
|
EXPECT_EQ(-1, WebRtcOpus_DecoderCreate(NULL, 1, 48000));
|
|
// Invalid channel number.
|
|
EXPECT_EQ(-1, WebRtcOpus_DecoderCreate(&opus_decoder, 257, 48000));
|
|
// Invalid sample rate.
|
|
EXPECT_EQ(-1, WebRtcOpus_DecoderCreate(&opus_decoder, 1, 12345));
|
|
}
|
|
|
|
// Test failing Free.
|
|
TEST(OpusTest, OpusFreeFail) {
|
|
// Test to see that an invalid pointer is caught.
|
|
EXPECT_EQ(-1, WebRtcOpus_EncoderFree(NULL));
|
|
EXPECT_EQ(-1, WebRtcOpus_DecoderFree(NULL));
|
|
}
|
|
|
|
// Test normal Create and Free.
|
|
TEST_P(OpusTest, OpusCreateFree) {
|
|
CreateSingleOrMultiStreamEncoder(&opus_encoder_, channels_, application_,
|
|
use_multistream_, encoder_sample_rate_hz_);
|
|
CreateSingleOrMultiStreamDecoder(&opus_decoder_, channels_, use_multistream_,
|
|
decoder_sample_rate_hz_);
|
|
EXPECT_TRUE(opus_encoder_ != NULL);
|
|
EXPECT_TRUE(opus_decoder_ != NULL);
|
|
// Free encoder and decoder memory.
|
|
EXPECT_EQ(0, WebRtcOpus_EncoderFree(opus_encoder_));
|
|
EXPECT_EQ(0, WebRtcOpus_DecoderFree(opus_decoder_));
|
|
}
|
|
|
|
#define ENCODER_CTL(inst, vargs) \
|
|
inst->encoder \
|
|
? opus_encoder_ctl(inst->encoder, vargs) \
|
|
: opus_multistream_encoder_ctl(inst->multistream_encoder, vargs)
|
|
|
|
TEST_P(OpusTest, OpusEncodeDecode) {
|
|
PrepareSpeechData(20, 20);
|
|
|
|
// Create encoder memory.
|
|
CreateSingleOrMultiStreamEncoder(&opus_encoder_, channels_, application_,
|
|
use_multistream_, encoder_sample_rate_hz_);
|
|
CreateSingleOrMultiStreamDecoder(&opus_decoder_, channels_, use_multistream_,
|
|
decoder_sample_rate_hz_);
|
|
|
|
// Set bitrate.
|
|
EXPECT_EQ(
|
|
0, WebRtcOpus_SetBitRate(opus_encoder_, channels_ == 1 ? 32000 : 64000));
|
|
|
|
// Check number of channels for decoder.
|
|
EXPECT_EQ(channels_, WebRtcOpus_DecoderChannels(opus_decoder_));
|
|
|
|
// Check application mode.
|
|
opus_int32 app;
|
|
ENCODER_CTL(opus_encoder_, OPUS_GET_APPLICATION(&app));
|
|
EXPECT_EQ(application_ == 0 ? OPUS_APPLICATION_VOIP : OPUS_APPLICATION_AUDIO,
|
|
app);
|
|
|
|
// Encode & decode.
|
|
int16_t audio_type;
|
|
const int decode_samples_per_channel =
|
|
SamplesPerChannel(decoder_sample_rate_hz_, /*ms=*/20);
|
|
int16_t* output_data_decode =
|
|
new int16_t[decode_samples_per_channel * channels_];
|
|
EXPECT_EQ(decode_samples_per_channel,
|
|
EncodeDecode(opus_encoder_, speech_data_.GetNextBlock(),
|
|
opus_decoder_, output_data_decode, &audio_type));
|
|
|
|
// Free memory.
|
|
delete[] output_data_decode;
|
|
EXPECT_EQ(0, WebRtcOpus_EncoderFree(opus_encoder_));
|
|
EXPECT_EQ(0, WebRtcOpus_DecoderFree(opus_decoder_));
|
|
}
|
|
|
|
TEST_P(OpusTest, OpusSetBitRate) {
|
|
// Test without creating encoder memory.
|
|
EXPECT_EQ(-1, WebRtcOpus_SetBitRate(opus_encoder_, 60000));
|
|
|
|
// Create encoder memory, try with different bitrates.
|
|
CreateSingleOrMultiStreamEncoder(&opus_encoder_, channels_, application_,
|
|
use_multistream_, encoder_sample_rate_hz_);
|
|
EXPECT_EQ(0, WebRtcOpus_SetBitRate(opus_encoder_, 30000));
|
|
EXPECT_EQ(0, WebRtcOpus_SetBitRate(opus_encoder_, 60000));
|
|
EXPECT_EQ(0, WebRtcOpus_SetBitRate(opus_encoder_, 300000));
|
|
EXPECT_EQ(0, WebRtcOpus_SetBitRate(opus_encoder_, 600000));
|
|
|
|
// Free memory.
|
|
EXPECT_EQ(0, WebRtcOpus_EncoderFree(opus_encoder_));
|
|
}
|
|
|
|
TEST_P(OpusTest, OpusSetComplexity) {
|
|
// Test without creating encoder memory.
|
|
EXPECT_EQ(-1, WebRtcOpus_SetComplexity(opus_encoder_, 9));
|
|
|
|
// Create encoder memory, try with different complexities.
|
|
CreateSingleOrMultiStreamEncoder(&opus_encoder_, channels_, application_,
|
|
use_multistream_, encoder_sample_rate_hz_);
|
|
|
|
EXPECT_EQ(0, WebRtcOpus_SetComplexity(opus_encoder_, 0));
|
|
EXPECT_EQ(0, WebRtcOpus_SetComplexity(opus_encoder_, 10));
|
|
EXPECT_EQ(-1, WebRtcOpus_SetComplexity(opus_encoder_, 11));
|
|
|
|
// Free memory.
|
|
EXPECT_EQ(0, WebRtcOpus_EncoderFree(opus_encoder_));
|
|
}
|
|
|
|
TEST_P(OpusTest, OpusSetBandwidth) {
|
|
if (channels_ > 2) {
|
|
// TODO(webrtc:10217): investigate why multi-stream Opus reports
|
|
// narrowband when it's configured with FULLBAND.
|
|
return;
|
|
}
|
|
PrepareSpeechData(20, 20);
|
|
|
|
int16_t audio_type;
|
|
const int decode_samples_per_channel =
|
|
SamplesPerChannel(decoder_sample_rate_hz_, /*ms=*/20);
|
|
std::unique_ptr<int16_t[]> output_data_decode(
|
|
new int16_t[decode_samples_per_channel * channels_]());
|
|
|
|
// Test without creating encoder memory.
|
|
EXPECT_EQ(-1,
|
|
WebRtcOpus_SetBandwidth(opus_encoder_, OPUS_BANDWIDTH_NARROWBAND));
|
|
EXPECT_EQ(-1, WebRtcOpus_GetBandwidth(opus_encoder_));
|
|
|
|
// Create encoder memory, try with different bandwidths.
|
|
CreateSingleOrMultiStreamEncoder(&opus_encoder_, channels_, application_,
|
|
use_multistream_, encoder_sample_rate_hz_);
|
|
CreateSingleOrMultiStreamDecoder(&opus_decoder_, channels_, use_multistream_,
|
|
decoder_sample_rate_hz_);
|
|
|
|
EXPECT_EQ(-1, WebRtcOpus_SetBandwidth(opus_encoder_,
|
|
OPUS_BANDWIDTH_NARROWBAND - 1));
|
|
EXPECT_EQ(0,
|
|
WebRtcOpus_SetBandwidth(opus_encoder_, OPUS_BANDWIDTH_NARROWBAND));
|
|
EncodeDecode(opus_encoder_, speech_data_.GetNextBlock(), opus_decoder_,
|
|
output_data_decode.get(), &audio_type);
|
|
EXPECT_EQ(OPUS_BANDWIDTH_NARROWBAND, WebRtcOpus_GetBandwidth(opus_encoder_));
|
|
EXPECT_EQ(0, WebRtcOpus_SetBandwidth(opus_encoder_, OPUS_BANDWIDTH_FULLBAND));
|
|
EncodeDecode(opus_encoder_, speech_data_.GetNextBlock(), opus_decoder_,
|
|
output_data_decode.get(), &audio_type);
|
|
EXPECT_EQ(encoder_sample_rate_hz_ == 16000 ? OPUS_BANDWIDTH_WIDEBAND
|
|
: OPUS_BANDWIDTH_FULLBAND,
|
|
WebRtcOpus_GetBandwidth(opus_encoder_));
|
|
EXPECT_EQ(
|
|
-1, WebRtcOpus_SetBandwidth(opus_encoder_, OPUS_BANDWIDTH_FULLBAND + 1));
|
|
EncodeDecode(opus_encoder_, speech_data_.GetNextBlock(), opus_decoder_,
|
|
output_data_decode.get(), &audio_type);
|
|
EXPECT_EQ(encoder_sample_rate_hz_ == 16000 ? OPUS_BANDWIDTH_WIDEBAND
|
|
: OPUS_BANDWIDTH_FULLBAND,
|
|
WebRtcOpus_GetBandwidth(opus_encoder_));
|
|
|
|
// Free memory.
|
|
EXPECT_EQ(0, WebRtcOpus_EncoderFree(opus_encoder_));
|
|
EXPECT_EQ(0, WebRtcOpus_DecoderFree(opus_decoder_));
|
|
}
|
|
|
|
TEST_P(OpusTest, OpusForceChannels) {
|
|
// Test without creating encoder memory.
|
|
EXPECT_EQ(-1, WebRtcOpus_SetForceChannels(opus_encoder_, 1));
|
|
|
|
CreateSingleOrMultiStreamEncoder(&opus_encoder_, channels_, application_,
|
|
use_multistream_, encoder_sample_rate_hz_);
|
|
ASSERT_NE(nullptr, opus_encoder_);
|
|
|
|
if (channels_ >= 2) {
|
|
EXPECT_EQ(-1, WebRtcOpus_SetForceChannels(opus_encoder_, 3));
|
|
EXPECT_EQ(0, WebRtcOpus_SetForceChannels(opus_encoder_, 2));
|
|
EXPECT_EQ(0, WebRtcOpus_SetForceChannels(opus_encoder_, 1));
|
|
EXPECT_EQ(0, WebRtcOpus_SetForceChannels(opus_encoder_, 0));
|
|
} else {
|
|
EXPECT_EQ(-1, WebRtcOpus_SetForceChannels(opus_encoder_, 2));
|
|
EXPECT_EQ(0, WebRtcOpus_SetForceChannels(opus_encoder_, 1));
|
|
EXPECT_EQ(0, WebRtcOpus_SetForceChannels(opus_encoder_, 0));
|
|
}
|
|
|
|
EXPECT_EQ(0, WebRtcOpus_EncoderFree(opus_encoder_));
|
|
}
|
|
|
|
// Encode and decode one frame, initialize the decoder and
|
|
// decode once more.
|
|
TEST_P(OpusTest, OpusDecodeInit) {
|
|
PrepareSpeechData(20, 20);
|
|
|
|
// Create encoder memory.
|
|
CreateSingleOrMultiStreamEncoder(&opus_encoder_, channels_, application_,
|
|
use_multistream_, encoder_sample_rate_hz_);
|
|
CreateSingleOrMultiStreamDecoder(&opus_decoder_, channels_, use_multistream_,
|
|
decoder_sample_rate_hz_);
|
|
|
|
// Encode & decode.
|
|
int16_t audio_type;
|
|
const int decode_samples_per_channel =
|
|
SamplesPerChannel(decoder_sample_rate_hz_, /*ms=*/20);
|
|
int16_t* output_data_decode =
|
|
new int16_t[decode_samples_per_channel * channels_];
|
|
EXPECT_EQ(decode_samples_per_channel,
|
|
EncodeDecode(opus_encoder_, speech_data_.GetNextBlock(),
|
|
opus_decoder_, output_data_decode, &audio_type));
|
|
|
|
WebRtcOpus_DecoderInit(opus_decoder_);
|
|
|
|
EXPECT_EQ(decode_samples_per_channel,
|
|
WebRtcOpus_Decode(opus_decoder_, bitstream_, encoded_bytes_,
|
|
output_data_decode, &audio_type));
|
|
|
|
// Free memory.
|
|
delete[] output_data_decode;
|
|
EXPECT_EQ(0, WebRtcOpus_EncoderFree(opus_encoder_));
|
|
EXPECT_EQ(0, WebRtcOpus_DecoderFree(opus_decoder_));
|
|
}
|
|
|
|
TEST_P(OpusTest, OpusEnableDisableFec) {
|
|
// Test without creating encoder memory.
|
|
EXPECT_EQ(-1, WebRtcOpus_EnableFec(opus_encoder_));
|
|
EXPECT_EQ(-1, WebRtcOpus_DisableFec(opus_encoder_));
|
|
|
|
// Create encoder memory.
|
|
CreateSingleOrMultiStreamEncoder(&opus_encoder_, channels_, application_,
|
|
use_multistream_, encoder_sample_rate_hz_);
|
|
|
|
EXPECT_EQ(0, WebRtcOpus_EnableFec(opus_encoder_));
|
|
EXPECT_EQ(0, WebRtcOpus_DisableFec(opus_encoder_));
|
|
|
|
// Free memory.
|
|
EXPECT_EQ(0, WebRtcOpus_EncoderFree(opus_encoder_));
|
|
}
|
|
|
|
TEST_P(OpusTest, OpusEnableDisableDtx) {
|
|
// Test without creating encoder memory.
|
|
EXPECT_EQ(-1, WebRtcOpus_EnableDtx(opus_encoder_));
|
|
EXPECT_EQ(-1, WebRtcOpus_DisableDtx(opus_encoder_));
|
|
|
|
// Create encoder memory.
|
|
CreateSingleOrMultiStreamEncoder(&opus_encoder_, channels_, application_,
|
|
use_multistream_, encoder_sample_rate_hz_);
|
|
|
|
opus_int32 dtx;
|
|
|
|
// DTX is off by default.
|
|
ENCODER_CTL(opus_encoder_, OPUS_GET_DTX(&dtx));
|
|
EXPECT_EQ(0, dtx);
|
|
|
|
// Test to enable DTX.
|
|
EXPECT_EQ(0, WebRtcOpus_EnableDtx(opus_encoder_));
|
|
ENCODER_CTL(opus_encoder_, OPUS_GET_DTX(&dtx));
|
|
EXPECT_EQ(1, dtx);
|
|
|
|
// Test to disable DTX.
|
|
EXPECT_EQ(0, WebRtcOpus_DisableDtx(opus_encoder_));
|
|
ENCODER_CTL(opus_encoder_, OPUS_GET_DTX(&dtx));
|
|
EXPECT_EQ(0, dtx);
|
|
|
|
// Free memory.
|
|
EXPECT_EQ(0, WebRtcOpus_EncoderFree(opus_encoder_));
|
|
}
|
|
|
|
TEST_P(OpusTest, OpusDtxOff) {
|
|
TestDtxEffect(false, 10);
|
|
TestDtxEffect(false, 20);
|
|
TestDtxEffect(false, 40);
|
|
}
|
|
|
|
TEST_P(OpusTest, OpusDtxOn) {
|
|
if (channels_ > 2) {
|
|
// TODO(webrtc:10218): adapt the test to the sizes and order of multi-stream
|
|
// DTX packets.
|
|
return;
|
|
}
|
|
TestDtxEffect(true, 10);
|
|
TestDtxEffect(true, 20);
|
|
TestDtxEffect(true, 40);
|
|
}
|
|
|
|
TEST_P(OpusTest, OpusCbrOff) {
|
|
TestCbrEffect(false, 10);
|
|
TestCbrEffect(false, 20);
|
|
TestCbrEffect(false, 40);
|
|
}
|
|
|
|
TEST_P(OpusTest, OpusCbrOn) {
|
|
TestCbrEffect(true, 10);
|
|
TestCbrEffect(true, 20);
|
|
TestCbrEffect(true, 40);
|
|
}
|
|
|
|
TEST_P(OpusTest, OpusSetPacketLossRate) {
|
|
// Test without creating encoder memory.
|
|
EXPECT_EQ(-1, WebRtcOpus_SetPacketLossRate(opus_encoder_, 50));
|
|
|
|
// Create encoder memory.
|
|
CreateSingleOrMultiStreamEncoder(&opus_encoder_, channels_, application_,
|
|
use_multistream_, encoder_sample_rate_hz_);
|
|
|
|
EXPECT_EQ(0, WebRtcOpus_SetPacketLossRate(opus_encoder_, 50));
|
|
EXPECT_EQ(-1, WebRtcOpus_SetPacketLossRate(opus_encoder_, -1));
|
|
EXPECT_EQ(-1, WebRtcOpus_SetPacketLossRate(opus_encoder_, 101));
|
|
|
|
// Free memory.
|
|
EXPECT_EQ(0, WebRtcOpus_EncoderFree(opus_encoder_));
|
|
}
|
|
|
|
TEST_P(OpusTest, OpusSetMaxPlaybackRate) {
|
|
// Test without creating encoder memory.
|
|
EXPECT_EQ(-1, WebRtcOpus_SetMaxPlaybackRate(opus_encoder_, 20000));
|
|
|
|
// Create encoder memory.
|
|
CreateSingleOrMultiStreamEncoder(&opus_encoder_, channels_, application_,
|
|
use_multistream_, encoder_sample_rate_hz_);
|
|
|
|
SetMaxPlaybackRate(opus_encoder_, OPUS_BANDWIDTH_FULLBAND, 48000);
|
|
SetMaxPlaybackRate(opus_encoder_, OPUS_BANDWIDTH_FULLBAND, 24001);
|
|
SetMaxPlaybackRate(opus_encoder_, OPUS_BANDWIDTH_SUPERWIDEBAND, 24000);
|
|
SetMaxPlaybackRate(opus_encoder_, OPUS_BANDWIDTH_SUPERWIDEBAND, 16001);
|
|
SetMaxPlaybackRate(opus_encoder_, OPUS_BANDWIDTH_WIDEBAND, 16000);
|
|
SetMaxPlaybackRate(opus_encoder_, OPUS_BANDWIDTH_WIDEBAND, 12001);
|
|
SetMaxPlaybackRate(opus_encoder_, OPUS_BANDWIDTH_MEDIUMBAND, 12000);
|
|
SetMaxPlaybackRate(opus_encoder_, OPUS_BANDWIDTH_MEDIUMBAND, 8001);
|
|
SetMaxPlaybackRate(opus_encoder_, OPUS_BANDWIDTH_NARROWBAND, 8000);
|
|
SetMaxPlaybackRate(opus_encoder_, OPUS_BANDWIDTH_NARROWBAND, 4000);
|
|
|
|
// Free memory.
|
|
EXPECT_EQ(0, WebRtcOpus_EncoderFree(opus_encoder_));
|
|
}
|
|
|
|
// Test PLC.
|
|
TEST_P(OpusTest, OpusDecodePlc) {
|
|
PrepareSpeechData(20, 20);
|
|
|
|
// Create encoder memory.
|
|
CreateSingleOrMultiStreamEncoder(&opus_encoder_, channels_, application_,
|
|
use_multistream_, encoder_sample_rate_hz_);
|
|
CreateSingleOrMultiStreamDecoder(&opus_decoder_, channels_, use_multistream_,
|
|
decoder_sample_rate_hz_);
|
|
|
|
// Set bitrate.
|
|
EXPECT_EQ(
|
|
0, WebRtcOpus_SetBitRate(opus_encoder_, channels_ == 1 ? 32000 : 64000));
|
|
|
|
// Check number of channels for decoder.
|
|
EXPECT_EQ(channels_, WebRtcOpus_DecoderChannels(opus_decoder_));
|
|
|
|
// Encode & decode.
|
|
int16_t audio_type;
|
|
const int decode_samples_per_channel =
|
|
SamplesPerChannel(decoder_sample_rate_hz_, /*ms=*/20);
|
|
int16_t* output_data_decode =
|
|
new int16_t[decode_samples_per_channel * channels_];
|
|
EXPECT_EQ(decode_samples_per_channel,
|
|
EncodeDecode(opus_encoder_, speech_data_.GetNextBlock(),
|
|
opus_decoder_, output_data_decode, &audio_type));
|
|
|
|
// Call decoder PLC.
|
|
constexpr int kPlcDurationMs = 10;
|
|
const int plc_samples = decoder_sample_rate_hz_ * kPlcDurationMs / 1000;
|
|
int16_t* plc_buffer = new int16_t[plc_samples * channels_];
|
|
EXPECT_EQ(plc_samples,
|
|
WebRtcOpus_Decode(opus_decoder_, NULL, 0, plc_buffer, &audio_type));
|
|
|
|
// Free memory.
|
|
delete[] plc_buffer;
|
|
delete[] output_data_decode;
|
|
EXPECT_EQ(0, WebRtcOpus_EncoderFree(opus_encoder_));
|
|
EXPECT_EQ(0, WebRtcOpus_DecoderFree(opus_decoder_));
|
|
}
|
|
|
|
// Duration estimation.
|
|
TEST_P(OpusTest, OpusDurationEstimation) {
|
|
PrepareSpeechData(20, 20);
|
|
|
|
// Create.
|
|
CreateSingleOrMultiStreamEncoder(&opus_encoder_, channels_, application_,
|
|
use_multistream_, encoder_sample_rate_hz_);
|
|
CreateSingleOrMultiStreamDecoder(&opus_decoder_, channels_, use_multistream_,
|
|
decoder_sample_rate_hz_);
|
|
|
|
// 10 ms. We use only first 10 ms of a 20 ms block.
|
|
auto speech_block = speech_data_.GetNextBlock();
|
|
int encoded_bytes_int = WebRtcOpus_Encode(
|
|
opus_encoder_, speech_block.data(),
|
|
rtc::CheckedDivExact(speech_block.size(), 2 * channels_), kMaxBytes,
|
|
bitstream_);
|
|
EXPECT_GE(encoded_bytes_int, 0);
|
|
EXPECT_EQ(SamplesPerChannel(decoder_sample_rate_hz_, /*ms=*/10),
|
|
WebRtcOpus_DurationEst(opus_decoder_, bitstream_,
|
|
static_cast<size_t>(encoded_bytes_int)));
|
|
|
|
// 20 ms
|
|
speech_block = speech_data_.GetNextBlock();
|
|
encoded_bytes_int =
|
|
WebRtcOpus_Encode(opus_encoder_, speech_block.data(),
|
|
rtc::CheckedDivExact(speech_block.size(), channels_),
|
|
kMaxBytes, bitstream_);
|
|
EXPECT_GE(encoded_bytes_int, 0);
|
|
EXPECT_EQ(SamplesPerChannel(decoder_sample_rate_hz_, /*ms=*/20),
|
|
WebRtcOpus_DurationEst(opus_decoder_, bitstream_,
|
|
static_cast<size_t>(encoded_bytes_int)));
|
|
|
|
// Free memory.
|
|
EXPECT_EQ(0, WebRtcOpus_EncoderFree(opus_encoder_));
|
|
EXPECT_EQ(0, WebRtcOpus_DecoderFree(opus_decoder_));
|
|
}
|
|
|
|
TEST_P(OpusTest, OpusDecodeRepacketized) {
|
|
if (channels_ > 2) {
|
|
// As per the Opus documentation
|
|
// https://mf4.xiph.org/jenkins/view/opus/job/opus/ws/doc/html/group__opus__repacketizer.html#details,
|
|
// multiple streams are not supported.
|
|
return;
|
|
}
|
|
constexpr size_t kPackets = 6;
|
|
|
|
PrepareSpeechData(20, 20 * kPackets);
|
|
|
|
// Create encoder memory.
|
|
CreateSingleOrMultiStreamEncoder(&opus_encoder_, channels_, application_,
|
|
use_multistream_, encoder_sample_rate_hz_);
|
|
ASSERT_NE(nullptr, opus_encoder_);
|
|
CreateSingleOrMultiStreamDecoder(&opus_decoder_, channels_, use_multistream_,
|
|
decoder_sample_rate_hz_);
|
|
ASSERT_NE(nullptr, opus_decoder_);
|
|
|
|
// Set bitrate.
|
|
EXPECT_EQ(
|
|
0, WebRtcOpus_SetBitRate(opus_encoder_, channels_ == 1 ? 32000 : 64000));
|
|
|
|
// Check number of channels for decoder.
|
|
EXPECT_EQ(channels_, WebRtcOpus_DecoderChannels(opus_decoder_));
|
|
|
|
// Encode & decode.
|
|
int16_t audio_type;
|
|
const int decode_samples_per_channel =
|
|
SamplesPerChannel(decoder_sample_rate_hz_, /*ms=*/20);
|
|
std::unique_ptr<int16_t[]> output_data_decode(
|
|
new int16_t[kPackets * decode_samples_per_channel * channels_]);
|
|
OpusRepacketizer* rp = opus_repacketizer_create();
|
|
|
|
size_t num_packets = 0;
|
|
constexpr size_t kMaxCycles = 100;
|
|
for (size_t idx = 0; idx < kMaxCycles; ++idx) {
|
|
auto speech_block = speech_data_.GetNextBlock();
|
|
encoded_bytes_ =
|
|
WebRtcOpus_Encode(opus_encoder_, speech_block.data(),
|
|
rtc::CheckedDivExact(speech_block.size(), channels_),
|
|
kMaxBytes, bitstream_);
|
|
if (opus_repacketizer_cat(rp, bitstream_,
|
|
rtc::checked_cast<opus_int32>(encoded_bytes_)) ==
|
|
OPUS_OK) {
|
|
++num_packets;
|
|
if (num_packets == kPackets) {
|
|
break;
|
|
}
|
|
} else {
|
|
// Opus repacketizer cannot guarantee a success. We try again if it fails.
|
|
opus_repacketizer_init(rp);
|
|
num_packets = 0;
|
|
}
|
|
}
|
|
EXPECT_EQ(kPackets, num_packets);
|
|
|
|
encoded_bytes_ = opus_repacketizer_out(rp, bitstream_, kMaxBytes);
|
|
|
|
EXPECT_EQ(decode_samples_per_channel * kPackets,
|
|
static_cast<size_t>(WebRtcOpus_DurationEst(
|
|
opus_decoder_, bitstream_, encoded_bytes_)));
|
|
|
|
EXPECT_EQ(decode_samples_per_channel * kPackets,
|
|
static_cast<size_t>(
|
|
WebRtcOpus_Decode(opus_decoder_, bitstream_, encoded_bytes_,
|
|
output_data_decode.get(), &audio_type)));
|
|
|
|
// Free memory.
|
|
opus_repacketizer_destroy(rp);
|
|
EXPECT_EQ(0, WebRtcOpus_EncoderFree(opus_encoder_));
|
|
EXPECT_EQ(0, WebRtcOpus_DecoderFree(opus_decoder_));
|
|
}
|
|
|
|
TEST(OpusVadTest, CeltUnknownStatus) {
|
|
const uint8_t celt[] = {0x80};
|
|
EXPECT_EQ(WebRtcOpus_PacketHasVoiceActivity(celt, 1), -1);
|
|
}
|
|
|
|
TEST(OpusVadTest, Mono20msVadSet) {
|
|
uint8_t silk20msMonoVad[] = {0x78, 0x80};
|
|
EXPECT_TRUE(WebRtcOpus_PacketHasVoiceActivity(silk20msMonoVad, 2));
|
|
}
|
|
|
|
TEST(OpusVadTest, Mono20MsVadUnset) {
|
|
uint8_t silk20msMonoSilence[] = {0x78, 0x00};
|
|
EXPECT_FALSE(WebRtcOpus_PacketHasVoiceActivity(silk20msMonoSilence, 2));
|
|
}
|
|
|
|
TEST(OpusVadTest, Stereo20MsVadOnSideChannel) {
|
|
uint8_t silk20msStereoVadSideChannel[] = {0x78 | 0x04, 0x20};
|
|
EXPECT_TRUE(
|
|
WebRtcOpus_PacketHasVoiceActivity(silk20msStereoVadSideChannel, 2));
|
|
}
|
|
|
|
TEST(OpusVadTest, TwoOpusMonoFramesVadOnSecond) {
|
|
uint8_t twoMonoFrames[] = {0x78 | 0x1, 0x00, 0x80};
|
|
EXPECT_TRUE(WebRtcOpus_PacketHasVoiceActivity(twoMonoFrames, 3));
|
|
}
|
|
|
|
TEST(OpusVadTest, DtxEmptyPacket) {
|
|
const uint8_t dtx[] = {0x78};
|
|
EXPECT_FALSE(WebRtcOpus_PacketHasVoiceActivity(dtx, 1));
|
|
}
|
|
|
|
TEST(OpusVadTest, DtxBackgroundNoisePacket) {
|
|
// DTX sends a frame coding background noise every 20 packets:
|
|
// https://tools.ietf.org/html/rfc6716#section-2.1.9
|
|
// The packet below represents such a frame and was captured using
|
|
// Wireshark while disabling encryption.
|
|
const uint8_t dtx[] = {0x78, 0x07, 0xc9, 0x79, 0xc8, 0xc9, 0x57, 0xc0, 0xa2,
|
|
0x12, 0x23, 0xfa, 0xef, 0x67, 0xf3, 0x2e, 0xe3, 0xd3,
|
|
0xd5, 0xe9, 0xec, 0xdb, 0x3e, 0xbc, 0x80, 0xb6, 0x6e,
|
|
0x2a, 0xb7, 0x8c, 0x83, 0xcd, 0x83, 0xcd, 0x00};
|
|
EXPECT_FALSE(WebRtcOpus_PacketHasVoiceActivity(dtx, 35));
|
|
}
|
|
|
|
} // namespace webrtc
|