Source code

Revision control

Copy as Markdown

Other Tools

/* -*- Mode: C++; tab-width: 2; indent-tabs-mode: nil; c-basic-offset: 2 -*- */
/* vim:set ts=2 sw=2 sts=2 et cindent: */
/* This Source Code Form is subject to the terms of the Mozilla Public
* License, v. 2.0. If a copy of the MPL was not distributed with this file,
* You can obtain one at http://mozilla.org/MPL/2.0/. */
#include <utility>
#include "AudioDecoderInputTrack.h"
#include "gmock/gmock.h"
#include "GraphDriver.h"
#include "gtest/gtest.h"
#include "MediaInfo.h"
#include "MediaTrackGraphImpl.h"
#include "mozilla/gtest/WaitFor.h"
#include "nsThreadUtils.h"
#include "VideoUtils.h"
using namespace mozilla;
using namespace mozilla::media;
using testing::AssertionResult;
using testing::NiceMock;
using testing::Return;
using ControlMessageInterface = MediaTrack::ControlMessageInterface;
constexpr uint32_t kNoFlags = 0;
constexpr TrackRate kRate = 44100;
constexpr uint32_t kChannels = 2;
class MockTestGraph : public MediaTrackGraphImpl {
public:
explicit MockTestGraph(TrackRate aRate)
: MediaTrackGraphImpl(0, aRate, nullptr, NS_GetCurrentThread()) {
ON_CALL(*this, OnGraphThread).WillByDefault(Return(true));
}
void Init(uint32_t aChannels) {
MediaTrackGraphImpl::Init(OFFLINE_THREAD_DRIVER, DIRECT_DRIVER, aChannels);
// We have to call `Destroy()` manually in order to break the reference.
// The reason we don't assign a null driver is because we would add a track
// to the graph, then it would trigger graph's `EnsureNextIteration()` that
// requires a non-null driver.
SetCurrentDriver(new NiceMock<MockDriver>());
}
MOCK_CONST_METHOD0(OnGraphThread, bool());
MOCK_METHOD1(AppendMessage, void(UniquePtr<ControlMessageInterface>));
protected:
~MockTestGraph() = default;
class MockDriver : public GraphDriver {
NS_INLINE_DECL_THREADSAFE_REFCOUNTING(MockDriver, override);
MockDriver() : GraphDriver(nullptr, nullptr, 0) {
ON_CALL(*this, OnThread).WillByDefault(Return(true));
ON_CALL(*this, ThreadRunning).WillByDefault(Return(true));
}
MOCK_METHOD0(Start, void());
MOCK_METHOD0(Shutdown, void());
MOCK_METHOD0(IterationDuration, uint32_t());
MOCK_METHOD0(EnsureNextIteration, void());
MOCK_CONST_METHOD0(OnThread, bool());
MOCK_CONST_METHOD0(ThreadRunning, bool());
protected:
~MockDriver() = default;
};
bool mEnableFakeAppend = false;
};
AudioData* CreateAudioDataFromInfo(uint32_t aFrames, const AudioInfo& aInfo) {
AlignedAudioBuffer samples(aFrames * aInfo.mChannels);
return new AudioData(0, TimeUnit::Zero(), std::move(samples), aInfo.mChannels,
aInfo.mRate);
}
AudioDecoderInputTrack* CreateTrack(MediaTrackGraph* aGraph,
nsISerialEventTarget* aThread,
const AudioInfo& aInfo,
float aPlaybackRate = 1.0,
float aVolume = 1.0,
bool aPreservesPitch = true) {
return AudioDecoderInputTrack::Create(aGraph, aThread, aInfo, aPlaybackRate,
aVolume, aPreservesPitch);
}
class TestAudioDecoderInputTrack : public testing::Test {
protected:
void SetUp() override {
mGraph = MakeRefPtr<NiceMock<MockTestGraph>>(kRate);
mGraph->Init(kChannels);
mInfo.mRate = kRate;
mInfo.mChannels = kChannels;
mTrack = CreateTrack(mGraph, NS_GetCurrentThread(), mInfo);
EXPECT_FALSE(mTrack->Ended());
}
void TearDown() override {
// This simulates the normal usage where the `Close()` is always be called
// before the `Destroy()`.
mTrack->Close();
mTrack->Destroy();
// Remove the reference of the track from the mock graph, and then release
// the self-reference of mock graph.
mGraph->RemoveTrackGraphThread(mTrack);
mGraph->Destroy();
}
AudioData* CreateAudioData(uint32_t aFrames) {
return CreateAudioDataFromInfo(aFrames, mInfo);
}
AudioSegment* GetTrackSegment() { return mTrack->GetData<AudioSegment>(); }
AssertionResult ExpectSegmentNonSilence(const char* aStartExpr,
const char* aEndExpr,
TrackTime aStart, TrackTime aEnd) {
AudioSegment checkedRange;
checkedRange.AppendSlice(*mTrack->GetData(), aStart, aEnd);
if (!checkedRange.IsNull()) {
return testing::AssertionSuccess();
}
return testing::AssertionFailure()
<< "segment [" << aStart << ":" << aEnd << "] should be non-silence";
}
AssertionResult ExpectSegmentSilence(const char* aStartExpr,
const char* aEndExpr, TrackTime aStart,
TrackTime aEnd) {
AudioSegment checkedRange;
checkedRange.AppendSlice(*mTrack->GetData(), aStart, aEnd);
if (checkedRange.IsNull()) {
return testing::AssertionSuccess();
}
return testing::AssertionFailure()
<< "segment [" << aStart << ":" << aEnd << "] should be silence";
}
RefPtr<MockTestGraph> mGraph;
RefPtr<AudioDecoderInputTrack> mTrack;
AudioInfo mInfo;
};
TEST_F(TestAudioDecoderInputTrack, BasicAppendData) {
// Start from [0:10] and each time we move the time by 10ms.
// Expected: outputDuration=10, outputFrames=0, outputSilence=10
TrackTime start = 0;
TrackTime end = 10;
mTrack->ProcessInput(start, end, kNoFlags);
EXPECT_EQ(mTrack->GetEnd(), end);
EXPECT_PRED_FORMAT2(ExpectSegmentSilence, start, end);
// Expected: outputDuration=20, outputFrames=5, outputSilence=15
RefPtr<AudioData> audio1 = CreateAudioData(5);
mTrack->AppendData(audio1, nullptr);
start = end;
end += 10;
mTrack->ProcessInput(start, end, kNoFlags);
EXPECT_EQ(mTrack->GetEnd(), end);
EXPECT_PRED_FORMAT2(ExpectSegmentNonSilence, start, start + audio1->Frames());
EXPECT_PRED_FORMAT2(ExpectSegmentSilence, start + audio1->Frames(), end);
// Expected: outputDuration=30, outputFrames=15, outputSilence=15
RefPtr<AudioData> audio2 = CreateAudioData(10);
mTrack->AppendData(audio2, nullptr);
start = end;
end += 10;
mTrack->ProcessInput(start, end, kNoFlags);
EXPECT_PRED_FORMAT2(ExpectSegmentNonSilence, start, end);
EXPECT_EQ(mTrack->GetEnd(), end);
// Expected : sent all data, track should be ended in the next iteration and
// fill slience in this iteration.
mTrack->NotifyEndOfStream();
start = end;
end += 10;
mTrack->ProcessInput(start, end, ProcessedMediaTrack::ALLOW_END);
EXPECT_PRED_FORMAT2(ExpectSegmentSilence, start, end);
EXPECT_EQ(mTrack->GetEnd(), end);
EXPECT_FALSE(mTrack->Ended());
// Expected : track ended
start = end;
end += 10;
mTrack->ProcessInput(start, end, ProcessedMediaTrack::ALLOW_END);
EXPECT_EQ(mTrack->WrittenFrames(), audio1->Frames() + audio2->Frames());
}
TEST_F(TestAudioDecoderInputTrack, ClearFuture) {
// Start from [0:10] and each time we move the time by 10ms.
// Expected: appended=30, expected duration=10
RefPtr<AudioData> audio1 = CreateAudioData(30);
mTrack->AppendData(audio1, nullptr);
TrackTime start = 0;
TrackTime end = 10;
mTrack->ProcessInput(start, end, kNoFlags);
EXPECT_PRED_FORMAT2(ExpectSegmentNonSilence, start, end);
// In next iteration [10:20], we would consume the remaining data that was
// appended in the previous iteration.
start = end;
end += 10;
mTrack->ProcessInput(start, end, kNoFlags);
EXPECT_PRED_FORMAT2(ExpectSegmentNonSilence, start, end);
// Clear future data which is the remaining 10 frames so the track would
// only output silence.
mTrack->ClearFutureData();
start = end;
end += 10;
mTrack->ProcessInput(start, end, kNoFlags);
EXPECT_PRED_FORMAT2(ExpectSegmentSilence, start, end);
// Test appending data again, to see if we can append data correctly after
// calling `ClearFutureData()`.
RefPtr<AudioData> audio2 = CreateAudioData(10);
mTrack->AppendData(audio2, nullptr);
start = end;
end += 10;
mTrack->ProcessInput(start, end, kNoFlags);
EXPECT_PRED_FORMAT2(ExpectSegmentNonSilence, start, end);
// Run another iteration that should only contains silence because the data
// we appended only enough for one iteration.
start = end;
end += 10;
mTrack->ProcessInput(start, end, kNoFlags);
EXPECT_PRED_FORMAT2(ExpectSegmentSilence, start, end);
// Clear future data would also remove the EOS.
mTrack->NotifyEndOfStream();
mTrack->ClearFutureData();
start = end;
end += 10;
mTrack->ProcessInput(start, end, ProcessedMediaTrack::ALLOW_END);
EXPECT_PRED_FORMAT2(ExpectSegmentSilence, start, end);
EXPECT_FALSE(mTrack->Ended());
// As EOS has been removed, in next iteration the track would still be
// running.
start = end;
end += 10;
mTrack->ProcessInput(start, end, ProcessedMediaTrack::ALLOW_END);
EXPECT_PRED_FORMAT2(ExpectSegmentSilence, start, end);
EXPECT_FALSE(mTrack->Ended());
EXPECT_EQ(mTrack->WrittenFrames(),
(audio1->Frames() - 10 /* got clear */) + audio2->Frames());
}
TEST_F(TestAudioDecoderInputTrack, InputRateChange) {
// Start from [0:10] and each time we move the time by 10ms.
// Expected: appended=10, expected duration=10
RefPtr<AudioData> audio1 = CreateAudioData(10);
mTrack->AppendData(audio1, nullptr);
TrackTime start = 0;
TrackTime end = 10;
mTrack->ProcessInput(start, end, kNoFlags);
EXPECT_PRED_FORMAT2(ExpectSegmentNonSilence, start, end);
// Change input sample rate to the half, input data should be resampled and
// its duration would become longer.
// Expected: appended=10 + 5,
// expected duration=10 + 5*2 (resampled)
mInfo.mRate = kRate / 2;
RefPtr<AudioData> audioHalfSampleRate = CreateAudioData(5);
mTrack->AppendData(audioHalfSampleRate, nullptr);
start = end;
end += 10;
mTrack->ProcessInput(start, end, kNoFlags);
EXPECT_PRED_FORMAT2(ExpectSegmentNonSilence, start, end);
// Change input sample rate to the double, input data should be resampled and
// its duration would become shorter.
// Expected: appended=10 + 10 + 10,
// expected duration=10 + 10 + 10/2(resampled) + 5(silence)
mInfo.mRate = kRate * 2;
RefPtr<AudioData> audioDoubleSampleRate = CreateAudioData(10);
TrackTime expectedDuration = audioDoubleSampleRate->Frames() / 2;
mTrack->AppendData(audioDoubleSampleRate, nullptr);
start = end;
end += 10;
mTrack->ProcessInput(start, end, kNoFlags);
EXPECT_PRED_FORMAT2(ExpectSegmentNonSilence, start, start + expectedDuration);
EXPECT_PRED_FORMAT2(ExpectSegmentSilence, start + expectedDuration, end);
EXPECT_EQ(mTrack->WrittenFrames(), audio1->Frames() +
audioHalfSampleRate->Frames() * 2 +
audioDoubleSampleRate->Frames() / 2);
}
TEST_F(TestAudioDecoderInputTrack, ChannelChange) {
// Start from [0:10] and each time we move the time by 10ms.
// Track was initialized in stero.
EXPECT_EQ(mTrack->NumberOfChannels(), uint32_t(2));
// But first audio data is mono, so the `NumberOfChannels()` changes to
// reflect the maximum channel in the audio segment.
mInfo.mChannels = 1;
RefPtr<AudioData> audioMono = CreateAudioData(10);
mTrack->AppendData(audioMono, nullptr);
TrackTime start = 0;
TrackTime end = 10;
mTrack->ProcessInput(start, end, kNoFlags);
EXPECT_PRED_FORMAT2(ExpectSegmentNonSilence, start, end);
EXPECT_EQ(mTrack->NumberOfChannels(), audioMono->mChannels);
// Then append audio data with 5 channels.
mInfo.mChannels = 5;
RefPtr<AudioData> audioWithFiveChannels = CreateAudioData(10);
mTrack->AppendData(audioWithFiveChannels, nullptr);
start = end;
end += 10;
mTrack->ProcessInput(start, end, kNoFlags);
EXPECT_PRED_FORMAT2(ExpectSegmentNonSilence, start, end);
EXPECT_EQ(mTrack->NumberOfChannels(), audioWithFiveChannels->mChannels);
EXPECT_EQ(mTrack->WrittenFrames(),
audioMono->Frames() + audioWithFiveChannels->Frames());
}
TEST_F(TestAudioDecoderInputTrack, VolumeChange) {
// In order to run the volume change directly without using a real graph.
// one for setting the track's volume, another for the track destruction.
EXPECT_CALL(*mGraph, AppendMessage)
.Times(2)
.WillOnce(
[](UniquePtr<ControlMessageInterface> aMessage) { aMessage->Run(); })
.WillOnce([](UniquePtr<ControlMessageInterface> aMessage) {});
// The default volume is 1.0.
float expectedVolume = 1.0;
RefPtr<AudioData> audio = CreateAudioData(20);
TrackTime start = 0;
TrackTime end = 10;
mTrack->AppendData(audio, nullptr);
mTrack->ProcessInput(start, end, kNoFlags);
EXPECT_PRED_FORMAT2(ExpectSegmentNonSilence, start, end);
EXPECT_TRUE(GetTrackSegment()->GetLastChunk()->mVolume == expectedVolume);
// After setting volume on the track, the data in the output chunk should be
// changed as well.
expectedVolume = 0.1;
mTrack->SetVolume(expectedVolume);
SpinEventLoopUntil<ProcessFailureBehavior::IgnoreAndContinue>(
"TEST_F(TestAudioDecoderInputTrack, VolumeChange)"_ns,
[&] { return mTrack->Volume() == expectedVolume; });
start = end;
end += 10;
mTrack->ProcessInput(start, end, kNoFlags);
EXPECT_PRED_FORMAT2(ExpectSegmentNonSilence, start, end);
EXPECT_TRUE(GetTrackSegment()->GetLastChunk()->mVolume == expectedVolume);
}
TEST_F(TestAudioDecoderInputTrack, BatchedData) {
uint32_t appendedFrames = 0;
RefPtr<AudioData> audio = CreateAudioData(10);
for (size_t idx = 0; idx < 50; idx++) {
mTrack->AppendData(audio, nullptr);
appendedFrames += audio->Frames();
}
// First we need to call `ProcessInput` at least once to drain the track's
// SPSC queue, otherwise we're not able to push the batched data later.
TrackTime start = 0;
TrackTime end = 10;
uint32_t expectedFrames = end - start;
mTrack->ProcessInput(start, end, kNoFlags);
EXPECT_PRED_FORMAT2(ExpectSegmentNonSilence, start, end);
// The batched data would be pushed to the graph thread in around 10ms after
// the track first time started to batch data, which we can't control here.
// Therefore, we need to wait until the batched data gets cleared.
SpinEventLoopUntil<ProcessFailureBehavior::IgnoreAndContinue>(
"TEST_F(TestAudioDecoderInputTrack, BatchedData)"_ns,
[&] { return !mTrack->HasBatchedData(); });
// Check that we received all the remainging data previously appended.
start = end;
end = start + (appendedFrames - expectedFrames);
mTrack->ProcessInput(start, end, kNoFlags);
EXPECT_PRED_FORMAT2(ExpectSegmentNonSilence, start, end);
// Check that we received no more data than previously appended.
start = end;
end += 10;
mTrack->ProcessInput(start, end, kNoFlags);
EXPECT_PRED_FORMAT2(ExpectSegmentSilence, start, end);
EXPECT_EQ(mTrack->WrittenFrames(), appendedFrames);
}
TEST_F(TestAudioDecoderInputTrack, OutputAndEndEvent) {
// Append an audio and EOS, the output event should notify the amount of
// frames that is equal to the amount of audio we appended.
RefPtr<AudioData> audio = CreateAudioData(10);
MozPromiseHolder<GenericPromise> holder;
RefPtr<GenericPromise> p = holder.Ensure(__func__);
MediaEventListener outputListener =
mTrack->OnOutput().Connect(NS_GetCurrentThread(), [&](TrackTime aFrame) {
EXPECT_EQ(aFrame, audio->Frames());
holder.Resolve(true, __func__);
});
mTrack->AppendData(audio, nullptr);
mTrack->NotifyEndOfStream();
TrackTime start = 0;
TrackTime end = 10;
mTrack->ProcessInput(start, end, ProcessedMediaTrack::ALLOW_END);
Unused << WaitFor(p);
// Track should end in this iteration, so the end event should be notified.
p = holder.Ensure(__func__);
MediaEventListener endListener = mTrack->OnEnd().Connect(
NS_GetCurrentThread(), [&]() { holder.Resolve(true, __func__); });
start = end;
end += 10;
mTrack->ProcessInput(start, end, ProcessedMediaTrack::ALLOW_END);
Unused << WaitFor(p);
outputListener.Disconnect();
endListener.Disconnect();
}
TEST_F(TestAudioDecoderInputTrack, PlaybackRateChange) {
// In order to run the playback change directly without using a real graph.
// one for setting the track's playback, another for the track destruction.
EXPECT_CALL(*mGraph, AppendMessage)
.Times(2)
.WillOnce(
[](UniquePtr<ControlMessageInterface> aMessage) { aMessage->Run(); })
.WillOnce([](UniquePtr<ControlMessageInterface> aMessage) {});
// Changing the playback rate.
float expectedPlaybackRate = 2.0;
mTrack->SetPlaybackRate(expectedPlaybackRate);
SpinEventLoopUntil<ProcessFailureBehavior::IgnoreAndContinue>(
"TEST_F(TestAudioDecoderInputTrack, PlaybackRateChange)"_ns,
[&] { return mTrack->PlaybackRate() == expectedPlaybackRate; });
// Time stretcher in the track would usually need certain amount of data
// before it outputs the time-stretched result. As we're in testing, we would
// only append data once, so signal an EOS after appending data, in order to
// ask the track to flush all samples from the time strecther.
RefPtr<AudioData> audio = CreateAudioData(100);
mTrack->AppendData(audio, nullptr);
mTrack->NotifyEndOfStream();
// Playback rate is 2x, so we should only get 1/2x sample frames, another 1/2
// should be silence.
TrackTime start = 0;
TrackTime end = audio->Frames();
mTrack->ProcessInput(start, end, kNoFlags);
EXPECT_PRED_FORMAT2(ExpectSegmentNonSilence, start, audio->Frames() / 2);
EXPECT_PRED_FORMAT2(ExpectSegmentSilence, start + audio->Frames() / 2, end);
}