228 lines
6.7 KiB
C++
228 lines
6.7 KiB
C++
// Copyright 2019 The Marl Authors.
|
|
//
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
// you may not use this file except in compliance with the License.
|
|
// You may obtain a copy of the License at
|
|
//
|
|
// https://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
|
|
#include "marl_test.h"
|
|
|
|
#include "marl/containers.h"
|
|
#include "marl/defer.h"
|
|
#include "marl/event.h"
|
|
#include "marl/waitgroup.h"
|
|
|
|
#include <atomic>
|
|
|
|
TEST_F(WithoutBoundScheduler, SchedulerConstructAndDestruct) {
|
|
auto scheduler = std::unique_ptr<marl::Scheduler>(
|
|
new marl::Scheduler(marl::Scheduler::Config()));
|
|
}
|
|
|
|
TEST_F(WithoutBoundScheduler, SchedulerBindGetUnbind) {
|
|
auto scheduler = std::unique_ptr<marl::Scheduler>(
|
|
new marl::Scheduler(marl::Scheduler::Config()));
|
|
scheduler->bind();
|
|
auto got = marl::Scheduler::get();
|
|
ASSERT_EQ(scheduler.get(), got);
|
|
scheduler->unbind();
|
|
got = marl::Scheduler::get();
|
|
ASSERT_EQ(got, nullptr);
|
|
}
|
|
|
|
TEST_F(WithoutBoundScheduler, CheckConfig) {
|
|
marl::Scheduler::Config cfg;
|
|
cfg.setAllocator(allocator).setWorkerThreadCount(10);
|
|
|
|
auto scheduler = std::unique_ptr<marl::Scheduler>(new marl::Scheduler(cfg));
|
|
|
|
auto gotCfg = scheduler->config();
|
|
ASSERT_EQ(gotCfg.allocator, allocator);
|
|
ASSERT_EQ(gotCfg.workerThread.count, 10);
|
|
}
|
|
|
|
TEST_P(WithBoundScheduler, DestructWithPendingTasks) {
|
|
std::atomic<int> counter = {0};
|
|
for (int i = 0; i < 1000; i++) {
|
|
marl::schedule([&] { counter++; });
|
|
}
|
|
|
|
auto scheduler = marl::Scheduler::get();
|
|
scheduler->unbind();
|
|
delete scheduler;
|
|
|
|
// All scheduled tasks should be completed before the scheduler is destructed.
|
|
ASSERT_EQ(counter.load(), 1000);
|
|
|
|
// Rebind a new scheduler so WithBoundScheduler::TearDown() is happy.
|
|
(new marl::Scheduler(marl::Scheduler::Config()))->bind();
|
|
}
|
|
|
|
TEST_P(WithBoundScheduler, DestructWithPendingFibers) {
|
|
std::atomic<int> counter = {0};
|
|
|
|
marl::WaitGroup wg(1);
|
|
for (int i = 0; i < 1000; i++) {
|
|
marl::schedule([&] {
|
|
wg.wait();
|
|
counter++;
|
|
});
|
|
}
|
|
|
|
// Schedule a task to unblock all the tasks scheduled above.
|
|
// We assume that some of these tasks will not finish before the scheduler
|
|
// destruction logic kicks in.
|
|
marl::schedule([=] {
|
|
wg.done(); // Ready, steady, go...
|
|
});
|
|
|
|
auto scheduler = marl::Scheduler::get();
|
|
scheduler->unbind();
|
|
delete scheduler;
|
|
|
|
// All scheduled tasks should be completed before the scheduler is destructed.
|
|
ASSERT_EQ(counter.load(), 1000);
|
|
|
|
// Rebind a new scheduler so WithBoundScheduler::TearDown() is happy.
|
|
(new marl::Scheduler(marl::Scheduler::Config()))->bind();
|
|
}
|
|
|
|
TEST_P(WithBoundScheduler, ScheduleWithArgs) {
|
|
std::string got;
|
|
marl::WaitGroup wg(1);
|
|
marl::schedule(
|
|
[wg, &got](std::string s, int i, bool b) {
|
|
got = "s: '" + s + "', i: " + std::to_string(i) +
|
|
", b: " + (b ? "true" : "false");
|
|
wg.done();
|
|
},
|
|
"a string", 42, true);
|
|
wg.wait();
|
|
ASSERT_EQ(got, "s: 'a string', i: 42, b: true");
|
|
}
|
|
|
|
TEST_P(WithBoundScheduler, FibersResumeOnSameThread) {
|
|
marl::WaitGroup fence(1);
|
|
marl::WaitGroup wg(1000);
|
|
for (int i = 0; i < 1000; i++) {
|
|
marl::schedule([=] {
|
|
auto threadID = std::this_thread::get_id();
|
|
fence.wait();
|
|
ASSERT_EQ(threadID, std::this_thread::get_id());
|
|
wg.done();
|
|
});
|
|
}
|
|
// just to try and get some tasks to yield.
|
|
std::this_thread::sleep_for(std::chrono::milliseconds(10));
|
|
fence.done();
|
|
wg.wait();
|
|
}
|
|
|
|
TEST_P(WithBoundScheduler, FibersResumeOnSameStdThread) {
|
|
auto scheduler = marl::Scheduler::get();
|
|
|
|
// on 32-bit OSs, excessive numbers of threads can run out of address space.
|
|
constexpr auto num_threads = sizeof(void*) > 4 ? 1000 : 100;
|
|
|
|
marl::WaitGroup fence(1);
|
|
marl::WaitGroup wg(num_threads);
|
|
|
|
marl::containers::vector<std::thread, 32> threads;
|
|
for (int i = 0; i < num_threads; i++) {
|
|
threads.emplace_back(std::thread([=] {
|
|
scheduler->bind();
|
|
defer(scheduler->unbind());
|
|
|
|
auto threadID = std::this_thread::get_id();
|
|
fence.wait();
|
|
ASSERT_EQ(threadID, std::this_thread::get_id());
|
|
wg.done();
|
|
}));
|
|
}
|
|
// just to try and get some tasks to yield.
|
|
std::this_thread::sleep_for(std::chrono::milliseconds(10));
|
|
fence.done();
|
|
wg.wait();
|
|
|
|
for (auto& thread : threads) {
|
|
thread.join();
|
|
}
|
|
}
|
|
|
|
TEST_F(WithoutBoundScheduler, TasksOnlyScheduledOnWorkerThreads) {
|
|
marl::Scheduler::Config cfg;
|
|
cfg.setWorkerThreadCount(8);
|
|
|
|
auto scheduler = std::unique_ptr<marl::Scheduler>(new marl::Scheduler(cfg));
|
|
scheduler->bind();
|
|
defer(scheduler->unbind());
|
|
|
|
std::mutex mutex;
|
|
marl::containers::unordered_set<std::thread::id> threads(allocator);
|
|
marl::WaitGroup wg;
|
|
for (int i = 0; i < 10000; i++) {
|
|
wg.add(1);
|
|
marl::schedule([&mutex, &threads, wg] {
|
|
defer(wg.done());
|
|
std::unique_lock<std::mutex> lock(mutex);
|
|
threads.emplace(std::this_thread::get_id());
|
|
});
|
|
}
|
|
wg.wait();
|
|
|
|
ASSERT_LE(threads.size(), 8U);
|
|
ASSERT_EQ(threads.count(std::this_thread::get_id()), 0U);
|
|
}
|
|
|
|
// Test that a marl::Scheduler *with dedicated worker threads* can be used
|
|
// without first binding to the scheduling thread.
|
|
TEST_F(WithoutBoundScheduler, ScheduleMTWWithNoBind) {
|
|
marl::Scheduler::Config cfg;
|
|
cfg.setWorkerThreadCount(8);
|
|
auto scheduler = std::unique_ptr<marl::Scheduler>(new marl::Scheduler(cfg));
|
|
|
|
marl::WaitGroup wg;
|
|
for (int i = 0; i < 100; i++) {
|
|
wg.add(1);
|
|
|
|
marl::Event event;
|
|
scheduler->enqueue(marl::Task([event, wg] {
|
|
event.wait(); // Test that tasks can wait on other tasks.
|
|
wg.done();
|
|
}));
|
|
|
|
scheduler->enqueue(marl::Task([event, &scheduler] {
|
|
// Despite the main thread never binding the scheduler, the scheduler
|
|
// should be automatically bound to worker threads.
|
|
ASSERT_EQ(marl::Scheduler::get(), scheduler.get());
|
|
|
|
event.signal();
|
|
}));
|
|
}
|
|
|
|
// As the scheduler has not been bound to the main thread, the wait() call
|
|
// here will block **without** fiber yielding.
|
|
wg.wait();
|
|
}
|
|
|
|
// Test that a marl::Scheduler *without dedicated worker threads* cannot be used
|
|
// without first binding to the scheduling thread.
|
|
TEST_F(WithoutBoundScheduler, ScheduleSTWWithNoBind) {
|
|
marl::Scheduler::Config cfg;
|
|
auto scheduler = std::unique_ptr<marl::Scheduler>(new marl::Scheduler(cfg));
|
|
|
|
#if MARL_DEBUG_ENABLED && GTEST_HAS_DEATH_TEST
|
|
EXPECT_DEATH(scheduler->enqueue(marl::Task([] {})),
|
|
"Did you forget to call marl::Scheduler::bind");
|
|
#elif !MARL_DEBUG_ENABLED
|
|
scheduler->enqueue(marl::Task([] { FAIL() << "Should not be called"; }));
|
|
#endif
|
|
}
|