d9f25cc668
Summary: Wal on workers didn't contain committed transactions ids, this is needed for distributed recovery so that the master may decide which transactions are present on all the workers. Reviewers: buda, msantl Reviewed By: buda Subscribers: pullbot, msantl, buda Differential Revision: https://phabricator.memgraph.io/D1440
259 lines
7.9 KiB
C++
259 lines
7.9 KiB
C++
#include <experimental/filesystem>
|
|
#include <memory>
|
|
#include <thread>
|
|
|
|
#include <gflags/gflags.h>
|
|
#include <gtest/gtest.h>
|
|
|
|
#include "database/graph_db.hpp"
|
|
#include "database/graph_db_accessor.hpp"
|
|
#include "distributed/data_manager.hpp"
|
|
#include "distributed/updates_rpc_server.hpp"
|
|
#include "storage/address_types.hpp"
|
|
#include "transactions/engine_master.hpp"
|
|
|
|
DECLARE_string(durability_directory);
|
|
|
|
namespace fs = std::experimental::filesystem;
|
|
|
|
class WorkerInThread {
|
|
public:
|
|
explicit WorkerInThread(database::Config config) : worker_(config) {
|
|
thread_ = std::thread([this, config] { worker_.WaitForShutdown(); });
|
|
}
|
|
|
|
~WorkerInThread() {
|
|
if (thread_.joinable()) thread_.join();
|
|
}
|
|
|
|
database::Worker *db() { return &worker_; }
|
|
|
|
database::Worker worker_;
|
|
std::thread thread_;
|
|
};
|
|
|
|
class DistributedGraphDbTest : public ::testing::Test {
|
|
const std::string kLocal = "127.0.0.1";
|
|
const int kWorkerCount = 2;
|
|
|
|
protected:
|
|
virtual int QueryExecutionTimeSec(int) { return 180; }
|
|
|
|
void Initialize(
|
|
std::function<database::Config(database::Config config)> modify_config) {
|
|
using namespace std::literals::chrono_literals;
|
|
const auto kInitTime = 200ms;
|
|
|
|
database::Config master_config;
|
|
master_config.master_endpoint = {kLocal, 0};
|
|
master_config.query_execution_time_sec = QueryExecutionTimeSec(0);
|
|
master_config.durability_directory = tmp_dir_;
|
|
// This is semantically wrong since this is not a cluster of size 1 but of
|
|
// size kWorkerCount+1, but it's hard to wait here for workers to recover
|
|
// and simultaneously assign the port to which the workers must connect
|
|
// TODO(dgleich): Fix sometime in the future - not mission critical
|
|
master_config.recovering_cluster_size = 1;
|
|
master_ = std::make_unique<database::Master>(modify_config(master_config));
|
|
|
|
std::this_thread::sleep_for(kInitTime);
|
|
auto worker_config = [this](int worker_id) {
|
|
database::Config config;
|
|
config.worker_id = worker_id;
|
|
config.master_endpoint = master_->endpoint();
|
|
config.durability_directory = tmp_dir_;
|
|
config.worker_endpoint = {kLocal, 0};
|
|
config.query_execution_time_sec = QueryExecutionTimeSec(worker_id);
|
|
return config;
|
|
};
|
|
|
|
// Flag needs to be updated due to props on disk storage.
|
|
FLAGS_durability_directory = tmp_dir_;
|
|
|
|
for (int i = 0; i < kWorkerCount; ++i) {
|
|
workers_.emplace_back(std::make_unique<WorkerInThread>(
|
|
modify_config(worker_config(i + 1))));
|
|
std::this_thread::sleep_for(kInitTime);
|
|
}
|
|
}
|
|
|
|
void SetUp() override {
|
|
Initialize([](database::Config config) { return config; });
|
|
}
|
|
|
|
void ShutDown() {
|
|
// Kill master first because it will expect a shutdown response from the
|
|
// workers.
|
|
auto t = std::thread([this]() { master_ = nullptr; });
|
|
workers_.clear();
|
|
if (t.joinable()) t.join();
|
|
}
|
|
|
|
void CleanDurability() {
|
|
if (fs::exists(tmp_dir_)) fs::remove_all(tmp_dir_);
|
|
}
|
|
|
|
void TearDown() override {
|
|
ShutDown();
|
|
CleanDurability();
|
|
}
|
|
|
|
database::Master &master() { return *master_; }
|
|
auto &master_tx_engine() {
|
|
return dynamic_cast<tx::MasterEngine &>(master_->tx_engine());
|
|
}
|
|
|
|
database::Worker &worker(int worker_id) {
|
|
return workers_[worker_id - 1]->worker_;
|
|
}
|
|
|
|
/// Inserts a vertex and returns it's global address. Does it in a new
|
|
/// transaction.
|
|
storage::VertexAddress InsertVertex(database::GraphDb &db) {
|
|
database::GraphDbAccessor dba{db};
|
|
auto r_val = dba.InsertVertex().GlobalAddress();
|
|
dba.Commit();
|
|
return r_val;
|
|
}
|
|
|
|
/// Inserts an edge (on the 'from' side) and returns it's global address.
|
|
auto InsertEdge(storage::VertexAddress from_addr,
|
|
storage::VertexAddress to_addr,
|
|
const std::string &edge_type_name) {
|
|
CHECK(from_addr.is_remote() && to_addr.is_remote())
|
|
<< "Distributed test InsertEdge only takes global addresses";
|
|
database::GraphDbAccessor dba{master()};
|
|
VertexAccessor from{from_addr, dba};
|
|
VertexAccessor to{to_addr, dba};
|
|
auto r_val =
|
|
dba.InsertEdge(from, to, dba.EdgeType(edge_type_name)).GlobalAddress();
|
|
master().updates_server().Apply(dba.transaction_id());
|
|
worker(1).updates_server().Apply(dba.transaction_id());
|
|
worker(2).updates_server().Apply(dba.transaction_id());
|
|
dba.Commit();
|
|
return r_val;
|
|
}
|
|
|
|
auto VertexCount(database::GraphDb &db) {
|
|
database::GraphDbAccessor dba{db};
|
|
auto vertices = dba.Vertices(false);
|
|
return std::distance(vertices.begin(), vertices.end());
|
|
};
|
|
|
|
auto EdgeCount(database::GraphDb &db) {
|
|
database::GraphDbAccessor dba(db);
|
|
auto edges = dba.Edges(false);
|
|
return std::distance(edges.begin(), edges.end());
|
|
};
|
|
|
|
fs::path tmp_dir_{fs::temp_directory_path() / "MG_test_unit_durability"};
|
|
|
|
public:
|
|
// Each test has to specify its own durability suffix to avoid conflicts
|
|
DistributedGraphDbTest() = delete;
|
|
|
|
DistributedGraphDbTest(const std::string &dir_suffix)
|
|
: dir_suffix_(dir_suffix) {
|
|
tmp_dir_ =
|
|
fs::temp_directory_path() / ("MG_test_unit_durability_" + dir_suffix_);
|
|
}
|
|
|
|
private:
|
|
std::string dir_suffix_{""};
|
|
std::unique_ptr<database::Master> master_;
|
|
std::vector<std::unique_ptr<WorkerInThread>> workers_;
|
|
};
|
|
|
|
enum class TestType { SINGLE_NODE, DISTRIBUTED };
|
|
|
|
// Class that can be used both in distributed and single node tests.
|
|
class Cluster {
|
|
public:
|
|
Cluster(TestType test_type, int num_workers = 0) : test_type_(test_type) {
|
|
using namespace std::literals::chrono_literals;
|
|
switch (test_type) {
|
|
case TestType::SINGLE_NODE:
|
|
master_ = std::make_unique<database::SingleNode>(database::Config{});
|
|
break;
|
|
case TestType::DISTRIBUTED:
|
|
database::Config master_config;
|
|
master_config.master_endpoint = {kLocal, 0};
|
|
|
|
auto master_tmp = std::make_unique<database::Master>(master_config);
|
|
auto master_endpoint = master_tmp->endpoint();
|
|
master_ = std::move(master_tmp);
|
|
|
|
const auto kInitTime = 200ms;
|
|
std::this_thread::sleep_for(kInitTime);
|
|
|
|
auto worker_config = [this, master_endpoint](int worker_id) {
|
|
database::Config config;
|
|
config.worker_id = worker_id;
|
|
config.master_endpoint = master_endpoint;
|
|
config.worker_endpoint = {kLocal, 0};
|
|
return config;
|
|
};
|
|
|
|
for (int i = 0; i < num_workers; ++i) {
|
|
workers_.emplace_back(
|
|
std::make_unique<WorkerInThread>(worker_config(i + 1)));
|
|
}
|
|
std::this_thread::sleep_for(kInitTime);
|
|
break;
|
|
}
|
|
}
|
|
|
|
~Cluster() {
|
|
auto t = std::thread([this] { master_ = nullptr; });
|
|
workers_.clear();
|
|
if (t.joinable()) t.join();
|
|
}
|
|
|
|
database::GraphDb *master() { return master_.get(); }
|
|
auto workers() {
|
|
return iter::imap([](auto &worker) { return worker->db(); }, workers_);
|
|
}
|
|
|
|
void ClearCache(tx::TransactionId tx_id) {
|
|
master()->data_manager().ClearCacheForSingleTransaction(tx_id);
|
|
for (auto member : workers()) {
|
|
member->data_manager().ClearCacheForSingleTransaction(tx_id);
|
|
}
|
|
}
|
|
|
|
void ApplyUpdates(tx::TransactionId tx_id) {
|
|
switch (test_type_) {
|
|
case TestType::SINGLE_NODE:
|
|
break;
|
|
case TestType::DISTRIBUTED:
|
|
master()->updates_server().Apply(tx_id);
|
|
for (auto member : workers()) {
|
|
member->updates_server().Apply(tx_id);
|
|
}
|
|
ClearCache(tx_id);
|
|
}
|
|
}
|
|
|
|
void AdvanceCommand(tx::TransactionId tx_id) {
|
|
switch (test_type_) {
|
|
case TestType::SINGLE_NODE: {
|
|
database::GraphDbAccessor dba{*master(), tx_id};
|
|
dba.AdvanceCommand();
|
|
break;
|
|
}
|
|
case TestType::DISTRIBUTED:
|
|
ApplyUpdates(tx_id);
|
|
master()->tx_engine().Advance(tx_id);
|
|
for (auto worker : workers()) worker->tx_engine().UpdateCommand(tx_id);
|
|
ClearCache(tx_id);
|
|
break;
|
|
}
|
|
}
|
|
|
|
private:
|
|
const std::string kLocal = "127.0.0.1";
|
|
|
|
TestType test_type_;
|
|
std::unique_ptr<database::GraphDb> master_;
|
|
std::vector<std::unique_ptr<WorkerInThread>> workers_;
|
|
};
|