Newer
Older
Marc Vef
committed
#include <global/rpc/rpc_types.hpp>
#include <global/rpc/distributor.hpp>
Marc Vef
committed
#include <daemon/handler/rpc_defs.hpp>
Marc Vef
committed
#include <daemon/adafs_ops/metadentry.hpp>
#include <daemon/backend/metadata/db.hpp>
#include <daemon/backend/data/chunk_storage.hpp>
#include <boost/filesystem.hpp>
#include <boost/program_options.hpp>
#include <boost/tokenizer.hpp>
#include <iostream>
#include <fstream>
Marc Vef
committed
#include <csignal>
Marc Vef
committed
#include <condition_variable>
#include <global/global_func.hpp>
Marc Vef
committed
using namespace std;
namespace po = boost::program_options;
Marc Vef
committed
static condition_variable shutdown_please;
static mutex mtx;
bool init_environment() {
// Initialize metadata db
std::string metadata_path = ADAFS_DATA->metadir() + "/rocksdb"s;
try {
ADAFS_DATA->mdb(std::make_shared<MetadataDB>(metadata_path));
} catch (const std::exception & e) {
ADAFS_DATA->spdlogger()->error("{}() unable to initialize metadata DB: {}", __func__, e.what());
}
// Initialize data backend
std::string chunk_storage_path = ADAFS_DATA->rootdir() + "/data/chunks"s;
ADAFS_DATA->spdlogger()->debug("{}() Creating chunk storage directory: '{}'", __func__, chunk_storage_path);
bfs::create_directories(chunk_storage_path);
try {
ADAFS_DATA->storage(std::make_shared<ChunkStorage>(chunk_storage_path, CHUNKSIZE));
} catch (const std::exception & e) {
ADAFS_DATA->spdlogger()->error("{}() unable to initialize storage backend: {}", __func__, e.what());
}
if (!init_rpc_server()) {
ADAFS_DATA->spdlogger()->error("{}() unable to initialize margo rpc server.", __func__);
return false;
}
// Init Argobots ESs to drive IO
if (!init_io_tasklet_pool()) {
ADAFS_DATA->spdlogger()->error("{}() Unable to initialize Argobots pool for I/O.", __func__);
return false;
}
/* Setup distributor */
auto simple_hash_dist = std::make_shared<SimpleHashDistributor>(ADAFS_DATA->host_id(), ADAFS_DATA->host_size());
ADAFS_DATA->distributor(simple_hash_dist);
// TODO set metadata configurations. these have to go into a user configurable file that is parsed here
ADAFS_DATA->atime_state(MDATA_USE_ATIME);
ADAFS_DATA->mtime_state(MDATA_USE_MTIME);
ADAFS_DATA->ctime_state(MDATA_USE_CTIME);
ADAFS_DATA->uid_state(MDATA_USE_UID);
ADAFS_DATA->gid_state(MDATA_USE_GID);
ADAFS_DATA->link_cnt_state(MDATA_USE_LINK_CNT);
ADAFS_DATA->blocks_state(MDATA_USE_BLOCKS);
Metadata root_md{S_IFDIR | S_IRWXU | S_IRWXG | S_IRWXO};
create_metadentry("/", root_md);
} catch (const std::exception& e ) {
ADAFS_DATA->spdlogger()->error("{}() Unable to write root metadentry to KV store: {}", __func__, e.what());
return false;
}
// Register daemon to system
if (!register_daemon_proc()) {
ADAFS_DATA->spdlogger()->error("{}() Unable to register the daemon process to the system.", __func__);
return false;
}
ADAFS_DATA->spdlogger()->info("Startup successful. Daemon is ready.");
return true;
}
/**
* Destroys the margo, argobots, and mercury environments
*/
void destroy_enviroment() {
#ifdef MARGODIAG
cout << "\n####################\n\nMargo RPC server stats: " << endl;
margo_diag_dump(RPC_DATA->server_rpc_mid(), "-", 0);
#endif
bfs::remove_all(ADAFS_DATA->mountdir());
for (unsigned int i = 0; i < RPC_DATA->io_streams().size(); i++) {
ABT_xstream_join(RPC_DATA->io_streams().at(i));
ABT_xstream_free(&RPC_DATA->io_streams().at(i));
}
ADAFS_DATA->spdlogger()->info("{}() Freeing I/O executions streams successful", __func__);
if (!deregister_daemon_proc())
ADAFS_DATA->spdlogger()->warn("{}() Unable to clean up auxiliary files", __func__);
else
ADAFS_DATA->spdlogger()->debug("{}() Cleaning auxiliary files successful", __func__);
if (RPC_DATA->server_rpc_mid() != nullptr) {
margo_finalize(RPC_DATA->server_rpc_mid());
ADAFS_DATA->spdlogger()->info("{}() Margo RPC server shut down successful", __func__);
}
ADAFS_DATA->spdlogger()->info("{}() Closing DB...", __func__);
ADAFS_DATA->close_mdb();
ADAFS_DATA->spdlogger()->info("All services shut down. ADA-FS shutdown complete.");
unsigned int xstreams_num = DAEMON_IO_XSTREAMS;
//retrieve the pool of the just created scheduler
ABT_pool pool;
auto ret = ABT_pool_create_basic(ABT_POOL_FIFO_WAIT, ABT_POOL_ACCESS_MPMC, ABT_TRUE, &pool);
ADAFS_DATA->spdlogger()->error("{}() Failed to create I/O tasks pool", __func__);
//create all subsequent xstream and the associated scheduler, all tapping into the same pool
vector<ABT_xstream> xstreams(xstreams_num);
for (unsigned int i = 0; i < xstreams_num; ++i) {
ret = ABT_xstream_create_basic(ABT_SCHED_BASIC_WAIT, 1, &pool,
ABT_SCHED_CONFIG_NULL, &xstreams[i]);
if (ret != ABT_SUCCESS) {
ADAFS_DATA->spdlogger()->error("{}() Failed to create task execution streams for I/O operations", __func__);
return false;
}
}
RPC_DATA->io_streams(xstreams);
RPC_DATA->io_pool(pool);
auto protocol_port = RPC_PROTOCOL + "://"s + get_my_hostname(false) + ":"s + to_string(RPC_PORT);
hg_addr_t addr_self;
hg_size_t addr_self_cstring_sz = 128;
char addr_self_cstring[128];
ADAFS_DATA->spdlogger()->debug("{}() Initializing Margo RPC server...", __func__);
// IMPORTANT: this struct needs to be zeroed before use
struct hg_init_info hg_options = {};
hg_options.auto_sm = HG_TRUE;
hg_options.stats = HG_FALSE;
hg_options.na_class = nullptr;
// Start Margo (this will also initialize Argobots and Mercury internally)
auto mid = margo_init_opt(protocol_port.c_str(),
MARGO_SERVER_MODE,
&hg_options,
HG_TRUE,
DAEMON_RPC_HANDLER_XSTREAMS);
ADAFS_DATA->spdlogger()->error("{}() margo_init failed to initialize the Margo RPC server", __func__);
return false;
}
#ifdef MARGODIAG
margo_diag_start(mid);
#endif
// Figure out what address this server is listening on (must be freed when finished)
auto hret = margo_addr_self(mid, &addr_self);
if (hret != HG_SUCCESS) {
ADAFS_DATA->spdlogger()->error("{}() margo_addr_self() Failed to retrieve server RPC address", __func__);
return false;
}
// Convert the address to a cstring (with \0 terminator).
hret = margo_addr_to_string(mid, addr_self_cstring, &addr_self_cstring_sz, addr_self);
if (hret != HG_SUCCESS) {
ADAFS_DATA->spdlogger()->error("{}() margo_addr_to_string Failed to convert address to cstring", __func__);
margo_addr_free(mid, addr_self);
margo_finalize(mid);
return false;
}
std::string addr_self_str(addr_self_cstring);
RPC_DATA->self_addr_str(addr_self_str);
ADAFS_DATA->spdlogger()->info("{}() Margo RPC server initialized. Accepting RPCs on address {}", __func__,
addr_self_cstring);
// Put context and class into RPC_data object
// register RPCs
return true;
}
* Registers RPC handlers to Margo instance
void register_server_rpcs(margo_instance_id mid) {
MARGO_REGISTER(mid, hg_tag::fs_config, rpc_config_in_t, rpc_config_out_t, rpc_srv_fs_config);
MARGO_REGISTER(mid, hg_tag::create, rpc_mk_node_in_t, rpc_err_out_t, rpc_srv_mk_node);
MARGO_REGISTER(mid, hg_tag::access, rpc_access_in_t, rpc_err_out_t, rpc_srv_access);
MARGO_REGISTER(mid, hg_tag::stat, rpc_path_only_in_t, rpc_stat_out_t, rpc_srv_stat);
MARGO_REGISTER(mid, hg_tag::decr_size, rpc_trunc_in_t, rpc_err_out_t, rpc_srv_decr_size);
MARGO_REGISTER(mid, hg_tag::remove, rpc_rm_node_in_t, rpc_err_out_t, rpc_srv_rm_node);
MARGO_REGISTER(mid, hg_tag::update_metadentry, rpc_update_metadentry_in_t, rpc_err_out_t,
MARGO_REGISTER(mid, hg_tag::get_metadentry_size, rpc_path_only_in_t, rpc_get_metadentry_size_out_t,
rpc_srv_get_metadentry_size);
MARGO_REGISTER(mid, hg_tag::update_metadentry_size, rpc_update_metadentry_size_in_t,
rpc_update_metadentry_size_out_t, rpc_srv_update_metadentry_size);
MARGO_REGISTER(mid, hg_tag::get_dirents, rpc_get_dirents_in_t, rpc_get_dirents_out_t,
rpc_srv_get_dirents);
MARGO_REGISTER(mid, hg_tag::write_data, rpc_write_data_in_t, rpc_data_out_t, rpc_srv_write_data);
MARGO_REGISTER(mid, hg_tag::read_data, rpc_read_data_in_t, rpc_data_out_t, rpc_srv_read_data);
MARGO_REGISTER(mid, hg_tag::trunc_data, rpc_trunc_in_t, rpc_err_out_t, rpc_srv_trunc_data);
MARGO_REGISTER(mid, hg_tag::chunk_stat, rpc_chunk_stat_in_t, rpc_chunk_stat_out_t, rpc_srv_chunk_stat);
}
/**
* Registers the daemon process to the system.
* This will create a file with additional information for clients started on the same node.
* @return
*/
bool register_daemon_proc() {
auto daemon_aux_path = DAEMON_AUX_PATH;
if (!bfs::exists(daemon_aux_path) && !bfs::create_directories(daemon_aux_path)) {
ADAFS_DATA->spdlogger()->error("{}() Unable to create adafs auxiliary directory in {}", __func__,
daemon_aux_path);
return false;
}
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
auto pid_file = daemon_pid_path();
// check if a pid file exists from another adafs_daemon
if (bfs::exists(pid_file)) {
// check if another daemon is still running
ifstream ifs(pid_file, ::ifstream::in);
if (ifs) {
string running_pid;
// first line is pid of daemon
if (getline(ifs, running_pid) && !running_pid.empty()) {
// check if pid is running and kill it
if (kill(::stoi(running_pid), 0) == 0) {
ADAFS_DATA->spdlogger()->warn("{}() Daemon with pid {} is already running. Proceed to force kill",
__func__, running_pid);
kill(::stoi(running_pid), SIGKILL);
sleep(1);
if (kill(::stoi(running_pid), 0) == 0) {
ADAFS_DATA->spdlogger()->error("{}() Running daemon with pid {} cannot be killed. Exiting ...",
__func__, running_pid);
ifs.close();
return false;
} else
ADAFS_DATA->spdlogger()->info("{}() Kill successful", __func__);
}
}
}
ifs.close();
}
auto my_pid = getpid();
if (my_pid == -1) {
ADAFS_DATA->spdlogger()->error("{}() Unable to get pid", __func__);
return false;
}
ofstream ofs(pid_file, ::ofstream::trunc);
if (ofs) {
ofs << to_string(my_pid) << std::endl;
ofs << RPC_DATA->self_addr_str() << std::endl;
ofs << ADAFS_DATA->mountdir() << std::endl;
} else {
cerr << "Unable to create daemon pid file at " << pid_file << endl;
ADAFS_DATA->spdlogger()->error("{}() Unable to create daemon pid file at {}. No permissions?", __func__,
pid_file);
return false;
}
ofs.close();
return true;
}
bool deregister_daemon_proc() {
return bfs::remove(daemon_pid_path());
Marc Vef
committed
}
void shutdown_handler(int dummy) {
shutdown_please.notify_all();
}
std::string path = DEFAULT_DAEMON_LOG_PATH;
// Try to get log path from env variable
std::string env_path_key = ENV_PREFIX;
env_path_key += "DAEMON_LOG_PATH";
char* env_path = getenv(env_path_key.c_str());
if (env_path != nullptr) {
path = env_path;
}
spdlog::level::level_enum level = get_spdlog_level(DEFAULT_DAEMON_LOG_LEVEL);
// Try to get log path from env variable
std::string env_level_key = ENV_PREFIX;
env_level_key += "LOG_LEVEL";
char* env_level = getenv(env_level_key.c_str());
if (env_level != nullptr) {
level = get_spdlog_level(env_level);
}
auto logger_names = std::vector<std::string>{
"main",
setup_loggers(logger_names, level, path);
Marc Vef
committed
Marc Vef
committed
initialize_loggers();
ADAFS_DATA->spdlogger(spdlog::get("main"));
// Parse input
Marc Vef
committed
po::options_description desc("Allowed options");
desc.add_options()
("help,h", "Help message")
("mountdir,m", po::value<string>()->required(), "User Fuse mountdir")
Marc Vef
committed
("rootdir,r", po::value<string>()->required(), "ADA-FS data directory")
("metadir,i", po::value<string>(), "ADA-FS metadata directory, if not set rootdir is used for metadata ")
Marc Vef
committed
("hostfile", po::value<string>(), "Path to the hosts_file for all fs participants")
("hosts,h", po::value<string>(), "Comma separated list of hosts_ for all fs participants");
po::variables_map vm;
po::store(po::parse_command_line(argc, argv, desc), vm);
if (vm.count("help")) {
cout << desc << "\n";
return 1;
}
Marc Vef
committed
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
try {
po::notify(vm);
} catch (po::required_option& e) {
std::cerr << "Error: " << e.what() << "\n";
return 1;
}
// parse host parameters
vector<string> hosts{};
if (vm.count("hostfile")) {
auto host_path = vm["hostfile"].as<string>();
fstream host_file(host_path);
if (host_file.is_open()) {
for (string line; getline(host_file, line);) {
if (line.at(0) != '#') {
auto subline = line.substr(0, line.find(' '));
hosts.push_back(subline);
}
}
} else {
cerr << "Hostfile path does not exist. Exiting ..." << endl;
ADAFS_DATA->spdlogger()->error("{}() Hostfile path does not exist. Exiting ...", __func__);
assert(host_file.is_open());
}
} else if (vm.count("hosts")) {
// split comma separated host string
boost::char_separator<char> sep(",");
boost::tokenizer<boost::char_separator<char>> tok(vm["hosts"].as<string>(), sep);
for (auto&& s : tok) {
hosts.push_back(s);
}
}
// convert host parameters into datastructures
std::map<uint64_t, std::string> hostmap;
auto hosts_raw = ""s;
if (!hosts.empty()) {
auto i = static_cast<uint64_t>(0);
auto found_hostname = false;
auto hostname = get_my_hostname(true);
if (hostname.empty()) {
cerr << "Unable to read the machine's hostname" << endl;
ADAFS_DATA->spdlogger()->error("{}() Unable to read the machine's hostname", __func__);
assert(!hostname.empty());
}
for (auto&& host : hosts) {
hostmap[i] = host;
hosts_raw += host + ","s;
if (hostname == host) {
ADAFS_DATA->host_id(i);
found_hostname = true;
}
i++;
}
if (!found_hostname) {
ADAFS_DATA->spdlogger()->error("{}() Hostname was not found in given parameters. Exiting ...", __func__);
cerr << "Hostname was not found in given parameters. Exiting ..." << endl;
assert(found_hostname);
}
hosts_raw = hosts_raw.substr(0, hosts_raw.size() - 1);
} else {
// single node mode
ADAFS_DATA->spdlogger()->info("{}() Single node mode set to self", __func__);
auto hostname = get_my_hostname(false);
hostmap[0] = hostname;
hosts_raw = hostname;
ADAFS_DATA->host_id(0);
}
ADAFS_DATA->hosts(hostmap);
ADAFS_DATA->host_size(hostmap.size());
ADAFS_DATA->rpc_port(fmt::format_int(RPC_PORT).str());
Marc Vef
committed
ADAFS_DATA->hosts_raw(hosts_raw);
ADAFS_DATA->spdlogger()->info("{}() Initializing environment. Hold on ...", __func__);
assert(vm.count("mountdir"));
auto mountdir = vm["mountdir"].as<string>();
Marc Vef
committed
// Create mountdir. We use this dir to get some information on the underlying fs with statfs in adafs_statfs
bfs::create_directories(mountdir);
ADAFS_DATA->mountdir(bfs::canonical(mountdir).native());
assert(vm.count("rootdir"));
auto rootdir = vm["rootdir"].as<string>();
bfs::create_directories(rootdir);
ADAFS_DATA->rootdir(bfs::canonical(rootdir).native());
if (vm.count("metadir")) {
auto metadir = vm["metadir"].as<string>();
bfs::create_directories(metadir);
ADAFS_DATA->metadir(bfs::canonical(metadir).native());
} else {
// use rootdir as metadata dir
ADAFS_DATA->metadir(ADAFS_DATA->rootdir());
}
Marc Vef
committed
if (init_environment()) {
signal(SIGINT, shutdown_handler);
signal(SIGTERM, shutdown_handler);
signal(SIGKILL, shutdown_handler);
unique_lock<mutex> lk(mtx);
// Wait for shutdown signal to initiate shutdown protocols
shutdown_please.wait(lk);
ADAFS_DATA->spdlogger()->info("{}() Shutting done signal encountered. Shutting down ...", __func__);
} else {
ADAFS_DATA->spdlogger()->info("{}() Starting up daemon environment failed. Shutting down ...", __func__);
}
destroy_enviroment();
return 0;