X-Git-Url: https://git.sesse.net/?p=cubemap;a=blobdiff_plain;f=main.cpp;h=5ec0cb487269649c59c32dde293b3c42133ab2a7;hp=1f7313ac0eb6c013648d7edafff1e93b99902460;hb=9abb89bcf7940e2ada9d708f86a218a56334f68d;hpb=d9b0536b128e10447a986f7b6ea2b5c93252c80f diff --git a/main.cpp b/main.cpp index 1f7313a..5ec0cb4 100644 --- a/main.cpp +++ b/main.cpp @@ -1,385 +1,436 @@ +#include +#include +#include +#include +#include +#include #include +#include #include -#include -#include -#include -#include -#include -#include -#include -#include #include -#include -#include -#include -#include -#include -#include +#include +#include #include #include +#include +#include +#include #include "acceptor.h" +#include "accesslog.h" +#include "config.h" +#include "input.h" +#include "log.h" #include "markpool.h" -#include "metacube.h" -#include "parse.h" -#include "server.h" #include "serverpool.h" -#include "input.h" -#include "stats.h" #include "state.pb.h" +#include "stats.h" +#include "stream.h" +#include "util.h" +#include "version.h" using namespace std; +AccessLogThread *access_log = NULL; ServerPool *servers = NULL; +vector mark_pools; volatile bool hupped = false; +volatile bool stopped = false; -void hup(int ignored) +struct InputWithRefcount { + Input *input; + int refcount; +}; + +void hup(int signum) { hupped = true; + if (signum == SIGINT) { + stopped = true; + } } -// Serialize the given state to a file descriptor, and return the (still open) -// descriptor. -int make_tempfile(const CubemapStateProto &state) +void do_nothing(int signum) { - char tmpl[] = "/tmp/cubemapstate.XXXXXX"; - int state_fd = mkstemp(tmpl); - if (state_fd == -1) { - perror("mkstemp"); - exit(1); - } - - string serialized; - state.SerializeToString(&serialized); - - const char *ptr = serialized.data(); - size_t to_write = serialized.size(); - while (to_write > 0) { - ssize_t ret = write(state_fd, ptr, to_write); - if (ret == -1) { - perror("write"); - exit(1); - } - - ptr += ret; - to_write -= ret; - } - - return state_fd; } CubemapStateProto collect_state(const timeval &serialize_start, - int server_sock, - int port, - const vector inputs, - ServerPool *servers, - int num_servers) + const vector acceptors, + const multimap inputs, + ServerPool *servers) { - CubemapStateProto state; + CubemapStateProto state = servers->serialize(); // Fills streams() and clients(). state.set_serialize_start_sec(serialize_start.tv_sec); state.set_serialize_start_usec(serialize_start.tv_usec); - state.set_server_sock(server_sock); - state.set_port(port); - - for (size_t i = 0; i < inputs.size(); ++i) { - state.add_inputs()->MergeFrom(inputs[i]->serialize()); + + for (size_t i = 0; i < acceptors.size(); ++i) { + state.add_acceptors()->MergeFrom(acceptors[i]->serialize()); } - for (int i = 0; i < num_servers; ++i) { - CubemapStateProto local_state = servers->get_server(i)->serialize(); - - // The stream state should be identical between the servers, so we only store it once. - if (i == 0) { - state.mutable_streams()->MergeFrom(local_state.streams()); - } - for (int j = 0; j < local_state.clients_size(); ++j) { - state.add_clients()->MergeFrom(local_state.clients(j)); - } + for (multimap::const_iterator input_it = inputs.begin(); + input_it != inputs.end(); + ++input_it) { + state.add_inputs()->MergeFrom(input_it->second.input->serialize()); } return state; } -// Read the state back from the file descriptor made by make_tempfile, -// and close it. -CubemapStateProto read_tempfile(int state_fd) +// Find all port statements in the configuration file, and create acceptors for htem. +vector create_acceptors( + const Config &config, + map *deserialized_acceptors) { - if (lseek(state_fd, 0, SEEK_SET) == -1) { - perror("lseek"); - exit(1); - } - - string serialized; - char buf[4096]; - for ( ;; ) { - ssize_t ret = read(state_fd, buf, sizeof(buf)); - if (ret == -1) { - perror("read"); - exit(1); + vector acceptors; + for (unsigned i = 0; i < config.acceptors.size(); ++i) { + const AcceptorConfig &acceptor_config = config.acceptors[i]; + Acceptor *acceptor = NULL; + map::iterator deserialized_acceptor_it = + deserialized_acceptors->find(acceptor_config.port); + if (deserialized_acceptor_it != deserialized_acceptors->end()) { + acceptor = deserialized_acceptor_it->second; + deserialized_acceptors->erase(deserialized_acceptor_it); + } else { + int server_sock = create_server_socket(acceptor_config.port, TCP_SOCKET); + acceptor = new Acceptor(server_sock, acceptor_config.port); } - if (ret == 0) { - // EOF. - break; - } - - serialized.append(string(buf, buf + ret)); + acceptor->run(); + acceptors.push_back(acceptor); } - close(state_fd); // Implicitly deletes the file. - - CubemapStateProto state; - if (!state.ParseFromString(serialized)) { - fprintf(stderr, "PANIC: Failed deserialization of state.\n"); - exit(1); + // Close all acceptors that are no longer in the configuration file. + for (map::iterator acceptor_it = deserialized_acceptors->begin(); + acceptor_it != deserialized_acceptors->end(); + ++acceptor_it) { + acceptor_it->second->close_socket(); + delete acceptor_it->second; } - return state; -} - -// Reuse mark pools if one already exists. -MarkPool *get_mark_pool(map, MarkPool *> *mark_pools, int from, int to) -{ - pair mark_range(from, to); - if (mark_pools->count(mark_range) != 0) { - return (*mark_pools)[mark_range]; - } - - // Check if we're overlapping some other mark pool. - for (map, MarkPool *>::const_iterator mp_it = mark_pools->begin(); - mp_it != mark_pools->end(); - ++mp_it) { - int other_from = mp_it->first.first; - int other_to = mp_it->first.second; - if ((from >= other_from && from < other_to) || - (to >= other_from && to < other_to)) { - fprintf(stderr, "WARNING: Mark pool %d-%d partially overlaps with %d-%d, you may get duplicate marks.\n", - from, to, other_from, other_to); - fprintf(stderr, " Mark pools must either be completely disjunct, or completely overlapping.\n"); - } - } - - MarkPool *mark_pool = new MarkPool(from, to); - mark_pools->insert(make_pair(mark_range, mark_pool)); - return mark_pool; + return acceptors; } - -MarkPool *parse_mark_pool(map, MarkPool *> *mark_pools, const string &mark_str) -{ - size_t split = mark_str.find_first_of('-'); - if (split == string::npos) { - fprintf(stderr, "WARNING: Invalid mark specification '%s' (expected 'X-Y'), ignoring.\n", - mark_str.c_str()); - return NULL; - } - - string from_str(mark_str.begin(), mark_str.begin() + split); - string to_str(mark_str.begin() + split + 1, mark_str.end()); - int from = atoi(from_str.c_str()); - int to = atoi(to_str.c_str()); - - if (from <= 0 || from >= 65536 || to <= 0 || to >= 65536) { - fprintf(stderr, "WARNING: Mark pool range %d-%d is outside legal range [1,65536>, ignoring.\n", - from, to); - return NULL; - } - return get_mark_pool(mark_pools, from, to); -} - // Find all streams in the configuration file, and create inputs for them. -vector create_inputs(const vector &config, - map *deserialized_inputs) +void create_config_inputs(const Config &config, multimap *inputs) { - vector inputs; - for (unsigned i = 0; i < config.size(); ++i) { - if (config[i].keyword != "stream") { + for (unsigned i = 0; i < config.streams.size(); ++i) { + const StreamConfig &stream_config = config.streams[i]; + if (stream_config.src.empty()) { continue; } - assert(config[i].arguments.size() == 1); - string stream_id = config[i].arguments[0]; - - map::const_iterator src_it = - config[i].parameters.find("src"); - if (src_it == config[i].parameters.end()) { - fprintf(stderr, "WARNING: stream '%s' has no src= attribute, clients will not get any data.\n", - stream_id.c_str()); + + string src = stream_config.src; + if (inputs->count(src) != 0) { continue; } - string src = src_it->second; - Input *input = NULL; - map::iterator deserialized_input_it = - deserialized_inputs->find(stream_id); - if (deserialized_input_it != deserialized_inputs->end()) { - input = deserialized_input_it->second; - if (input->get_url() != src) { - fprintf(stderr, "INFO: Stream '%s' has changed URL from '%s' to '%s', restarting input.\n", - stream_id.c_str(), input->get_url().c_str(), src.c_str()); - delete input; - input = NULL; - } - deserialized_inputs->erase(deserialized_input_it); - } - if (input == NULL) { - input = new Input(stream_id, src); + InputWithRefcount iwr; + iwr.input = create_input(src); + if (iwr.input == NULL) { + log(ERROR, "did not understand URL '%s', clients will not get any data.", + src.c_str()); + continue; } - input->run(); - inputs.push_back(input); + iwr.refcount = 0; + inputs->insert(make_pair(src, iwr)); } - return inputs; } -void create_streams(const vector &config, - const set &deserialized_stream_ids, - map *deserialized_inputs) +void create_streams(const Config &config, + const set &deserialized_urls, + multimap *inputs) { - set expecting_stream_ids = deserialized_stream_ids; - map, MarkPool *> mark_pools; - for (unsigned i = 0; i < config.size(); ++i) { - if (config[i].keyword != "stream") { - continue; - } - if (config[i].arguments.size() != 1) { - fprintf(stderr, "ERROR: 'stream' takes exactly one argument\n"); - exit(1); + for (unsigned i = 0; i < config.mark_pools.size(); ++i) { + const MarkPoolConfig &mp_config = config.mark_pools[i]; + mark_pools.push_back(new MarkPool(mp_config.from, mp_config.to)); + } + + set expecting_urls = deserialized_urls; + for (unsigned i = 0; i < config.streams.size(); ++i) { + const StreamConfig &stream_config = config.streams[i]; + int stream_index; + if (deserialized_urls.count(stream_config.url) == 0) { + stream_index = servers->add_stream(stream_config.url, + stream_config.backlog_size, + Stream::Encoding(stream_config.encoding)); + } else { + stream_index = servers->lookup_stream_by_url(stream_config.url); + assert(stream_index != -1); + servers->set_backlog_size(stream_index, stream_config.backlog_size); + servers->set_encoding(stream_index, + Stream::Encoding(stream_config.encoding)); } - string stream_id = config[i].arguments[0]; - if (deserialized_stream_ids.count(stream_id) == 0) { - servers->add_stream(stream_id); + expecting_urls.erase(stream_config.url); + + if (stream_config.mark_pool != -1) { + servers->set_mark_pool(stream_index, mark_pools[stream_config.mark_pool]); } - expecting_stream_ids.erase(stream_id); - - // Set up marks, if so desired. - map::const_iterator mark_parm_it = - config[i].parameters.find("mark"); - if (mark_parm_it != config[i].parameters.end()) { - MarkPool *mark_pool = parse_mark_pool(&mark_pools, mark_parm_it->second); - servers->set_mark_pool(stream_id, mark_pool); + + string src = stream_config.src; + if (!src.empty()) { + multimap::iterator input_it = inputs->find(src); + assert(input_it != inputs->end()); + input_it->second.input->add_destination(stream_index); + ++input_it->second.refcount; } } // Warn about any servers we've lost. // TODO: Make an option (delete=yes?) to actually shut down streams. - for (set::const_iterator stream_it = expecting_stream_ids.begin(); - stream_it != expecting_stream_ids.end(); + for (set::const_iterator stream_it = expecting_urls.begin(); + stream_it != expecting_urls.end(); ++stream_it) { - string stream_id = *stream_it; - fprintf(stderr, "WARNING: stream '%s' disappeared from the configuration file.\n", - stream_id.c_str()); - fprintf(stderr, " It will not be deleted, but clients will not get any new inputs.\n"); - if (deserialized_inputs->count(stream_id) != 0) { - delete (*deserialized_inputs)[stream_id]; - deserialized_inputs->erase(stream_id); + string url = *stream_it; + log(WARNING, "stream '%s' disappeared from the configuration file. " + "It will not be deleted, but clients will not get any new inputs.", + url.c_str()); + } +} + +void open_logs(const vector &log_destinations) +{ + for (size_t i = 0; i < log_destinations.size(); ++i) { + if (log_destinations[i].type == LogConfig::LOG_TYPE_FILE) { + add_log_destination_file(log_destinations[i].filename); + } else if (log_destinations[i].type == LogConfig::LOG_TYPE_CONSOLE) { + add_log_destination_console(); + } else if (log_destinations[i].type == LogConfig::LOG_TYPE_SYSLOG) { + add_log_destination_syslog(); + } else { + assert(false); } } + start_logging(); +} + +bool dry_run_config(const std::string &argv0, const std::string &config_filename) +{ + char *argv0_copy = strdup(argv0.c_str()); + char *config_filename_copy = strdup(config_filename.c_str()); + + pid_t pid = fork(); + switch (pid) { + case -1: + log_perror("fork()"); + free(argv0_copy); + free(config_filename_copy); + return false; + case 0: + // Child. + execlp(argv0_copy, argv0_copy, "--test-config", config_filename_copy, NULL); + log_perror(argv0_copy); + _exit(1); + default: + // Parent. + break; + } + + free(argv0_copy); + free(config_filename_copy); + + int status; + pid_t err; + do { + err = waitpid(pid, &status, 0); + } while (err == -1 && errno == EINTR); + + if (err == -1) { + log_perror("waitpid()"); + return false; + } + + return (WIFEXITED(status) && WEXITSTATUS(status) == 0); } int main(int argc, char **argv) { - fprintf(stderr, "\nCubemap starting.\n"); + signal(SIGHUP, hup); + signal(SIGINT, hup); + signal(SIGUSR1, do_nothing); // Used in internal signalling. + signal(SIGPIPE, SIG_IGN); + + // Parse options. + int state_fd = -1; + bool test_config = false; + for ( ;; ) { + static const option long_options[] = { + { "state", required_argument, 0, 's' }, + { "test-config", no_argument, 0, 't' }, + { 0, 0, 0, 0 } + }; + int option_index = 0; + int c = getopt_long(argc, argv, "s:t", long_options, &option_index); + + if (c == -1) { + break; + } + switch (c) { + case 's': + state_fd = atoi(optarg); + break; + case 't': + test_config = true; + break; + default: + fprintf(stderr, "Unknown option '%s'\n", argv[option_index]); + exit(1); + } + } - struct timeval serialize_start; - bool is_reexec = false; + string config_filename = "cubemap.config"; + if (optind < argc) { + config_filename = argv[optind++]; + } - string config_filename = (argc == 1) ? "cubemap.config" : argv[1]; - vector config = parse_config(config_filename); + // Canonicalize argv[0] and config_filename. + char argv0_canon[PATH_MAX]; + char config_filename_canon[PATH_MAX]; - int port = fetch_config_int(config, "port", 1, 65535, PARAMATER_MANDATORY); - int num_servers = fetch_config_int(config, "num_servers", 1, 20000, PARAMATER_MANDATORY); // Insanely high max limit. + if (realpath(argv[0], argv0_canon) == NULL) { + log_perror(argv[0]); + exit(1); + } + if (realpath(config_filename.c_str(), config_filename_canon) == NULL) { + log_perror(config_filename.c_str()); + exit(1); + } - servers = new ServerPool(num_servers); + // Now parse the configuration file. + Config config; + if (!parse_config(config_filename_canon, &config)) { + exit(1); + } + if (test_config) { + exit(0); + } + + // Ideally we'd like to daemonize only when we've started up all threads etc., + // but daemon() forks, which is not good in multithreaded software, so we'll + // have to do it here. + if (config.daemonize) { + if (daemon(0, 0) == -1) { + log_perror("daemon"); + exit(1); + } + } - CubemapStateProto loaded_state; - int server_sock = -1, old_port = -1; - set deserialized_stream_ids; - map deserialized_inputs; - if (argc == 4 && strcmp(argv[2], "-state") == 0) { - is_reexec = true; +start: + // Open logs as soon as possible. + open_logs(config.log_destinations); + + log(INFO, "Cubemap " SERVER_VERSION " starting."); + if (config.access_log_file.empty()) { + // Create a dummy logger. + access_log = new AccessLogThread(); + } else { + access_log = new AccessLogThread(config.access_log_file); + } + access_log->run(); + + servers = new ServerPool(config.num_servers); - fprintf(stderr, "Deserializing state from previous process... "); - int state_fd = atoi(argv[3]); - loaded_state = read_tempfile(state_fd); + CubemapStateProto loaded_state; + struct timeval serialize_start; + set deserialized_urls; + map deserialized_acceptors; + multimap inputs; // multimap due to older versions without deduplication. + if (state_fd != -1) { + log(INFO, "Deserializing state from previous process..."); + string serialized; + if (!read_tempfile(state_fd, &serialized)) { + exit(1); + } + if (!loaded_state.ParseFromString(serialized)) { + log(ERROR, "Failed deserialization of state."); + exit(1); + } serialize_start.tv_sec = loaded_state.serialize_start_sec(); serialize_start.tv_usec = loaded_state.serialize_start_usec(); // Deserialize the streams. for (int i = 0; i < loaded_state.streams_size(); ++i) { - servers->add_stream_from_serialized(loaded_state.streams(i)); - deserialized_stream_ids.insert(loaded_state.streams(i).stream_id()); + const StreamProto &stream = loaded_state.streams(i); + + vector data_fds; + for (int j = 0; j < stream.data_fds_size(); ++j) { + data_fds.push_back(stream.data_fds(j)); + } + + // Older versions stored the data once in the protobuf instead of + // sending around file descriptors. + if (data_fds.empty() && stream.has_data()) { + data_fds.push_back(make_tempfile(stream.data())); + } + + servers->add_stream_from_serialized(stream, data_fds); + deserialized_urls.insert(stream.url()); } - // Deserialize the inputs. Note that we don't actually add them to any state yet. + // Deserialize the inputs. Note that we don't actually add them to any stream yet. for (int i = 0; i < loaded_state.inputs_size(); ++i) { - deserialized_inputs.insert(make_pair( - loaded_state.inputs(i).stream_id(), - new Input(loaded_state.inputs(i)))); + InputWithRefcount iwr; + iwr.input = create_input(loaded_state.inputs(i)); + iwr.refcount = 0; + inputs.insert(make_pair(loaded_state.inputs(i).url(), iwr)); } - // Deserialize the server socket. - server_sock = loaded_state.server_sock(); - old_port = loaded_state.port(); - - fprintf(stderr, "done.\n"); - } - - // Find all streams in the configuration file, and create them. - create_streams(config, deserialized_stream_ids, &deserialized_inputs); + // Deserialize the acceptors. + for (int i = 0; i < loaded_state.acceptors_size(); ++i) { + deserialized_acceptors.insert(make_pair( + loaded_state.acceptors(i).port(), + new Acceptor(loaded_state.acceptors(i)))); + } - // Open a new server socket if we do not already have one, or if we changed ports. - if (server_sock != -1 && port != old_port) { - fprintf(stderr, "NOTE: Port changed from %d to %d; opening new socket.\n", old_port, port); - close(server_sock); - server_sock = -1; - } - if (server_sock == -1) { - server_sock = create_server_socket(port); + log(INFO, "Deserialization done."); } - // See if the user wants stats. - string stats_file = fetch_config_string(config, "stats_file", PARAMETER_OPTIONAL); - int stats_interval = fetch_config_int(config, "stats_interval", 1, INT_MAX, PARAMETER_OPTIONAL, -1); - if (stats_interval != -1 && stats_file.empty()) { - fprintf(stderr, "WARNING: 'stats_interval' given, but no 'stats_file'. No statistics will be written.\n"); + // Add any new inputs coming from the config. + create_config_inputs(config, &inputs); + + // Find all streams in the configuration file, create them, and connect to the inputs. + create_streams(config, deserialized_urls, &inputs); + vector acceptors = create_acceptors(config, &deserialized_acceptors); + + // Put back the existing clients. It doesn't matter which server we + // allocate them to, so just do round-robin. However, we need to add + // them after the mark pools have been set up. + for (int i = 0; i < loaded_state.clients_size(); ++i) { + servers->add_client_from_serialized(loaded_state.clients(i)); } - + servers->run(); - AcceptorThread acceptor_thread(server_sock); - acceptor_thread.run(); - - vector inputs = create_inputs(config, &deserialized_inputs); - - // All deserialized inputs should now have been taken care of, one way or the other. - assert(deserialized_inputs.empty()); - - if (is_reexec) { - // Put back the existing clients. It doesn't matter which server we - // allocate them to, so just do round-robin. However, we need to add - // them after the mark pools have been set up. - for (int i = 0; i < loaded_state.clients_size(); ++i) { - servers->add_client_from_serialized(loaded_state.clients(i)); + // Now delete all inputs that are longer in use, and start the others. + for (multimap::iterator input_it = inputs.begin(); + input_it != inputs.end(); ) { + if (input_it->second.refcount == 0) { + log(WARNING, "Input '%s' no longer in use, closing.", + input_it->first.c_str()); + input_it->second.input->close_socket(); + delete input_it->second.input; + inputs.erase(input_it++); + } else { + input_it->second.input->run(); + ++input_it; } } // Start writing statistics. StatsThread *stats_thread = NULL; - if (!stats_file.empty()) { - stats_thread = new StatsThread(stats_file, stats_interval); + if (!config.stats_file.empty()) { + stats_thread = new StatsThread(config.stats_file, config.stats_interval); stats_thread->run(); } - signal(SIGHUP, hup); - struct timeval server_start; gettimeofday(&server_start, NULL); - if (is_reexec) { + if (state_fd != -1) { // Measure time from we started deserializing (below) to now, when basically everything // is up and running. This is, in other words, a conservative estimate of how long our // “glitch” period was, not counting of course reconnects if the configuration changed. double glitch_time = server_start.tv_sec - serialize_start.tv_sec + 1e-6 * (server_start.tv_usec - serialize_start.tv_usec); - fprintf(stderr, "Re-exec happened in approx. %.0f ms.\n", glitch_time * 1000.0); + log(INFO, "Re-exec happened in approx. %.0f ms.", glitch_time * 1000.0); } while (!hupped) { @@ -391,25 +442,65 @@ int main(int argc, char **argv) if (stats_thread != NULL) { stats_thread->stop(); + delete stats_thread; } - acceptor_thread.stop(); - for (size_t i = 0; i < inputs.size(); ++i) { - inputs[i]->stop(); + for (size_t i = 0; i < acceptors.size(); ++i) { + acceptors[i]->stop(); + } + for (multimap::iterator input_it = inputs.begin(); + input_it != inputs.end(); + ++input_it) { + input_it->second.input->stop(); } servers->stop(); - fprintf(stderr, "Serializing state and re-execing...\n"); - int state_fd = make_tempfile(collect_state( - serialize_start, server_sock, port, inputs, servers, num_servers)); + CubemapStateProto state; + if (stopped) { + log(INFO, "Shutting down."); + } else { + log(INFO, "Serializing state and re-execing..."); + state = collect_state( + serialize_start, acceptors, inputs, servers); + string serialized; + state.SerializeToString(&serialized); + state_fd = make_tempfile(serialized); + if (state_fd == -1) { + exit(1); + } + } delete servers; + + for (unsigned i = 0; i < mark_pools.size(); ++i) { + delete mark_pools[i]; + } + mark_pools.clear(); + + access_log->stop(); + delete access_log; + shut_down_logging(); + + if (stopped) { + exit(0); + } + + // OK, so the signal was SIGHUP. Check that the new config is okay, then exec the new binary. + if (!dry_run_config(argv0_canon, config_filename_canon)) { + open_logs(config.log_destinations); + log(ERROR, "%s --test-config failed. Restarting old version instead of new.", argv[0]); + hupped = false; + shut_down_logging(); + goto start; + } char buf[16]; sprintf(buf, "%d", state_fd); for ( ;; ) { - execlp(argv[0], argv[0], config_filename.c_str(), "-state", buf, NULL); - perror("execlp"); - fprintf(stderr, "PANIC: re-exec of %s failed. Waiting 0.2 seconds and trying again...\n", argv[0]); + execlp(argv0_canon, argv0_canon, config_filename_canon, "--state", buf, NULL); + open_logs(config.log_destinations); + log_perror("execlp"); + log(ERROR, "re-exec of %s failed. Waiting 0.2 seconds and trying again...", argv0_canon); + shut_down_logging(); usleep(200000); } }