X-Git-Url: https://git.sesse.net/?p=cubemap;a=blobdiff_plain;f=main.cpp;h=0b8825a9da595b6f36bf7247cd75cfd25bdb619a;hp=1369c9faeb8f42394e1afe4b0878f4831592f3a7;hb=029aca758029094d6e9d59e8d88babce94c412e4;hpb=f51b3892514540ff3f08ab052296091f3a6f7a93 diff --git a/main.cpp b/main.cpp index 1369c9f..0b8825a 100644 --- a/main.cpp +++ b/main.cpp @@ -1,65 +1,92 @@ -#include -#include -#include -#include #include +#include #include -#include -#include -#include -#include -#include -#include +#include +#include +#include +#include +#include +#include #include -#include #include -#include -#include -#include -#include -#include -#include +#include +#include +#include #include #include +#include +#include +#include #include "acceptor.h" +#include "accesslog.h" #include "config.h" -#include "markpool.h" -#include "metacube.h" -#include "parse.h" -#include "server.h" -#include "serverpool.h" #include "input.h" +#include "input_stats.h" +#include "log.h" +#include "sa_compare.h" +#include "serverpool.h" +#include "state.pb.h" #include "stats.h" +#include "stream.h" #include "util.h" #include "version.h" -#include "state.pb.h" using namespace std; +AccessLogThread *access_log = NULL; ServerPool *servers = NULL; volatile bool hupped = false; +volatile bool stopped = false; + +typedef pair InputKey; + +namespace { + +struct OrderByConnectionTime { + bool operator() (const ClientProto &a, const ClientProto &b) const { + if (a.connect_time_sec() != b.connect_time_sec()) + return a.connect_time_sec() < b.connect_time_sec(); + return a.connect_time_nsec() < b.connect_time_nsec(); + } +}; + +} // namespace + +struct InputWithRefcount { + Input *input; + int refcount; +}; -void hup(int ignored) +void hup(int signum) { hupped = true; + if (signum == SIGINT) { + stopped = true; + } +} + +void do_nothing(int signum) +{ } -CubemapStateProto collect_state(const timeval &serialize_start, +CubemapStateProto collect_state(const timespec &serialize_start, const vector acceptors, - const vector inputs, + const multimap inputs, ServerPool *servers) { CubemapStateProto state = servers->serialize(); // Fills streams() and clients(). state.set_serialize_start_sec(serialize_start.tv_sec); - state.set_serialize_start_usec(serialize_start.tv_usec); + state.set_serialize_start_usec(serialize_start.tv_nsec / 1000); for (size_t i = 0; i < acceptors.size(); ++i) { state.add_acceptors()->MergeFrom(acceptors[i]->serialize()); } - for (size_t i = 0; i < inputs.size(); ++i) { - state.add_inputs()->MergeFrom(inputs[i]->serialize()); + for (multimap::const_iterator input_it = inputs.begin(); + input_it != inputs.end(); + ++input_it) { + state.add_inputs()->MergeFrom(input_it->second.input->serialize()); } return state; @@ -68,27 +95,28 @@ CubemapStateProto collect_state(const timeval &serialize_start, // Find all port statements in the configuration file, and create acceptors for htem. vector create_acceptors( const Config &config, - map *deserialized_acceptors) + map *deserialized_acceptors) { vector acceptors; for (unsigned i = 0; i < config.acceptors.size(); ++i) { const AcceptorConfig &acceptor_config = config.acceptors[i]; Acceptor *acceptor = NULL; - map::iterator deserialized_acceptor_it = - deserialized_acceptors->find(acceptor_config.port); + map::iterator deserialized_acceptor_it = + deserialized_acceptors->find(acceptor_config.addr); if (deserialized_acceptor_it != deserialized_acceptors->end()) { acceptor = deserialized_acceptor_it->second; deserialized_acceptors->erase(deserialized_acceptor_it); } else { - int server_sock = create_server_socket(acceptor_config.port, TCP_SOCKET); - acceptor = new Acceptor(server_sock, acceptor_config.port); + int server_sock = create_server_socket(acceptor_config.addr, TCP_SOCKET); + acceptor = new Acceptor(server_sock, acceptor_config.addr); } acceptor->run(); acceptors.push_back(acceptor); } // Close all acceptors that are no longer in the configuration file. - for (map::iterator acceptor_it = deserialized_acceptors->begin(); + for (map::iterator + acceptor_it = deserialized_acceptors->begin(); acceptor_it != deserialized_acceptors->end(); ++acceptor_it) { acceptor_it->second->close_socket(); @@ -98,89 +126,143 @@ vector create_acceptors( return acceptors; } +void create_config_input(const string &src, Input::Encoding encoding, multimap *inputs) +{ + if (src.empty()) { + return; + } + InputKey key(src, encoding); + if (inputs->count(key) != 0) { + return; + } + + InputWithRefcount iwr; + iwr.input = create_input(src, encoding); + if (iwr.input == NULL) { + log(ERROR, "did not understand URL '%s' or source encoding was invalid, clients will not get any data.", + src.c_str()); + return; + } + iwr.refcount = 0; + inputs->insert(make_pair(key, iwr)); +} + // Find all streams in the configuration file, and create inputs for them. -vector create_inputs(const Config &config, - map *deserialized_inputs) +void create_config_inputs(const Config &config, multimap *inputs) { - vector inputs; for (unsigned i = 0; i < config.streams.size(); ++i) { const StreamConfig &stream_config = config.streams[i]; - if (stream_config.src.empty()) { - continue; - } - - string stream_id = stream_config.stream_id; - string src = stream_config.src; - - Input *input = NULL; - map::iterator deserialized_input_it = - deserialized_inputs->find(stream_id); - if (deserialized_input_it != deserialized_inputs->end()) { - input = deserialized_input_it->second; - if (input->get_url() != src) { - fprintf(stderr, "INFO: Stream '%s' has changed URL from '%s' to '%s', restarting input.\n", - stream_id.c_str(), input->get_url().c_str(), src.c_str()); - input->close_socket(); - delete input; - input = NULL; - } - deserialized_inputs->erase(deserialized_input_it); - } - if (input == NULL) { - input = create_input(stream_id, src); - if (input == NULL) { - fprintf(stderr, "ERROR: did not understand URL '%s', clients will not get any data.\n", - src.c_str()); - continue; - } + if (stream_config.src != "delete") { + create_config_input(stream_config.src, Input::Encoding(stream_config.src_encoding), inputs); } - input->run(); - inputs.push_back(input); } - return inputs; + for (unsigned i = 0; i < config.udpstreams.size(); ++i) { + const UDPStreamConfig &udpstream_config = config.udpstreams[i]; + create_config_input(udpstream_config.src, Input::INPUT_ENCODING_RAW, inputs); + } } void create_streams(const Config &config, - const set &deserialized_stream_ids, - map *deserialized_inputs) + const set &deserialized_urls, + multimap *inputs) { - vector mark_pools; // FIXME: leak - for (unsigned i = 0; i < config.mark_pools.size(); ++i) { - const MarkPoolConfig &mp_config = config.mark_pools[i]; - mark_pools.push_back(new MarkPool(mp_config.from, mp_config.to)); - } - - set expecting_stream_ids = deserialized_stream_ids; + // HTTP streams. + set expecting_urls = deserialized_urls; for (unsigned i = 0; i < config.streams.size(); ++i) { const StreamConfig &stream_config = config.streams[i]; - if (deserialized_stream_ids.count(stream_config.stream_id) == 0) { - servers->add_stream(stream_config.stream_id, stream_config.backlog_size); + int stream_index; + + expecting_urls.erase(stream_config.url); + + // Special-case deleted streams; they were never deserialized in the first place, + // so just ignore them. + if (stream_config.src == "delete") { + continue; + } + + if (deserialized_urls.count(stream_config.url) == 0) { + stream_index = servers->add_stream(stream_config.url, + stream_config.backlog_size, + stream_config.prebuffering_bytes, + Stream::Encoding(stream_config.encoding), + Stream::Encoding(stream_config.src_encoding)); + } else { + stream_index = servers->lookup_stream_by_url(stream_config.url); + assert(stream_index != -1); + servers->set_backlog_size(stream_index, stream_config.backlog_size); + servers->set_prebuffering_bytes(stream_index, stream_config.prebuffering_bytes); + servers->set_encoding(stream_index, + Stream::Encoding(stream_config.encoding)); + servers->set_src_encoding(stream_index, + Stream::Encoding(stream_config.src_encoding)); } - expecting_stream_ids.erase(stream_config.stream_id); - if (stream_config.mark_pool != -1) { - servers->set_mark_pool(stream_config.stream_id, - mark_pools[stream_config.mark_pool]); + servers->set_pacing_rate(stream_index, stream_config.pacing_rate); + + string src = stream_config.src; + Input::Encoding src_encoding = Input::Encoding(stream_config.src_encoding); + if (!src.empty()) { + multimap::iterator input_it = inputs->find(make_pair(src, src_encoding)); + if (input_it != inputs->end()) { + input_it->second.input->add_destination(stream_index); + ++input_it->second.refcount; + } } } - // Warn about any servers we've lost. - // TODO: Make an option (delete=yes?) to actually shut down streams. - for (set::const_iterator stream_it = expecting_stream_ids.begin(); - stream_it != expecting_stream_ids.end(); + // Warn about any streams servers we've lost. + for (set::const_iterator stream_it = expecting_urls.begin(); + stream_it != expecting_urls.end(); ++stream_it) { - string stream_id = *stream_it; - fprintf(stderr, "WARNING: stream '%s' disappeared from the configuration file.\n", - stream_id.c_str()); - fprintf(stderr, " It will not be deleted, but clients will not get any new inputs.\n"); - if (deserialized_inputs->count(stream_id) != 0) { - delete (*deserialized_inputs)[stream_id]; - deserialized_inputs->erase(stream_id); + string url = *stream_it; + log(WARNING, "stream '%s' disappeared from the configuration file. " + "It will not be deleted, but clients will not get any new inputs. " + "If you really meant to delete it, set src=delete and reload.", + url.c_str()); + } + + // UDP streams. + for (unsigned i = 0; i < config.udpstreams.size(); ++i) { + const UDPStreamConfig &udpstream_config = config.udpstreams[i]; + int stream_index = servers->add_udpstream( + udpstream_config.dst, + udpstream_config.pacing_rate, + udpstream_config.ttl, + udpstream_config.multicast_iface_index); + + string src = udpstream_config.src; + if (!src.empty()) { + multimap::iterator input_it = inputs->find(make_pair(src, Input::INPUT_ENCODING_RAW)); + assert(input_it != inputs->end()); + input_it->second.input->add_destination(stream_index); + ++input_it->second.refcount; } } + + // HTTP gen204 endpoints. + for (unsigned i = 0; i < config.pings.size(); ++i) { + const Gen204Config &ping_config = config.pings[i]; + servers->add_gen204(ping_config.url, ping_config.allow_origin); + } } -bool dry_run_config(const std::string &argv0, const std::string &config_filename) +void open_logs(const vector &log_destinations) +{ + for (size_t i = 0; i < log_destinations.size(); ++i) { + if (log_destinations[i].type == LogConfig::LOG_TYPE_FILE) { + add_log_destination_file(log_destinations[i].filename); + } else if (log_destinations[i].type == LogConfig::LOG_TYPE_CONSOLE) { + add_log_destination_console(); + } else if (log_destinations[i].type == LogConfig::LOG_TYPE_SYSLOG) { + add_log_destination_syslog(); + } else { + assert(false); + } + } + start_logging(); +} + +bool dry_run_config(const string &argv0, const string &config_filename) { char *argv0_copy = strdup(argv0.c_str()); char *config_filename_copy = strdup(config_filename.c_str()); @@ -188,14 +270,14 @@ bool dry_run_config(const std::string &argv0, const std::string &config_filename pid_t pid = fork(); switch (pid) { case -1: - perror("fork()"); + log_perror("fork()"); free(argv0_copy); free(config_filename_copy); return false; case 0: // Child. execlp(argv0_copy, argv0_copy, "--test-config", config_filename_copy, NULL); - perror(argv0_copy); + log_perror(argv0_copy); _exit(1); default: // Parent. @@ -212,15 +294,31 @@ bool dry_run_config(const std::string &argv0, const std::string &config_filename } while (err == -1 && errno == EINTR); if (err == -1) { - perror("waitpid()"); + log_perror("waitpid()"); return false; } return (WIFEXITED(status) && WEXITSTATUS(status) == 0); } +void find_deleted_streams(const Config &config, set *deleted_urls) +{ + for (unsigned i = 0; i < config.streams.size(); ++i) { + const StreamConfig &stream_config = config.streams[i]; + if (stream_config.src == "delete") { + log(INFO, "Deleting stream '%s'.", stream_config.url.c_str()); + deleted_urls->insert(stream_config.url); + } + } +} + int main(int argc, char **argv) { + signal(SIGHUP, hup); + signal(SIGINT, hup); + signal(SIGUSR1, do_nothing); // Used in internal signalling. + signal(SIGPIPE, SIG_IGN); + // Parse options. int state_fd = -1; bool test_config = false; @@ -228,9 +326,10 @@ int main(int argc, char **argv) static const option long_options[] = { { "state", required_argument, 0, 's' }, { "test-config", no_argument, 0, 't' }, + { 0, 0, 0, 0 } }; int option_index = 0; - int c = getopt_long (argc, argv, "s:t", long_options, &option_index); + int c = getopt_long(argc, argv, "s:t", long_options, &option_index); if (c == -1) { break; @@ -243,7 +342,8 @@ int main(int argc, char **argv) test_config = true; break; default: - assert(false); + fprintf(stderr, "Unknown option '%s'\n", argv[option_index]); + exit(1); } } @@ -252,76 +352,167 @@ int main(int argc, char **argv) config_filename = argv[optind++]; } + // Canonicalize argv[0] and config_filename. + char argv0_canon[PATH_MAX]; + char config_filename_canon[PATH_MAX]; + + if (realpath("/proc/self/exe", argv0_canon) == NULL) { + log_perror(argv[0]); + exit(1); + } + if (realpath(config_filename.c_str(), config_filename_canon) == NULL) { + log_perror(config_filename.c_str()); + exit(1); + } + + // Now parse the configuration file. Config config; - if (!parse_config(config_filename, &config)) { + if (!parse_config(config_filename_canon, &config)) { exit(1); } if (test_config) { exit(0); } + + // Ideally we'd like to daemonize only when we've started up all threads etc., + // but daemon() forks, which is not good in multithreaded software, so we'll + // have to do it here. + if (config.daemonize) { + if (daemon(0, 0) == -1) { + log_perror("daemon"); + exit(1); + } + } start: - fprintf(stderr, "\nCubemap " SERVER_VERSION " starting.\n"); + // Open logs as soon as possible. + open_logs(config.log_destinations); + + log(INFO, "Cubemap " SERVER_VERSION " starting."); + if (config.access_log_file.empty()) { + // Create a dummy logger. + access_log = new AccessLogThread(); + } else { + access_log = new AccessLogThread(config.access_log_file); + } + access_log->run(); + servers = new ServerPool(config.num_servers); + // Find all the streams that are to be deleted. + set deleted_urls; + find_deleted_streams(config, &deleted_urls); + CubemapStateProto loaded_state; - struct timeval serialize_start; - set deserialized_stream_ids; - map deserialized_inputs; - map deserialized_acceptors; + timespec serialize_start; + set deserialized_urls; + map deserialized_acceptors; + multimap inputs; // multimap due to older versions without deduplication. if (state_fd != -1) { - fprintf(stderr, "Deserializing state from previous process... "); + log(INFO, "Deserializing state from previous process..."); string serialized; - if (!read_tempfile(state_fd, &serialized)) { + if (!read_tempfile_and_close(state_fd, &serialized)) { exit(1); } if (!loaded_state.ParseFromString(serialized)) { - fprintf(stderr, "ERROR: Failed deserialization of state.\n"); + log(ERROR, "Failed deserialization of state."); exit(1); } serialize_start.tv_sec = loaded_state.serialize_start_sec(); - serialize_start.tv_usec = loaded_state.serialize_start_usec(); + serialize_start.tv_nsec = loaded_state.serialize_start_usec() * 1000ull; // Deserialize the streams. + map stream_headers_for_url; // See below. for (int i = 0; i < loaded_state.streams_size(); ++i) { - servers->add_stream_from_serialized(loaded_state.streams(i)); - deserialized_stream_ids.insert(loaded_state.streams(i).stream_id()); + const StreamProto &stream = loaded_state.streams(i); + + if (deleted_urls.count(stream.url()) != 0) { + // Delete the stream backlogs. + for (int j = 0; j < stream.data_fds_size(); ++j) { + safe_close(stream.data_fds(j)); + } + } else { + vector data_fds; + for (int j = 0; j < stream.data_fds_size(); ++j) { + data_fds.push_back(stream.data_fds(j)); + } + + servers->add_stream_from_serialized(stream, data_fds); + deserialized_urls.insert(stream.url()); + + stream_headers_for_url.insert(make_pair(stream.url(), stream.stream_header())); + } } - // Deserialize the inputs. Note that we don't actually add them to any state yet. + // Deserialize the inputs. Note that we don't actually add them to any stream yet. for (int i = 0; i < loaded_state.inputs_size(); ++i) { - deserialized_inputs.insert(make_pair( - loaded_state.inputs(i).stream_id(), - create_input(loaded_state.inputs(i)))); + InputProto serialized_input = loaded_state.inputs(i); + + InputWithRefcount iwr; + iwr.input = create_input(serialized_input); + iwr.refcount = 0; + + Input::Encoding src_encoding = serialized_input.is_metacube_encoded() ? + Input::INPUT_ENCODING_METACUBE : + Input::INPUT_ENCODING_RAW; + InputKey key(serialized_input.url(), src_encoding); + inputs.insert(make_pair(key, iwr)); } // Deserialize the acceptors. for (int i = 0; i < loaded_state.acceptors_size(); ++i) { + sockaddr_in6 sin6 = extract_address_from_acceptor_proto(loaded_state.acceptors(i)); deserialized_acceptors.insert(make_pair( - loaded_state.acceptors(i).port(), + sin6, new Acceptor(loaded_state.acceptors(i)))); } - fprintf(stderr, "done.\n"); + log(INFO, "Deserialization done."); } - // Find all streams in the configuration file, and create them. - create_streams(config, deserialized_stream_ids, &deserialized_inputs); - - servers->run(); - - vector acceptors = create_acceptors(config, &deserialized_acceptors); - vector inputs = create_inputs(config, &deserialized_inputs); + // Add any new inputs coming from the config. + create_config_inputs(config, &inputs); - // All deserialized inputs should now have been taken care of, one way or the other. - assert(deserialized_inputs.empty()); + // Find all streams in the configuration file, create them, and connect to the inputs. + create_streams(config, deserialized_urls, &inputs); + vector acceptors = create_acceptors(config, &deserialized_acceptors); // Put back the existing clients. It doesn't matter which server we - // allocate them to, so just do round-robin. However, we need to add - // them after the mark pools have been set up. + // allocate them to, so just do round-robin. However, we need to sort them + // by connection time first, since add_client_serialized() expects that. + sort(loaded_state.mutable_clients()->begin(), + loaded_state.mutable_clients()->end(), + OrderByConnectionTime()); for (int i = 0; i < loaded_state.clients_size(); ++i) { - servers->add_client_from_serialized(loaded_state.clients(i)); + if (deleted_urls.count(loaded_state.clients(i).url()) != 0) { + safe_close(loaded_state.clients(i).sock()); + } else { + servers->add_client_from_serialized(loaded_state.clients(i)); + } + } + + servers->run(); + + // Now delete all inputs that are longer in use, and start the others. + for (multimap::iterator input_it = inputs.begin(); + input_it != inputs.end(); ) { + if (input_it->second.refcount == 0) { + if (input_it->first.second == Input::INPUT_ENCODING_RAW) { + log(WARNING, "Raw input '%s' no longer in use, closing.", + input_it->first.first.c_str()); + } else { + assert(input_it->first.second == Input::INPUT_ENCODING_METACUBE); + log(WARNING, "Metacube input '%s' no longer in use, closing.", + input_it->first.first.c_str()); + } + input_it->second.input->close_socket(); + delete input_it->second.input; + inputs.erase(input_it++); + } else { + input_it->second.input->run(); + ++input_it; + } } // Start writing statistics. @@ -331,51 +522,94 @@ start: stats_thread->run(); } - signal(SIGHUP, hup); - - struct timeval server_start; - gettimeofday(&server_start, NULL); + InputStatsThread *input_stats_thread = NULL; + if (!config.input_stats_file.empty()) { + vector inputs_no_refcount; + for (multimap::iterator input_it = inputs.begin(); + input_it != inputs.end(); ++input_it) { + inputs_no_refcount.push_back(input_it->second.input); + } + + input_stats_thread = new InputStatsThread(config.input_stats_file, config.input_stats_interval, inputs_no_refcount); + input_stats_thread->run(); + } + + timespec server_start; + int err = clock_gettime(CLOCK_MONOTONIC, &server_start); + assert(err != -1); if (state_fd != -1) { // Measure time from we started deserializing (below) to now, when basically everything // is up and running. This is, in other words, a conservative estimate of how long our // “glitch” period was, not counting of course reconnects if the configuration changed. double glitch_time = server_start.tv_sec - serialize_start.tv_sec + - 1e-6 * (server_start.tv_usec - serialize_start.tv_usec); - fprintf(stderr, "Re-exec happened in approx. %.0f ms.\n", glitch_time * 1000.0); + 1e-9 * (server_start.tv_nsec - serialize_start.tv_nsec); + log(INFO, "Re-exec happened in approx. %.0f ms.", glitch_time * 1000.0); } + sd_notify(0, "READY=1"); + while (!hupped) { usleep(100000); } + if (stopped) { + sd_notify(0, "STOPPING=1"); + } else { + sd_notify(0, "RELOADING=1"); + } + // OK, we've been HUPed. Time to shut down everything, serialize, and re-exec. - gettimeofday(&serialize_start, NULL); + err = clock_gettime(CLOCK_MONOTONIC, &serialize_start); + assert(err != -1); + if (input_stats_thread != NULL) { + input_stats_thread->stop(); + delete input_stats_thread; + } if (stats_thread != NULL) { stats_thread->stop(); + delete stats_thread; } for (size_t i = 0; i < acceptors.size(); ++i) { acceptors[i]->stop(); } - for (size_t i = 0; i < inputs.size(); ++i) { - inputs[i]->stop(); + for (multimap::iterator input_it = inputs.begin(); + input_it != inputs.end(); + ++input_it) { + input_it->second.input->stop(); } servers->stop(); - fprintf(stderr, "Serializing state and re-execing...\n"); - CubemapStateProto state = collect_state( - serialize_start, acceptors, inputs, servers); - string serialized; - state.SerializeToString(&serialized); - state_fd = make_tempfile(serialized); - if (state_fd == -1) { - exit(1); + CubemapStateProto state; + if (stopped) { + log(INFO, "Shutting down."); + } else { + log(INFO, "Serializing state and re-execing..."); + state = collect_state( + serialize_start, acceptors, inputs, servers); + string serialized; + state.SerializeToString(&serialized); + state_fd = make_tempfile(serialized); + if (state_fd == -1) { + exit(1); + } } delete servers; - if (!dry_run_config(argv[0], config_filename)) { - fprintf(stderr, "ERROR: %s --test-config failed. Restarting old version instead of new.\n", argv[0]); + access_log->stop(); + delete access_log; + shut_down_logging(); + + if (stopped) { + exit(0); + } + + // OK, so the signal was SIGHUP. Check that the new config is okay, then exec the new binary. + if (!dry_run_config(argv0_canon, config_filename_canon)) { + open_logs(config.log_destinations); + log(ERROR, "%s --test-config failed. Restarting old version instead of new.", argv[0]); hupped = false; + shut_down_logging(); goto start; } @@ -383,9 +617,11 @@ start: sprintf(buf, "%d", state_fd); for ( ;; ) { - execlp(argv[0], argv[0], config_filename.c_str(), "--state", buf, NULL); - perror("execlp"); - fprintf(stderr, "PANIC: re-exec of %s failed. Waiting 0.2 seconds and trying again...\n", argv[0]); + execlp(argv0_canon, argv0_canon, config_filename_canon, "--state", buf, NULL); + open_logs(config.log_destinations); + log_perror("execlp"); + log(ERROR, "re-exec of %s failed. Waiting 0.2 seconds and trying again...", argv0_canon); + shut_down_logging(); usleep(200000); } }