X-Git-Url: https://git.sesse.net/?p=cubemap;a=blobdiff_plain;f=main.cpp;h=88c49f4513bf60521fe782a387cddaf53999506c;hp=0218623c0ed6ea670182e54d17094a94f4e0f996;hb=061988af511f42da3cd584b4d983177504ddc177;hpb=0e96bbf9ee0fbebd5fe3fba4d186c0e0d73c9a32 diff --git a/main.cpp b/main.cpp index 0218623..88c49f4 100644 --- a/main.cpp +++ b/main.cpp @@ -9,13 +9,17 @@ #include #include #include +#include #include +#include #include #include #include #include #include +#include "tlse.h" + #include "acceptor.h" #include "accesslog.h" #include "config.h" @@ -32,11 +36,42 @@ using namespace std; -AccessLogThread *access_log = NULL; -ServerPool *servers = NULL; +AccessLogThread *access_log = nullptr; +ServerPool *servers = nullptr; volatile bool hupped = false; volatile bool stopped = false; +typedef pair InputKey; + +namespace { + +struct OrderByConnectionTime { + bool operator() (const ClientProto &a, const ClientProto &b) const { + if (a.connect_time_sec() != b.connect_time_sec()) + return a.connect_time_sec() < b.connect_time_sec(); + return a.connect_time_nsec() < b.connect_time_nsec(); + } +}; + +// An arbitrary ordering. +struct AcceptorConfigCompare { + bool operator() (const AcceptorConfig &a, const AcceptorConfig &b) const { + int cmp = a.certificate_chain.compare(b.certificate_chain); + if (cmp != 0) { + return cmp < 0; + } + + cmp = a.private_key.compare(b.private_key); + if (cmp != 0) { + return cmp < 0; + } + + return Sockaddr6Compare()(a.addr, b.addr); + } +}; + +} // namespace + struct InputWithRefcount { Input *input; int refcount; @@ -54,112 +89,106 @@ void do_nothing(int signum) { } -CubemapStateProto collect_state(const timeval &serialize_start, +CubemapStateProto collect_state(const timespec &serialize_start, const vector acceptors, - const multimap inputs, + const multimap inputs, ServerPool *servers) { CubemapStateProto state = servers->serialize(); // Fills streams() and clients(). state.set_serialize_start_sec(serialize_start.tv_sec); - state.set_serialize_start_usec(serialize_start.tv_usec); - - for (size_t i = 0; i < acceptors.size(); ++i) { - state.add_acceptors()->MergeFrom(acceptors[i]->serialize()); + state.set_serialize_start_usec(serialize_start.tv_nsec / 1000); + + for (Acceptor *acceptor : acceptors) { + state.add_acceptors()->MergeFrom(acceptor->serialize()); } - for (multimap::const_iterator input_it = inputs.begin(); - input_it != inputs.end(); - ++input_it) { - state.add_inputs()->MergeFrom(input_it->second.input->serialize()); + for (const auto &key_and_input_with_refcount : inputs) { + state.add_inputs()->MergeFrom(key_and_input_with_refcount.second.input->serialize()); } return state; } -// Find all port statements in the configuration file, and create acceptors for htem. +// Find all port statements in the configuration file, and create acceptors for them. vector create_acceptors( const Config &config, - map *deserialized_acceptors) + map *deserialized_acceptors) { vector acceptors; - for (unsigned i = 0; i < config.acceptors.size(); ++i) { - const AcceptorConfig &acceptor_config = config.acceptors[i]; - Acceptor *acceptor = NULL; - map::iterator deserialized_acceptor_it = - deserialized_acceptors->find(acceptor_config.addr); + for (const AcceptorConfig &acceptor_config : config.acceptors) { + Acceptor *acceptor = nullptr; + const auto deserialized_acceptor_it = deserialized_acceptors->find(acceptor_config); if (deserialized_acceptor_it != deserialized_acceptors->end()) { acceptor = deserialized_acceptor_it->second; deserialized_acceptors->erase(deserialized_acceptor_it); } else { int server_sock = create_server_socket(acceptor_config.addr, TCP_SOCKET); - acceptor = new Acceptor(server_sock, acceptor_config.addr); + acceptor = new Acceptor(server_sock, acceptor_config.addr, + acceptor_config.certificate_chain, + acceptor_config.private_key); } acceptor->run(); acceptors.push_back(acceptor); } // Close all acceptors that are no longer in the configuration file. - for (map::iterator - acceptor_it = deserialized_acceptors->begin(); - acceptor_it != deserialized_acceptors->end(); - ++acceptor_it) { - acceptor_it->second->close_socket(); - delete acceptor_it->second; + for (auto &config_and_acceptor : *deserialized_acceptors) { + config_and_acceptor.second->close_socket(); + delete config_and_acceptor.second; } return acceptors; } -void create_config_input(const string &src, multimap *inputs) +void create_config_input(const string &src, Input::Encoding encoding, multimap *inputs) { if (src.empty()) { return; } - if (inputs->count(src) != 0) { + InputKey key(src, encoding); + if (inputs->count(key) != 0) { return; } InputWithRefcount iwr; - iwr.input = create_input(src); - if (iwr.input == NULL) { - log(ERROR, "did not understand URL '%s', clients will not get any data.", + iwr.input = create_input(src, encoding); + if (iwr.input == nullptr) { + log(ERROR, "did not understand URL '%s' or source encoding was invalid, clients will not get any data.", src.c_str()); return; } iwr.refcount = 0; - inputs->insert(make_pair(src, iwr)); + inputs->insert(make_pair(key, iwr)); } // Find all streams in the configuration file, and create inputs for them. -void create_config_inputs(const Config &config, multimap *inputs) +void create_config_inputs(const Config &config, multimap *inputs) { - for (unsigned i = 0; i < config.streams.size(); ++i) { - const StreamConfig &stream_config = config.streams[i]; - if (stream_config.src != "delete") { - create_config_input(stream_config.src, inputs); + for (const StreamConfig &stream_config : config.streams) { + if (stream_config.src == "delete") { + // Ignored for pre-1.4.0 configuration compatibility. + continue; } + create_config_input(stream_config.src, Input::Encoding(stream_config.src_encoding), inputs); } - for (unsigned i = 0; i < config.udpstreams.size(); ++i) { - const UDPStreamConfig &udpstream_config = config.udpstreams[i]; - create_config_input(udpstream_config.src, inputs); + for (const UDPStreamConfig &udpstream_config : config.udpstreams) { + create_config_input(udpstream_config.src, Input::INPUT_ENCODING_RAW, inputs); } } void create_streams(const Config &config, const set &deserialized_urls, - multimap *inputs) + multimap *inputs) { // HTTP streams. set expecting_urls = deserialized_urls; - for (unsigned i = 0; i < config.streams.size(); ++i) { - const StreamConfig &stream_config = config.streams[i]; + for (const StreamConfig &stream_config : config.streams) { int stream_index; expecting_urls.erase(stream_config.url); - // Special-case deleted streams; they were never deserialized in the first place, - // so just ignore them. if (stream_config.src == "delete") { + // Ignored for pre-1.4.0 configuration compatibility. continue; } @@ -167,20 +196,25 @@ void create_streams(const Config &config, stream_index = servers->add_stream(stream_config.url, stream_config.backlog_size, stream_config.prebuffering_bytes, - Stream::Encoding(stream_config.encoding)); + Stream::Encoding(stream_config.encoding), + Stream::Encoding(stream_config.src_encoding)); } else { stream_index = servers->lookup_stream_by_url(stream_config.url); assert(stream_index != -1); servers->set_backlog_size(stream_index, stream_config.backlog_size); + servers->set_prebuffering_bytes(stream_index, stream_config.prebuffering_bytes); servers->set_encoding(stream_index, Stream::Encoding(stream_config.encoding)); + servers->set_src_encoding(stream_index, + Stream::Encoding(stream_config.src_encoding)); } servers->set_pacing_rate(stream_index, stream_config.pacing_rate); string src = stream_config.src; + Input::Encoding src_encoding = Input::Encoding(stream_config.src_encoding); if (!src.empty()) { - multimap::iterator input_it = inputs->find(src); + const auto input_it = inputs->find(make_pair(src, src_encoding)); if (input_it != inputs->end()) { input_it->second.input->add_destination(stream_index); ++input_it->second.refcount; @@ -189,10 +223,7 @@ void create_streams(const Config &config, } // Warn about any streams servers we've lost. - for (set::const_iterator stream_it = expecting_urls.begin(); - stream_it != expecting_urls.end(); - ++stream_it) { - string url = *stream_it; + for (const string &url : expecting_urls) { log(WARNING, "stream '%s' disappeared from the configuration file. " "It will not be deleted, but clients will not get any new inputs. " "If you really meant to delete it, set src=delete and reload.", @@ -200,8 +231,7 @@ void create_streams(const Config &config, } // UDP streams. - for (unsigned i = 0; i < config.udpstreams.size(); ++i) { - const UDPStreamConfig &udpstream_config = config.udpstreams[i]; + for (const UDPStreamConfig &udpstream_config : config.udpstreams) { int stream_index = servers->add_udpstream( udpstream_config.dst, udpstream_config.pacing_rate, @@ -210,22 +240,27 @@ void create_streams(const Config &config, string src = udpstream_config.src; if (!src.empty()) { - multimap::iterator input_it = inputs->find(src); + auto input_it = inputs->find(make_pair(src, Input::INPUT_ENCODING_RAW)); assert(input_it != inputs->end()); input_it->second.input->add_destination(stream_index); ++input_it->second.refcount; } } + + // HTTP gen204 endpoints. + for (const Gen204Config &ping_config : config.pings) { + servers->add_gen204(ping_config.url, ping_config.allow_origin); + } } void open_logs(const vector &log_destinations) { - for (size_t i = 0; i < log_destinations.size(); ++i) { - if (log_destinations[i].type == LogConfig::LOG_TYPE_FILE) { - add_log_destination_file(log_destinations[i].filename); - } else if (log_destinations[i].type == LogConfig::LOG_TYPE_CONSOLE) { + for (const LogConfig &log_destination : log_destinations) { + if (log_destination.type == LogConfig::LOG_TYPE_FILE) { + add_log_destination_file(log_destination.filename); + } else if (log_destination.type == LogConfig::LOG_TYPE_CONSOLE) { add_log_destination_console(); - } else if (log_destinations[i].type == LogConfig::LOG_TYPE_SYSLOG) { + } else if (log_destination.type == LogConfig::LOG_TYPE_SYSLOG) { add_log_destination_syslog(); } else { assert(false); @@ -234,7 +269,7 @@ void open_logs(const vector &log_destinations) start_logging(); } -bool dry_run_config(const std::string &argv0, const std::string &config_filename) +bool dry_run_config(const string &argv0, const string &config_filename) { char *argv0_copy = strdup(argv0.c_str()); char *config_filename_copy = strdup(config_filename.c_str()); @@ -248,7 +283,7 @@ bool dry_run_config(const std::string &argv0, const std::string &config_filename return false; case 0: // Child. - execlp(argv0_copy, argv0_copy, "--test-config", config_filename_copy, NULL); + execlp(argv0_copy, argv0_copy, "--test-config", config_filename_copy, nullptr); log_perror(argv0_copy); _exit(1); default: @@ -273,13 +308,16 @@ bool dry_run_config(const std::string &argv0, const std::string &config_filename return (WIFEXITED(status) && WEXITSTATUS(status) == 0); } -void find_deleted_streams(const Config &config, set *deleted_urls) +void find_all_streams(const Config &config, set *all_urls) { - for (unsigned i = 0; i < config.streams.size(); ++i) { - const StreamConfig &stream_config = config.streams[i]; + for (const StreamConfig &stream_config : config.streams) { if (stream_config.src == "delete") { - log(INFO, "Deleting stream '%s'.", stream_config.url.c_str()); - deleted_urls->insert(stream_config.url); + log(WARNING, "stream '%s' has src=delete; ignoring it. Since Cubemap 1.4.0, you do not " + "need to set src=delete to delete streams anymore; just delete them from " + "the configuration file.", + stream_config.url.c_str()); + } else { + all_urls->insert(stream_config.url); } } } @@ -290,6 +328,8 @@ int main(int argc, char **argv) signal(SIGINT, hup); signal(SIGUSR1, do_nothing); // Used in internal signalling. signal(SIGPIPE, SIG_IGN); + + tls_init(); // Parse options. int state_fd = -1; @@ -328,11 +368,11 @@ int main(int argc, char **argv) char argv0_canon[PATH_MAX]; char config_filename_canon[PATH_MAX]; - if (realpath("/proc/self/exe", argv0_canon) == NULL) { + if (realpath("/proc/self/exe", argv0_canon) == nullptr) { log_perror(argv[0]); exit(1); } - if (realpath(config_filename.c_str(), config_filename_canon) == NULL) { + if (realpath(config_filename.c_str(), config_filename_canon) == nullptr) { log_perror(config_filename.c_str()); exit(1); } @@ -371,15 +411,15 @@ start: servers = new ServerPool(config.num_servers); - // Find all the streams that are to be deleted. - set deleted_urls; - find_deleted_streams(config, &deleted_urls); + // Find all the streams that are to be kept. + set all_urls; + find_all_streams(config, &all_urls); CubemapStateProto loaded_state; - struct timeval serialize_start; + timespec serialize_start; set deserialized_urls; - map deserialized_acceptors; - multimap inputs; // multimap due to older versions without deduplication. + map deserialized_acceptors; + multimap inputs; // multimap due to older versions without deduplication. if (state_fd != -1) { log(INFO, "Deserializing state from previous process..."); string serialized; @@ -392,22 +432,21 @@ start: } serialize_start.tv_sec = loaded_state.serialize_start_sec(); - serialize_start.tv_usec = loaded_state.serialize_start_usec(); + serialize_start.tv_nsec = loaded_state.serialize_start_usec() * 1000ull; // Deserialize the streams. map stream_headers_for_url; // See below. - for (int i = 0; i < loaded_state.streams_size(); ++i) { - const StreamProto &stream = loaded_state.streams(i); - - if (deleted_urls.count(stream.url()) != 0) { + for (const StreamProto &stream : loaded_state.streams()) { + if (all_urls.count(stream.url()) == 0) { // Delete the stream backlogs. - for (int j = 0; j < stream.data_fds_size(); ++j) { - safe_close(stream.data_fds(j)); + log(INFO, "Deleting stream '%s'.", stream.url().c_str()); + for (const int fd : stream.data_fds()) { + safe_close(fd); } } else { vector data_fds; - for (int j = 0; j < stream.data_fds_size(); ++j) { - data_fds.push_back(stream.data_fds(j)); + for (const int fd : stream.data_fds()) { + data_fds.push_back(fd); } servers->add_stream_from_serialized(stream, data_fds); @@ -418,21 +457,27 @@ start: } // Deserialize the inputs. Note that we don't actually add them to any stream yet. - for (int i = 0; i < loaded_state.inputs_size(); ++i) { - InputProto serialized_input = loaded_state.inputs(i); - + for (const InputProto &serialized_input : loaded_state.inputs()) { InputWithRefcount iwr; iwr.input = create_input(serialized_input); iwr.refcount = 0; - inputs.insert(make_pair(serialized_input.url(), iwr)); + + Input::Encoding src_encoding = serialized_input.is_metacube_encoded() ? + Input::INPUT_ENCODING_METACUBE : + Input::INPUT_ENCODING_RAW; + InputKey key(serialized_input.url(), src_encoding); + inputs.insert(make_pair(key, iwr)); } // Deserialize the acceptors. - for (int i = 0; i < loaded_state.acceptors_size(); ++i) { - sockaddr_in6 sin6 = extract_address_from_acceptor_proto(loaded_state.acceptors(i)); + for (const AcceptorProto &serialized_acceptor : loaded_state.acceptors()) { + AcceptorConfig config; + config.addr = extract_address_from_acceptor_proto(serialized_acceptor); + config.certificate_chain = serialized_acceptor.certificate_chain(); + config.private_key = serialized_acceptor.private_key(); deserialized_acceptors.insert(make_pair( - sin6, - new Acceptor(loaded_state.acceptors(i)))); + config, + new Acceptor(serialized_acceptor))); } log(INFO, "Deserialization done."); @@ -444,25 +489,49 @@ start: // Find all streams in the configuration file, create them, and connect to the inputs. create_streams(config, deserialized_urls, &inputs); vector acceptors = create_acceptors(config, &deserialized_acceptors); + + // Make all the servers create TLS contexts for every TLS keypair we have. + for (Acceptor *acceptor : acceptors) { + if (acceptor->is_tls()) { + servers->create_tls_context_for_acceptor(acceptor); + } + } + + // Allocate strings for the short responses. + vector> short_response_pool; + for (const ShortResponsePool &str : loaded_state.short_response_pool()) { + short_response_pool.emplace_back(new string(str.header_or_short_response())); + } // Put back the existing clients. It doesn't matter which server we - // allocate them to, so just do round-robin. + // allocate them to, so just do round-robin. However, we need to sort them + // by connection time first, since add_client_serialized() expects that. + sort(loaded_state.mutable_clients()->begin(), + loaded_state.mutable_clients()->end(), + OrderByConnectionTime()); for (int i = 0; i < loaded_state.clients_size(); ++i) { - if (deleted_urls.count(loaded_state.clients(i).url()) != 0) { + if (all_urls.count(loaded_state.clients(i).url()) == 0) { safe_close(loaded_state.clients(i).sock()); } else { - servers->add_client_from_serialized(loaded_state.clients(i)); + servers->add_client_from_serialized(loaded_state.clients(i), short_response_pool); } } + short_response_pool.clear(); // No longer needed; the clients have their own refcounts now. + servers->run(); // Now delete all inputs that are longer in use, and start the others. - for (multimap::iterator input_it = inputs.begin(); - input_it != inputs.end(); ) { + for (auto input_it = inputs.begin(); input_it != inputs.end(); ) { if (input_it->second.refcount == 0) { - log(WARNING, "Input '%s' no longer in use, closing.", - input_it->first.c_str()); + if (input_it->first.second == Input::INPUT_ENCODING_RAW) { + log(WARNING, "Raw input '%s' no longer in use, closing.", + input_it->first.first.c_str()); + } else { + assert(input_it->first.second == Input::INPUT_ENCODING_METACUBE); + log(WARNING, "Metacube input '%s' no longer in use, closing.", + input_it->first.first.c_str()); + } input_it->second.input->close_socket(); delete input_it->second.input; inputs.erase(input_it++); @@ -473,57 +542,64 @@ start: } // Start writing statistics. - StatsThread *stats_thread = NULL; + unique_ptr stats_thread; if (!config.stats_file.empty()) { - stats_thread = new StatsThread(config.stats_file, config.stats_interval); + stats_thread.reset(new StatsThread(config.stats_file, config.stats_interval)); stats_thread->run(); } - InputStatsThread *input_stats_thread = NULL; + unique_ptr input_stats_thread; if (!config.input_stats_file.empty()) { vector inputs_no_refcount; - for (multimap::iterator input_it = inputs.begin(); - input_it != inputs.end(); ++input_it) { - inputs_no_refcount.push_back(input_it->second.input); + for (const auto &key_and_input_with_refcount : inputs) { + inputs_no_refcount.push_back(key_and_input_with_refcount.second.input); } - input_stats_thread = new InputStatsThread(config.input_stats_file, config.input_stats_interval, inputs_no_refcount); + input_stats_thread.reset(new InputStatsThread(config.input_stats_file, config.input_stats_interval, inputs_no_refcount)); input_stats_thread->run(); } - struct timeval server_start; - gettimeofday(&server_start, NULL); + timespec server_start; + int err = clock_gettime(CLOCK_MONOTONIC, &server_start); + assert(err != -1); if (state_fd != -1) { // Measure time from we started deserializing (below) to now, when basically everything // is up and running. This is, in other words, a conservative estimate of how long our // “glitch” period was, not counting of course reconnects if the configuration changed. double glitch_time = server_start.tv_sec - serialize_start.tv_sec + - 1e-6 * (server_start.tv_usec - serialize_start.tv_usec); + 1e-9 * (server_start.tv_nsec - serialize_start.tv_nsec); log(INFO, "Re-exec happened in approx. %.0f ms.", glitch_time * 1000.0); } + sd_notify(0, "READY=1"); + while (!hupped) { usleep(100000); } + if (stopped) { + sd_notify(0, "STOPPING=1"); + } else { + sd_notify(0, "RELOADING=1"); + } + // OK, we've been HUPed. Time to shut down everything, serialize, and re-exec. - gettimeofday(&serialize_start, NULL); + err = clock_gettime(CLOCK_MONOTONIC, &serialize_start); + assert(err != -1); - if (input_stats_thread != NULL) { + if (input_stats_thread != nullptr) { input_stats_thread->stop(); - delete input_stats_thread; + input_stats_thread.reset(); } - if (stats_thread != NULL) { + if (stats_thread != nullptr) { stats_thread->stop(); - delete stats_thread; + stats_thread.reset(); } - for (size_t i = 0; i < acceptors.size(); ++i) { - acceptors[i]->stop(); + for (Acceptor *acceptor : acceptors) { + acceptor->stop(); } - for (multimap::iterator input_it = inputs.begin(); - input_it != inputs.end(); - ++input_it) { - input_it->second.input->stop(); + for (const auto &key_and_input_with_refcount : inputs) { + key_and_input_with_refcount.second.input->stop(); } servers->stop(); @@ -564,7 +640,7 @@ start: sprintf(buf, "%d", state_fd); for ( ;; ) { - execlp(argv0_canon, argv0_canon, config_filename_canon, "--state", buf, NULL); + execlp(argv0_canon, argv0_canon, config_filename_canon, "--state", buf, nullptr); open_logs(config.log_destinations); log_perror("execlp"); log(ERROR, "re-exec of %s failed. Waiting 0.2 seconds and trying again...", argv0_canon);