X-Git-Url: https://git.sesse.net/?p=cubemap;a=blobdiff_plain;f=main.cpp;h=737f0c3573acea88fa7d9f0585aee69d95ec0838;hp=c208b2d19f381146df11e3322f86a4c34302862a;hb=ed218ece51e5be77fd8bd8f014b5f05708e4be5b;hpb=26fe3ab755034ea3be8321ec0af548670f8c3bd8 diff --git a/main.cpp b/main.cpp index c208b2d..737f0c3 100644 --- a/main.cpp +++ b/main.cpp @@ -15,6 +15,7 @@ #include #include #include +#include #include #include @@ -36,8 +37,8 @@ using namespace std; -AccessLogThread *access_log = NULL; -ServerPool *servers = NULL; +AccessLogThread *access_log = nullptr; +ServerPool *servers = nullptr; volatile bool hupped = false; volatile bool stopped = false; @@ -116,7 +117,7 @@ vector create_acceptors( { vector acceptors; for (const AcceptorConfig &acceptor_config : config.acceptors) { - Acceptor *acceptor = NULL; + Acceptor *acceptor = nullptr; const auto deserialized_acceptor_it = deserialized_acceptors->find(acceptor_config); if (deserialized_acceptor_it != deserialized_acceptors->end()) { acceptor = deserialized_acceptor_it->second; @@ -152,7 +153,7 @@ void create_config_input(const string &src, Input::Encoding encoding, multimap *inputs) { for (const StreamConfig &stream_config : config.streams) { - if (stream_config.src != "delete") { - create_config_input(stream_config.src, Input::Encoding(stream_config.src_encoding), inputs); + if (stream_config.src == "delete") { + // Ignored for pre-1.4.0 configuration compatibility. + continue; } + create_config_input(stream_config.src, Input::Encoding(stream_config.src_encoding), inputs); } for (const UDPStreamConfig &udpstream_config : config.udpstreams) { create_config_input(udpstream_config.src, Input::INPUT_ENCODING_RAW, inputs); @@ -185,27 +188,36 @@ void create_streams(const Config &config, expecting_urls.erase(stream_config.url); - // Special-case deleted streams; they were never deserialized in the first place, - // so just ignore them. if (stream_config.src == "delete") { + // Ignored for pre-1.4.0 configuration compatibility. continue; } if (deserialized_urls.count(stream_config.url) == 0) { stream_index = servers->add_stream(stream_config.url, + stream_config.hls_url, stream_config.backlog_size, stream_config.prebuffering_bytes, Stream::Encoding(stream_config.encoding), - Stream::Encoding(stream_config.src_encoding)); + Stream::Encoding(stream_config.src_encoding), + stream_config.hls_frag_duration, + stream_config.hls_backlog_margin, + stream_config.allow_origin); } else { stream_index = servers->lookup_stream_by_url(stream_config.url); assert(stream_index != -1); servers->set_backlog_size(stream_index, stream_config.backlog_size); + if (!stream_config.hls_url.empty()) { + servers->register_hls_url(stream_index, stream_config.hls_url); + } servers->set_prebuffering_bytes(stream_index, stream_config.prebuffering_bytes); servers->set_encoding(stream_index, Stream::Encoding(stream_config.encoding)); servers->set_src_encoding(stream_index, Stream::Encoding(stream_config.src_encoding)); + servers->set_hls_frag_duration(stream_index, stream_config.hls_frag_duration); + servers->set_hls_backlog_margin(stream_index, stream_config.hls_backlog_margin); + servers->set_allow_origin(stream_index, stream_config.allow_origin); } servers->set_pacing_rate(stream_index, stream_config.pacing_rate); @@ -282,7 +294,7 @@ bool dry_run_config(const string &argv0, const string &config_filename) return false; case 0: // Child. - execlp(argv0_copy, argv0_copy, "--test-config", config_filename_copy, NULL); + execlp(argv0_copy, argv0_copy, "--test-config", config_filename_copy, nullptr); log_perror(argv0_copy); _exit(1); default: @@ -307,12 +319,16 @@ bool dry_run_config(const string &argv0, const string &config_filename) return (WIFEXITED(status) && WEXITSTATUS(status) == 0); } -void find_deleted_streams(const Config &config, set *deleted_urls) +void find_all_streams(const Config &config, set *all_urls) { for (const StreamConfig &stream_config : config.streams) { if (stream_config.src == "delete") { - log(INFO, "Deleting stream '%s'.", stream_config.url.c_str()); - deleted_urls->insert(stream_config.url); + log(WARNING, "stream '%s' has src=delete; ignoring it. Since Cubemap 1.4.0, you do not " + "need to set src=delete to delete streams anymore; just delete them from " + "the configuration file.", + stream_config.url.c_str()); + } else { + all_urls->insert(stream_config.url); } } } @@ -363,11 +379,11 @@ int main(int argc, char **argv) char argv0_canon[PATH_MAX]; char config_filename_canon[PATH_MAX]; - if (realpath("/proc/self/exe", argv0_canon) == NULL) { + if (realpath("/proc/self/exe", argv0_canon) == nullptr) { log_perror(argv[0]); exit(1); } - if (realpath(config_filename.c_str(), config_filename_canon) == NULL) { + if (realpath(config_filename.c_str(), config_filename_canon) == nullptr) { log_perror(config_filename.c_str()); exit(1); } @@ -406,9 +422,9 @@ start: servers = new ServerPool(config.num_servers); - // Find all the streams that are to be deleted. - set deleted_urls; - find_deleted_streams(config, &deleted_urls); + // Find all the streams that are to be kept. + set all_urls; + find_all_streams(config, &all_urls); CubemapStateProto loaded_state; timespec serialize_start; @@ -430,10 +446,11 @@ start: serialize_start.tv_nsec = loaded_state.serialize_start_usec() * 1000ull; // Deserialize the streams. - map stream_headers_for_url; // See below. + unordered_map stream_headers_for_url; // See below. for (const StreamProto &stream : loaded_state.streams()) { - if (deleted_urls.count(stream.url()) != 0) { + if (all_urls.count(stream.url()) == 0) { // Delete the stream backlogs. + log(INFO, "Deleting stream '%s'.", stream.url().c_str()); for (const int fd : stream.data_fds()) { safe_close(fd); } @@ -490,6 +507,12 @@ start: servers->create_tls_context_for_acceptor(acceptor); } } + + // Allocate strings for the short responses. + vector> short_response_pool; + for (const ShortResponsePool &str : loaded_state.short_response_pool()) { + short_response_pool.emplace_back(new string(str.header_or_short_response())); + } // Put back the existing clients. It doesn't matter which server we // allocate them to, so just do round-robin. However, we need to sort them @@ -498,13 +521,26 @@ start: loaded_state.mutable_clients()->end(), OrderByConnectionTime()); for (int i = 0; i < loaded_state.clients_size(); ++i) { - if (deleted_urls.count(loaded_state.clients(i).url()) != 0) { + if (!loaded_state.clients(i).url().empty() && + all_urls.count(loaded_state.clients(i).url()) == 0) { + // Belongs to a dead stream (not keepalive), so we just have to close. safe_close(loaded_state.clients(i).sock()); } else { - servers->add_client_from_serialized(loaded_state.clients(i)); + servers->add_client_from_serialized(loaded_state.clients(i), short_response_pool); } } + short_response_pool.clear(); // No longer needed; the clients have their own refcounts now. + + // Put back the HLS zombies. There's no really good allocation here + // except round-robin; it would be marginally more efficient to match it + // to the client (since that would have them deleted immediately when + // the client requests the next fragment, instead of being later weeded + // out during statistics collection), but it's not a big deal. + for (const HLSZombieProto &zombie_proto : loaded_state.hls_zombies()) { + servers->add_hls_zombie_from_serialized(zombie_proto); + } + servers->run(); // Now delete all inputs that are longer in use, and start the others. @@ -573,11 +609,11 @@ start: err = clock_gettime(CLOCK_MONOTONIC, &serialize_start); assert(err != -1); - if (input_stats_thread != NULL) { + if (input_stats_thread != nullptr) { input_stats_thread->stop(); input_stats_thread.reset(); } - if (stats_thread != NULL) { + if (stats_thread != nullptr) { stats_thread->stop(); stats_thread.reset(); } @@ -626,7 +662,7 @@ start: sprintf(buf, "%d", state_fd); for ( ;; ) { - execlp(argv0_canon, argv0_canon, config_filename_canon, "--state", buf, NULL); + execlp(argv0_canon, argv0_canon, config_filename_canon, "--state", buf, nullptr); open_logs(config.log_destinations); log_perror("execlp"); log(ERROR, "re-exec of %s failed. Waiting 0.2 seconds and trying again...", argv0_canon);