X-Git-Url: https://git.sesse.net/?p=cubemap;a=blobdiff_plain;f=main.cpp;h=0218623c0ed6ea670182e54d17094a94f4e0f996;hp=46eddaf834ce7aacfa3c3e518d70618747731246;hb=6889a665614e926437484a556124a5ff60363568;hpb=5cd46e39b4063d94f6dc559ae350beeb8406a8f9 diff --git a/main.cpp b/main.cpp index 46eddaf..0218623 100644 --- a/main.cpp +++ b/main.cpp @@ -20,8 +20,9 @@ #include "accesslog.h" #include "config.h" #include "input.h" +#include "input_stats.h" #include "log.h" -#include "markpool.h" +#include "sa_compare.h" #include "serverpool.h" #include "state.pb.h" #include "stats.h" @@ -33,7 +34,6 @@ using namespace std; AccessLogThread *access_log = NULL; ServerPool *servers = NULL; -vector mark_pools; volatile bool hupped = false; volatile bool stopped = false; @@ -50,6 +50,10 @@ void hup(int signum) } } +void do_nothing(int signum) +{ +} + CubemapStateProto collect_state(const timeval &serialize_start, const vector acceptors, const multimap inputs, @@ -75,27 +79,28 @@ CubemapStateProto collect_state(const timeval &serialize_start, // Find all port statements in the configuration file, and create acceptors for htem. vector create_acceptors( const Config &config, - map *deserialized_acceptors) + map *deserialized_acceptors) { vector acceptors; for (unsigned i = 0; i < config.acceptors.size(); ++i) { const AcceptorConfig &acceptor_config = config.acceptors[i]; Acceptor *acceptor = NULL; - map::iterator deserialized_acceptor_it = - deserialized_acceptors->find(acceptor_config.port); + map::iterator deserialized_acceptor_it = + deserialized_acceptors->find(acceptor_config.addr); if (deserialized_acceptor_it != deserialized_acceptors->end()) { acceptor = deserialized_acceptor_it->second; deserialized_acceptors->erase(deserialized_acceptor_it); } else { - int server_sock = create_server_socket(acceptor_config.port, TCP_SOCKET); - acceptor = new Acceptor(server_sock, acceptor_config.port); + int server_sock = create_server_socket(acceptor_config.addr, TCP_SOCKET); + acceptor = new Acceptor(server_sock, acceptor_config.addr); } acceptor->run(); acceptors.push_back(acceptor); } // Close all acceptors that are no longer in the configuration file. - for (map::iterator acceptor_it = deserialized_acceptors->begin(); + for (map::iterator + acceptor_it = deserialized_acceptors->begin(); acceptor_it != deserialized_acceptors->end(); ++acceptor_it) { acceptor_it->second->close_socket(); @@ -105,78 +110,111 @@ vector create_acceptors( return acceptors; } +void create_config_input(const string &src, multimap *inputs) +{ + if (src.empty()) { + return; + } + if (inputs->count(src) != 0) { + return; + } + + InputWithRefcount iwr; + iwr.input = create_input(src); + if (iwr.input == NULL) { + log(ERROR, "did not understand URL '%s', clients will not get any data.", + src.c_str()); + return; + } + iwr.refcount = 0; + inputs->insert(make_pair(src, iwr)); +} + // Find all streams in the configuration file, and create inputs for them. void create_config_inputs(const Config &config, multimap *inputs) { for (unsigned i = 0; i < config.streams.size(); ++i) { const StreamConfig &stream_config = config.streams[i]; - if (stream_config.src.empty()) { - continue; - } - - string src = stream_config.src; - if (inputs->count(src) != 0) { - continue; + if (stream_config.src != "delete") { + create_config_input(stream_config.src, inputs); } - - InputWithRefcount iwr; - iwr.input = create_input(src); - if (iwr.input == NULL) { - log(ERROR, "did not understand URL '%s', clients will not get any data.", - src.c_str()); - continue; - } - iwr.refcount = 0; - inputs->insert(make_pair(src, iwr)); + } + for (unsigned i = 0; i < config.udpstreams.size(); ++i) { + const UDPStreamConfig &udpstream_config = config.udpstreams[i]; + create_config_input(udpstream_config.src, inputs); } } void create_streams(const Config &config, - const set &deserialized_stream_ids, + const set &deserialized_urls, multimap *inputs) { - for (unsigned i = 0; i < config.mark_pools.size(); ++i) { - const MarkPoolConfig &mp_config = config.mark_pools[i]; - mark_pools.push_back(new MarkPool(mp_config.from, mp_config.to)); - } - - set expecting_stream_ids = deserialized_stream_ids; + // HTTP streams. + set expecting_urls = deserialized_urls; for (unsigned i = 0; i < config.streams.size(); ++i) { const StreamConfig &stream_config = config.streams[i]; - if (deserialized_stream_ids.count(stream_config.stream_id) == 0) { - servers->add_stream(stream_config.stream_id, - stream_config.backlog_size, - Stream::Encoding(stream_config.encoding)); + int stream_index; + + expecting_urls.erase(stream_config.url); + + // Special-case deleted streams; they were never deserialized in the first place, + // so just ignore them. + if (stream_config.src == "delete") { + continue; + } + + if (deserialized_urls.count(stream_config.url) == 0) { + stream_index = servers->add_stream(stream_config.url, + stream_config.backlog_size, + stream_config.prebuffering_bytes, + Stream::Encoding(stream_config.encoding)); } else { - servers->set_backlog_size(stream_config.stream_id, stream_config.backlog_size); - servers->set_encoding(stream_config.stream_id, + stream_index = servers->lookup_stream_by_url(stream_config.url); + assert(stream_index != -1); + servers->set_backlog_size(stream_index, stream_config.backlog_size); + servers->set_encoding(stream_index, Stream::Encoding(stream_config.encoding)); } - expecting_stream_ids.erase(stream_config.stream_id); - if (stream_config.mark_pool != -1) { - servers->set_mark_pool(stream_config.stream_id, - mark_pools[stream_config.mark_pool]); - } + servers->set_pacing_rate(stream_index, stream_config.pacing_rate); string src = stream_config.src; if (!src.empty()) { multimap::iterator input_it = inputs->find(src); - assert(input_it != inputs->end()); - input_it->second.input->add_destination(stream_config.stream_id); - ++input_it->second.refcount; + if (input_it != inputs->end()) { + input_it->second.input->add_destination(stream_index); + ++input_it->second.refcount; + } } } - // Warn about any servers we've lost. - // TODO: Make an option (delete=yes?) to actually shut down streams. - for (set::const_iterator stream_it = expecting_stream_ids.begin(); - stream_it != expecting_stream_ids.end(); + // Warn about any streams servers we've lost. + for (set::const_iterator stream_it = expecting_urls.begin(); + stream_it != expecting_urls.end(); ++stream_it) { - string stream_id = *stream_it; + string url = *stream_it; log(WARNING, "stream '%s' disappeared from the configuration file. " - "It will not be deleted, but clients will not get any new inputs.", - stream_id.c_str()); + "It will not be deleted, but clients will not get any new inputs. " + "If you really meant to delete it, set src=delete and reload.", + url.c_str()); + } + + // UDP streams. + for (unsigned i = 0; i < config.udpstreams.size(); ++i) { + const UDPStreamConfig &udpstream_config = config.udpstreams[i]; + int stream_index = servers->add_udpstream( + udpstream_config.dst, + udpstream_config.pacing_rate, + udpstream_config.ttl, + udpstream_config.multicast_iface_index); + + string src = udpstream_config.src; + if (!src.empty()) { + multimap::iterator input_it = inputs->find(src); + assert(input_it != inputs->end()); + input_it->second.input->add_destination(stream_index); + ++input_it->second.refcount; + } } } @@ -235,10 +273,22 @@ bool dry_run_config(const std::string &argv0, const std::string &config_filename return (WIFEXITED(status) && WEXITSTATUS(status) == 0); } +void find_deleted_streams(const Config &config, set *deleted_urls) +{ + for (unsigned i = 0; i < config.streams.size(); ++i) { + const StreamConfig &stream_config = config.streams[i]; + if (stream_config.src == "delete") { + log(INFO, "Deleting stream '%s'.", stream_config.url.c_str()); + deleted_urls->insert(stream_config.url); + } + } +} + int main(int argc, char **argv) { signal(SIGHUP, hup); signal(SIGINT, hup); + signal(SIGUSR1, do_nothing); // Used in internal signalling. signal(SIGPIPE, SIG_IGN); // Parse options. @@ -278,7 +328,7 @@ int main(int argc, char **argv) char argv0_canon[PATH_MAX]; char config_filename_canon[PATH_MAX]; - if (realpath(argv[0], argv0_canon) == NULL) { + if (realpath("/proc/self/exe", argv0_canon) == NULL) { log_perror(argv[0]); exit(1); } @@ -321,15 +371,19 @@ start: servers = new ServerPool(config.num_servers); + // Find all the streams that are to be deleted. + set deleted_urls; + find_deleted_streams(config, &deleted_urls); + CubemapStateProto loaded_state; struct timeval serialize_start; - set deserialized_stream_ids; - map deserialized_acceptors; + set deserialized_urls; + map deserialized_acceptors; multimap inputs; // multimap due to older versions without deduplication. if (state_fd != -1) { log(INFO, "Deserializing state from previous process..."); string serialized; - if (!read_tempfile(state_fd, &serialized)) { + if (!read_tempfile_and_close(state_fd, &serialized)) { exit(1); } if (!loaded_state.ParseFromString(serialized)) { @@ -341,36 +395,43 @@ start: serialize_start.tv_usec = loaded_state.serialize_start_usec(); // Deserialize the streams. + map stream_headers_for_url; // See below. for (int i = 0; i < loaded_state.streams_size(); ++i) { const StreamProto &stream = loaded_state.streams(i); - vector data_fds; - for (int j = 0; j < stream.data_fds_size(); ++j) { - data_fds.push_back(stream.data_fds(j)); - } - - // Older versions stored the data once in the protobuf instead of - // sending around file descriptors. - if (data_fds.empty() && stream.has_data()) { - data_fds.push_back(make_tempfile(stream.data())); + if (deleted_urls.count(stream.url()) != 0) { + // Delete the stream backlogs. + for (int j = 0; j < stream.data_fds_size(); ++j) { + safe_close(stream.data_fds(j)); + } + } else { + vector data_fds; + for (int j = 0; j < stream.data_fds_size(); ++j) { + data_fds.push_back(stream.data_fds(j)); + } + + servers->add_stream_from_serialized(stream, data_fds); + deserialized_urls.insert(stream.url()); + + stream_headers_for_url.insert(make_pair(stream.url(), stream.stream_header())); } - - servers->add_stream_from_serialized(stream, data_fds); - deserialized_stream_ids.insert(stream.stream_id()); } // Deserialize the inputs. Note that we don't actually add them to any stream yet. for (int i = 0; i < loaded_state.inputs_size(); ++i) { + InputProto serialized_input = loaded_state.inputs(i); + InputWithRefcount iwr; - iwr.input = create_input(loaded_state.inputs(i)); + iwr.input = create_input(serialized_input); iwr.refcount = 0; - inputs.insert(make_pair(loaded_state.inputs(i).url(), iwr)); + inputs.insert(make_pair(serialized_input.url(), iwr)); } // Deserialize the acceptors. for (int i = 0; i < loaded_state.acceptors_size(); ++i) { + sockaddr_in6 sin6 = extract_address_from_acceptor_proto(loaded_state.acceptors(i)); deserialized_acceptors.insert(make_pair( - loaded_state.acceptors(i).port(), + sin6, new Acceptor(loaded_state.acceptors(i)))); } @@ -381,14 +442,17 @@ start: create_config_inputs(config, &inputs); // Find all streams in the configuration file, create them, and connect to the inputs. - create_streams(config, deserialized_stream_ids, &inputs); + create_streams(config, deserialized_urls, &inputs); vector acceptors = create_acceptors(config, &deserialized_acceptors); // Put back the existing clients. It doesn't matter which server we - // allocate them to, so just do round-robin. However, we need to add - // them after the mark pools have been set up. + // allocate them to, so just do round-robin. for (int i = 0; i < loaded_state.clients_size(); ++i) { - servers->add_client_from_serialized(loaded_state.clients(i)); + if (deleted_urls.count(loaded_state.clients(i).url()) != 0) { + safe_close(loaded_state.clients(i).sock()); + } else { + servers->add_client_from_serialized(loaded_state.clients(i)); + } } servers->run(); @@ -415,6 +479,18 @@ start: stats_thread->run(); } + InputStatsThread *input_stats_thread = NULL; + if (!config.input_stats_file.empty()) { + vector inputs_no_refcount; + for (multimap::iterator input_it = inputs.begin(); + input_it != inputs.end(); ++input_it) { + inputs_no_refcount.push_back(input_it->second.input); + } + + input_stats_thread = new InputStatsThread(config.input_stats_file, config.input_stats_interval, inputs_no_refcount); + input_stats_thread->run(); + } + struct timeval server_start; gettimeofday(&server_start, NULL); if (state_fd != -1) { @@ -433,6 +509,10 @@ start: // OK, we've been HUPed. Time to shut down everything, serialize, and re-exec. gettimeofday(&serialize_start, NULL); + if (input_stats_thread != NULL) { + input_stats_thread->stop(); + delete input_stats_thread; + } if (stats_thread != NULL) { stats_thread->stop(); delete stats_thread; @@ -463,11 +543,6 @@ start: } delete servers; - for (unsigned i = 0; i < mark_pools.size(); ++i) { - delete mark_pools[i]; - } - mark_pools.clear(); - access_log->stop(); delete access_log; shut_down_logging();