]> git.sesse.net Git - cubemap/blobdiff - main.cpp
Keep information about HLS downloads around for some time afterwards.
[cubemap] / main.cpp
index 8ac7802bb30454f690822e1d7761ed06b8080bf9..737f0c3573acea88fa7d9f0585aee69d95ec0838 100644 (file)
--- a/main.cpp
+++ b/main.cpp
@@ -1,7 +1,7 @@
 #include <assert.h>
 #include <errno.h>
 #include <getopt.h>
-#include <google/protobuf/stubs/common.h>
+#include <limits.h>
 #include <signal.h>
 #include <stddef.h>
 #include <stdio.h>
 #include <string.h>
 #include <sys/time.h>
 #include <sys/wait.h>
+#include <systemd/sd-daemon.h>
 #include <unistd.h>
+#include <algorithm>
 #include <map>
 #include <set>
 #include <string>
+#include <unordered_map>
 #include <utility>
 #include <vector>
 
+#include "tlse.h"
+
 #include "acceptor.h"
+#include "accesslog.h"
 #include "config.h"
 #include "input.h"
-#include "markpool.h"
+#include "input_stats.h"
+#include "log.h"
+#include "sa_compare.h"
 #include "serverpool.h"
 #include "state.pb.h"
 #include "stats.h"
+#include "stream.h"
 #include "util.h"
 #include "version.h"
 
 using namespace std;
 
-ServerPool *servers = NULL;
+AccessLogThread *access_log = nullptr;
+ServerPool *servers = nullptr;
 volatile bool hupped = false;
+volatile bool stopped = false;
+
+typedef pair<string, Input::Encoding> InputKey;
+
+namespace {
+
+struct OrderByConnectionTime {
+       bool operator() (const ClientProto &a, const ClientProto &b) const {
+               if (a.connect_time_sec() != b.connect_time_sec())
+                       return a.connect_time_sec() < b.connect_time_sec();
+               return a.connect_time_nsec() < b.connect_time_nsec();
+       }
+};
+
+// An arbitrary ordering.
+struct AcceptorConfigCompare {
+       bool operator() (const AcceptorConfig &a, const AcceptorConfig &b) const {
+               int cmp = a.certificate_chain.compare(b.certificate_chain);
+               if (cmp != 0) {
+                       return cmp < 0;
+               }
+
+               cmp = a.private_key.compare(b.private_key);
+               if (cmp != 0) {
+                       return cmp < 0;
+               }
+
+               return Sockaddr6Compare()(a.addr, b.addr);
+       }
+};
+
+}  // namespace
 
-void hup(int ignored)
+struct InputWithRefcount {
+       Input *input;
+       int refcount;
+};
+
+void hup(int signum)
 {
        hupped = true;
+       if (signum == SIGINT) {
+               stopped = true;
+       }
+}
+
+void do_nothing(int signum)
+{
 }
 
-CubemapStateProto collect_state(const timeval &serialize_start,
+CubemapStateProto collect_state(const timespec &serialize_start,
                                 const vector<Acceptor *> acceptors,
-                                const vector<Input *> inputs,
+                                const multimap<InputKey, InputWithRefcount> inputs,
                                 ServerPool *servers)
 {
        CubemapStateProto state = servers->serialize();  // Fills streams() and clients().
        state.set_serialize_start_sec(serialize_start.tv_sec);
-       state.set_serialize_start_usec(serialize_start.tv_usec);
-       
-       for (size_t i = 0; i < acceptors.size(); ++i) {
-               state.add_acceptors()->MergeFrom(acceptors[i]->serialize());
+       state.set_serialize_start_usec(serialize_start.tv_nsec / 1000);
+
+       for (Acceptor *acceptor : acceptors) {  
+               state.add_acceptors()->MergeFrom(acceptor->serialize());
        }
 
-       for (size_t i = 0; i < inputs.size(); ++i) {
-               state.add_inputs()->MergeFrom(inputs[i]->serialize());
+       for (const auto &key_and_input_with_refcount : inputs) {
+               state.add_inputs()->MergeFrom(key_and_input_with_refcount.second.input->serialize());
        }
 
        return state;
 }
 
-// Find all port statements in the configuration file, and create acceptors for htem.
+// Find all port statements in the configuration file, and create acceptors for them.
 vector<Acceptor *> create_acceptors(
        const Config &config,
-       map<int, Acceptor *> *deserialized_acceptors)
+       map<AcceptorConfig, Acceptor *, AcceptorConfigCompare> *deserialized_acceptors)
 {
        vector<Acceptor *> acceptors;
-       for (unsigned i = 0; i < config.acceptors.size(); ++i) {
-               const AcceptorConfig &acceptor_config = config.acceptors[i];
-               Acceptor *acceptor = NULL;
-               map<int, Acceptor *>::iterator deserialized_acceptor_it =
-                       deserialized_acceptors->find(acceptor_config.port);
+       for (const AcceptorConfig &acceptor_config : config.acceptors) {
+               Acceptor *acceptor = nullptr;
+               const auto deserialized_acceptor_it = deserialized_acceptors->find(acceptor_config);
                if (deserialized_acceptor_it != deserialized_acceptors->end()) {
                        acceptor = deserialized_acceptor_it->second;
                        deserialized_acceptors->erase(deserialized_acceptor_it);
                } else {
-                       int server_sock = create_server_socket(acceptor_config.port, TCP_SOCKET);
-                       acceptor = new Acceptor(server_sock, acceptor_config.port);
+                       int server_sock = create_server_socket(acceptor_config.addr, TCP_SOCKET);
+                       acceptor = new Acceptor(server_sock, acceptor_config.addr,
+                                               acceptor_config.certificate_chain,
+                                               acceptor_config.private_key);
                }
                acceptor->run();
                acceptors.push_back(acceptor);
        }
 
        // Close all acceptors that are no longer in the configuration file.
-       for (map<int, Acceptor *>::iterator acceptor_it = deserialized_acceptors->begin();
-            acceptor_it != deserialized_acceptors->end();
-            ++acceptor_it) {
-               acceptor_it->second->close_socket();
-               delete acceptor_it->second;
+       for (auto &config_and_acceptor : *deserialized_acceptors) {
+               config_and_acceptor.second->close_socket();
+               delete config_and_acceptor.second;
        }
 
        return acceptors;
 }
 
+void create_config_input(const string &src, Input::Encoding encoding, multimap<InputKey, InputWithRefcount> *inputs)
+{
+       if (src.empty()) {
+               return;
+       }
+       InputKey key(src, encoding);
+       if (inputs->count(key) != 0) {
+               return;
+       }
+
+       InputWithRefcount iwr;
+       iwr.input = create_input(src, encoding);
+       if (iwr.input == nullptr) {
+               log(ERROR, "did not understand URL '%s' or source encoding was invalid, clients will not get any data.",
+                       src.c_str());
+               return;
+       }
+       iwr.refcount = 0;
+       inputs->insert(make_pair(key, iwr));
+}
+
 // Find all streams in the configuration file, and create inputs for them.
-vector<Input *> create_inputs(const Config &config,
-                              map<string, Input *> *deserialized_inputs)
+void create_config_inputs(const Config &config, multimap<InputKey, InputWithRefcount> *inputs)
 {
-       vector<Input *> inputs;
-       for (unsigned i = 0; i < config.streams.size(); ++i) {
-               const StreamConfig &stream_config = config.streams[i];
-               if (stream_config.src.empty()) {
+       for (const StreamConfig &stream_config : config.streams) {
+               if (stream_config.src == "delete") {
+                       // Ignored for pre-1.4.0 configuration compatibility.
                        continue;
                }
+               create_config_input(stream_config.src, Input::Encoding(stream_config.src_encoding), inputs);
+       }
+       for (const UDPStreamConfig &udpstream_config : config.udpstreams) {
+               create_config_input(udpstream_config.src, Input::INPUT_ENCODING_RAW, inputs);
+       }
+}
 
-               string stream_id = stream_config.stream_id;
-               string src = stream_config.src;
+void create_streams(const Config &config,
+                    const set<string> &deserialized_urls,
+                    multimap<InputKey, InputWithRefcount> *inputs)
+{
+       // HTTP streams.
+       set<string> expecting_urls = deserialized_urls;
+       for (const StreamConfig &stream_config : config.streams) {
+               int stream_index;
+
+               expecting_urls.erase(stream_config.url);
+
+               if (stream_config.src == "delete") {
+                       // Ignored for pre-1.4.0 configuration compatibility.
+                       continue;
+               }
 
-               Input *input = NULL;
-               map<string, Input *>::iterator deserialized_input_it =
-                       deserialized_inputs->find(stream_id);
-               if (deserialized_input_it != deserialized_inputs->end()) {
-                       input = deserialized_input_it->second;
-                       if (input->get_url() != src) {
-                               fprintf(stderr, "INFO: Stream '%s' has changed URL from '%s' to '%s', restarting input.\n",
-                                       stream_id.c_str(), input->get_url().c_str(), src.c_str());
-                               input->close_socket();
-                               delete input;
-                               input = NULL;
+               if (deserialized_urls.count(stream_config.url) == 0) {
+                       stream_index = servers->add_stream(stream_config.url,
+                                                          stream_config.hls_url,
+                                                          stream_config.backlog_size,
+                                                          stream_config.prebuffering_bytes,
+                                                          Stream::Encoding(stream_config.encoding),
+                                                          Stream::Encoding(stream_config.src_encoding),
+                                                          stream_config.hls_frag_duration,
+                                                          stream_config.hls_backlog_margin,
+                                                          stream_config.allow_origin);
+               } else {
+                       stream_index = servers->lookup_stream_by_url(stream_config.url);
+                       assert(stream_index != -1);
+                       servers->set_backlog_size(stream_index, stream_config.backlog_size);
+                       if (!stream_config.hls_url.empty()) {
+                               servers->register_hls_url(stream_index, stream_config.hls_url);
                        }
-                       deserialized_inputs->erase(deserialized_input_it);
+                       servers->set_prebuffering_bytes(stream_index, stream_config.prebuffering_bytes);
+                       servers->set_encoding(stream_index,
+                                             Stream::Encoding(stream_config.encoding));
+                       servers->set_src_encoding(stream_index,
+                                                 Stream::Encoding(stream_config.src_encoding));
+                       servers->set_hls_frag_duration(stream_index, stream_config.hls_frag_duration);
+                       servers->set_hls_backlog_margin(stream_index, stream_config.hls_backlog_margin);
+                       servers->set_allow_origin(stream_index, stream_config.allow_origin);
                }
-               if (input == NULL) {
-                       input = create_input(stream_id, src);
-                       if (input == NULL) {
-                               fprintf(stderr, "ERROR: did not understand URL '%s', clients will not get any data.\n",
-                                       src.c_str());
-                               continue;
+
+               servers->set_pacing_rate(stream_index, stream_config.pacing_rate);
+
+               string src = stream_config.src;
+               Input::Encoding src_encoding = Input::Encoding(stream_config.src_encoding);
+               if (!src.empty()) {
+                       const auto input_it = inputs->find(make_pair(src, src_encoding));
+                       if (input_it != inputs->end()) {
+                               input_it->second.input->add_destination(stream_index);
+                               ++input_it->second.refcount;
                        }
                }
-               input->run();
-               inputs.push_back(input);
        }
-       return inputs;
-}
 
-void create_streams(const Config &config,
-                    const set<string> &deserialized_stream_ids,
-                   map<string, Input *> *deserialized_inputs)
-{
-       vector<MarkPool *> mark_pools;  // FIXME: leak
-       for (unsigned i = 0; i < config.mark_pools.size(); ++i) {
-               const MarkPoolConfig &mp_config = config.mark_pools[i];
-               mark_pools.push_back(new MarkPool(mp_config.from, mp_config.to));
+       // Warn about any streams servers we've lost.
+       for (const string &url : expecting_urls) {
+               log(WARNING, "stream '%s' disappeared from the configuration file. "
+                            "It will not be deleted, but clients will not get any new inputs. "
+                            "If you really meant to delete it, set src=delete and reload.",
+                            url.c_str());
        }
 
-       set<string> expecting_stream_ids = deserialized_stream_ids;
-       for (unsigned i = 0; i < config.streams.size(); ++i) {
-               const StreamConfig &stream_config = config.streams[i];
-               if (deserialized_stream_ids.count(stream_config.stream_id) == 0) {
-                       servers->add_stream(stream_config.stream_id, stream_config.backlog_size);
-               }
-               expecting_stream_ids.erase(stream_config.stream_id);
-
-               if (stream_config.mark_pool != -1) {
-                       servers->set_mark_pool(stream_config.stream_id,
-                                              mark_pools[stream_config.mark_pool]);
+       // UDP streams.
+       for (const UDPStreamConfig &udpstream_config : config.udpstreams) {
+               int stream_index = servers->add_udpstream(
+                       udpstream_config.dst,
+                       udpstream_config.pacing_rate,
+                       udpstream_config.ttl,
+                       udpstream_config.multicast_iface_index);
+
+               string src = udpstream_config.src;
+               if (!src.empty()) {
+                       auto input_it = inputs->find(make_pair(src, Input::INPUT_ENCODING_RAW));
+                       assert(input_it != inputs->end());
+                       input_it->second.input->add_destination(stream_index);
+                       ++input_it->second.refcount;
                }
        }
 
-       // Warn about any servers we've lost.
-       // TODO: Make an option (delete=yes?) to actually shut down streams.
-       for (set<string>::const_iterator stream_it = expecting_stream_ids.begin();
-            stream_it != expecting_stream_ids.end();
-            ++stream_it) {
-               string stream_id = *stream_it;
-               fprintf(stderr, "WARNING: stream '%s' disappeared from the configuration file.\n",
-                       stream_id.c_str());
-               fprintf(stderr, "         It will not be deleted, but clients will not get any new inputs.\n");
-               if (deserialized_inputs->count(stream_id) != 0) {
-                       delete (*deserialized_inputs)[stream_id];
-                       deserialized_inputs->erase(stream_id);
+       // HTTP gen204 endpoints.
+       for (const Gen204Config &ping_config : config.pings) {
+               servers->add_gen204(ping_config.url, ping_config.allow_origin);
+       }
+}
+       
+void open_logs(const vector<LogConfig> &log_destinations)
+{
+       for (const LogConfig &log_destination : log_destinations) {
+               if (log_destination.type == LogConfig::LOG_TYPE_FILE) {
+                       add_log_destination_file(log_destination.filename);
+               } else if (log_destination.type == LogConfig::LOG_TYPE_CONSOLE) {
+                       add_log_destination_console();
+               } else if (log_destination.type == LogConfig::LOG_TYPE_SYSLOG) {
+                       add_log_destination_syslog();
+               } else {
+                       assert(false);
                }
        }
+       start_logging();
 }
        
-bool dry_run_config(const std::string &argv0, const std::string &config_filename)
+bool dry_run_config(const string &argv0, const string &config_filename)
 {
        char *argv0_copy = strdup(argv0.c_str());
        char *config_filename_copy = strdup(config_filename.c_str());
@@ -179,14 +288,14 @@ bool dry_run_config(const std::string &argv0, const std::string &config_filename
        pid_t pid = fork();
        switch (pid) {
        case -1:
-               perror("fork()");
+               log_perror("fork()");
                free(argv0_copy);
                free(config_filename_copy);
                return false;
        case 0:
                // Child.
-               execlp(argv0_copy, argv0_copy, "--test-config", config_filename_copy, NULL);
-               perror(argv0_copy);
+               execlp(argv0_copy, argv0_copy, "--test-config", config_filename_copy, nullptr);
+               log_perror(argv0_copy);
                _exit(1);
        default:
                // Parent.
@@ -203,15 +312,36 @@ bool dry_run_config(const std::string &argv0, const std::string &config_filename
        } while (err == -1 && errno == EINTR);
 
        if (err == -1) {
-               perror("waitpid()");
+               log_perror("waitpid()");
                return false;
        }       
 
        return (WIFEXITED(status) && WEXITSTATUS(status) == 0);
 }
 
+void find_all_streams(const Config &config, set<string> *all_urls)
+{
+       for (const StreamConfig &stream_config : config.streams) {
+               if (stream_config.src == "delete") {
+                       log(WARNING, "stream '%s' has src=delete; ignoring it. Since Cubemap 1.4.0, you do not "
+                                    "need to set src=delete to delete streams anymore; just delete them from "
+                                    "the configuration file.",
+                                    stream_config.url.c_str());
+               } else {
+                       all_urls->insert(stream_config.url);
+               }
+       }
+}
+
 int main(int argc, char **argv)
 {
+       signal(SIGHUP, hup);
+       signal(SIGINT, hup);
+       signal(SIGUSR1, do_nothing);  // Used in internal signalling.
+       signal(SIGPIPE, SIG_IGN);
+
+       tls_init();
+       
        // Parse options.
        int state_fd = -1;
        bool test_config = false;
@@ -219,9 +349,10 @@ int main(int argc, char **argv)
                static const option long_options[] = {
                        { "state", required_argument, 0, 's' },
                        { "test-config", no_argument, 0, 't' },
+                       { 0, 0, 0, 0 }
                };
                int option_index = 0;
-               int c = getopt_long (argc, argv, "s:t", long_options, &option_index);
+               int c = getopt_long(argc, argv, "s:t", long_options, &option_index);
      
                if (c == -1) {
                        break;
@@ -234,7 +365,8 @@ int main(int argc, char **argv)
                        test_config = true;
                        break;
                default:
-                       assert(false);
+                       fprintf(stderr, "Unknown option '%s'\n", argv[option_index]);
+                       exit(1);
                }
        }
 
@@ -243,130 +375,286 @@ int main(int argc, char **argv)
                config_filename = argv[optind++];
        }
 
+       // Canonicalize argv[0] and config_filename.
+       char argv0_canon[PATH_MAX];
+       char config_filename_canon[PATH_MAX];
+
+       if (realpath("/proc/self/exe", argv0_canon) == nullptr) {
+               log_perror(argv[0]);
+               exit(1);
+       }
+       if (realpath(config_filename.c_str(), config_filename_canon) == nullptr) {
+               log_perror(config_filename.c_str());
+               exit(1);
+       }
+
+       // Now parse the configuration file.
        Config config;
-       if (!parse_config(config_filename, &config)) {
+       if (!parse_config(config_filename_canon, &config)) {
                exit(1);
        }
        if (test_config) {
                exit(0);
        }
+       
+       // Ideally we'd like to daemonize only when we've started up all threads etc.,
+       // but daemon() forks, which is not good in multithreaded software, so we'll
+       // have to do it here.
+       if (config.daemonize) {
+               if (daemon(0, 0) == -1) {
+                       log_perror("daemon");
+                       exit(1);
+               }
+       }
 
 start:
-       fprintf(stderr, "\nCubemap " SERVER_VERSION " starting.\n");
+       // Open logs as soon as possible.
+       open_logs(config.log_destinations);
+
+       log(INFO, "Cubemap " SERVER_VERSION " starting.");
+       if (config.access_log_file.empty()) {
+               // Create a dummy logger.
+               access_log = new AccessLogThread();
+       } else {
+               access_log = new AccessLogThread(config.access_log_file);
+       }
+       access_log->run();
+
        servers = new ServerPool(config.num_servers);
 
+       // Find all the streams that are to be kept.
+       set<string> all_urls;
+       find_all_streams(config, &all_urls);
+
        CubemapStateProto loaded_state;
-       struct timeval serialize_start;
-       set<string> deserialized_stream_ids;
-       map<string, Input *> deserialized_inputs;
-       map<int, Acceptor *> deserialized_acceptors;
+       timespec serialize_start;
+       set<string> deserialized_urls;
+       map<AcceptorConfig, Acceptor *, AcceptorConfigCompare> deserialized_acceptors;
+       multimap<InputKey, InputWithRefcount> inputs;  // multimap due to older versions without deduplication.
        if (state_fd != -1) {
-               fprintf(stderr, "Deserializing state from previous process... ");
+               log(INFO, "Deserializing state from previous process...");
                string serialized;
-               if (!read_tempfile(state_fd, &serialized)) {
+               if (!read_tempfile_and_close(state_fd, &serialized)) {
                        exit(1);
                }
                if (!loaded_state.ParseFromString(serialized)) {
-                       fprintf(stderr, "ERROR: Failed deserialization of state.\n");
+                       log(ERROR, "Failed deserialization of state.");
                        exit(1);
                }
 
                serialize_start.tv_sec = loaded_state.serialize_start_sec();
-               serialize_start.tv_usec = loaded_state.serialize_start_usec();
+               serialize_start.tv_nsec = loaded_state.serialize_start_usec() * 1000ull;
 
                // Deserialize the streams.
-               for (int i = 0; i < loaded_state.streams_size(); ++i) {
-                       servers->add_stream_from_serialized(loaded_state.streams(i));
-                       deserialized_stream_ids.insert(loaded_state.streams(i).stream_id());
+               unordered_map<string, string> stream_headers_for_url;  // See below.
+               for (const StreamProto &stream : loaded_state.streams()) {
+                       if (all_urls.count(stream.url()) == 0) {
+                               // Delete the stream backlogs.
+                               log(INFO, "Deleting stream '%s'.", stream.url().c_str());
+                               for (const int fd : stream.data_fds()) {
+                                       safe_close(fd);
+                               }
+                       } else {
+                               vector<int> data_fds;
+                               for (const int fd : stream.data_fds()) {
+                                       data_fds.push_back(fd);
+                               }
+
+                               servers->add_stream_from_serialized(stream, data_fds);
+                               deserialized_urls.insert(stream.url());
+
+                               stream_headers_for_url.insert(make_pair(stream.url(), stream.stream_header()));
+                       }
                }
 
-               // Deserialize the inputs. Note that we don't actually add them to any state yet.
-               for (int i = 0; i < loaded_state.inputs_size(); ++i) {
-                       deserialized_inputs.insert(make_pair(
-                               loaded_state.inputs(i).stream_id(),
-                               create_input(loaded_state.inputs(i))));
+               // Deserialize the inputs. Note that we don't actually add them to any stream yet.
+               for (const InputProto &serialized_input : loaded_state.inputs()) {
+                       InputWithRefcount iwr;
+                       iwr.input = create_input(serialized_input);
+                       iwr.refcount = 0;
+
+                       Input::Encoding src_encoding = serialized_input.is_metacube_encoded() ?
+                               Input::INPUT_ENCODING_METACUBE :
+                               Input::INPUT_ENCODING_RAW;
+                       InputKey key(serialized_input.url(), src_encoding);
+                       inputs.insert(make_pair(key, iwr));
                } 
 
                // Deserialize the acceptors.
-               for (int i = 0; i < loaded_state.acceptors_size(); ++i) {
+               for (const AcceptorProto &serialized_acceptor : loaded_state.acceptors()) {
+                       AcceptorConfig config;
+                       config.addr = extract_address_from_acceptor_proto(serialized_acceptor);
+                       config.certificate_chain = serialized_acceptor.certificate_chain();
+                       config.private_key = serialized_acceptor.private_key();
                        deserialized_acceptors.insert(make_pair(
-                               loaded_state.acceptors(i).port(),
-                               new Acceptor(loaded_state.acceptors(i))));
+                               config,
+                               new Acceptor(serialized_acceptor)));
                }
 
-               fprintf(stderr, "done.\n");
+               log(INFO, "Deserialization done.");
        }
 
-       // Find all streams in the configuration file, and create them.
-       create_streams(config, deserialized_stream_ids, &deserialized_inputs);
+       // Add any new inputs coming from the config.
+       create_config_inputs(config, &inputs);
+       
+       // Find all streams in the configuration file, create them, and connect to the inputs.
+       create_streams(config, deserialized_urls, &inputs);
+       vector<Acceptor *> acceptors = create_acceptors(config, &deserialized_acceptors);
 
-       servers->run();
+       // Make all the servers create TLS contexts for every TLS keypair we have.
+       for (Acceptor *acceptor : acceptors) {
+               if (acceptor->is_tls()) {
+                       servers->create_tls_context_for_acceptor(acceptor);
+               }
+       }
 
-       vector<Acceptor *> acceptors = create_acceptors(config, &deserialized_acceptors);
-       vector<Input *> inputs = create_inputs(config, &deserialized_inputs);
-       
-       // All deserialized inputs should now have been taken care of, one way or the other.
-       assert(deserialized_inputs.empty());
+       // Allocate strings for the short responses.
+       vector<shared_ptr<const string>> short_response_pool;
+       for (const ShortResponsePool &str : loaded_state.short_response_pool()) {
+               short_response_pool.emplace_back(new string(str.header_or_short_response()));
+       }
        
        // Put back the existing clients. It doesn't matter which server we
-       // allocate them to, so just do round-robin. However, we need to add
-       // them after the mark pools have been set up.
+       // allocate them to, so just do round-robin. However, we need to sort them
+       // by connection time first, since add_client_serialized() expects that.
+       sort(loaded_state.mutable_clients()->begin(),
+            loaded_state.mutable_clients()->end(),
+            OrderByConnectionTime());
        for (int i = 0; i < loaded_state.clients_size(); ++i) {
-               servers->add_client_from_serialized(loaded_state.clients(i));
+               if (!loaded_state.clients(i).url().empty() &&
+                   all_urls.count(loaded_state.clients(i).url()) == 0) {
+                       // Belongs to a dead stream (not keepalive), so we just have to close.
+                       safe_close(loaded_state.clients(i).sock());
+               } else {
+                       servers->add_client_from_serialized(loaded_state.clients(i), short_response_pool);
+               }
+       }
+       
+       short_response_pool.clear();  // No longer needed; the clients have their own refcounts now.
+
+       // Put back the HLS zombies. There's no really good allocation here
+       // except round-robin; it would be marginally more efficient to match it
+       // to the client (since that would have them deleted immediately when
+       // the client requests the next fragment, instead of being later weeded
+       // out during statistics collection), but it's not a big deal.
+       for (const HLSZombieProto &zombie_proto : loaded_state.hls_zombies()) {
+               servers->add_hls_zombie_from_serialized(zombie_proto);
+       }
+
+       servers->run();
+
+       // Now delete all inputs that are longer in use, and start the others.
+       for (auto input_it = inputs.begin(); input_it != inputs.end(); ) {
+               if (input_it->second.refcount == 0) {
+                       if (input_it->first.second == Input::INPUT_ENCODING_RAW) {
+                               log(WARNING, "Raw input '%s' no longer in use, closing.",
+                                   input_it->first.first.c_str());
+                       } else {
+                               assert(input_it->first.second == Input::INPUT_ENCODING_METACUBE);
+                               log(WARNING, "Metacube input '%s' no longer in use, closing.",
+                                   input_it->first.first.c_str());
+                       }
+                       input_it->second.input->close_socket();
+                       delete input_it->second.input;
+                       inputs.erase(input_it++);
+               } else {
+                       input_it->second.input->run();
+                       ++input_it;
+               }
        }
 
        // Start writing statistics.
-       StatsThread *stats_thread = NULL;
+       unique_ptr<StatsThread> stats_thread;
        if (!config.stats_file.empty()) {
-               stats_thread = new StatsThread(config.stats_file, config.stats_interval);
+               stats_thread.reset(new StatsThread(config.stats_file, config.stats_interval));
                stats_thread->run();
        }
 
-       signal(SIGHUP, hup);
-       
-       struct timeval server_start;
-       gettimeofday(&server_start, NULL);
+       unique_ptr<InputStatsThread> input_stats_thread;
+       if (!config.input_stats_file.empty()) {
+               vector<Input*> inputs_no_refcount;
+               for (const auto &key_and_input_with_refcount : inputs) {
+                       inputs_no_refcount.push_back(key_and_input_with_refcount.second.input);
+               }
+
+               input_stats_thread.reset(new InputStatsThread(config.input_stats_file, config.input_stats_interval, inputs_no_refcount));
+               input_stats_thread->run();
+       }
+
+       timespec server_start;
+       int err = clock_gettime(CLOCK_MONOTONIC, &server_start);
+       assert(err != -1);
        if (state_fd != -1) {
                // Measure time from we started deserializing (below) to now, when basically everything
                // is up and running. This is, in other words, a conservative estimate of how long our
                // “glitch” period was, not counting of course reconnects if the configuration changed.
                double glitch_time = server_start.tv_sec - serialize_start.tv_sec +
-                       1e-6 * (server_start.tv_usec - serialize_start.tv_usec);
-               fprintf(stderr, "Re-exec happened in approx. %.0f ms.\n", glitch_time * 1000.0);
+                       1e-9 * (server_start.tv_nsec - serialize_start.tv_nsec);
+               log(INFO, "Re-exec happened in approx. %.0f ms.", glitch_time * 1000.0);
        }
 
+       sd_notify(0, "READY=1");
+
        while (!hupped) {
                usleep(100000);
        }
 
+       if (stopped) {
+               sd_notify(0, "STOPPING=1");
+       } else {
+               sd_notify(0, "RELOADING=1");
+       }
+
        // OK, we've been HUPed. Time to shut down everything, serialize, and re-exec.
-       gettimeofday(&serialize_start, NULL);
+       err = clock_gettime(CLOCK_MONOTONIC, &serialize_start);
+       assert(err != -1);
 
-       if (stats_thread != NULL) {
+       if (input_stats_thread != nullptr) {
+               input_stats_thread->stop();
+               input_stats_thread.reset();
+       }
+       if (stats_thread != nullptr) {
                stats_thread->stop();
+               stats_thread.reset();
        }
-       for (size_t i = 0; i < acceptors.size(); ++i) {
-               acceptors[i]->stop();
+       for (Acceptor *acceptor : acceptors) {
+               acceptor->stop();
        }
-       for (size_t i = 0; i < inputs.size(); ++i) {
-               inputs[i]->stop();
+       for (const auto &key_and_input_with_refcount : inputs) {
+               key_and_input_with_refcount.second.input->stop();
        }
        servers->stop();
 
-       fprintf(stderr, "Serializing state and re-execing...\n");
-       CubemapStateProto state = collect_state(
-                serialize_start, acceptors, inputs, servers);
-       string serialized;
-       state.SerializeToString(&serialized);
-       state_fd = make_tempfile(serialized);
-       if (state_fd == -1) {
-               exit(1);
+       CubemapStateProto state;
+       if (stopped) {
+               log(INFO, "Shutting down.");
+       } else {
+               log(INFO, "Serializing state and re-execing...");
+               state = collect_state(
+                       serialize_start, acceptors, inputs, servers);
+               string serialized;
+               state.SerializeToString(&serialized);
+               state_fd = make_tempfile(serialized);
+               if (state_fd == -1) {
+                       exit(1);
+               }
        }
        delete servers;
 
-       if (!dry_run_config(argv[0], config_filename)) {
-               fprintf(stderr, "ERROR: %s --test-config failed. Restarting old version instead of new.\n", argv[0]);
+       access_log->stop();
+       delete access_log;
+       shut_down_logging();
+
+       if (stopped) {
+               exit(0);
+       }
+
+       // OK, so the signal was SIGHUP. Check that the new config is okay, then exec the new binary.
+       if (!dry_run_config(argv0_canon, config_filename_canon)) {
+               open_logs(config.log_destinations);
+               log(ERROR, "%s --test-config failed. Restarting old version instead of new.", argv[0]);
                hupped = false;
+               shut_down_logging();
                goto start;
        }
         
@@ -374,9 +662,11 @@ start:
        sprintf(buf, "%d", state_fd);
 
        for ( ;; ) {
-               execlp(argv[0], argv[0], config_filename.c_str(), "--state", buf, NULL);
-               perror("execlp");
-               fprintf(stderr, "PANIC: re-exec of %s failed. Waiting 0.2 seconds and trying again...\n", argv[0]);
+               execlp(argv0_canon, argv0_canon, config_filename_canon, "--state", buf, nullptr);
+               open_logs(config.log_destinations);
+               log_perror("execlp");
+               log(ERROR, "re-exec of %s failed. Waiting 0.2 seconds and trying again...", argv0_canon);
+               shut_down_logging();
                usleep(200000);
        }
 }