X-Git-Url: https://git.sesse.net/?p=cubemap;a=blobdiff_plain;f=main.cpp;h=737f0c3573acea88fa7d9f0585aee69d95ec0838;hp=858a6d0867000bf9ec9fcb456fa66f621bd4544d;hb=ed218ece51e5be77fd8bd8f014b5f05708e4be5b;hpb=2494f1e73d7910cfba3808be0bce9697f28df094 diff --git a/main.cpp b/main.cpp index 858a6d0..737f0c3 100644 --- a/main.cpp +++ b/main.cpp @@ -15,6 +15,7 @@ #include #include #include +#include #include #include @@ -194,19 +195,29 @@ void create_streams(const Config &config, if (deserialized_urls.count(stream_config.url) == 0) { stream_index = servers->add_stream(stream_config.url, + stream_config.hls_url, stream_config.backlog_size, stream_config.prebuffering_bytes, Stream::Encoding(stream_config.encoding), - Stream::Encoding(stream_config.src_encoding)); + Stream::Encoding(stream_config.src_encoding), + stream_config.hls_frag_duration, + stream_config.hls_backlog_margin, + stream_config.allow_origin); } else { stream_index = servers->lookup_stream_by_url(stream_config.url); assert(stream_index != -1); servers->set_backlog_size(stream_index, stream_config.backlog_size); + if (!stream_config.hls_url.empty()) { + servers->register_hls_url(stream_index, stream_config.hls_url); + } servers->set_prebuffering_bytes(stream_index, stream_config.prebuffering_bytes); servers->set_encoding(stream_index, Stream::Encoding(stream_config.encoding)); servers->set_src_encoding(stream_index, Stream::Encoding(stream_config.src_encoding)); + servers->set_hls_frag_duration(stream_index, stream_config.hls_frag_duration); + servers->set_hls_backlog_margin(stream_index, stream_config.hls_backlog_margin); + servers->set_allow_origin(stream_index, stream_config.allow_origin); } servers->set_pacing_rate(stream_index, stream_config.pacing_rate); @@ -435,7 +446,7 @@ start: serialize_start.tv_nsec = loaded_state.serialize_start_usec() * 1000ull; // Deserialize the streams. - map stream_headers_for_url; // See below. + unordered_map stream_headers_for_url; // See below. for (const StreamProto &stream : loaded_state.streams()) { if (all_urls.count(stream.url()) == 0) { // Delete the stream backlogs. @@ -496,6 +507,12 @@ start: servers->create_tls_context_for_acceptor(acceptor); } } + + // Allocate strings for the short responses. + vector> short_response_pool; + for (const ShortResponsePool &str : loaded_state.short_response_pool()) { + short_response_pool.emplace_back(new string(str.header_or_short_response())); + } // Put back the existing clients. It doesn't matter which server we // allocate them to, so just do round-robin. However, we need to sort them @@ -504,13 +521,26 @@ start: loaded_state.mutable_clients()->end(), OrderByConnectionTime()); for (int i = 0; i < loaded_state.clients_size(); ++i) { - if (all_urls.count(loaded_state.clients(i).url()) == 0) { + if (!loaded_state.clients(i).url().empty() && + all_urls.count(loaded_state.clients(i).url()) == 0) { + // Belongs to a dead stream (not keepalive), so we just have to close. safe_close(loaded_state.clients(i).sock()); } else { - servers->add_client_from_serialized(loaded_state.clients(i)); + servers->add_client_from_serialized(loaded_state.clients(i), short_response_pool); } } + short_response_pool.clear(); // No longer needed; the clients have their own refcounts now. + + // Put back the HLS zombies. There's no really good allocation here + // except round-robin; it would be marginally more efficient to match it + // to the client (since that would have them deleted immediately when + // the client requests the next fragment, instead of being later weeded + // out during statistics collection), but it's not a big deal. + for (const HLSZombieProto &zombie_proto : loaded_state.hls_zombies()) { + servers->add_hls_zombie_from_serialized(zombie_proto); + } + servers->run(); // Now delete all inputs that are longer in use, and start the others.