void create_config_inputs(const Config &config, multimap<InputKey, InputWithRefcount> *inputs)
{
for (const StreamConfig &stream_config : config.streams) {
- if (stream_config.src != "delete") {
- create_config_input(stream_config.src, Input::Encoding(stream_config.src_encoding), inputs);
+ if (stream_config.src == "delete") {
+ // Ignored for pre-1.4.0 configuration compatibility.
+ continue;
}
+ create_config_input(stream_config.src, Input::Encoding(stream_config.src_encoding), inputs);
}
for (const UDPStreamConfig &udpstream_config : config.udpstreams) {
create_config_input(udpstream_config.src, Input::INPUT_ENCODING_RAW, inputs);
expecting_urls.erase(stream_config.url);
- // Special-case deleted streams; they were never deserialized in the first place,
- // so just ignore them.
if (stream_config.src == "delete") {
+ // Ignored for pre-1.4.0 configuration compatibility.
continue;
}
return (WIFEXITED(status) && WEXITSTATUS(status) == 0);
}
-void find_deleted_streams(const Config &config, set<string> *deleted_urls)
+void find_all_streams(const Config &config, set<string> *all_urls)
{
for (const StreamConfig &stream_config : config.streams) {
if (stream_config.src == "delete") {
- log(INFO, "Deleting stream '%s'.", stream_config.url.c_str());
- deleted_urls->insert(stream_config.url);
+ log(WARNING, "stream '%s' has src=delete; ignoring it. Since Cubemap 1.4.0, you do not "
+ "need to set src=delete to delete streams anymore; just delete them from "
+ "the configuration file.",
+ stream_config.url.c_str());
+ } else {
+ all_urls->insert(stream_config.url);
}
}
}
servers = new ServerPool(config.num_servers);
- // Find all the streams that are to be deleted.
- set<string> deleted_urls;
- find_deleted_streams(config, &deleted_urls);
+ // Find all the streams that are to be kept.
+ set<string> all_urls;
+ find_all_streams(config, &all_urls);
CubemapStateProto loaded_state;
timespec serialize_start;
// Deserialize the streams.
map<string, string> stream_headers_for_url; // See below.
for (const StreamProto &stream : loaded_state.streams()) {
- if (deleted_urls.count(stream.url()) != 0) {
+ if (all_urls.count(stream.url()) == 0) {
// Delete the stream backlogs.
+ log(INFO, "Deleting stream '%s'.", stream.url().c_str());
for (const int fd : stream.data_fds()) {
safe_close(fd);
}
servers->create_tls_context_for_acceptor(acceptor);
}
}
+
+ // Allocate strings for the short responses.
+ vector<shared_ptr<const string>> short_response_pool;
+ for (const ShortResponsePool &str : loaded_state.short_response_pool()) {
+ short_response_pool.emplace_back(new string(str.header_or_short_response()));
+ }
// Put back the existing clients. It doesn't matter which server we
// allocate them to, so just do round-robin. However, we need to sort them
loaded_state.mutable_clients()->end(),
OrderByConnectionTime());
for (int i = 0; i < loaded_state.clients_size(); ++i) {
- if (deleted_urls.count(loaded_state.clients(i).url()) != 0) {
+ if (all_urls.count(loaded_state.clients(i).url()) == 0) {
safe_close(loaded_state.clients(i).sock());
} else {
- servers->add_client_from_serialized(loaded_state.clients(i));
+ servers->add_client_from_serialized(loaded_state.clients(i), short_response_pool);
}
}
+ short_response_pool.clear(); // No longer needed; the clients have their own refcounts now.
+
servers->run();
// Now delete all inputs that are longer in use, and start the others.