3 #include <google/protobuf/repeated_field.h>
10 #include "serverpool.h"
16 ServerPool::ServerPool(int size)
17 : servers(new Server[size]),
23 ServerPool::~ServerPool()
28 CubemapStateProto ServerPool::serialize()
30 CubemapStateProto state;
32 for (int i = 0; i < num_servers; ++i) {
33 CubemapStateProto local_state = servers[i].serialize();
35 // The stream state should be identical between the servers, so we only store it once,
36 // save for the fds, which we keep around to distribute to the servers after re-exec.
38 state.mutable_streams()->MergeFrom(local_state.streams());
40 assert(state.streams_size() == local_state.streams_size());
41 for (int j = 0; j < local_state.streams_size(); ++j) {
42 assert(local_state.streams(j).data_fds_size() == 1);
43 state.mutable_streams(j)->add_data_fds(local_state.streams(j).data_fds(0));
46 for (int j = 0; j < local_state.clients_size(); ++j) {
47 state.add_clients()->MergeFrom(local_state.clients(j));
54 void ServerPool::add_client(int sock)
56 servers[clients_added++ % num_servers].add_client_deferred(sock);
59 void ServerPool::add_client_from_serialized(const ClientProto &client)
61 servers[clients_added++ % num_servers].add_client_from_serialized(client);
64 void ServerPool::add_stream(const string &stream_id, size_t backlog_size, Stream::Encoding encoding)
66 for (int i = 0; i < num_servers; ++i) {
67 servers[i].add_stream(stream_id, backlog_size, encoding);
71 void ServerPool::add_stream_from_serialized(const StreamProto &stream, const vector<int> &data_fds)
73 assert(!data_fds.empty());
75 for (int i = 0; i < num_servers; ++i) {
77 if (i < int(data_fds.size())) {
78 // Reuse one of the existing file descriptors.
79 data_fd = data_fds[i];
81 // Clone the first one.
82 if (contents.empty()) {
83 if (!read_tempfile(data_fds[0], &contents)) {
87 data_fd = make_tempfile(contents);
90 servers[i].add_stream_from_serialized(stream, data_fd);
93 // Close and delete any leftovers, if the number of servers was reduced.
94 for (size_t i = num_servers; i < data_fds.size(); ++i) {
97 ret = close(data_fds[i]); // Implicitly deletes the file.
98 } while (ret == -1 && errno == EINTR);
102 // Can still continue.
107 void ServerPool::set_header(const string &stream_id, const string &http_header, const string &stream_header)
109 for (int i = 0; i < num_servers; ++i) {
110 servers[i].set_header(stream_id, http_header, stream_header);
114 void ServerPool::add_data(const string &stream_id, const char *data, size_t bytes)
116 for (int i = 0; i < num_servers; ++i) {
117 servers[i].add_data_deferred(stream_id, data, bytes);
121 void ServerPool::run()
123 for (int i = 0; i < num_servers; ++i) {
128 void ServerPool::stop()
130 for (int i = 0; i < num_servers; ++i) {
135 vector<ClientStats> ServerPool::get_client_stats() const
137 vector<ClientStats> ret;
138 for (int i = 0; i < num_servers; ++i) {
139 vector<ClientStats> stats = servers[i].get_client_stats();
140 ret.insert(ret.end(), stats.begin(), stats.end());
145 void ServerPool::set_mark_pool(const string &stream_id, MarkPool *mark_pool)
147 for (int i = 0; i < num_servers; ++i) {
148 servers[i].set_mark_pool(stream_id, mark_pool);
152 void ServerPool::set_backlog_size(const string &stream_id, size_t new_size)
154 for (int i = 0; i < num_servers; ++i) {
155 servers[i].set_backlog_size(stream_id, new_size);
159 void ServerPool::set_encoding(const string &stream_id, Stream::Encoding encoding)
161 for (int i = 0; i < num_servers; ++i) {
162 servers[i].set_encoding(stream_id, encoding);