#include <assert.h>
#include <errno.h>
+#include <limits.h>
#include <netinet/in.h>
#include <stdlib.h>
#include <string.h>
data_fd(make_tempfile("")),
backlog_size(backlog_size),
bytes_received(0),
- mark_pool(NULL)
+ last_suitable_starting_point(-1),
+ mark_pool(NULL),
+ queued_data_last_starting_point(-1)
{
if (data_fd == -1) {
exit(1);
data_fd(data_fd),
backlog_size(serialized.backlog_size()),
bytes_received(serialized.bytes_received()),
- mark_pool(NULL)
+ mark_pool(NULL),
+ queued_data_last_starting_point(-1)
{
if (data_fd == -1) {
exit(1);
stream_header = header.substr(split, string::npos);
}
}
+
+ // Older versions did not set last_suitable_starting_point.
+ if (serialized.has_last_suitable_starting_point()) {
+ last_suitable_starting_point = serialized.last_suitable_starting_point();
+ } else {
+ last_suitable_starting_point = bytes_received;
+ }
}
StreamProto Stream::serialize()
serialized.add_data_fds(data_fd);
serialized.set_backlog_size(backlog_size);
serialized.set_bytes_received(bytes_received);
+ serialized.set_last_suitable_starting_point(last_suitable_starting_point);
serialized.set_url(url);
data_fd = -1;
return serialized;
vector<iovec> collect_iovecs(const vector<iovec> &data, size_t bytes_wanted)
{
vector<iovec> ret;
- for (size_t i = 0; i < data.size() && bytes_wanted > 0; ++i) {
+ size_t max_iovecs = std::min<size_t>(data.size(), IOV_MAX);
+ for (size_t i = 0; i < max_iovecs && bytes_wanted > 0; ++i) {
if (data[i].iov_len <= bytes_wanted) {
// Consume the entire iovec.
ret.push_back(data[i]);
} while (ret == -1 && errno == EINTR);
if (ret == -1) {
- log_perror("pwrite");
+ log_perror("pwritev");
// Dazed and confused, but trying to continue...
return;
}
}
}
-void Stream::add_data_deferred(const char *data, size_t bytes)
+void Stream::add_data_deferred(const char *data, size_t bytes, StreamStartSuitability suitable_for_stream_start)
{
+ assert(suitable_for_stream_start == SUITABLE_FOR_STREAM_START ||
+ suitable_for_stream_start == NOT_SUITABLE_FOR_STREAM_START);
+ if (suitable_for_stream_start == SUITABLE_FOR_STREAM_START) {
+ queued_data_last_starting_point = queued_data.size();
+ }
+
if (encoding == Stream::STREAM_ENCODING_METACUBE) {
// Add a Metacube block header before the data.
metacube_block_header hdr;
memcpy(hdr.sync, METACUBE_SYNC, sizeof(hdr.sync));
hdr.size = htonl(bytes);
hdr.flags = htonl(0);
+ if (suitable_for_stream_start == NOT_SUITABLE_FOR_STREAM_START) {
+ hdr.flags |= htonl(METACUBE_FLAGS_NOT_SUITABLE_FOR_STREAM_START);
+ }
iovec iov;
iov.iov_base = new char[bytes + sizeof(hdr)];
return;
}
+ // Update the last suitable starting point for the stream,
+ // if the queued data contains such a starting point.
+ assert(queued_data_last_starting_point < ssize_t(queued_data.size()));
+ if (queued_data_last_starting_point >= 0) {
+ last_suitable_starting_point = bytes_received;
+ for (int i = 0; i < queued_data_last_starting_point; ++i) {
+ last_suitable_starting_point += queued_data[i].iov_len;
+ }
+ }
+
add_data_raw(queued_data);
for (size_t i = 0; i < queued_data.size(); ++i) {
char *data = reinterpret_cast<char *>(queued_data[i].iov_base);
delete[] data;
}
queued_data.clear();
+ queued_data_last_starting_point = -1;
// We have more data, so wake up all clients.
if (to_process.empty()) {