summaryrefslogtreecommitdiff
path: root/thirdparty/libwebsockets/output.c
diff options
context:
space:
mode:
authorRémi Verschelde <rverschelde@gmail.com>2018-06-07 10:46:05 +0200
committerRémi Verschelde <rverschelde@gmail.com>2018-06-07 10:46:05 +0200
commitd0811ed3f9604b2dab755ab6df0a2950fd1ccede (patch)
treeebb74de9c5c8c2b9eecb898121307755ca5ba0a7 /thirdparty/libwebsockets/output.c
parent7d392cd6f496374aea98e8f5c85792acb6f8bc7a (diff)
Thirdparty: Fill copyright for lws, miniupnpc, clipper
Rename `lws` to `libwebsockets` which is its library name. Add missing license file for mbedtls.
Diffstat (limited to 'thirdparty/libwebsockets/output.c')
-rw-r--r--thirdparty/libwebsockets/output.c883
1 files changed, 883 insertions, 0 deletions
diff --git a/thirdparty/libwebsockets/output.c b/thirdparty/libwebsockets/output.c
new file mode 100644
index 0000000000..375ff3ef99
--- /dev/null
+++ b/thirdparty/libwebsockets/output.c
@@ -0,0 +1,883 @@
+/*
+ * libwebsockets - small server side websockets and web server implementation
+ *
+ * Copyright (C) 2010-2017 Andy Green <andy@warmcat.com>
+ *
+ * This library is free software; you can redistribute it and/or
+ * modify it under the terms of the GNU Lesser General Public
+ * License as published by the Free Software Foundation:
+ * version 2.1 of the License.
+ *
+ * This library is distributed in the hope that it will be useful,
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
+ * Lesser General Public License for more details.
+ *
+ * You should have received a copy of the GNU Lesser General Public
+ * License along with this library; if not, write to the Free Software
+ * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston,
+ * MA 02110-1301 USA
+ */
+
+#include "private-libwebsockets.h"
+
+static int
+lws_0405_frame_mask_generate(struct lws *wsi)
+{
+#if 0
+ wsi->u.ws.mask[0] = 0;
+ wsi->u.ws.mask[1] = 0;
+ wsi->u.ws.mask[2] = 0;
+ wsi->u.ws.mask[3] = 0;
+#else
+ int n;
+ /* fetch the per-frame nonce */
+
+ n = lws_get_random(lws_get_context(wsi), wsi->u.ws.mask, 4);
+ if (n != 4) {
+ lwsl_parser("Unable to read from random device %s %d\n",
+ SYSTEM_RANDOM_FILEPATH, n);
+ return 1;
+ }
+#endif
+ /* start masking from first byte of masking key buffer */
+ wsi->u.ws.mask_idx = 0;
+
+ return 0;
+}
+
+/*
+ * notice this returns number of bytes consumed, or -1
+ */
+int lws_issue_raw(struct lws *wsi, unsigned char *buf, size_t len)
+{
+ struct lws_context *context = lws_get_context(wsi);
+ struct lws_context_per_thread *pt = &wsi->context->pt[(int)wsi->tsi];
+ size_t real_len = len;
+ unsigned int n;
+ int m;
+
+ lws_stats_atomic_bump(wsi->context, pt, LWSSTATS_C_API_WRITE, 1);
+
+ if (!len)
+ return 0;
+ /* just ignore sends after we cleared the truncation buffer */
+ if (wsi->state == LWSS_FLUSHING_STORED_SEND_BEFORE_CLOSE &&
+ !wsi->trunc_len)
+ return len;
+
+ if (wsi->trunc_len && (buf < wsi->trunc_alloc ||
+ buf > (wsi->trunc_alloc + wsi->trunc_len + wsi->trunc_offset))) {
+ char dump[20];
+ strncpy(dump, (char *)buf, sizeof(dump) - 1);
+ dump[sizeof(dump) - 1] = '\0';
+#if defined(LWS_WITH_ESP8266)
+ lwsl_err("****** %p: Sending new %lu (%s), pending truncated ...\n",
+ wsi, (unsigned long)len, dump);
+#else
+ lwsl_err("****** %p: Sending new %lu (%s), pending truncated ...\n"
+ " It's illegal to do an lws_write outside of\n"
+ " the writable callback: fix your code\n",
+ wsi, (unsigned long)len, dump);
+#endif
+ assert(0);
+
+ return -1;
+ }
+
+ m = lws_ext_cb_active(wsi, LWS_EXT_CB_PACKET_TX_DO_SEND, &buf, len);
+ if (m < 0)
+ return -1;
+ if (m) /* handled */ {
+ n = m;
+ goto handle_truncated_send;
+ }
+
+ if (!wsi->http2_substream && !lws_socket_is_valid(wsi->desc.sockfd))
+ lwsl_warn("** error invalid sock but expected to send\n");
+
+ /* limit sending */
+ if (wsi->protocol->tx_packet_size)
+ n = wsi->protocol->tx_packet_size;
+ else {
+ n = wsi->protocol->rx_buffer_size;
+ if (!n)
+ n = context->pt_serv_buf_size;
+ }
+ n += LWS_PRE + 4;
+ if (n > len)
+ n = len;
+#if defined(LWS_WITH_ESP8266)
+ if (wsi->pending_send_completion) {
+ n = 0;
+ goto handle_truncated_send;
+ }
+#endif
+
+ /* nope, send it on the socket directly */
+ lws_latency_pre(context, wsi);
+ n = lws_ssl_capable_write(wsi, buf, n);
+ lws_latency(context, wsi, "send lws_issue_raw", n, n == len);
+
+ switch (n) {
+ case LWS_SSL_CAPABLE_ERROR:
+ /* we're going to close, let close know sends aren't possible */
+ wsi->socket_is_permanently_unusable = 1;
+ return -1;
+ case LWS_SSL_CAPABLE_MORE_SERVICE:
+ /* nothing got sent, not fatal, retry the whole thing later */
+ n = 0;
+ break;
+ }
+
+handle_truncated_send:
+ /*
+ * we were already handling a truncated send?
+ */
+ if (wsi->trunc_len) {
+ lwsl_info("%p partial adv %d (vs %ld)\n", wsi, n, (long)real_len);
+ wsi->trunc_offset += n;
+ wsi->trunc_len -= n;
+
+ if (!wsi->trunc_len) {
+ lwsl_info("***** %p partial send completed\n", wsi);
+ /* done with it, but don't free it */
+ n = real_len;
+ if (wsi->state == LWSS_FLUSHING_STORED_SEND_BEFORE_CLOSE) {
+ lwsl_info("***** %p signalling to close now\n", wsi);
+ return -1; /* retry closing now */
+ }
+ }
+ /* always callback on writeable */
+ lws_callback_on_writable(wsi);
+
+ return n;
+ }
+
+ if ((unsigned int)n == real_len)
+ /* what we just sent went out cleanly */
+ return n;
+
+ /*
+ * Newly truncated send. Buffer the remainder (it will get
+ * first priority next time the socket is writable)
+ */
+ lwsl_debug("%p new partial sent %d from %lu total\n", wsi, n,
+ (unsigned long)real_len);
+
+ lws_stats_atomic_bump(wsi->context, pt, LWSSTATS_C_WRITE_PARTIALS, 1);
+ lws_stats_atomic_bump(wsi->context, pt, LWSSTATS_B_PARTIALS_ACCEPTED_PARTS, n);
+
+ /*
+ * - if we still have a suitable malloc lying around, use it
+ * - or, if too small, reallocate it
+ * - or, if no buffer, create it
+ */
+ if (!wsi->trunc_alloc || real_len - n > wsi->trunc_alloc_len) {
+ lws_free(wsi->trunc_alloc);
+
+ wsi->trunc_alloc_len = real_len - n;
+ wsi->trunc_alloc = lws_malloc(real_len - n, "truncated send alloc");
+ if (!wsi->trunc_alloc) {
+ lwsl_err("truncated send: unable to malloc %lu\n",
+ (unsigned long)(real_len - n));
+ return -1;
+ }
+ }
+ wsi->trunc_offset = 0;
+ wsi->trunc_len = real_len - n;
+ memcpy(wsi->trunc_alloc, buf + n, real_len - n);
+
+ /* since something buffered, force it to get another chance to send */
+ lws_callback_on_writable(wsi);
+
+ return real_len;
+}
+
+LWS_VISIBLE int lws_write(struct lws *wsi, unsigned char *buf, size_t len,
+ enum lws_write_protocol wp)
+{
+ struct lws_context_per_thread *pt = &wsi->context->pt[(int)wsi->tsi];
+ int masked7 = (wsi->mode == LWSCM_WS_CLIENT);
+ unsigned char is_masked_bit = 0;
+ unsigned char *dropmask = NULL;
+ struct lws_tokens eff_buf;
+ size_t orig_len = len;
+ int pre = 0, n;
+
+ if (wsi->parent_carries_io) {
+ struct lws_write_passthru pas;
+
+ pas.buf = buf;
+ pas.len = len;
+ pas.wp = wp;
+ pas.wsi = wsi;
+
+ if (wsi->parent->protocol->callback(wsi->parent,
+ LWS_CALLBACK_CHILD_WRITE_VIA_PARENT,
+ wsi->parent->user_space,
+ (void *)&pas, 0))
+ return 1;
+
+ return len;
+ }
+
+ lws_stats_atomic_bump(wsi->context, pt, LWSSTATS_C_API_LWS_WRITE, 1);
+
+ if ((int)len < 0) {
+ lwsl_err("%s: suspicious len int %d, ulong %lu\n", __func__,
+ (int)len, (unsigned long)len);
+ return -1;
+ }
+
+ lws_stats_atomic_bump(wsi->context, pt, LWSSTATS_B_WRITE, len);
+
+#ifdef LWS_WITH_ACCESS_LOG
+ wsi->access_log.sent += len;
+#endif
+ if (wsi->vhost)
+ wsi->vhost->conn_stats.tx += len;
+
+ if (wsi->state == LWSS_ESTABLISHED && wsi->u.ws.tx_draining_ext) {
+ /* remove us from the list */
+ struct lws **w = &pt->tx_draining_ext_list;
+
+ wsi->u.ws.tx_draining_ext = 0;
+ /* remove us from context draining ext list */
+ while (*w) {
+ if (*w == wsi) {
+ *w = wsi->u.ws.tx_draining_ext_list;
+ break;
+ }
+ w = &((*w)->u.ws.tx_draining_ext_list);
+ }
+ wsi->u.ws.tx_draining_ext_list = NULL;
+ wp = (wsi->u.ws.tx_draining_stashed_wp & 0xc0) |
+ LWS_WRITE_CONTINUATION;
+
+ lwsl_ext("FORCED draining wp to 0x%02X\n", wp);
+ }
+
+ lws_restart_ws_ping_pong_timer(wsi);
+
+ if ((wp & 0x1f) == LWS_WRITE_HTTP ||
+ (wp & 0x1f) == LWS_WRITE_HTTP_FINAL ||
+ (wp & 0x1f) == LWS_WRITE_HTTP_HEADERS_CONTINUATION ||
+ (wp & 0x1f) == LWS_WRITE_HTTP_HEADERS)
+ goto send_raw;
+
+ /* if not in a state to send stuff, then just send nothing */
+
+ if (wsi->state != LWSS_ESTABLISHED &&
+ ((wsi->state != LWSS_RETURNED_CLOSE_ALREADY &&
+ wsi->state != LWSS_WAITING_TO_SEND_CLOSE_NOTIFICATION &&
+ wsi->state != LWSS_AWAITING_CLOSE_ACK) ||
+ wp != LWS_WRITE_CLOSE)) {
+ lwsl_debug("binning\n");
+ return 0;
+ }
+
+ /* if we are continuing a frame that already had its header done */
+
+ if (wsi->u.ws.inside_frame) {
+ lwsl_debug("INSIDE FRAME\n");
+ goto do_more_inside_frame;
+ }
+
+ wsi->u.ws.clean_buffer = 1;
+
+ /*
+ * give a chance to the extensions to modify payload
+ * the extension may decide to produce unlimited payload erratically
+ * (eg, compression extension), so we require only that if he produces
+ * something, it will be a complete fragment of the length known at
+ * the time (just the fragment length known), and if he has
+ * more we will come back next time he is writeable and allow him to
+ * produce more fragments until he's drained.
+ *
+ * This allows what is sent each time it is writeable to be limited to
+ * a size that can be sent without partial sends or blocking, allows
+ * interleaving of control frames and other connection service.
+ */
+ eff_buf.token = (char *)buf;
+ eff_buf.token_len = len;
+
+ switch ((int)wp) {
+ case LWS_WRITE_PING:
+ case LWS_WRITE_PONG:
+ case LWS_WRITE_CLOSE:
+ break;
+ default:
+ lwsl_debug("LWS_EXT_CB_PAYLOAD_TX\n");
+ n = lws_ext_cb_active(wsi, LWS_EXT_CB_PAYLOAD_TX, &eff_buf, wp);
+ if (n < 0)
+ return -1;
+
+ if (n && eff_buf.token_len) {
+ lwsl_debug("drain len %d\n", (int)eff_buf.token_len);
+ /* extension requires further draining */
+ wsi->u.ws.tx_draining_ext = 1;
+ wsi->u.ws.tx_draining_ext_list = pt->tx_draining_ext_list;
+ pt->tx_draining_ext_list = wsi;
+ /* we must come back to do more */
+ lws_callback_on_writable(wsi);
+ /*
+ * keep a copy of the write type for the overall
+ * action that has provoked generation of these
+ * fragments, so the last guy can use its FIN state.
+ */
+ wsi->u.ws.tx_draining_stashed_wp = wp;
+ /* this is definitely not actually the last fragment
+ * because the extension asserted he has more coming
+ * So make sure this intermediate one doesn't go out
+ * with a FIN.
+ */
+ wp |= LWS_WRITE_NO_FIN;
+ }
+
+ if (eff_buf.token_len && wsi->u.ws.stashed_write_pending) {
+ wsi->u.ws.stashed_write_pending = 0;
+ wp = (wp &0xc0) | (int)wsi->u.ws.stashed_write_type;
+ }
+ }
+
+ /*
+ * an extension did something we need to keep... for example, if
+ * compression extension, it has already updated its state according
+ * to this being issued
+ */
+ if ((char *)buf != eff_buf.token) {
+ /*
+ * ext might eat it, but not have anything to issue yet.
+ * In that case we have to follow his lead, but stash and
+ * replace the write type that was lost here the first time.
+ */
+ if (len && !eff_buf.token_len) {
+ if (!wsi->u.ws.stashed_write_pending)
+ wsi->u.ws.stashed_write_type = (char)wp & 0x3f;
+ wsi->u.ws.stashed_write_pending = 1;
+ return len;
+ }
+ /*
+ * extension recreated it:
+ * need to buffer this if not all sent
+ */
+ wsi->u.ws.clean_buffer = 0;
+ }
+
+ buf = (unsigned char *)eff_buf.token;
+ len = eff_buf.token_len;
+
+ if (!buf) {
+ lwsl_err("null buf (%d)\n", (int)len);
+ return -1;
+ }
+
+ switch (wsi->ietf_spec_revision) {
+ case 13:
+ if (masked7) {
+ pre += 4;
+ dropmask = &buf[0 - pre];
+ is_masked_bit = 0x80;
+ }
+
+ switch (wp & 0xf) {
+ case LWS_WRITE_TEXT:
+ n = LWSWSOPC_TEXT_FRAME;
+ break;
+ case LWS_WRITE_BINARY:
+ n = LWSWSOPC_BINARY_FRAME;
+ break;
+ case LWS_WRITE_CONTINUATION:
+ n = LWSWSOPC_CONTINUATION;
+ break;
+
+ case LWS_WRITE_CLOSE:
+ n = LWSWSOPC_CLOSE;
+ break;
+ case LWS_WRITE_PING:
+ n = LWSWSOPC_PING;
+ break;
+ case LWS_WRITE_PONG:
+ n = LWSWSOPC_PONG;
+ break;
+ default:
+ lwsl_warn("lws_write: unknown write opc / wp\n");
+ return -1;
+ }
+
+ if (!(wp & LWS_WRITE_NO_FIN))
+ n |= 1 << 7;
+
+ if (len < 126) {
+ pre += 2;
+ buf[-pre] = n;
+ buf[-pre + 1] = (unsigned char)(len | is_masked_bit);
+ } else {
+ if (len < 65536) {
+ pre += 4;
+ buf[-pre] = n;
+ buf[-pre + 1] = 126 | is_masked_bit;
+ buf[-pre + 2] = (unsigned char)(len >> 8);
+ buf[-pre + 3] = (unsigned char)len;
+ } else {
+ pre += 10;
+ buf[-pre] = n;
+ buf[-pre + 1] = 127 | is_masked_bit;
+#if defined __LP64__
+ buf[-pre + 2] = (len >> 56) & 0x7f;
+ buf[-pre + 3] = len >> 48;
+ buf[-pre + 4] = len >> 40;
+ buf[-pre + 5] = len >> 32;
+#else
+ buf[-pre + 2] = 0;
+ buf[-pre + 3] = 0;
+ buf[-pre + 4] = 0;
+ buf[-pre + 5] = 0;
+#endif
+ buf[-pre + 6] = (unsigned char)(len >> 24);
+ buf[-pre + 7] = (unsigned char)(len >> 16);
+ buf[-pre + 8] = (unsigned char)(len >> 8);
+ buf[-pre + 9] = (unsigned char)len;
+ }
+ }
+ break;
+ }
+
+do_more_inside_frame:
+
+ /*
+ * Deal with masking if we are in client -> server direction and
+ * the wp demands it
+ */
+
+ if (masked7) {
+ if (!wsi->u.ws.inside_frame)
+ if (lws_0405_frame_mask_generate(wsi)) {
+ lwsl_err("frame mask generation failed\n");
+ return -1;
+ }
+
+ /*
+ * in v7, just mask the payload
+ */
+ if (dropmask) { /* never set if already inside frame */
+ for (n = 4; n < (int)len + 4; n++)
+ dropmask[n] = dropmask[n] ^ wsi->u.ws.mask[
+ (wsi->u.ws.mask_idx++) & 3];
+
+ /* copy the frame nonce into place */
+ memcpy(dropmask, wsi->u.ws.mask, 4);
+ }
+ }
+
+send_raw:
+ switch ((int)(wp & 0x1f)) {
+ case LWS_WRITE_CLOSE:
+/* lwsl_hexdump(&buf[-pre], len); */
+ case LWS_WRITE_HTTP:
+ case LWS_WRITE_HTTP_FINAL:
+ case LWS_WRITE_HTTP_HEADERS:
+ case LWS_WRITE_HTTP_HEADERS_CONTINUATION:
+ case LWS_WRITE_PONG:
+ case LWS_WRITE_PING:
+#ifdef LWS_WITH_HTTP2
+ if (wsi->mode == LWSCM_HTTP2_SERVING) {
+ unsigned char flags = 0;
+
+ n = LWS_H2_FRAME_TYPE_DATA;
+ if ((wp & 0x1f) == LWS_WRITE_HTTP_HEADERS) {
+ n = LWS_H2_FRAME_TYPE_HEADERS;
+ if (!(wp & LWS_WRITE_NO_FIN))
+ flags = LWS_H2_FLAG_END_HEADERS;
+ if (wsi->u.h2.send_END_STREAM || (wp & LWS_WRITE_H2_STREAM_END)) {
+ flags |= LWS_H2_FLAG_END_STREAM;
+ wsi->u.h2.send_END_STREAM = 1;
+ }
+ }
+
+ if ((wp & 0x1f) == LWS_WRITE_HTTP_HEADERS_CONTINUATION) {
+ n = LWS_H2_FRAME_TYPE_CONTINUATION;
+ if (!(wp & LWS_WRITE_NO_FIN))
+ flags = LWS_H2_FLAG_END_HEADERS;
+ if (wsi->u.h2.send_END_STREAM || (wp & LWS_WRITE_H2_STREAM_END)) {
+ flags |= LWS_H2_FLAG_END_STREAM;
+ wsi->u.h2.send_END_STREAM = 1;
+ }
+ }
+
+ if (((wp & 0x1f) == LWS_WRITE_HTTP ||
+ (wp & 0x1f) == LWS_WRITE_HTTP_FINAL) &&
+ wsi->u.http.tx_content_length) {
+ wsi->u.http.tx_content_remain -= len;
+ lwsl_info("%s: wsi %p: tx_content_remain = %llu\n", __func__, wsi,
+ (unsigned long long)wsi->u.http.tx_content_remain);
+ if (!wsi->u.http.tx_content_remain) {
+ lwsl_info("%s: selecting final write mode\n", __func__);
+ wp = LWS_WRITE_HTTP_FINAL;
+ }
+ }
+
+ if ((wp & 0x1f) == LWS_WRITE_HTTP_FINAL || (wp & LWS_WRITE_H2_STREAM_END)) {
+ //lws_get_network_wsi(wsi)->u.h2.END_STREAM) {
+ lwsl_info("%s: setting END_STREAM\n", __func__);
+ flags |= LWS_H2_FLAG_END_STREAM;
+ wsi->u.h2.send_END_STREAM = 1;
+ }
+
+ return lws_h2_frame_write(wsi, n, flags,
+ wsi->u.h2.my_sid, len, buf);
+ }
+#endif
+ return lws_issue_raw(wsi, (unsigned char *)buf - pre, len + pre);
+ default:
+ break;
+ }
+
+ /*
+ * give any active extensions a chance to munge the buffer
+ * before send. We pass in a pointer to an lws_tokens struct
+ * prepared with the default buffer and content length that's in
+ * there. Rather than rewrite the default buffer, extensions
+ * that expect to grow the buffer can adapt .token to
+ * point to their own per-connection buffer in the extension
+ * user allocation. By default with no extensions or no
+ * extension callback handling, just the normal input buffer is
+ * used then so it is efficient.
+ *
+ * callback returns 1 in case it wants to spill more buffers
+ *
+ * This takes care of holding the buffer if send is incomplete, ie,
+ * if wsi->u.ws.clean_buffer is 0 (meaning an extension meddled with
+ * the buffer). If wsi->u.ws.clean_buffer is 1, it will instead
+ * return to the user code how much OF THE USER BUFFER was consumed.
+ */
+
+ n = lws_issue_raw_ext_access(wsi, buf - pre, len + pre);
+ wsi->u.ws.inside_frame = 1;
+ if (n <= 0)
+ return n;
+
+ if (n == (int)len + pre) {
+ /* everything in the buffer was handled (or rebuffered...) */
+ wsi->u.ws.inside_frame = 0;
+ return orig_len;
+ }
+
+ /*
+ * it is how many bytes of user buffer got sent... may be < orig_len
+ * in which case callback when writable has already been arranged
+ * and user code can call lws_write() again with the rest
+ * later.
+ */
+
+ return n - pre;
+}
+
+LWS_VISIBLE int lws_serve_http_file_fragment(struct lws *wsi)
+{
+ struct lws_context *context = wsi->context;
+ struct lws_context_per_thread *pt = &context->pt[(int)wsi->tsi];
+ struct lws_process_html_args args;
+ lws_filepos_t amount, poss;
+ unsigned char *p, *pstart;
+#if defined(LWS_WITH_RANGES)
+ unsigned char finished = 0;
+#endif
+ int n, m;
+
+ lwsl_debug("wsi->http2_substream %d\n", wsi->http2_substream);
+
+ while (!lws_send_pipe_choked(wsi)) {
+
+ if (wsi->trunc_len) {
+ if (lws_issue_raw(wsi, wsi->trunc_alloc +
+ wsi->trunc_offset,
+ wsi->trunc_len) < 0) {
+ lwsl_info("%s: closing\n", __func__);
+ goto file_had_it;
+ }
+ continue;
+ }
+
+ if (wsi->u.http.filepos == wsi->u.http.filelen)
+ goto all_sent;
+
+ n = 0;
+
+ pstart = pt->serv_buf + LWS_H2_FRAME_HEADER_LENGTH;
+
+ p = pstart;
+
+#if defined(LWS_WITH_RANGES)
+ if (wsi->u.http.range.count_ranges && !wsi->u.http.range.inside) {
+
+ lwsl_notice("%s: doing range start %llu\n", __func__, wsi->u.http.range.start);
+
+ if ((long long)lws_vfs_file_seek_cur(wsi->u.http.fop_fd,
+ wsi->u.http.range.start -
+ wsi->u.http.filepos) < 0)
+ goto file_had_it;
+
+ wsi->u.http.filepos = wsi->u.http.range.start;
+
+ if (wsi->u.http.range.count_ranges > 1) {
+ n = lws_snprintf((char *)p, context->pt_serv_buf_size - LWS_H2_FRAME_HEADER_LENGTH,
+ "_lws\x0d\x0a"
+ "Content-Type: %s\x0d\x0a"
+ "Content-Range: bytes %llu-%llu/%llu\x0d\x0a"
+ "\x0d\x0a",
+ wsi->u.http.multipart_content_type,
+ wsi->u.http.range.start,
+ wsi->u.http.range.end,
+ wsi->u.http.range.extent);
+ p += n;
+ }
+
+ wsi->u.http.range.budget = wsi->u.http.range.end -
+ wsi->u.http.range.start + 1;
+ wsi->u.http.range.inside = 1;
+ }
+#endif
+
+ poss = context->pt_serv_buf_size - n - LWS_H2_FRAME_HEADER_LENGTH;
+
+ if (poss > wsi->u.http.tx_content_remain)
+ poss = wsi->u.http.tx_content_remain;
+
+ /*
+ * if there is a hint about how much we will do well to send at one time,
+ * restrict ourselves to only trying to send that.
+ */
+ if (wsi->protocol->tx_packet_size &&
+ poss > wsi->protocol->tx_packet_size)
+ poss = wsi->protocol->tx_packet_size;
+
+#if defined(LWS_WITH_HTTP2)
+ m = lws_h2_tx_cr_get(wsi);
+ if (!m) {
+ lwsl_info("%s: came here with no tx credit", __func__);
+ return 0;
+ }
+ if (m < poss)
+ poss = m;
+ /*
+ * consumption of the actual payload amount sent will be handled
+ * when the http2 data frame is sent
+ */
+#endif
+
+#if defined(LWS_WITH_RANGES)
+ if (wsi->u.http.range.count_ranges) {
+ if (wsi->u.http.range.count_ranges > 1)
+ poss -= 7; /* allow for final boundary */
+ if (poss > wsi->u.http.range.budget)
+ poss = wsi->u.http.range.budget;
+ }
+#endif
+ if (wsi->sending_chunked) {
+ /* we need to drop the chunk size in here */
+ p += 10;
+ /* allow for the chunk to grow by 128 in translation */
+ poss -= 10 + 128;
+ }
+
+ if (lws_vfs_file_read(wsi->u.http.fop_fd, &amount, p, poss) < 0)
+ goto file_had_it; /* caller will close */
+
+ if (wsi->sending_chunked)
+ n = (int)amount;
+ else
+ n = (p - pstart) + (int)amount;
+
+ lwsl_debug("%s: sending %d\n", __func__, n);
+
+ if (n) {
+ lws_set_timeout(wsi, PENDING_TIMEOUT_HTTP_CONTENT,
+ context->timeout_secs);
+
+ if (wsi->sending_chunked) {
+ args.p = (char *)p;
+ args.len = n;
+ args.max_len = (unsigned int)poss + 128;
+ args.final = wsi->u.http.filepos + n ==
+ wsi->u.http.filelen;
+ if (user_callback_handle_rxflow(
+ wsi->vhost->protocols[(int)wsi->protocol_interpret_idx].callback, wsi,
+ LWS_CALLBACK_PROCESS_HTML,
+ wsi->user_space, &args, 0) < 0)
+ goto file_had_it;
+ n = args.len;
+ p = (unsigned char *)args.p;
+ } else
+ p = pstart;
+
+#if defined(LWS_WITH_RANGES)
+ if (wsi->u.http.range.send_ctr + 1 ==
+ wsi->u.http.range.count_ranges && // last range
+ wsi->u.http.range.count_ranges > 1 && // was 2+ ranges (ie, multipart)
+ wsi->u.http.range.budget - amount == 0) {// final part
+ n += lws_snprintf((char *)pstart + n, 6,
+ "_lws\x0d\x0a"); // append trailing boundary
+ lwsl_debug("added trailing boundary\n");
+ }
+#endif
+ m = lws_write(wsi, p, n,
+ wsi->u.http.filepos == wsi->u.http.filelen ?
+ LWS_WRITE_HTTP_FINAL :
+ LWS_WRITE_HTTP
+ );
+ if (m < 0)
+ goto file_had_it;
+
+ wsi->u.http.filepos += amount;
+
+#if defined(LWS_WITH_RANGES)
+ if (wsi->u.http.range.count_ranges >= 1) {
+ wsi->u.http.range.budget -= amount;
+ if (wsi->u.http.range.budget == 0) {
+ lwsl_notice("range budget exhausted\n");
+ wsi->u.http.range.inside = 0;
+ wsi->u.http.range.send_ctr++;
+
+ if (lws_ranges_next(&wsi->u.http.range) < 1) {
+ finished = 1;
+ goto all_sent;
+ }
+ }
+ }
+#endif
+
+ if (m != n) {
+ /* adjust for what was not sent */
+ if (lws_vfs_file_seek_cur(wsi->u.http.fop_fd,
+ m - n) ==
+ (unsigned long)-1)
+ goto file_had_it;
+ }
+ }
+
+all_sent:
+ if ((!wsi->trunc_len && wsi->u.http.filepos >= wsi->u.http.filelen)
+#if defined(LWS_WITH_RANGES)
+ || finished)
+#else
+ )
+#endif
+ {
+ wsi->state = LWSS_HTTP;
+ /* we might be in keepalive, so close it off here */
+ lws_vfs_file_close(&wsi->u.http.fop_fd);
+
+ lwsl_debug("file completed\n");
+
+ if (wsi->protocol->callback &&
+ user_callback_handle_rxflow(wsi->protocol->callback,
+ wsi, LWS_CALLBACK_HTTP_FILE_COMPLETION,
+ wsi->user_space, NULL,
+ 0) < 0) {
+ /*
+ * For http/1.x, the choices from
+ * transaction_completed are either
+ * 0 to use the connection for pipelined
+ * or nonzero to hang it up.
+ *
+ * However for http/2. while we are
+ * still interested in hanging up the
+ * nwsi if there was a network-level
+ * fatal error, simply completing the
+ * transaction is a matter of the stream
+ * state, not the root connection at the
+ * network level
+ */
+ if (wsi->http2_substream)
+ return 1;
+ else
+ return -1;
+ }
+
+ return 1; /* >0 indicates completed */
+ }
+ }
+
+ lws_callback_on_writable(wsi);
+
+ return 0; /* indicates further processing must be done */
+
+file_had_it:
+ lws_vfs_file_close(&wsi->u.http.fop_fd);
+
+ return -1;
+}
+
+#if LWS_POSIX
+LWS_VISIBLE int
+lws_ssl_capable_read_no_ssl(struct lws *wsi, unsigned char *buf, int len)
+{
+ struct lws_context *context = wsi->context;
+ struct lws_context_per_thread *pt = &context->pt[(int)wsi->tsi];
+ int n;
+
+ lws_stats_atomic_bump(context, pt, LWSSTATS_C_API_READ, 1);
+
+ n = recv(wsi->desc.sockfd, (char *)buf, len, 0);
+ if (n >= 0) {
+ if (wsi->vhost)
+ wsi->vhost->conn_stats.rx += n;
+ lws_stats_atomic_bump(context, pt, LWSSTATS_B_READ, n);
+ lws_restart_ws_ping_pong_timer(wsi);
+ return n;
+ }
+#if LWS_POSIX
+ if (LWS_ERRNO == LWS_EAGAIN ||
+ LWS_ERRNO == LWS_EWOULDBLOCK ||
+ LWS_ERRNO == LWS_EINTR)
+ return LWS_SSL_CAPABLE_MORE_SERVICE;
+#endif
+ lwsl_notice("error on reading from skt : %d\n", LWS_ERRNO);
+ return LWS_SSL_CAPABLE_ERROR;
+}
+
+LWS_VISIBLE int
+lws_ssl_capable_write_no_ssl(struct lws *wsi, unsigned char *buf, int len)
+{
+ int n = 0;
+
+#if LWS_POSIX
+ n = send(wsi->desc.sockfd, (char *)buf, len, MSG_NOSIGNAL);
+// lwsl_info("%s: sent len %d result %d", __func__, len, n);
+ if (n >= 0)
+ return n;
+
+ if (LWS_ERRNO == LWS_EAGAIN ||
+ LWS_ERRNO == LWS_EWOULDBLOCK ||
+ LWS_ERRNO == LWS_EINTR) {
+ if (LWS_ERRNO == LWS_EWOULDBLOCK) {
+ lws_set_blocking_send(wsi);
+ }
+
+ return LWS_SSL_CAPABLE_MORE_SERVICE;
+ }
+#else
+ (void)n;
+ (void)wsi;
+ (void)buf;
+ (void)len;
+ // !!!
+#endif
+
+ lwsl_debug("ERROR writing len %d to skt fd %d err %d / errno %d\n",
+ len, wsi->desc.sockfd, n, LWS_ERRNO);
+ return LWS_SSL_CAPABLE_ERROR;
+}
+#endif
+LWS_VISIBLE int
+lws_ssl_pending_no_ssl(struct lws *wsi)
+{
+ (void)wsi;
+#if defined(LWS_WITH_ESP32)
+ return 100;
+#else
+ return 0;
+#endif
+}