reverseproxy: Separate ignore_client_gone option

Sponsor request is to terminate streaming connections with backend when client leaves, but wants immediate flushing.

See #4922 and #4952.
This commit is contained in:
Matthew Holt 2024-10-28 15:21:47 -06:00
parent c5d6b22f37
commit 8c9e87d0a6
No known key found for this signature in database
GPG Key ID: 2A349DD577D586A5

View File

@ -106,13 +106,16 @@ type Handler struct {
// response is recognized as a streaming response, or if its
// content length is -1; for such responses, writes are flushed
// to the client immediately.
//
// Normally, a request will be canceled if the client disconnects
// before the response is received from the backend. If explicitly
// set to -1, client disconnection will be ignored and the request
// will be completed to help facilitate low-latency streaming.
FlushInterval caddy.Duration `json:"flush_interval,omitempty"`
// Normally, a request will be canceled if the client disconnects
// before the response is received from the backend. If enabled,
// client disconnection will be ignored and the request with the
// backend will carry on until the backend terminates it. This
// can help facilitate low-latency streaming. See #4922 and #4952.
// EXPERIMENTAL: Will likely be removed in the future.
IgnoreClientGone bool `json:"ignore_client_gone,omitempty"`
// A list of IP ranges (supports CIDR notation) from which
// X-Forwarded-* header values should be trusted. By default,
// no proxies are trusted, so existing values will be ignored
@ -773,12 +776,15 @@ func (h *Handler) reverseProxy(rw http.ResponseWriter, req *http.Request, origRe
}
req = req.WithContext(httptrace.WithClientTrace(req.Context(), trace))
// if FlushInterval is explicitly configured to -1 (i.e. flush continuously to achieve
// low-latency streaming), don't let the transport cancel the request if the client
// disconnects: user probably wants us to finish sending the data to the upstream
// regardless, and we should expect client disconnection in low-latency streaming
// scenarios (see issue #4922)
if h.FlushInterval == -1 {
// if enabled, don't let the transport cancel the request if the client disconnects:
// user probably wants us to finish sending the data to the upstream regardless,
// and we should expect client disconnection in low-latency streaming scenarios
// (see issue #4922)
// TODO: An ideal solution, if the client disconnects before the backend is done
// receiving data from the proxy, is to wait until the baxkend is done receiving
// the data and then close the connection with the backend, rather than an explicit
// option to always leave it open...
if h.IgnoreClientGone {
req = req.WithContext(ignoreClientGoneContext{req.Context()})
}