| // Copyright 2015 The Go Authors. All rights reserved. |
| // Use of this source code is governed by a BSD-style |
| // license that can be found in the LICENSE file. |
| |
| // Transport code. |
| |
| package http2 |
| |
| import ( |
| "bufio" |
| "bytes" |
| "compress/gzip" |
| "context" |
| "crypto/rand" |
| "crypto/tls" |
| "errors" |
| "fmt" |
| "io" |
| "io/fs" |
| "log" |
| "math" |
| "math/bits" |
| mathrand "math/rand" |
| "net" |
| "net/http" |
| "net/http/httptrace" |
| "net/textproto" |
| "sort" |
| "strconv" |
| "strings" |
| "sync" |
| "sync/atomic" |
| "time" |
| |
| "golang.org/x/net/http/httpguts" |
| "golang.org/x/net/http2/hpack" |
| "golang.org/x/net/idna" |
| ) |
| |
| const ( |
| // transportDefaultConnFlow is how many connection-level flow control |
| // tokens we give the server at start-up, past the default 64k. |
| transportDefaultConnFlow = 1 << 30 |
| |
| // transportDefaultStreamFlow is how many stream-level flow |
| // control tokens we announce to the peer, and how many bytes |
| // we buffer per stream. |
| transportDefaultStreamFlow = 4 << 20 |
| |
| defaultUserAgent = "Go-http-client/2.0" |
| |
| // initialMaxConcurrentStreams is a connections maxConcurrentStreams until |
| // it's received servers initial SETTINGS frame, which corresponds with the |
| // spec's minimum recommended value. |
| initialMaxConcurrentStreams = 100 |
| |
| // defaultMaxConcurrentStreams is a connections default maxConcurrentStreams |
| // if the server doesn't include one in its initial SETTINGS frame. |
| defaultMaxConcurrentStreams = 1000 |
| ) |
| |
| // Transport is an HTTP/2 Transport. |
| // |
| // A Transport internally caches connections to servers. It is safe |
| // for concurrent use by multiple goroutines. |
| type Transport struct { |
| // DialTLSContext specifies an optional dial function with context for |
| // creating TLS connections for requests. |
| // |
| // If DialTLSContext and DialTLS is nil, tls.Dial is used. |
| // |
| // If the returned net.Conn has a ConnectionState method like tls.Conn, |
| // it will be used to set http.Response.TLS. |
| DialTLSContext func(ctx context.Context, network, addr string, cfg *tls.Config) (net.Conn, error) |
| |
| // DialTLS specifies an optional dial function for creating |
| // TLS connections for requests. |
| // |
| // If DialTLSContext and DialTLS is nil, tls.Dial is used. |
| // |
| // Deprecated: Use DialTLSContext instead, which allows the transport |
| // to cancel dials as soon as they are no longer needed. |
| // If both are set, DialTLSContext takes priority. |
| DialTLS func(network, addr string, cfg *tls.Config) (net.Conn, error) |
| |
| // TLSClientConfig specifies the TLS configuration to use with |
| // tls.Client. If nil, the default configuration is used. |
| TLSClientConfig *tls.Config |
| |
| // ConnPool optionally specifies an alternate connection pool to use. |
| // If nil, the default is used. |
| ConnPool ClientConnPool |
| |
| // DisableCompression, if true, prevents the Transport from |
| // requesting compression with an "Accept-Encoding: gzip" |
| // request header when the Request contains no existing |
| // Accept-Encoding value. If the Transport requests gzip on |
| // its own and gets a gzipped response, it's transparently |
| // decoded in the Response.Body. However, if the user |
| // explicitly requested gzip it is not automatically |
| // uncompressed. |
| DisableCompression bool |
| |
| // AllowHTTP, if true, permits HTTP/2 requests using the insecure, |
| // plain-text "http" scheme. Note that this does not enable h2c support. |
| AllowHTTP bool |
| |
| // MaxHeaderListSize is the http2 SETTINGS_MAX_HEADER_LIST_SIZE to |
| // send in the initial settings frame. It is how many bytes |
| // of response headers are allowed. Unlike the http2 spec, zero here |
| // means to use a default limit (currently 10MB). If you actually |
| // want to advertise an unlimited value to the peer, Transport |
| // interprets the highest possible value here (0xffffffff or 1<<32-1) |
| // to mean no limit. |
| MaxHeaderListSize uint32 |
| |
| // MaxReadFrameSize is the http2 SETTINGS_MAX_FRAME_SIZE to send in the |
| // initial settings frame. It is the size in bytes of the largest frame |
| // payload that the sender is willing to receive. If 0, no setting is |
| // sent, and the value is provided by the peer, which should be 16384 |
| // according to the spec: |
| // https://datatracker.ietf.org/doc/html/rfc7540#section-6.5.2. |
| // Values are bounded in the range 16k to 16M. |
| MaxReadFrameSize uint32 |
| |
| // MaxDecoderHeaderTableSize optionally specifies the http2 |
| // SETTINGS_HEADER_TABLE_SIZE to send in the initial settings frame. It |
| // informs the remote endpoint of the maximum size of the header compression |
| // table used to decode header blocks, in octets. If zero, the default value |
| // of 4096 is used. |
| MaxDecoderHeaderTableSize uint32 |
| |
| // MaxEncoderHeaderTableSize optionally specifies an upper limit for the |
| // header compression table used for encoding request headers. Received |
| // SETTINGS_HEADER_TABLE_SIZE settings are capped at this limit. If zero, |
| // the default value of 4096 is used. |
| MaxEncoderHeaderTableSize uint32 |
| |
| // StrictMaxConcurrentStreams controls whether the server's |
| // SETTINGS_MAX_CONCURRENT_STREAMS should be respected |
| // globally. If false, new TCP connections are created to the |
| // server as needed to keep each under the per-connection |
| // SETTINGS_MAX_CONCURRENT_STREAMS limit. If true, the |
| // server's SETTINGS_MAX_CONCURRENT_STREAMS is interpreted as |
| // a global limit and callers of RoundTrip block when needed, |
| // waiting for their turn. |
| StrictMaxConcurrentStreams bool |
| |
| // IdleConnTimeout is the maximum amount of time an idle |
| // (keep-alive) connection will remain idle before closing |
| // itself. |
| // Zero means no limit. |
| IdleConnTimeout time.Duration |
| |
| // ReadIdleTimeout is the timeout after which a health check using ping |
| // frame will be carried out if no frame is received on the connection. |
| // Note that a ping response will is considered a received frame, so if |
| // there is no other traffic on the connection, the health check will |
| // be performed every ReadIdleTimeout interval. |
| // If zero, no health check is performed. |
| ReadIdleTimeout time.Duration |
| |
| // PingTimeout is the timeout after which the connection will be closed |
| // if a response to Ping is not received. |
| // Defaults to 15s. |
| PingTimeout time.Duration |
| |
| // WriteByteTimeout is the timeout after which the connection will be |
| // closed no data can be written to it. The timeout begins when data is |
| // available to write, and is extended whenever any bytes are written. |
| WriteByteTimeout time.Duration |
| |
| // CountError, if non-nil, is called on HTTP/2 transport errors. |
| // It's intended to increment a metric for monitoring, such |
| // as an expvar or Prometheus metric. |
| // The errType consists of only ASCII word characters. |
| CountError func(errType string) |
| |
| // t1, if non-nil, is the standard library Transport using |
| // this transport. Its settings are used (but not its |
| // RoundTrip method, etc). |
| t1 *http.Transport |
| |
| connPoolOnce sync.Once |
| connPoolOrDef ClientConnPool // non-nil version of ConnPool |
| |
| *transportTestHooks |
| } |
| |
| // Hook points used for testing. |
| // Outside of tests, t.transportTestHooks is nil and these all have minimal implementations. |
| // Inside tests, see the testSyncHooks function docs. |
| |
| type transportTestHooks struct { |
| newclientconn func(*ClientConn) |
| group synctestGroupInterface |
| } |
| |
| func (t *Transport) markNewGoroutine() { |
| if t != nil && t.transportTestHooks != nil { |
| t.transportTestHooks.group.Join() |
| } |
| } |
| |
| // newTimer creates a new time.Timer, or a synthetic timer in tests. |
| func (t *Transport) newTimer(d time.Duration) timer { |
| if t.transportTestHooks != nil { |
| return t.transportTestHooks.group.NewTimer(d) |
| } |
| return timeTimer{time.NewTimer(d)} |
| } |
| |
| // afterFunc creates a new time.AfterFunc timer, or a synthetic timer in tests. |
| func (t *Transport) afterFunc(d time.Duration, f func()) timer { |
| if t.transportTestHooks != nil { |
| return t.transportTestHooks.group.AfterFunc(d, f) |
| } |
| return timeTimer{time.AfterFunc(d, f)} |
| } |
| |
| func (t *Transport) contextWithTimeout(ctx context.Context, d time.Duration) (context.Context, context.CancelFunc) { |
| if t.transportTestHooks != nil { |
| return t.transportTestHooks.group.ContextWithTimeout(ctx, d) |
| } |
| return context.WithTimeout(ctx, d) |
| } |
| |
| func (t *Transport) maxHeaderListSize() uint32 { |
| n := int64(t.MaxHeaderListSize) |
| if t.t1 != nil && t.t1.MaxResponseHeaderBytes != 0 { |
| n = t.t1.MaxResponseHeaderBytes |
| if n > 0 { |
| n = adjustHTTP1MaxHeaderSize(n) |
| } |
| } |
| if n <= 0 { |
| return 10 << 20 |
| } |
| if n >= 0xffffffff { |
| return 0 |
| } |
| return uint32(n) |
| } |
| |
| func (t *Transport) disableCompression() bool { |
| return t.DisableCompression || (t.t1 != nil && t.t1.DisableCompression) |
| } |
| |
| // ConfigureTransport configures a net/http HTTP/1 Transport to use HTTP/2. |
| // It returns an error if t1 has already been HTTP/2-enabled. |
| // |
| // Use ConfigureTransports instead to configure the HTTP/2 Transport. |
| func ConfigureTransport(t1 *http.Transport) error { |
| _, err := ConfigureTransports(t1) |
| return err |
| } |
| |
| // ConfigureTransports configures a net/http HTTP/1 Transport to use HTTP/2. |
| // It returns a new HTTP/2 Transport for further configuration. |
| // It returns an error if t1 has already been HTTP/2-enabled. |
| func ConfigureTransports(t1 *http.Transport) (*Transport, error) { |
| return configureTransports(t1) |
| } |
| |
| func configureTransports(t1 *http.Transport) (*Transport, error) { |
| connPool := new(clientConnPool) |
| t2 := &Transport{ |
| ConnPool: noDialClientConnPool{connPool}, |
| t1: t1, |
| } |
| connPool.t = t2 |
| if err := registerHTTPSProtocol(t1, noDialH2RoundTripper{t2}); err != nil { |
| return nil, err |
| } |
| if t1.TLSClientConfig == nil { |
| t1.TLSClientConfig = new(tls.Config) |
| } |
| if !strSliceContains(t1.TLSClientConfig.NextProtos, "h2") { |
| t1.TLSClientConfig.NextProtos = append([]string{"h2"}, t1.TLSClientConfig.NextProtos...) |
| } |
| if !strSliceContains(t1.TLSClientConfig.NextProtos, "http/1.1") { |
| t1.TLSClientConfig.NextProtos = append(t1.TLSClientConfig.NextProtos, "http/1.1") |
| } |
| upgradeFn := func(authority string, c *tls.Conn) http.RoundTripper { |
| addr := authorityAddr("https", authority) |
| if used, err := connPool.addConnIfNeeded(addr, t2, c); err != nil { |
| go c.Close() |
| return erringRoundTripper{err} |
| } else if !used { |
| // Turns out we don't need this c. |
| // For example, two goroutines made requests to the same host |
| // at the same time, both kicking off TCP dials. (since protocol |
| // was unknown) |
| go c.Close() |
| } |
| return t2 |
| } |
| if m := t1.TLSNextProto; len(m) == 0 { |
| t1.TLSNextProto = map[string]func(string, *tls.Conn) http.RoundTripper{ |
| "h2": upgradeFn, |
| } |
| } else { |
| m["h2"] = upgradeFn |
| } |
| return t2, nil |
| } |
| |
| func (t *Transport) connPool() ClientConnPool { |
| t.connPoolOnce.Do(t.initConnPool) |
| return t.connPoolOrDef |
| } |
| |
| func (t *Transport) initConnPool() { |
| if t.ConnPool != nil { |
| t.connPoolOrDef = t.ConnPool |
| } else { |
| t.connPoolOrDef = &clientConnPool{t: t} |
| } |
| } |
| |
| // ClientConn is the state of a single HTTP/2 client connection to an |
| // HTTP/2 server. |
| type ClientConn struct { |
| t *Transport |
| tconn net.Conn // usually *tls.Conn, except specialized impls |
| tlsState *tls.ConnectionState // nil only for specialized impls |
| reused uint32 // whether conn is being reused; atomic |
| singleUse bool // whether being used for a single http.Request |
| getConnCalled bool // used by clientConnPool |
| |
| // readLoop goroutine fields: |
| readerDone chan struct{} // closed on error |
| readerErr error // set before readerDone is closed |
| |
| idleTimeout time.Duration // or 0 for never |
| idleTimer timer |
| |
| mu sync.Mutex // guards following |
| cond *sync.Cond // hold mu; broadcast on flow/closed changes |
| flow outflow // our conn-level flow control quota (cs.outflow is per stream) |
| inflow inflow // peer's conn-level flow control |
| doNotReuse bool // whether conn is marked to not be reused for any future requests |
| closing bool |
| closed bool |
| seenSettings bool // true if we've seen a settings frame, false otherwise |
| wantSettingsAck bool // we sent a SETTINGS frame and haven't heard back |
| goAway *GoAwayFrame // if non-nil, the GoAwayFrame we received |
| goAwayDebug string // goAway frame's debug data, retained as a string |
| streams map[uint32]*clientStream // client-initiated |
| streamsReserved int // incr by ReserveNewRequest; decr on RoundTrip |
| nextStreamID uint32 |
| pendingRequests int // requests blocked and waiting to be sent because len(streams) == maxConcurrentStreams |
| pings map[[8]byte]chan struct{} // in flight ping data to notification channel |
| br *bufio.Reader |
| lastActive time.Time |
| lastIdle time.Time // time last idle |
| // Settings from peer: (also guarded by wmu) |
| maxFrameSize uint32 |
| maxConcurrentStreams uint32 |
| peerMaxHeaderListSize uint64 |
| peerMaxHeaderTableSize uint32 |
| initialWindowSize uint32 |
| initialStreamRecvWindowSize int32 |
| readIdleTimeout time.Duration |
| pingTimeout time.Duration |
| |
| // pendingResets is the number of RST_STREAM frames we have sent to the peer, |
| // without confirming that the peer has received them. When we send a RST_STREAM, |
| // we bundle it with a PING frame, unless a PING is already in flight. We count |
| // the reset stream against the connection's concurrency limit until we get |
| // a PING response. This limits the number of requests we'll try to send to a |
| // completely unresponsive connection. |
| pendingResets int |
| |
| // reqHeaderMu is a 1-element semaphore channel controlling access to sending new requests. |
| // Write to reqHeaderMu to lock it, read from it to unlock. |
| // Lock reqmu BEFORE mu or wmu. |
| reqHeaderMu chan struct{} |
| |
| // wmu is held while writing. |
| // Acquire BEFORE mu when holding both, to avoid blocking mu on network writes. |
| // Only acquire both at the same time when changing peer settings. |
| wmu sync.Mutex |
| bw *bufio.Writer |
| fr *Framer |
| werr error // first write error that has occurred |
| hbuf bytes.Buffer // HPACK encoder writes into this |
| henc *hpack.Encoder |
| } |
| |
| // clientStream is the state for a single HTTP/2 stream. One of these |
| // is created for each Transport.RoundTrip call. |
| type clientStream struct { |
| cc *ClientConn |
| |
| // Fields of Request that we may access even after the response body is closed. |
| ctx context.Context |
| reqCancel <-chan struct{} |
| |
| trace *httptrace.ClientTrace // or nil |
| ID uint32 |
| bufPipe pipe // buffered pipe with the flow-controlled response payload |
| requestedGzip bool |
| isHead bool |
| |
| abortOnce sync.Once |
| abort chan struct{} // closed to signal stream should end immediately |
| abortErr error // set if abort is closed |
| |
| peerClosed chan struct{} // closed when the peer sends an END_STREAM flag |
| donec chan struct{} // closed after the stream is in the closed state |
| on100 chan struct{} // buffered; written to if a 100 is received |
| |
| respHeaderRecv chan struct{} // closed when headers are received |
| res *http.Response // set if respHeaderRecv is closed |
| |
| flow outflow // guarded by cc.mu |
| inflow inflow // guarded by cc.mu |
| bytesRemain int64 // -1 means unknown; owned by transportResponseBody.Read |
| readErr error // sticky read error; owned by transportResponseBody.Read |
| |
| reqBody io.ReadCloser |
| reqBodyContentLength int64 // -1 means unknown |
| reqBodyClosed chan struct{} // guarded by cc.mu; non-nil on Close, closed when done |
| |
| // owned by writeRequest: |
| sentEndStream bool // sent an END_STREAM flag to the peer |
| sentHeaders bool |
| |
| // owned by clientConnReadLoop: |
| firstByte bool // got the first response byte |
| pastHeaders bool // got first MetaHeadersFrame (actual headers) |
| pastTrailers bool // got optional second MetaHeadersFrame (trailers) |
| readClosed bool // peer sent an END_STREAM flag |
| readAborted bool // read loop reset the stream |
| totalHeaderSize int64 // total size of 1xx headers seen |
| |
| trailer http.Header // accumulated trailers |
| resTrailer *http.Header // client's Response.Trailer |
| } |
| |
| var got1xxFuncForTests func(int, textproto.MIMEHeader) error |
| |
| // get1xxTraceFunc returns the value of request's httptrace.ClientTrace.Got1xxResponse func, |
| // if any. It returns nil if not set or if the Go version is too old. |
| func (cs *clientStream) get1xxTraceFunc() func(int, textproto.MIMEHeader) error { |
| if fn := got1xxFuncForTests; fn != nil { |
| return fn |
| } |
| return traceGot1xxResponseFunc(cs.trace) |
| } |
| |
| func (cs *clientStream) abortStream(err error) { |
| cs.cc.mu.Lock() |
| defer cs.cc.mu.Unlock() |
| cs.abortStreamLocked(err) |
| } |
| |
| func (cs *clientStream) abortStreamLocked(err error) { |
| cs.abortOnce.Do(func() { |
| cs.abortErr = err |
| close(cs.abort) |
| }) |
| if cs.reqBody != nil { |
| cs.closeReqBodyLocked() |
| } |
| // TODO(dneil): Clean up tests where cs.cc.cond is nil. |
| if cs.cc.cond != nil { |
| // Wake up writeRequestBody if it is waiting on flow control. |
| cs.cc.cond.Broadcast() |
| } |
| } |
| |
| func (cs *clientStream) abortRequestBodyWrite() { |
| cc := cs.cc |
| cc.mu.Lock() |
| defer cc.mu.Unlock() |
| if cs.reqBody != nil && cs.reqBodyClosed == nil { |
| cs.closeReqBodyLocked() |
| cc.cond.Broadcast() |
| } |
| } |
| |
| func (cs *clientStream) closeReqBodyLocked() { |
| if cs.reqBodyClosed != nil { |
| return |
| } |
| cs.reqBodyClosed = make(chan struct{}) |
| reqBodyClosed := cs.reqBodyClosed |
| go func() { |
| cs.cc.t.markNewGoroutine() |
| cs.reqBody.Close() |
| close(reqBodyClosed) |
| }() |
| } |
| |
| type stickyErrWriter struct { |
| group synctestGroupInterface |
| conn net.Conn |
| timeout time.Duration |
| err *error |
| } |
| |
| func (sew stickyErrWriter) Write(p []byte) (n int, err error) { |
| if *sew.err != nil { |
| return 0, *sew.err |
| } |
| n, err = writeWithByteTimeout(sew.group, sew.conn, sew.timeout, p) |
| *sew.err = err |
| return n, err |
| } |
| |
| // noCachedConnError is the concrete type of ErrNoCachedConn, which |
| // needs to be detected by net/http regardless of whether it's its |
| // bundled version (in h2_bundle.go with a rewritten type name) or |
| // from a user's x/net/http2. As such, as it has a unique method name |
| // (IsHTTP2NoCachedConnError) that net/http sniffs for via func |
| // isNoCachedConnError. |
| type noCachedConnError struct{} |
| |
| func (noCachedConnError) IsHTTP2NoCachedConnError() {} |
| func (noCachedConnError) Error() string { return "http2: no cached connection was available" } |
| |
| // isNoCachedConnError reports whether err is of type noCachedConnError |
| // or its equivalent renamed type in net/http2's h2_bundle.go. Both types |
| // may coexist in the same running program. |
| func isNoCachedConnError(err error) bool { |
| _, ok := err.(interface{ IsHTTP2NoCachedConnError() }) |
| return ok |
| } |
| |
| var ErrNoCachedConn error = noCachedConnError{} |
| |
| // RoundTripOpt are options for the Transport.RoundTripOpt method. |
| type RoundTripOpt struct { |
| // OnlyCachedConn controls whether RoundTripOpt may |
| // create a new TCP connection. If set true and |
| // no cached connection is available, RoundTripOpt |
| // will return ErrNoCachedConn. |
| OnlyCachedConn bool |
| } |
| |
| func (t *Transport) RoundTrip(req *http.Request) (*http.Response, error) { |
| return t.RoundTripOpt(req, RoundTripOpt{}) |
| } |
| |
| // authorityAddr returns a given authority (a host/IP, or host:port / ip:port) |
| // and returns a host:port. The port 443 is added if needed. |
| func authorityAddr(scheme string, authority string) (addr string) { |
| host, port, err := net.SplitHostPort(authority) |
| if err != nil { // authority didn't have a port |
| host = authority |
| port = "" |
| } |
| if port == "" { // authority's port was empty |
| port = "443" |
| if scheme == "http" { |
| port = "80" |
| } |
| } |
| if a, err := idna.ToASCII(host); err == nil { |
| host = a |
| } |
| // IPv6 address literal, without a port: |
| if strings.HasPrefix(host, "[") && strings.HasSuffix(host, "]") { |
| return host + ":" + port |
| } |
| return net.JoinHostPort(host, port) |
| } |
| |
| // RoundTripOpt is like RoundTrip, but takes options. |
| func (t *Transport) RoundTripOpt(req *http.Request, opt RoundTripOpt) (*http.Response, error) { |
| if !(req.URL.Scheme == "https" || (req.URL.Scheme == "http" && t.AllowHTTP)) { |
| return nil, errors.New("http2: unsupported scheme") |
| } |
| |
| addr := authorityAddr(req.URL.Scheme, req.URL.Host) |
| for retry := 0; ; retry++ { |
| cc, err := t.connPool().GetClientConn(req, addr) |
| if err != nil { |
| t.vlogf("http2: Transport failed to get client conn for %s: %v", addr, err) |
| return nil, err |
| } |
| reused := !atomic.CompareAndSwapUint32(&cc.reused, 0, 1) |
| traceGotConn(req, cc, reused) |
| res, err := cc.RoundTrip(req) |
| if err != nil && retry <= 6 { |
| roundTripErr := err |
| if req, err = shouldRetryRequest(req, err); err == nil { |
| // After the first retry, do exponential backoff with 10% jitter. |
| if retry == 0 { |
| t.vlogf("RoundTrip retrying after failure: %v", roundTripErr) |
| continue |
| } |
| backoff := float64(uint(1) << (uint(retry) - 1)) |
| backoff += backoff * (0.1 * mathrand.Float64()) |
| d := time.Second * time.Duration(backoff) |
| tm := t.newTimer(d) |
| select { |
| case <-tm.C(): |
| t.vlogf("RoundTrip retrying after failure: %v", roundTripErr) |
| continue |
| case <-req.Context().Done(): |
| tm.Stop() |
| err = req.Context().Err() |
| } |
| } |
| } |
| if err != nil { |
| t.vlogf("RoundTrip failure: %v", err) |
| return nil, err |
| } |
| return res, nil |
| } |
| } |
| |
| // CloseIdleConnections closes any connections which were previously |
| // connected from previous requests but are now sitting idle. |
| // It does not interrupt any connections currently in use. |
| func (t *Transport) CloseIdleConnections() { |
| if cp, ok := t.connPool().(clientConnPoolIdleCloser); ok { |
| cp.closeIdleConnections() |
| } |
| } |
| |
| var ( |
| errClientConnClosed = errors.New("http2: client conn is closed") |
| errClientConnUnusable = errors.New("http2: client conn not usable") |
| errClientConnGotGoAway = errors.New("http2: Transport received Server's graceful shutdown GOAWAY") |
| ) |
| |
| // shouldRetryRequest is called by RoundTrip when a request fails to get |
| // response headers. It is always called with a non-nil error. |
| // It returns either a request to retry (either the same request, or a |
| // modified clone), or an error if the request can't be replayed. |
| func shouldRetryRequest(req *http.Request, err error) (*http.Request, error) { |
| if !canRetryError(err) { |
| return nil, err |
| } |
| // If the Body is nil (or http.NoBody), it's safe to reuse |
| // this request and its Body. |
| if req.Body == nil || req.Body == http.NoBody { |
| return req, nil |
| } |
| |
| // If the request body can be reset back to its original |
| // state via the optional req.GetBody, do that. |
| if req.GetBody != nil { |
| body, err := req.GetBody() |
| if err != nil { |
| return nil, err |
| } |
| newReq := *req |
| newReq.Body = body |
| return &newReq, nil |
| } |
| |
| // The Request.Body can't reset back to the beginning, but we |
| // don't seem to have started to read from it yet, so reuse |
| // the request directly. |
| if err == errClientConnUnusable { |
| return req, nil |
| } |
| |
| return nil, fmt.Errorf("http2: Transport: cannot retry err [%v] after Request.Body was written; define Request.GetBody to avoid this error", err) |
| } |
| |
| func canRetryError(err error) bool { |
| if err == errClientConnUnusable || err == errClientConnGotGoAway { |
| return true |
| } |
| if se, ok := err.(StreamError); ok { |
| if se.Code == ErrCodeProtocol && se.Cause == errFromPeer { |
| // See golang/go#47635, golang/go#42777 |
| return true |
| } |
| return se.Code == ErrCodeRefusedStream |
| } |
| return false |
| } |
| |
| func (t *Transport) dialClientConn(ctx context.Context, addr string, singleUse bool) (*ClientConn, error) { |
| if t.transportTestHooks != nil { |
| return t.newClientConn(nil, singleUse) |
| } |
| host, _, err := net.SplitHostPort(addr) |
| if err != nil { |
| return nil, err |
| } |
| tconn, err := t.dialTLS(ctx, "tcp", addr, t.newTLSConfig(host)) |
| if err != nil { |
| return nil, err |
| } |
| return t.newClientConn(tconn, singleUse) |
| } |
| |
| func (t *Transport) newTLSConfig(host string) *tls.Config { |
| cfg := new(tls.Config) |
| if t.TLSClientConfig != nil { |
| *cfg = *t.TLSClientConfig.Clone() |
| } |
| if !strSliceContains(cfg.NextProtos, NextProtoTLS) { |
| cfg.NextProtos = append([]string{NextProtoTLS}, cfg.NextProtos...) |
| } |
| if cfg.ServerName == "" { |
| cfg.ServerName = host |
| } |
| return cfg |
| } |
| |
| func (t *Transport) dialTLS(ctx context.Context, network, addr string, tlsCfg *tls.Config) (net.Conn, error) { |
| if t.DialTLSContext != nil { |
| return t.DialTLSContext(ctx, network, addr, tlsCfg) |
| } else if t.DialTLS != nil { |
| return t.DialTLS(network, addr, tlsCfg) |
| } |
| |
| tlsCn, err := t.dialTLSWithContext(ctx, network, addr, tlsCfg) |
| if err != nil { |
| return nil, err |
| } |
| state := tlsCn.ConnectionState() |
| if p := state.NegotiatedProtocol; p != NextProtoTLS { |
| return nil, fmt.Errorf("http2: unexpected ALPN protocol %q; want %q", p, NextProtoTLS) |
| } |
| if !state.NegotiatedProtocolIsMutual { |
| return nil, errors.New("http2: could not negotiate protocol mutually") |
| } |
| return tlsCn, nil |
| } |
| |
| // disableKeepAlives reports whether connections should be closed as |
| // soon as possible after handling the first request. |
| func (t *Transport) disableKeepAlives() bool { |
| return t.t1 != nil && t.t1.DisableKeepAlives |
| } |
| |
| func (t *Transport) expectContinueTimeout() time.Duration { |
| if t.t1 == nil { |
| return 0 |
| } |
| return t.t1.ExpectContinueTimeout |
| } |
| |
| func (t *Transport) NewClientConn(c net.Conn) (*ClientConn, error) { |
| return t.newClientConn(c, t.disableKeepAlives()) |
| } |
| |
| func (t *Transport) newClientConn(c net.Conn, singleUse bool) (*ClientConn, error) { |
| conf := configFromTransport(t) |
| cc := &ClientConn{ |
| t: t, |
| tconn: c, |
| readerDone: make(chan struct{}), |
| nextStreamID: 1, |
| maxFrameSize: 16 << 10, // spec default |
| initialWindowSize: 65535, // spec default |
| initialStreamRecvWindowSize: conf.MaxUploadBufferPerStream, |
| maxConcurrentStreams: initialMaxConcurrentStreams, // "infinite", per spec. Use a smaller value until we have received server settings. |
| peerMaxHeaderListSize: 0xffffffffffffffff, // "infinite", per spec. Use 2^64-1 instead. |
| streams: make(map[uint32]*clientStream), |
| singleUse: singleUse, |
| wantSettingsAck: true, |
| readIdleTimeout: conf.SendPingTimeout, |
| pingTimeout: conf.PingTimeout, |
| pings: make(map[[8]byte]chan struct{}), |
| reqHeaderMu: make(chan struct{}, 1), |
| } |
| var group synctestGroupInterface |
| if t.transportTestHooks != nil { |
| t.markNewGoroutine() |
| t.transportTestHooks.newclientconn(cc) |
| c = cc.tconn |
| group = t.group |
| } |
| if VerboseLogs { |
| t.vlogf("http2: Transport creating client conn %p to %v", cc, c.RemoteAddr()) |
| } |
| |
| cc.cond = sync.NewCond(&cc.mu) |
| cc.flow.add(int32(initialWindowSize)) |
| |
| // TODO: adjust this writer size to account for frame size + |
| // MTU + crypto/tls record padding. |
| cc.bw = bufio.NewWriter(stickyErrWriter{ |
| group: group, |
| conn: c, |
| timeout: conf.WriteByteTimeout, |
| err: &cc.werr, |
| }) |
| cc.br = bufio.NewReader(c) |
| cc.fr = NewFramer(cc.bw, cc.br) |
| cc.fr.SetMaxReadFrameSize(conf.MaxReadFrameSize) |
| if t.CountError != nil { |
| cc.fr.countError = t.CountError |
| } |
| maxHeaderTableSize := conf.MaxDecoderHeaderTableSize |
| cc.fr.ReadMetaHeaders = hpack.NewDecoder(maxHeaderTableSize, nil) |
| cc.fr.MaxHeaderListSize = t.maxHeaderListSize() |
| |
| cc.henc = hpack.NewEncoder(&cc.hbuf) |
| cc.henc.SetMaxDynamicTableSizeLimit(conf.MaxEncoderHeaderTableSize) |
| cc.peerMaxHeaderTableSize = initialHeaderTableSize |
| |
| if cs, ok := c.(connectionStater); ok { |
| state := cs.ConnectionState() |
| cc.tlsState = &state |
| } |
| |
| initialSettings := []Setting{ |
| {ID: SettingEnablePush, Val: 0}, |
| {ID: SettingInitialWindowSize, Val: uint32(cc.initialStreamRecvWindowSize)}, |
| } |
| initialSettings = append(initialSettings, Setting{ID: SettingMaxFrameSize, Val: conf.MaxReadFrameSize}) |
| if max := t.maxHeaderListSize(); max != 0 { |
| initialSettings = append(initialSettings, Setting{ID: SettingMaxHeaderListSize, Val: max}) |
| } |
| if maxHeaderTableSize != initialHeaderTableSize { |
| initialSettings = append(initialSettings, Setting{ID: SettingHeaderTableSize, Val: maxHeaderTableSize}) |
| } |
| |
| cc.bw.Write(clientPreface) |
| cc.fr.WriteSettings(initialSettings...) |
| cc.fr.WriteWindowUpdate(0, uint32(conf.MaxUploadBufferPerConnection)) |
| cc.inflow.init(conf.MaxUploadBufferPerConnection + initialWindowSize) |
| cc.bw.Flush() |
| if cc.werr != nil { |
| cc.Close() |
| return nil, cc.werr |
| } |
| |
| // Start the idle timer after the connection is fully initialized. |
| if d := t.idleConnTimeout(); d != 0 { |
| cc.idleTimeout = d |
| cc.idleTimer = t.afterFunc(d, cc.onIdleTimeout) |
| } |
| |
| go cc.readLoop() |
| return cc, nil |
| } |
| |
| func (cc *ClientConn) healthCheck() { |
| pingTimeout := cc.pingTimeout |
| // We don't need to periodically ping in the health check, because the readLoop of ClientConn will |
| // trigger the healthCheck again if there is no frame received. |
| ctx, cancel := cc.t.contextWithTimeout(context.Background(), pingTimeout) |
| defer cancel() |
| cc.vlogf("http2: Transport sending health check") |
| err := cc.Ping(ctx) |
| if err != nil { |
| cc.vlogf("http2: Transport health check failure: %v", err) |
| cc.closeForLostPing() |
| } else { |
| cc.vlogf("http2: Transport health check success") |
| } |
| } |
| |
| // SetDoNotReuse marks cc as not reusable for future HTTP requests. |
| func (cc *ClientConn) SetDoNotReuse() { |
| cc.mu.Lock() |
| defer cc.mu.Unlock() |
| cc.doNotReuse = true |
| } |
| |
| func (cc *ClientConn) setGoAway(f *GoAwayFrame) { |
| cc.mu.Lock() |
| defer cc.mu.Unlock() |
| |
| old := cc.goAway |
| cc.goAway = f |
| |
| // Merge the previous and current GoAway error frames. |
| if cc.goAwayDebug == "" { |
| cc.goAwayDebug = string(f.DebugData()) |
| } |
| if old != nil && old.ErrCode != ErrCodeNo { |
| cc.goAway.ErrCode = old.ErrCode |
| } |
| last := f.LastStreamID |
| for streamID, cs := range cc.streams { |
| if streamID <= last { |
| // The server's GOAWAY indicates that it received this stream. |
| // It will either finish processing it, or close the connection |
| // without doing so. Either way, leave the stream alone for now. |
| continue |
| } |
| if streamID == 1 && cc.goAway.ErrCode != ErrCodeNo { |
| // Don't retry the first stream on a connection if we get a non-NO error. |
| // If the server is sending an error on a new connection, |
| // retrying the request on a new one probably isn't going to work. |
| cs.abortStreamLocked(fmt.Errorf("http2: Transport received GOAWAY from server ErrCode:%v", cc.goAway.ErrCode)) |
| } else { |
| // Aborting the stream with errClentConnGotGoAway indicates that |
| // the request should be retried on a new connection. |
| cs.abortStreamLocked(errClientConnGotGoAway) |
| } |
| } |
| } |
| |
| // CanTakeNewRequest reports whether the connection can take a new request, |
| // meaning it has not been closed or received or sent a GOAWAY. |
| // |
| // If the caller is going to immediately make a new request on this |
| // connection, use ReserveNewRequest instead. |
| func (cc *ClientConn) CanTakeNewRequest() bool { |
| cc.mu.Lock() |
| defer cc.mu.Unlock() |
| return cc.canTakeNewRequestLocked() |
| } |
| |
| // ReserveNewRequest is like CanTakeNewRequest but also reserves a |
| // concurrent stream in cc. The reservation is decremented on the |
| // next call to RoundTrip. |
| func (cc *ClientConn) ReserveNewRequest() bool { |
| cc.mu.Lock() |
| defer cc.mu.Unlock() |
| if st := cc.idleStateLocked(); !st.canTakeNewRequest { |
| return false |
| } |
| cc.streamsReserved++ |
| return true |
| } |
| |
| // ClientConnState describes the state of a ClientConn. |
| type ClientConnState struct { |
| // Closed is whether the connection is closed. |
| Closed bool |
| |
| // Closing is whether the connection is in the process of |
| // closing. It may be closing due to shutdown, being a |
| // single-use connection, being marked as DoNotReuse, or |
| // having received a GOAWAY frame. |
| Closing bool |
| |
| // StreamsActive is how many streams are active. |
| StreamsActive int |
| |
| // StreamsReserved is how many streams have been reserved via |
| // ClientConn.ReserveNewRequest. |
| StreamsReserved int |
| |
| // StreamsPending is how many requests have been sent in excess |
| // of the peer's advertised MaxConcurrentStreams setting and |
| // are waiting for other streams to complete. |
| StreamsPending int |
| |
| // MaxConcurrentStreams is how many concurrent streams the |
| // peer advertised as acceptable. Zero means no SETTINGS |
| // frame has been received yet. |
| MaxConcurrentStreams uint32 |
| |
| // LastIdle, if non-zero, is when the connection last |
| // transitioned to idle state. |
| LastIdle time.Time |
| } |
| |
| // State returns a snapshot of cc's state. |
| func (cc *ClientConn) State() ClientConnState { |
| cc.wmu.Lock() |
| maxConcurrent := cc.maxConcurrentStreams |
| if !cc.seenSettings { |
| maxConcurrent = 0 |
| } |
| cc.wmu.Unlock() |
| |
| cc.mu.Lock() |
| defer cc.mu.Unlock() |
| return ClientConnState{ |
| Closed: cc.closed, |
| Closing: cc.closing || cc.singleUse || cc.doNotReuse || cc.goAway != nil, |
| StreamsActive: len(cc.streams) + cc.pendingResets, |
| StreamsReserved: cc.streamsReserved, |
| StreamsPending: cc.pendingRequests, |
| LastIdle: cc.lastIdle, |
| MaxConcurrentStreams: maxConcurrent, |
| } |
| } |
| |
| // clientConnIdleState describes the suitability of a client |
| // connection to initiate a new RoundTrip request. |
| type clientConnIdleState struct { |
| canTakeNewRequest bool |
| } |
| |
| func (cc *ClientConn) idleState() clientConnIdleState { |
| cc.mu.Lock() |
| defer cc.mu.Unlock() |
| return cc.idleStateLocked() |
| } |
| |
| func (cc *ClientConn) idleStateLocked() (st clientConnIdleState) { |
| if cc.singleUse && cc.nextStreamID > 1 { |
| return |
| } |
| var maxConcurrentOkay bool |
| if cc.t.StrictMaxConcurrentStreams { |
| // We'll tell the caller we can take a new request to |
| // prevent the caller from dialing a new TCP |
| // connection, but then we'll block later before |
| // writing it. |
| maxConcurrentOkay = true |
| } else { |
| // We can take a new request if the total of |
| // - active streams; |
| // - reservation slots for new streams; and |
| // - streams for which we have sent a RST_STREAM and a PING, |
| // but received no subsequent frame |
| // is less than the concurrency limit. |
| maxConcurrentOkay = cc.currentRequestCountLocked() < int(cc.maxConcurrentStreams) |
| } |
| |
| st.canTakeNewRequest = cc.goAway == nil && !cc.closed && !cc.closing && maxConcurrentOkay && |
| !cc.doNotReuse && |
| int64(cc.nextStreamID)+2*int64(cc.pendingRequests) < math.MaxInt32 && |
| !cc.tooIdleLocked() |
| return |
| } |
| |
| // currentRequestCountLocked reports the number of concurrency slots currently in use, |
| // including active streams, reserved slots, and reset streams waiting for acknowledgement. |
| func (cc *ClientConn) currentRequestCountLocked() int { |
| return len(cc.streams) + cc.streamsReserved + cc.pendingResets |
| } |
| |
| func (cc *ClientConn) canTakeNewRequestLocked() bool { |
| st := cc.idleStateLocked() |
| return st.canTakeNewRequest |
| } |
| |
| // tooIdleLocked reports whether this connection has been been sitting idle |
| // for too much wall time. |
| func (cc *ClientConn) tooIdleLocked() bool { |
| // The Round(0) strips the monontonic clock reading so the |
| // times are compared based on their wall time. We don't want |
| // to reuse a connection that's been sitting idle during |
| // VM/laptop suspend if monotonic time was also frozen. |
| return cc.idleTimeout != 0 && !cc.lastIdle.IsZero() && time.Since(cc.lastIdle.Round(0)) > cc.idleTimeout |
| } |
| |
| // onIdleTimeout is called from a time.AfterFunc goroutine. It will |
| // only be called when we're idle, but because we're coming from a new |
| // goroutine, there could be a new request coming in at the same time, |
| // so this simply calls the synchronized closeIfIdle to shut down this |
| // connection. The timer could just call closeIfIdle, but this is more |
| // clear. |
| func (cc *ClientConn) onIdleTimeout() { |
| cc.closeIfIdle() |
| } |
| |
| func (cc *ClientConn) closeConn() { |
| t := time.AfterFunc(250*time.Millisecond, cc.forceCloseConn) |
| defer t.Stop() |
| cc.tconn.Close() |
| } |
| |
| // A tls.Conn.Close can hang for a long time if the peer is unresponsive. |
| // Try to shut it down more aggressively. |
| func (cc *ClientConn) forceCloseConn() { |
| tc, ok := cc.tconn.(*tls.Conn) |
| if !ok { |
| return |
| } |
| if nc := tc.NetConn(); nc != nil { |
| nc.Close() |
| } |
| } |
| |
| func (cc *ClientConn) closeIfIdle() { |
| cc.mu.Lock() |
| if len(cc.streams) > 0 || cc.streamsReserved > 0 { |
| cc.mu.Unlock() |
| return |
| } |
| cc.closed = true |
| nextID := cc.nextStreamID |
| // TODO: do clients send GOAWAY too? maybe? Just Close: |
| cc.mu.Unlock() |
| |
| if VerboseLogs { |
| cc.vlogf("http2: Transport closing idle conn %p (forSingleUse=%v, maxStream=%v)", cc, cc.singleUse, nextID-2) |
| } |
| cc.closeConn() |
| } |
| |
| func (cc *ClientConn) isDoNotReuseAndIdle() bool { |
| cc.mu.Lock() |
| defer cc.mu.Unlock() |
| return cc.doNotReuse && len(cc.streams) == 0 |
| } |
| |
| var shutdownEnterWaitStateHook = func() {} |
| |
| // Shutdown gracefully closes the client connection, waiting for running streams to complete. |
| func (cc *ClientConn) Shutdown(ctx context.Context) error { |
| if err := cc.sendGoAway(); err != nil { |
| return err |
| } |
| // Wait for all in-flight streams to complete or connection to close |
| done := make(chan struct{}) |
| cancelled := false // guarded by cc.mu |
| go func() { |
| cc.t.markNewGoroutine() |
| cc.mu.Lock() |
| defer cc.mu.Unlock() |
| for { |
| if len(cc.streams) == 0 || cc.closed { |
| cc.closed = true |
| close(done) |
| break |
| } |
| if cancelled { |
| break |
| } |
| cc.cond.Wait() |
| } |
| }() |
| shutdownEnterWaitStateHook() |
| select { |
| case <-done: |
| cc.closeConn() |
| return nil |
| case <-ctx.Done(): |
| cc.mu.Lock() |
| // Free the goroutine above |
| cancelled = true |
| cc.cond.Broadcast() |
| cc.mu.Unlock() |
| return ctx.Err() |
| } |
| } |
| |
| func (cc *ClientConn) sendGoAway() error { |
| cc.mu.Lock() |
| closing := cc.closing |
| cc.closing = true |
| maxStreamID := cc.nextStreamID |
| cc.mu.Unlock() |
| if closing { |
| // GOAWAY sent already |
| return nil |
| } |
| |
| cc.wmu.Lock() |
| defer cc.wmu.Unlock() |
| // Send a graceful shutdown frame to server |
| if err := cc.fr.WriteGoAway(maxStreamID, ErrCodeNo, nil); err != nil { |
| return err |
| } |
| if err := cc.bw.Flush(); err != nil { |
| return err |
| } |
| // Prevent new requests |
| return nil |
| } |
| |
| // closes the client connection immediately. In-flight requests are interrupted. |
| // err is sent to streams. |
| func (cc *ClientConn) closeForError(err error) { |
| cc.mu.Lock() |
| cc.closed = true |
| for _, cs := range cc.streams { |
| cs.abortStreamLocked(err) |
| } |
| cc.cond.Broadcast() |
| cc.mu.Unlock() |
| cc.closeConn() |
| } |
| |
| // Close closes the client connection immediately. |
| // |
| // In-flight requests are interrupted. For a graceful shutdown, use Shutdown instead. |
| func (cc *ClientConn) Close() error { |
| err := errors.New("http2: client connection force closed via ClientConn.Close") |
| cc.closeForError(err) |
| return nil |
| } |
| |
| // closes the client connection immediately. In-flight requests are interrupted. |
| func (cc *ClientConn) closeForLostPing() { |
| err := errors.New("http2: client connection lost") |
| if f := cc.t.CountError; f != nil { |
| f("conn_close_lost_ping") |
| } |
| cc.closeForError(err) |
| } |
| |
| // errRequestCanceled is a copy of net/http's errRequestCanceled because it's not |
| // exported. At least they'll be DeepEqual for h1-vs-h2 comparisons tests. |
| var errRequestCanceled = errors.New("net/http: request canceled") |
| |
| func commaSeparatedTrailers(req *http.Request) (string, error) { |
| keys := make([]string, 0, len(req.Trailer)) |
| for k := range req.Trailer { |
| k = canonicalHeader(k) |
| switch k { |
| case "Transfer-Encoding", "Trailer", "Content-Length": |
| return "", fmt.Errorf("invalid Trailer key %q", k) |
| } |
| keys = append(keys, k) |
| } |
| if len(keys) > 0 { |
| sort.Strings(keys) |
| return strings.Join(keys, ","), nil |
| } |
| return "", nil |
| } |
| |
| func (cc *ClientConn) responseHeaderTimeout() time.Duration { |
| if cc.t.t1 != nil { |
| return cc.t.t1.ResponseHeaderTimeout |
| } |
| // No way to do this (yet?) with just an http2.Transport. Probably |
| // no need. Request.Cancel this is the new way. We only need to support |
| // this for compatibility with the old http.Transport fields when |
| // we're doing transparent http2. |
| return 0 |
| } |
| |
| // checkConnHeaders checks whether req has any invalid connection-level headers. |
| // per RFC 7540 section 8.1.2.2: Connection-Specific Header Fields. |
| // Certain headers are special-cased as okay but not transmitted later. |
| func checkConnHeaders(req *http.Request) error { |
| if v := req.Header.Get("Upgrade"); v != "" { |
| return fmt.Errorf("http2: invalid Upgrade request header: %q", req.Header["Upgrade"]) |
| } |
| if vv := req.Header["Transfer-Encoding"]; len(vv) > 0 && (len(vv) > 1 || vv[0] != "" && vv[0] != "chunked") { |
| return fmt.Errorf("http2: invalid Transfer-Encoding request header: %q", vv) |
| } |
| if vv := req.Header["Connection"]; len(vv) > 0 && (len(vv) > 1 || vv[0] != "" && !asciiEqualFold(vv[0], "close") && !asciiEqualFold(vv[0], "keep-alive")) { |
| return fmt.Errorf("http2: invalid Connection request header: %q", vv) |
| } |
| return nil |
| } |
| |
| // actualContentLength returns a sanitized version of |
| // req.ContentLength, where 0 actually means zero (not unknown) and -1 |
| // means unknown. |
| func actualContentLength(req *http.Request) int64 { |
| if req.Body == nil || req.Body == http.NoBody { |
| return 0 |
| } |
| if req.ContentLength != 0 { |
| return req.ContentLength |
| } |
| return -1 |
| } |
| |
| func (cc *ClientConn) decrStreamReservations() { |
| cc.mu.Lock() |
| defer cc.mu.Unlock() |
| cc.decrStreamReservationsLocked() |
| } |
| |
| func (cc *ClientConn) decrStreamReservationsLocked() { |
| if cc.streamsReserved > 0 { |
| cc.streamsReserved-- |
| } |
| } |
| |
| func (cc *ClientConn) RoundTrip(req *http.Request) (*http.Response, error) { |
| return cc.roundTrip(req, nil) |
| } |
| |
| func (cc *ClientConn) roundTrip(req *http.Request, streamf func(*clientStream)) (*http.Response, error) { |
| ctx := req.Context() |
| cs := &clientStream{ |
| cc: cc, |
| ctx: ctx, |
| reqCancel: req.Cancel, |
| isHead: req.Method == "HEAD", |
| reqBody: req.Body, |
| reqBodyContentLength: actualContentLength(req), |
| trace: httptrace.ContextClientTrace(ctx), |
| peerClosed: make(chan struct{}), |
| abort: make(chan struct{}), |
| respHeaderRecv: make(chan struct{}), |
| donec: make(chan struct{}), |
| } |
| |
| // TODO(bradfitz): this is a copy of the logic in net/http. Unify somewhere? |
| if !cc.t.disableCompression() && |
| req.Header.Get("Accept-Encoding") == "" && |
| req.Header.Get("Range") == "" && |
| !cs.isHead { |
| // Request gzip only, not deflate. Deflate is ambiguous and |
| // not as universally supported anyway. |
| // See: https://zlib.net/zlib_faq.html#faq39 |
| // |
| // Note that we don't request this for HEAD requests, |
| // due to a bug in nginx: |
| // http://trac.nginx.org/nginx/ticket/358 |
| // https://golang.org/issue/5522 |
| // |
| // We don't request gzip if the request is for a range, since |
| // auto-decoding a portion of a gzipped document will just fail |
| // anyway. See https://golang.org/issue/8923 |
| cs.requestedGzip = true |
| } |
| |
| go cs.doRequest(req, streamf) |
| |
| waitDone := func() error { |
| select { |
| case <-cs.donec: |
| return nil |
| case <-ctx.Done(): |
| return ctx.Err() |
| case <-cs.reqCancel: |
| return errRequestCanceled |
| } |
| } |
| |
| handleResponseHeaders := func() (*http.Response, error) { |
| res := cs.res |
| if res.StatusCode > 299 { |
| // On error or status code 3xx, 4xx, 5xx, etc abort any |
| // ongoing write, assuming that the server doesn't care |
| // about our request body. If the server replied with 1xx or |
| // 2xx, however, then assume the server DOES potentially |
| // want our body (e.g. full-duplex streaming: |
| // golang.org/issue/13444). If it turns out the server |
| // doesn't, they'll RST_STREAM us soon enough. This is a |
| // heuristic to avoid adding knobs to Transport. Hopefully |
| // we can keep it. |
| cs.abortRequestBodyWrite() |
| } |
| res.Request = req |
| res.TLS = cc.tlsState |
| if res.Body == noBody && actualContentLength(req) == 0 { |
| // If there isn't a request or response body still being |
| // written, then wait for the stream to be closed before |
| // RoundTrip returns. |
| if err := waitDone(); err != nil { |
| return nil, err |
| } |
| } |
| return res, nil |
| } |
| |
| cancelRequest := func(cs *clientStream, err error) error { |
| cs.cc.mu.Lock() |
| bodyClosed := cs.reqBodyClosed |
| cs.cc.mu.Unlock() |
| // Wait for the request body to be closed. |
| // |
| // If nothing closed the body before now, abortStreamLocked |
| // will have started a goroutine to close it. |
| // |
| // Closing the body before returning avoids a race condition |
| // with net/http checking its readTrackingBody to see if the |
| // body was read from or closed. See golang/go#60041. |
| // |
| // The body is closed in a separate goroutine without the |
| // connection mutex held, but dropping the mutex before waiting |
| // will keep us from holding it indefinitely if the body |
| // close is slow for some reason. |
| if bodyClosed != nil { |
| <-bodyClosed |
| } |
| return err |
| } |
| |
| for { |
| select { |
| case <-cs.respHeaderRecv: |
| return handleResponseHeaders() |
| case <-cs.abort: |
| select { |
| case <-cs.respHeaderRecv: |
| // If both cs.respHeaderRecv and cs.abort are signaling, |
| // pick respHeaderRecv. The server probably wrote the |
| // response and immediately reset the stream. |
| // golang.org/issue/49645 |
| return handleResponseHeaders() |
| default: |
| waitDone() |
| return nil, cs.abortErr |
| } |
| case <-ctx.Done(): |
| err := ctx.Err() |
| cs.abortStream(err) |
| return nil, cancelRequest(cs, err) |
| case <-cs.reqCancel: |
| cs.abortStream(errRequestCanceled) |
| return nil, cancelRequest(cs, errRequestCanceled) |
| } |
| } |
| } |
| |
| // doRequest runs for the duration of the request lifetime. |
| // |
| // It sends the request and performs post-request cleanup (closing Request.Body, etc.). |
| func (cs *clientStream) doRequest(req *http.Request, streamf func(*clientStream)) { |
| cs.cc.t.markNewGoroutine() |
| err := cs.writeRequest(req, streamf) |
| cs.cleanupWriteRequest(err) |
| } |
| |
| // writeRequest sends a request. |
| // |
| // It returns nil after the request is written, the response read, |
| // and the request stream is half-closed by the peer. |
| // |
| // It returns non-nil if the request ends otherwise. |
| // If the returned error is StreamError, the error Code may be used in resetting the stream. |
| func (cs *clientStream) writeRequest(req *http.Request, streamf func(*clientStream)) (err error) { |
| cc := cs.cc |
| ctx := cs.ctx |
| |
| if err := checkConnHeaders(req); err != nil { |
| return err |
| } |
| |
| // Acquire the new-request lock by writing to reqHeaderMu. |
| // This lock guards the critical section covering allocating a new stream ID |
| // (requires mu) and creating the stream (requires wmu). |
| if cc.reqHeaderMu == nil { |
| panic("RoundTrip on uninitialized ClientConn") // for tests |
| } |
| select { |
| case cc.reqHeaderMu <- struct{}{}: |
| case <-cs.reqCancel: |
| return errRequestCanceled |
| case <-ctx.Done(): |
| return ctx.Err() |
| } |
| |
| cc.mu.Lock() |
| if cc.idleTimer != nil { |
| cc.idleTimer.Stop() |
| } |
| cc.decrStreamReservationsLocked() |
| if err := cc.awaitOpenSlotForStreamLocked(cs); err != nil { |
| cc.mu.Unlock() |
| <-cc.reqHeaderMu |
| return err |
| } |
| cc.addStreamLocked(cs) // assigns stream ID |
| if isConnectionCloseRequest(req) { |
| cc.doNotReuse = true |
| } |
| cc.mu.Unlock() |
| |
| if streamf != nil { |
| streamf(cs) |
| } |
| |
| continueTimeout := cc.t.expectContinueTimeout() |
| if continueTimeout != 0 { |
| if !httpguts.HeaderValuesContainsToken(req.Header["Expect"], "100-continue") { |
| continueTimeout = 0 |
| } else { |
| cs.on100 = make(chan struct{}, 1) |
| } |
| } |
| |
| // Past this point (where we send request headers), it is possible for |
| // RoundTrip to return successfully. Since the RoundTrip contract permits |
| // the caller to "mutate or reuse" the Request after closing the Response's Body, |
| // we must take care when referencing the Request from here on. |
| err = cs.encodeAndWriteHeaders(req) |
| <-cc.reqHeaderMu |
| if err != nil { |
| return err |
| } |
| |
| hasBody := cs.reqBodyContentLength != 0 |
| if !hasBody { |
| cs.sentEndStream = true |
| } else { |
| if continueTimeout != 0 { |
| traceWait100Continue(cs.trace) |
| timer := time.NewTimer(continueTimeout) |
| select { |
| case <-timer.C: |
| err = nil |
| case <-cs.on100: |
| err = nil |
| case <-cs.abort: |
| err = cs.abortErr |
| case <-ctx.Done(): |
| err = ctx.Err() |
| case <-cs.reqCancel: |
| err = errRequestCanceled |
| } |
| timer.Stop() |
| if err != nil { |
| traceWroteRequest(cs.trace, err) |
| return err |
| } |
| } |
| |
| if err = cs.writeRequestBody(req); err != nil { |
| if err != errStopReqBodyWrite { |
| traceWroteRequest(cs.trace, err) |
| return err |
| } |
| } else { |
| cs.sentEndStream = true |
| } |
| } |
| |
| traceWroteRequest(cs.trace, err) |
| |
| var respHeaderTimer <-chan time.Time |
| var respHeaderRecv chan struct{} |
| if d := cc.responseHeaderTimeout(); d != 0 { |
| timer := cc.t.newTimer(d) |
| defer timer.Stop() |
| respHeaderTimer = timer.C() |
| respHeaderRecv = cs.respHeaderRecv |
| } |
| // Wait until the peer half-closes its end of the stream, |
| // or until the request is aborted (via context, error, or otherwise), |
| // whichever comes first. |
| for { |
| select { |
| case <-cs.peerClosed: |
| return nil |
| case <-respHeaderTimer: |
| return errTimeout |
| case <-respHeaderRecv: |
| respHeaderRecv = nil |
| respHeaderTimer = nil // keep waiting for END_STREAM |
| case <-cs.abort: |
| return cs.abortErr |
| case <-ctx.Done(): |
| return ctx.Err() |
| case <-cs.reqCancel: |
| return errRequestCanceled |
| } |
| } |
| } |
| |
| func (cs *clientStream) encodeAndWriteHeaders(req *http.Request) error { |
| cc := cs.cc |
| ctx := cs.ctx |
| |
| cc.wmu.Lock() |
| defer cc.wmu.Unlock() |
| |
| // If the request was canceled while waiting for cc.mu, just quit. |
| select { |
| case <-cs.abort: |
| return cs.abortErr |
| case <-ctx.Done(): |
| return ctx.Err() |
| case <-cs.reqCancel: |
| return errRequestCanceled |
| default: |
| } |
| |
| // Encode headers. |
| // |
| // we send: HEADERS{1}, CONTINUATION{0,} + DATA{0,} (DATA is |
| // sent by writeRequestBody below, along with any Trailers, |
| // again in form HEADERS{1}, CONTINUATION{0,}) |
| trailers, err := commaSeparatedTrailers(req) |
| if err != nil { |
| return err |
| } |
| hasTrailers := trailers != "" |
| contentLen := actualContentLength(req) |
| hasBody := contentLen != 0 |
| hdrs, err := cc.encodeHeaders(req, cs.requestedGzip, trailers, contentLen) |
| if err != nil { |
| return err |
| } |
| |
| // Write the request. |
| endStream := !hasBody && !hasTrailers |
| cs.sentHeaders = true |
| err = cc.writeHeaders(cs.ID, endStream, int(cc.maxFrameSize), hdrs) |
| traceWroteHeaders(cs.trace) |
| return err |
| } |
| |
| // cleanupWriteRequest performs post-request tasks. |
| // |
| // If err (the result of writeRequest) is non-nil and the stream is not closed, |
| // cleanupWriteRequest will send a reset to the peer. |
| func (cs *clientStream) cleanupWriteRequest(err error) { |
| cc := cs.cc |
| |
| if cs.ID == 0 { |
| // We were canceled before creating the stream, so return our reservation. |
| cc.decrStreamReservations() |
| } |
| |
| // TODO: write h12Compare test showing whether |
| // Request.Body is closed by the Transport, |
| // and in multiple cases: server replies <=299 and >299 |
| // while still writing request body |
| cc.mu.Lock() |
| mustCloseBody := false |
| if cs.reqBody != nil && cs.reqBodyClosed == nil { |
| mustCloseBody = true |
| cs.reqBodyClosed = make(chan struct{}) |
| } |
| bodyClosed := cs.reqBodyClosed |
| closeOnIdle := cc.singleUse || cc.doNotReuse || cc.t.disableKeepAlives() || cc.goAway != nil |
| cc.mu.Unlock() |
| if mustCloseBody { |
| cs.reqBody.Close() |
| close(bodyClosed) |
| } |
| if bodyClosed != nil { |
| <-bodyClosed |
| } |
| |
| if err != nil && cs.sentEndStream { |
| // If the connection is closed immediately after the response is read, |
| // we may be aborted before finishing up here. If the stream was closed |
| // cleanly on both sides, there is no error. |
| select { |
| case <-cs.peerClosed: |
| err = nil |
| default: |
| } |
| } |
| if err != nil { |
| cs.abortStream(err) // possibly redundant, but harmless |
| if cs.sentHeaders { |
| if se, ok := err.(StreamError); ok { |
| if se.Cause != errFromPeer { |
| cc.writeStreamReset(cs.ID, se.Code, false, err) |
| } |
| } else { |
| // We're cancelling an in-flight request. |
| // |
| // This could be due to the server becoming unresponsive. |
| // To avoid sending too many requests on a dead connection, |
| // we let the request continue to consume a concurrency slot |
| // until we can confirm the server is still responding. |
| // We do this by sending a PING frame along with the RST_STREAM |
| // (unless a ping is already in flight). |
| // |
| // For simplicity, we don't bother tracking the PING payload: |
| // We reset cc.pendingResets any time we receive a PING ACK. |
| // |
| // We skip this if the conn is going to be closed on idle, |
| // because it's short lived and will probably be closed before |
| // we get the ping response. |
| ping := false |
| if !closeOnIdle { |
| cc.mu.Lock() |
| if cc.pendingResets == 0 { |
| ping = true |
| } |
| cc.pendingResets++ |
| cc.mu.Unlock() |
| } |
| cc.writeStreamReset(cs.ID, ErrCodeCancel, ping, err) |
| } |
| } |
| cs.bufPipe.CloseWithError(err) // no-op if already closed |
| } else { |
| if cs.sentHeaders && !cs.sentEndStream { |
| cc.writeStreamReset(cs.ID, ErrCodeNo, false, nil) |
| } |
| cs.bufPipe.CloseWithError(errRequestCanceled) |
| } |
| if cs.ID != 0 { |
| cc.forgetStreamID(cs.ID) |
| } |
| |
| cc.wmu.Lock() |
| werr := cc.werr |
| cc.wmu.Unlock() |
| if werr != nil { |
| cc.Close() |
| } |
| |
| close(cs.donec) |
| } |
| |
| // awaitOpenSlotForStreamLocked waits until len(streams) < maxConcurrentStreams. |
| // Must hold cc.mu. |
| func (cc *ClientConn) awaitOpenSlotForStreamLocked(cs *clientStream) error { |
| for { |
| cc.lastActive = time.Now() |
| if cc.closed || !cc.canTakeNewRequestLocked() { |
| return errClientConnUnusable |
| } |
| cc.lastIdle = time.Time{} |
| if cc.currentRequestCountLocked() < int(cc.maxConcurrentStreams) { |
| return nil |
| } |
| cc.pendingRequests++ |
| cc.cond.Wait() |
| cc.pendingRequests-- |
| select { |
| case <-cs.abort: |
| return cs.abortErr |
| default: |
| } |
| } |
| } |
| |
| // requires cc.wmu be held |
| func (cc *ClientConn) writeHeaders(streamID uint32, endStream bool, maxFrameSize int, hdrs []byte) error { |
| first := true // first frame written (HEADERS is first, then CONTINUATION) |
| for len(hdrs) > 0 && cc.werr == nil { |
| chunk := hdrs |
| if len(chunk) > maxFrameSize { |
| chunk = chunk[:maxFrameSize] |
| } |
| hdrs = hdrs[len(chunk):] |
| endHeaders := len(hdrs) == 0 |
| if first { |
| cc.fr.WriteHeaders(HeadersFrameParam{ |
| StreamID: streamID, |
| BlockFragment: chunk, |
| EndStream: endStream, |
| EndHeaders: endHeaders, |
| }) |
| first = false |
| } else { |
| cc.fr.WriteContinuation(streamID, endHeaders, chunk) |
| } |
| } |
| cc.bw.Flush() |
| return cc.werr |
| } |
| |
| // internal error values; they don't escape to callers |
| var ( |
| // abort request body write; don't send cancel |
| errStopReqBodyWrite = errors.New("http2: aborting request body write") |
| |
| // abort request body write, but send stream reset of cancel. |
| errStopReqBodyWriteAndCancel = errors.New("http2: canceling request") |
| |
| errReqBodyTooLong = errors.New("http2: request body larger than specified content length") |
| ) |
| |
| // frameScratchBufferLen returns the length of a buffer to use for |
| // outgoing request bodies to read/write to/from. |
| // |
| // It returns max(1, min(peer's advertised max frame size, |
| // Request.ContentLength+1, 512KB)). |
| func (cs *clientStream) frameScratchBufferLen(maxFrameSize int) int { |
| const max = 512 << 10 |
| n := int64(maxFrameSize) |
| if n > max { |
| n = max |
| } |
| if cl := cs.reqBodyContentLength; cl != -1 && cl+1 < n { |
| // Add an extra byte past the declared content-length to |
| // give the caller's Request.Body io.Reader a chance to |
| // give us more bytes than they declared, so we can catch it |
| // early. |
| n = cl + 1 |
| } |
| if n < 1 { |
| return 1 |
| } |
| return int(n) // doesn't truncate; max is 512K |
| } |
| |
| // Seven bufPools manage different frame sizes. This helps to avoid scenarios where long-running |
| // streaming requests using small frame sizes occupy large buffers initially allocated for prior |
| // requests needing big buffers. The size ranges are as follows: |
| // {0 KB, 16 KB], {16 KB, 32 KB], {32 KB, 64 KB], {64 KB, 128 KB], {128 KB, 256 KB], |
| // {256 KB, 512 KB], {512 KB, infinity} |
| // In practice, the maximum scratch buffer size should not exceed 512 KB due to |
| // frameScratchBufferLen(maxFrameSize), thus the "infinity pool" should never be used. |
| // It exists mainly as a safety measure, for potential future increases in max buffer size. |
| var bufPools [7]sync.Pool // of *[]byte |
| func bufPoolIndex(size int) int { |
| if size <= 16384 { |
| return 0 |
| } |
| size -= 1 |
| bits := bits.Len(uint(size)) |
| index := bits - 14 |
| if index >= len(bufPools) { |
| return len(bufPools) - 1 |
| } |
| return index |
| } |
| |
| func (cs *clientStream) writeRequestBody(req *http.Request) (err error) { |
| cc := cs.cc |
| body := cs.reqBody |
| sentEnd := false // whether we sent the final DATA frame w/ END_STREAM |
| |
| hasTrailers := req.Trailer != nil |
| remainLen := cs.reqBodyContentLength |
| hasContentLen := remainLen != -1 |
| |
| cc.mu.Lock() |
| maxFrameSize := int(cc.maxFrameSize) |
| cc.mu.Unlock() |
| |
| // Scratch buffer for reading into & writing from. |
| scratchLen := cs.frameScratchBufferLen(maxFrameSize) |
| var buf []byte |
| index := bufPoolIndex(scratchLen) |
| if bp, ok := bufPools[index].Get().(*[]byte); ok && len(*bp) >= scratchLen { |
| defer bufPools[index].Put(bp) |
| buf = *bp |
| } else { |
| buf = make([]byte, scratchLen) |
| defer bufPools[index].Put(&buf) |
| } |
| |
| var sawEOF bool |
| for !sawEOF { |
| n, err := body.Read(buf) |
| if hasContentLen { |
| remainLen -= int64(n) |
| if remainLen == 0 && err == nil { |
| // The request body's Content-Length was predeclared and |
| // we just finished reading it all, but the underlying io.Reader |
| // returned the final chunk with a nil error (which is one of |
| // the two valid things a Reader can do at EOF). Because we'd prefer |
| // to send the END_STREAM bit early, double-check that we're actually |
| // at EOF. Subsequent reads should return (0, EOF) at this point. |
| // If either value is different, we return an error in one of two ways below. |
| var scratch [1]byte |
| var n1 int |
| n1, err = body.Read(scratch[:]) |
| remainLen -= int64(n1) |
| } |
| if remainLen < 0 { |
| err = errReqBodyTooLong |
| return err |
| } |
| } |
| if err != nil { |
| cc.mu.Lock() |
| bodyClosed := cs.reqBodyClosed != nil |
| cc.mu.Unlock() |
| switch { |
| case bodyClosed: |
| return errStopReqBodyWrite |
| case err == io.EOF: |
| sawEOF = true |
| err = nil |
| default: |
| return err |
| } |
| } |
| |
| remain := buf[:n] |
| for len(remain) > 0 && err == nil { |
| var allowed int32 |
| allowed, err = cs.awaitFlowControl(len(remain)) |
| if err != nil { |
| return err |
| } |
| cc.wmu.Lock() |
| data := remain[:allowed] |
| remain = remain[allowed:] |
| sentEnd = sawEOF && len(remain) == 0 && !hasTrailers |
| err = cc.fr.WriteData(cs.ID, sentEnd, data) |
| if err == nil { |
| // TODO(bradfitz): this flush is for latency, not bandwidth. |
| // Most requests won't need this. Make this opt-in or |
| // opt-out? Use some heuristic on the body type? Nagel-like |
| // timers? Based on 'n'? Only last chunk of this for loop, |
| // unless flow control tokens are low? For now, always. |
| // If we change this, see comment below. |
| err = cc.bw.Flush() |
| } |
| cc.wmu.Unlock() |
| } |
| if err != nil { |
| return err |
| } |
| } |
| |
| if sentEnd { |
| // Already sent END_STREAM (which implies we have no |
| // trailers) and flushed, because currently all |
| // WriteData frames above get a flush. So we're done. |
| return nil |
| } |
| |
| // Since the RoundTrip contract permits the caller to "mutate or reuse" |
| // a request after the Response's Body is closed, verify that this hasn't |
| // happened before accessing the trailers. |
| cc.mu.Lock() |
| trailer := req.Trailer |
| err = cs.abortErr |
| cc.mu.Unlock() |
| if err != nil { |
| return err |
| } |
| |
| cc.wmu.Lock() |
| defer cc.wmu.Unlock() |
| var trls []byte |
| if len(trailer) > 0 { |
| trls, err = cc.encodeTrailers(trailer) |
| if err != nil { |
| return err |
| } |
| } |
| |
| // Two ways to send END_STREAM: either with trailers, or |
| // with an empty DATA frame. |
| if len(trls) > 0 { |
| err = cc.writeHeaders(cs.ID, true, maxFrameSize, trls) |
| } else { |
| err = cc.fr.WriteData(cs.ID, true, nil) |
| } |
| if ferr := cc.bw.Flush(); ferr != nil && err == nil { |
| err = ferr |
| } |
| return err |
| } |
| |
| // awaitFlowControl waits for [1, min(maxBytes, cc.cs.maxFrameSize)] flow |
| // control tokens from the server. |
| // It returns either the non-zero number of tokens taken or an error |
| // if the stream is dead. |
| func (cs *clientStream) awaitFlowControl(maxBytes int) (taken int32, err error) { |
| cc := cs.cc |
| ctx := cs.ctx |
| cc.mu.Lock() |
| defer cc.mu.Unlock() |
| for { |
| if cc.closed { |
| return 0, errClientConnClosed |
| } |
| if cs.reqBodyClosed != nil { |
| return 0, errStopReqBodyWrite |
| } |
| select { |
| case <-cs.abort: |
| return 0, cs.abortErr |
| case <-ctx.Done(): |
| return 0, ctx.Err() |
| case <-cs.reqCancel: |
| return 0, errRequestCanceled |
| default: |
| } |
| if a := cs.flow.available(); a > 0 { |
| take := a |
| if int(take) > maxBytes { |
| |
| take = int32(maxBytes) // can't truncate int; take is int32 |
| } |
| if take > int32(cc.maxFrameSize) { |
| take = int32(cc.maxFrameSize) |
| } |
| cs.flow.take(take) |
| return take, nil |
| } |
| cc.cond.Wait() |
| } |
| } |
| |
| func validateHeaders(hdrs http.Header) string { |
| for k, vv := range hdrs { |
| if !httpguts.ValidHeaderFieldName(k) { |
| return fmt.Sprintf("name %q", k) |
| } |
| for _, v := range vv { |
| if !httpguts.ValidHeaderFieldValue(v) { |
| // Don't include the value in the error, |
| // because it may be sensitive. |
| return fmt.Sprintf("value for header %q", k) |
| } |
| } |
| } |
| return "" |
| } |
| |
| var errNilRequestURL = errors.New("http2: Request.URI is nil") |
| |
| // requires cc.wmu be held. |
| func (cc *ClientConn) encodeHeaders(req *http.Request, addGzipHeader bool, trailers string, contentLength int64) ([]byte, error) { |
| cc.hbuf.Reset() |
| if req.URL == nil { |
| return nil, errNilRequestURL |
| } |
| |
| host := req.Host |
| if host == "" { |
| host = req.URL.Host |
| } |
| host, err := httpguts.PunycodeHostPort(host) |
| if err != nil { |
| return nil, err |
| } |
| if !httpguts.ValidHostHeader(host) { |
| return nil, errors.New("http2: invalid Host header") |
| } |
| |
| var path string |
| if req.Method != "CONNECT" { |
| path = req.URL.RequestURI() |
| if !validPseudoPath(path) { |
| orig := path |
| path = strings.TrimPrefix(path, req.URL.Scheme+"://"+host) |
| if !validPseudoPath(path) { |
| if req.URL.Opaque != "" { |
| return nil, fmt.Errorf("invalid request :path %q from URL.Opaque = %q", orig, req.URL.Opaque) |
| } else { |
| return nil, fmt.Errorf("invalid request :path %q", orig) |
| } |
| } |
| } |
| } |
| |
| // Check for any invalid headers+trailers and return an error before we |
| // potentially pollute our hpack state. (We want to be able to |
| // continue to reuse the hpack encoder for future requests) |
| if err := validateHeaders(req.Header); err != "" { |
| return nil, fmt.Errorf("invalid HTTP header %s", err) |
| } |
| if err := validateHeaders(req.Trailer); err != "" { |
| return nil, fmt.Errorf("invalid HTTP trailer %s", err) |
| } |
| |
| enumerateHeaders := func(f func(name, value string)) { |
| // 8.1.2.3 Request Pseudo-Header Fields |
| // The :path pseudo-header field includes the path and query parts of the |
| // target URI (the path-absolute production and optionally a '?' character |
| // followed by the query production, see Sections 3.3 and 3.4 of |
| // [RFC3986]). |
| f(":authority", host) |
| m := req.Method |
| if m == "" { |
| m = http.MethodGet |
| } |
| f(":method", m) |
| if req.Method != "CONNECT" { |
| f(":path", path) |
| f(":scheme", req.URL.Scheme) |
| } |
| if trailers != "" { |
| f("trailer", trailers) |
| } |
| |
| var didUA bool |
| for k, vv := range req.Header { |
| if asciiEqualFold(k, "host") || asciiEqualFold(k, "content-length") { |
| // Host is :authority, already sent. |
| // Content-Length is automatic, set below. |
| continue |
| } else if asciiEqualFold(k, "connection") || |
| asciiEqualFold(k, "proxy-connection") || |
| asciiEqualFold(k, "transfer-encoding") || |
| asciiEqualFold(k, "upgrade") || |
| asciiEqualFold(k, "keep-alive") { |
| // Per 8.1.2.2 Connection-Specific Header |
| // Fields, don't send connection-specific |
| // fields. We have already checked if any |
| // are error-worthy so just ignore the rest. |
| continue |
| } else if asciiEqualFold(k, "user-agent") { |
| // Match Go's http1 behavior: at most one |
| // User-Agent. If set to nil or empty string, |
| // then omit it. Otherwise if not mentioned, |
| // include the default (below). |
| didUA = true |
| if len(vv) < 1 { |
| continue |
| } |
| vv = vv[:1] |
| if vv[0] == "" { |
| continue |
| } |
| } else if asciiEqualFold(k, "cookie") { |
| // Per 8.1.2.5 To allow for better compression efficiency, the |
| // Cookie header field MAY be split into separate header fields, |
| // each with one or more cookie-pairs. |
| for _, v := range vv { |
| for { |
| p := strings.IndexByte(v, ';') |
| if p < 0 { |
| break |
| } |
| f("cookie", v[:p]) |
| p++ |
| // strip space after semicolon if any. |
| for p+1 <= len(v) && v[p] == ' ' { |
| p++ |
| } |
| v = v[p:] |
| } |
| if len(v) > 0 { |
| f("cookie", v) |
| } |
| } |
| continue |
| } |
| |
| for _, v := range vv { |
| f(k, v) |
| } |
| } |
| if shouldSendReqContentLength(req.Method, contentLength) { |
| f("content-length", strconv.FormatInt(contentLength, 10)) |
| } |
| if addGzipHeader { |
| f("accept-encoding", "gzip") |
| } |
| if !didUA { |
| f("user-agent", defaultUserAgent) |
| } |
| } |
| |
| // Do a first pass over the headers counting bytes to ensure |
| // we don't exceed cc.peerMaxHeaderListSize. This is done as a |
| // separate pass before encoding the headers to prevent |
| // modifying the hpack state. |
| hlSize := uint64(0) |
| enumerateHeaders(func(name, value string) { |
| hf := hpack.HeaderField{Name: name, Value: value} |
| hlSize += uint64(hf.Size()) |
| }) |
| |
| if hlSize > cc.peerMaxHeaderListSize { |
| return nil, errRequestHeaderListSize |
| } |
| |
| trace := httptrace.ContextClientTrace(req.Context()) |
| traceHeaders := traceHasWroteHeaderField(trace) |
| |
| // Header list size is ok. Write the headers. |
| enumerateHeaders(func(name, value string) { |
| name, ascii := lowerHeader(name) |
| if !ascii { |
| // Skip writing invalid headers. Per RFC 7540, Section 8.1.2, header |
| // field names have to be ASCII characters (just as in HTTP/1.x). |
| return |
| } |
| cc.writeHeader(name, value) |
| if traceHeaders { |
| traceWroteHeaderField(trace, name, value) |
| } |
| }) |
| |
| return cc.hbuf.Bytes(), nil |
| } |
| |
| // shouldSendReqContentLength reports whether the http2.Transport should send |
| // a "content-length" request header. This logic is basically a copy of the net/http |
| // transferWriter.shouldSendContentLength. |
| // The contentLength is the corrected contentLength (so 0 means actually 0, not unknown). |
| // -1 means unknown. |
| func shouldSendReqContentLength(method string, contentLength int64) bool { |
| if contentLength > 0 { |
| return true |
| } |
| if contentLength < 0 { |
| return false |
| } |
| // For zero bodies, whether we send a content-length depends on the method. |
| // It also kinda doesn't matter for http2 either way, with END_STREAM. |
| switch method { |
| case "POST", "PUT", "PATCH": |
| return true |
| default: |
| return false |
| } |
| } |
| |
| // requires cc.wmu be held. |
| func (cc *ClientConn) encodeTrailers(trailer http.Header) ([]byte, error) { |
| cc.hbuf.Reset() |
| |
| hlSize := uint64(0) |
| for k, vv := range trailer { |
| for _, v := range vv { |
| hf := hpack.HeaderField{Name: k, Value: v} |
| hlSize += uint64(hf.Size()) |
| } |
| } |
| if hlSize > cc.peerMaxHeaderListSize { |
| return nil, errRequestHeaderListSize |
| } |
| |
| for k, vv := range trailer { |
| lowKey, ascii := lowerHeader(k) |
| if !ascii { |
| // Skip writing invalid headers. Per RFC 7540, Section 8.1.2, header |
| // field names have to be ASCII characters (just as in HTTP/1.x). |
| continue |
| } |
| // Transfer-Encoding, etc.. have already been filtered at the |
| // start of RoundTrip |
| for _, v := range vv { |
| cc.writeHeader(lowKey, v) |
| } |
| } |
| return cc.hbuf.Bytes(), nil |
| } |
| |
| func (cc *ClientConn) writeHeader(name, value string) { |
| if VerboseLogs { |
| log.Printf("http2: Transport encoding header %q = %q", name, value) |
| } |
| cc.henc.WriteField(hpack.HeaderField{Name: name, Value: value}) |
| } |
| |
| type resAndError struct { |
| _ incomparable |
| res *http.Response |
| err error |
| } |
| |
| // requires cc.mu be held. |
| func (cc *ClientConn) addStreamLocked(cs *clientStream) { |
| cs.flow.add(int32(cc.initialWindowSize)) |
| cs.flow.setConnFlow(&cc.flow) |
| cs.inflow.init(cc.initialStreamRecvWindowSize) |
| cs.ID = cc.nextStreamID |
| cc.nextStreamID += 2 |
| cc.streams[cs.ID] = cs |
| if cs.ID == 0 { |
| panic("assigned stream ID 0") |
| } |
| } |
| |
| func (cc *ClientConn) forgetStreamID(id uint32) { |
| cc.mu.Lock() |
| slen := len(cc.streams) |
| delete(cc.streams, id) |
| if len(cc.streams) != slen-1 { |
| panic("forgetting unknown stream id") |
| } |
| cc.lastActive = time.Now() |
| if len(cc.streams) == 0 && cc.idleTimer != nil { |
| cc.idleTimer.Reset(cc.idleTimeout) |
| cc.lastIdle = time.Now() |
| } |
| // Wake up writeRequestBody via clientStream.awaitFlowControl and |
| // wake up RoundTrip if there is a pending request. |
| cc.cond.Broadcast() |
| |
| closeOnIdle := cc.singleUse || cc.doNotReuse || cc.t.disableKeepAlives() || cc.goAway != nil |
| if closeOnIdle && cc.streamsReserved == 0 && len(cc.streams) == 0 { |
| if VerboseLogs { |
| cc.vlogf("http2: Transport closing idle conn %p (forSingleUse=%v, maxStream=%v)", cc, cc.singleUse, cc.nextStreamID-2) |
| } |
| cc.closed = true |
| defer cc.closeConn() |
| } |
| |
| cc.mu.Unlock() |
| } |
| |
| // clientConnReadLoop is the state owned by the clientConn's frame-reading readLoop. |
| type clientConnReadLoop struct { |
| _ incomparable |
| cc *ClientConn |
| } |
| |
| // readLoop runs in its own goroutine and reads and dispatches frames. |
| func (cc *ClientConn) readLoop() { |
| cc.t.markNewGoroutine() |
| rl := &clientConnReadLoop{cc: cc} |
| defer rl.cleanup() |
| cc.readerErr = rl.run() |
| if ce, ok := cc.readerErr.(ConnectionError); ok { |
| cc.wmu.Lock() |
| cc.fr.WriteGoAway(0, ErrCode(ce), nil) |
| cc.wmu.Unlock() |
| } |
| } |
| |
| // GoAwayError is returned by the Transport when the server closes the |
| // TCP connection after sending a GOAWAY frame. |
| type GoAwayError struct { |
| LastStreamID uint32 |
| ErrCode ErrCode |
| DebugData string |
| } |
| |
| func (e GoAwayError) Error() string { |
| return fmt.Sprintf("http2: server sent GOAWAY and closed the connection; LastStreamID=%v, ErrCode=%v, debug=%q", |
| e.LastStreamID, e.ErrCode, e.DebugData) |
| } |
| |
| func isEOFOrNetReadError(err error) bool { |
| if err == io.EOF { |
| return true |
| } |
| ne, ok := err.(*net.OpError) |
| return ok && ne.Op == "read" |
| } |
| |
| func (rl *clientConnReadLoop) cleanup() { |
| cc := rl.cc |
| cc.t.connPool().MarkDead(cc) |
| defer cc.closeConn() |
| defer close(cc.readerDone) |
| |
| if cc.idleTimer != nil { |
| cc.idleTimer.Stop() |
| } |
| |
| // Close any response bodies if the server closes prematurely. |
| // TODO: also do this if we've written the headers but not |
| // gotten a response yet. |
| err := cc.readerErr |
| cc.mu.Lock() |
| if cc.goAway != nil && isEOFOrNetReadError(err) { |
| err = GoAwayError{ |
| LastStreamID: cc.goAway.LastStreamID, |
| ErrCode: cc.goAway.ErrCode, |
| DebugData: cc.goAwayDebug, |
| } |
| } else if err == io.EOF { |
| err = io.ErrUnexpectedEOF |
| } |
| cc.closed = true |
| |
| for _, cs := range cc.streams { |
| select { |
| case <-cs.peerClosed: |
| // The server closed the stream before closing the conn, |
| // so no need to interrupt it. |
| default: |
| cs.abortStreamLocked(err) |
| } |
| } |
| cc.cond.Broadcast() |
| cc.mu.Unlock() |
| } |
| |
| // countReadFrameError calls Transport.CountError with a string |
| // representing err. |
| func (cc *ClientConn) countReadFrameError(err error) { |
| f := cc.t.CountError |
| if f == nil || err == nil { |
| return |
| } |
| if ce, ok := err.(ConnectionError); ok { |
| errCode := ErrCode(ce) |
| f(fmt.Sprintf("read_frame_conn_error_%s", errCode.stringToken())) |
| return |
| } |
| if errors.Is(err, io.EOF) { |
| f("read_frame_eof") |
| return |
| } |
| if errors.Is(err, io.ErrUnexpectedEOF) { |
| f("read_frame_unexpected_eof") |
| return |
| } |
| if errors.Is(err, ErrFrameTooLarge) { |
| f("read_frame_too_large") |
| return |
| } |
| f("read_frame_other") |
| } |
| |
| func (rl *clientConnReadLoop) run() error { |
| cc := rl.cc |
| gotSettings := false |
| readIdleTimeout := cc.readIdleTimeout |
| var t timer |
| if readIdleTimeout != 0 { |
| t = cc.t.afterFunc(readIdleTimeout, cc.healthCheck) |
| } |
| for { |
| f, err := cc.fr.ReadFrame() |
| if t != nil { |
| t.Reset(readIdleTimeout) |
| } |
| if err != nil { |
| cc.vlogf("http2: Transport readFrame error on conn %p: (%T) %v", cc, err, err) |
| } |
| if se, ok := err.(StreamError); ok { |
| if cs := rl.streamByID(se.StreamID); cs != nil { |
| if se.Cause == nil { |
| se.Cause = cc.fr.errDetail |
| } |
| rl.endStreamError(cs, se) |
| } |
| continue |
| } else if err != nil { |
| cc.countReadFrameError(err) |
| return err |
| } |
| if VerboseLogs { |
| cc.vlogf("http2: Transport received %s", summarizeFrame(f)) |
| } |
| if !gotSettings { |
| if _, ok := f.(*SettingsFrame); !ok { |
| cc.logf("protocol error: received %T before a SETTINGS frame", f) |
| return ConnectionError(ErrCodeProtocol) |
| } |
| gotSettings = true |
| } |
| |
| switch f := f.(type) { |
| case *MetaHeadersFrame: |
| err = rl.processHeaders(f) |
| case *DataFrame: |
| err = rl.processData(f) |
| case *GoAwayFrame: |
| err = rl.processGoAway(f) |
| case *RSTStreamFrame: |
| err = rl.processResetStream(f) |
| case *SettingsFrame: |
| err = rl.processSettings(f) |
| case *PushPromiseFrame: |
| err = rl.processPushPromise(f) |
| case *WindowUpdateFrame: |
| err = rl.processWindowUpdate(f) |
| case *PingFrame: |
| err = rl.processPing(f) |
| default: |
| cc.logf("Transport: unhandled response frame type %T", f) |
| } |
| if err != nil { |
| if VerboseLogs { |
| cc.vlogf("http2: Transport conn %p received error from processing frame %v: %v", cc, summarizeFrame(f), err) |
| } |
| return err |
| } |
| } |
| } |
| |
| func (rl *clientConnReadLoop) processHeaders(f *MetaHeadersFrame) error { |
| cs := rl.streamByID(f.StreamID) |
| if cs == nil { |
| // We'd get here if we canceled a request while the |
| // server had its response still in flight. So if this |
| // was just something we canceled, ignore it. |
| return nil |
| } |
| if cs.readClosed { |
| rl.endStreamError(cs, StreamError{ |
| StreamID: f.StreamID, |
| Code: ErrCodeProtocol, |
| Cause: errors.New("protocol error: headers after END_STREAM"), |
| }) |
| return nil |
| } |
| if !cs.firstByte { |
| if cs.trace != nil { |
| // TODO(bradfitz): move first response byte earlier, |
| // when we first read the 9 byte header, not waiting |
| // until all the HEADERS+CONTINUATION frames have been |
| // merged. This works for now. |
| traceFirstResponseByte(cs.trace) |
| } |
| cs.firstByte = true |
| } |
| if !cs.pastHeaders { |
| cs.pastHeaders = true |
| } else { |
| return rl.processTrailers(cs, f) |
| } |
| |
| res, err := rl.handleResponse(cs, f) |
| if err != nil { |
| if _, ok := err.(ConnectionError); ok { |
| return err |
| } |
| // Any other error type is a stream error. |
| rl.endStreamError(cs, StreamError{ |
| StreamID: f.StreamID, |
| Code: ErrCodeProtocol, |
| Cause: err, |
| }) |
| return nil // return nil from process* funcs to keep conn alive |
| } |
| if res == nil { |
| // (nil, nil) special case. See handleResponse docs. |
| return nil |
| } |
| cs.resTrailer = &res.Trailer |
| cs.res = res |
| close(cs.respHeaderRecv) |
| if f.StreamEnded() { |
| rl.endStream(cs) |
| } |
| return nil |
| } |
| |
| // may return error types nil, or ConnectionError. Any other error value |
| // is a StreamError of type ErrCodeProtocol. The returned error in that case |
| // is the detail. |
| // |
| // As a special case, handleResponse may return (nil, nil) to skip the |
| // frame (currently only used for 1xx responses). |
| func (rl *clientConnReadLoop) handleResponse(cs *clientStream, f *MetaHeadersFrame) (*http.Response, error) { |
| if f.Truncated { |
| return nil, errResponseHeaderListSize |
| } |
| |
| status := f.PseudoValue("status") |
| if status == "" { |
| return nil, errors.New("malformed response from server: missing status pseudo header") |
| } |
| statusCode, err := strconv.Atoi(status) |
| if err != nil { |
| return nil, errors.New("malformed response from server: malformed non-numeric status pseudo header") |
| } |
| |
| regularFields := f.RegularFields() |
| strs := make([]string, len(regularFields)) |
| header := make(http.Header, len(regularFields)) |
| res := &http.Response{ |
| Proto: "HTTP/2.0", |
| ProtoMajor: 2, |
| Header: header, |
| StatusCode: statusCode, |
| Status: status + " " + http.StatusText(statusCode), |
| } |
| for _, hf := range regularFields { |
| key := canonicalHeader(hf.Name) |
| if key == "Trailer" { |
| t := res.Trailer |
| if t == nil { |
| t = make(http.Header) |
| res.Trailer = t |
| } |
| foreachHeaderElement(hf.Value, func(v string) { |
| t[canonicalHeader(v)] = nil |
| }) |
| } else { |
| vv := header[key] |
| if vv == nil && len(strs) > 0 { |
| // More than likely this will be a single-element key. |
| // Most headers aren't multi-valued. |
| // Set the capacity on strs[0] to 1, so any future append |
| // won't extend the slice into the other strings. |
| vv, strs = strs[:1:1], strs[1:] |
| vv[0] = hf.Value |
| header[key] = vv |
| } else { |
| header[key] = append(vv, hf.Value) |
| } |
| } |
| } |
| |
| if statusCode >= 100 && statusCode <= 199 { |
| if f.StreamEnded() { |
| return nil, errors.New("1xx informational response with END_STREAM flag") |
| } |
| if fn := cs.get1xxTraceFunc(); fn != nil { |
| // If the 1xx response is being delivered to the user, |
| // then they're responsible for limiting the number |
| // of responses. |
| if err := fn(statusCode, textproto.MIMEHeader(header)); err != nil { |
| return nil, err |
| } |
| } else { |
| // If the user didn't examine the 1xx response, then we |
| // limit the size of all 1xx headers. |
| // |
| // This differs a bit from the HTTP/1 implementation, which |
| // limits the size of all 1xx headers plus the final response. |
| // Use the larger limit of MaxHeaderListSize and |
| // net/http.Transport.MaxResponseHeaderBytes. |
| limit := int64(cs.cc.t.maxHeaderListSize()) |
| if t1 := cs.cc.t.t1; t1 != nil && t1.MaxResponseHeaderBytes > limit { |
| limit = t1.MaxResponseHeaderBytes |
| } |
| for _, h := range f.Fields { |
| cs.totalHeaderSize += int64(h.Size()) |
| } |
| if cs.totalHeaderSize > limit { |
| if VerboseLogs { |
| log.Printf("http2: 1xx informational responses too large") |
| } |
| return nil, errors.New("header list too large") |
| } |
| } |
| if statusCode == 100 { |
| traceGot100Continue(cs.trace) |
| select { |
| case cs.on100 <- struct{}{}: |
| default: |
| } |
| } |
| cs.pastHeaders = false // do it all again |
| return nil, nil |
| } |
| |
| res.ContentLength = -1 |
| if clens := res.Header["Content-Length"]; len(clens) == 1 { |
| if cl, err := strconv.ParseUint(clens[0], 10, 63); err == nil { |
| res.ContentLength = int64(cl) |
| } else { |
| // TODO: care? unlike http/1, it won't mess up our framing, so it's |
| // more safe smuggling-wise to ignore. |
| } |
| } else if len(clens) > 1 { |
| // TODO: care? unlike http/1, it won't mess up our framing, so it's |
| // more safe smuggling-wise to ignore. |
| } else if f.StreamEnded() && !cs.isHead { |
| res.ContentLength = 0 |
| } |
| |
| if cs.isHead { |
| res.Body = noBody |
| return res, nil |
| } |
| |
| if f.StreamEnded() { |
| if res.ContentLength > 0 { |
| res.Body = missingBody{} |
| } else { |
| res.Body = noBody |
| } |
| return res, nil |
| } |
| |
| cs.bufPipe.setBuffer(&dataBuffer{expected: res.ContentLength}) |
| cs.bytesRemain = res.ContentLength |
| res.Body = transportResponseBody{cs} |
| |
| if cs.requestedGzip && asciiEqualFold(res.Header.Get("Content-Encoding"), "gzip") { |
| res.Header.Del("Content-Encoding") |
| res.Header.Del("Content-Length") |
| res.ContentLength = -1 |
| res.Body = &gzipReader{body: res.Body} |
| res.Uncompressed = true |
| } |
| return res, nil |
| } |
| |
| func (rl *clientConnReadLoop) processTrailers(cs *clientStream, f *MetaHeadersFrame) error { |
| if cs.pastTrailers { |
| // Too many HEADERS frames for this stream. |
| return ConnectionError(ErrCodeProtocol) |
| } |
| cs.pastTrailers = true |
| if !f.StreamEnded() { |
| // We expect that any headers for trailers also |
| // has END_STREAM. |
| return ConnectionError(ErrCodeProtocol) |
| } |
| if len(f.PseudoFields()) > 0 { |
| // No pseudo header fields are defined for trailers. |
| // TODO: ConnectionError might be overly harsh? Check. |
| return ConnectionError(ErrCodeProtocol) |
| } |
| |
| trailer := make(http.Header) |
| for _, hf := range f.RegularFields() { |
| key := canonicalHeader(hf.Name) |
| trailer[key] = append(trailer[key], hf.Value) |
| } |
| cs.trailer = trailer |
| |
| rl.endStream(cs) |
| return nil |
| } |
| |
| // transportResponseBody is the concrete type of Transport.RoundTrip's |
| // Response.Body. It is an io.ReadCloser. |
| type transportResponseBody struct { |
| cs *clientStream |
| } |
| |
| func (b transportResponseBody) Read(p []byte) (n int, err error) { |
| cs := b.cs |
| cc := cs.cc |
| |
| if cs.readErr != nil { |
| return 0, cs.readErr |
| } |
| n, err = b.cs.bufPipe.Read(p) |
| if cs.bytesRemain != -1 { |
| if int64(n) > cs.bytesRemain { |
| n = int(cs.bytesRemain) |
| if err == nil { |
| err = errors.New("net/http: server replied with more than declared Content-Length; truncated") |
| cs.abortStream(err) |
| } |
| cs.readErr = err |
| return int(cs.bytesRemain), err |
| } |
| cs.bytesRemain -= int64(n) |
| if err == io.EOF && cs.bytesRemain > 0 { |
| err = io.ErrUnexpectedEOF |
| cs.readErr = err |
| return n, err |
| } |
| } |
| if n == 0 { |
| // No flow control tokens to send back. |
| return |
| } |
| |
| cc.mu.Lock() |
| connAdd := cc.inflow.add(n) |
| var streamAdd int32 |
| if err == nil { // No need to refresh if the stream is over or failed. |
| streamAdd = cs.inflow.add(n) |
| } |
| cc.mu.Unlock() |
| |
| if connAdd != 0 || streamAdd != 0 { |
| cc.wmu.Lock() |
| defer cc.wmu.Unlock() |
| if connAdd != 0 { |
| cc.fr.WriteWindowUpdate(0, mustUint31(connAdd)) |
| } |
| if streamAdd != 0 { |
| cc.fr.WriteWindowUpdate(cs.ID, mustUint31(streamAdd)) |
| } |
| cc.bw.Flush() |
| } |
| return |
| } |
| |
| var errClosedResponseBody = errors.New("http2: response body closed") |
| |
| func (b transportResponseBody) Close() error { |
| cs := b.cs |
| cc := cs.cc |
| |
| cs.bufPipe.BreakWithError(errClosedResponseBody) |
| cs.abortStream(errClosedResponseBody) |
| |
| unread := cs.bufPipe.Len() |
| if unread > 0 { |
| cc.mu.Lock() |
| // Return connection-level flow control. |
| connAdd := cc.inflow.add(unread) |
| cc.mu.Unlock() |
| |
| // TODO(dneil): Acquiring this mutex can block indefinitely. |
| // Move flow control return to a goroutine? |
| cc.wmu.Lock() |
| // Return connection-level flow control. |
| if connAdd > 0 { |
| cc.fr.WriteWindowUpdate(0, uint32(connAdd)) |
| } |
| cc.bw.Flush() |
| cc.wmu.Unlock() |
| } |
| |
| select { |
| case <-cs.donec: |
| case <-cs.ctx.Done(): |
| // See golang/go#49366: The net/http package can cancel the |
| // request context after the response body is fully read. |
| // Don't treat this as an error. |
| return nil |
| case <-cs.reqCancel: |
| return errRequestCanceled |
| } |
| return nil |
| } |
| |
| func (rl *clientConnReadLoop) processData(f *DataFrame) error { |
| cc := rl.cc |
| cs := rl.streamByID(f.StreamID) |
| data := f.Data() |
| if cs == nil { |
| cc.mu.Lock() |
| neverSent := cc.nextStreamID |
| cc.mu.Unlock() |
| if f.StreamID >= neverSent { |
| // We never asked for this. |
| cc.logf("http2: Transport received unsolicited DATA frame; closing connection") |
| return ConnectionError(ErrCodeProtocol) |
| } |
| // We probably did ask for this, but canceled. Just ignore it. |
| // TODO: be stricter here? only silently ignore things which |
| // we canceled, but not things which were closed normally |
| // by the peer? Tough without accumulating too much state. |
| |
| // But at least return their flow control: |
| if f.Length > 0 { |
| cc.mu.Lock() |
| ok := cc.inflow.take(f.Length) |
| connAdd := cc.inflow.add(int(f.Length)) |
| cc.mu.Unlock() |
| if !ok { |
| return ConnectionError(ErrCodeFlowControl) |
| } |
| if connAdd > 0 { |
| cc.wmu.Lock() |
| cc.fr.WriteWindowUpdate(0, uint32(connAdd)) |
| cc.bw.Flush() |
| cc.wmu.Unlock() |
| } |
| } |
| return nil |
| } |
| if cs.readClosed { |
| cc.logf("protocol error: received DATA after END_STREAM") |
| rl.endStreamError(cs, StreamError{ |
| StreamID: f.StreamID, |
| Code: ErrCodeProtocol, |
| }) |
| return nil |
| } |
| if !cs.pastHeaders { |
| cc.logf("protocol error: received DATA before a HEADERS frame") |
| rl.endStreamError(cs, StreamError{ |
| StreamID: f.StreamID, |
| Code: ErrCodeProtocol, |
| }) |
| return nil |
| } |
| if f.Length > 0 { |
| if cs.isHead && len(data) > 0 { |
| cc.logf("protocol error: received DATA on a HEAD request") |
| rl.endStreamError(cs, StreamError{ |
| StreamID: f.StreamID, |
| Code: ErrCodeProtocol, |
| }) |
| return nil |
| } |
| // Check connection-level flow control. |
| cc.mu.Lock() |
| if !takeInflows(&cc.inflow, &cs.inflow, f.Length) { |
| cc.mu.Unlock() |
| return ConnectionError(ErrCodeFlowControl) |
| } |
| // Return any padded flow control now, since we won't |
| // refund it later on body reads. |
| var refund int |
| if pad := int(f.Length) - len(data); pad > 0 { |
| refund += pad |
| } |
| |
| didReset := false |
| var err error |
| if len(data) > 0 { |
| if _, err = cs.bufPipe.Write(data); err != nil { |
| // Return len(data) now if the stream is already closed, |
| // since data will never be read. |
| didReset = true |
| refund += len(data) |
| } |
| } |
| |
| sendConn := cc.inflow.add(refund) |
| var sendStream int32 |
| if !didReset { |
| sendStream = cs.inflow.add(refund) |
| } |
| cc.mu.Unlock() |
| |
| if sendConn > 0 || sendStream > 0 { |
| cc.wmu.Lock() |
| if sendConn > 0 { |
| cc.fr.WriteWindowUpdate(0, uint32(sendConn)) |
| } |
| if sendStream > 0 { |
| cc.fr.WriteWindowUpdate(cs.ID, uint32(sendStream)) |
| } |
| cc.bw.Flush() |
| cc.wmu.Unlock() |
| } |
| |
| if err != nil { |
| rl.endStreamError(cs, err) |
| return nil |
| } |
| } |
| |
| if f.StreamEnded() { |
| rl.endStream(cs) |
| } |
| return nil |
| } |
| |
| func (rl *clientConnReadLoop) endStream(cs *clientStream) { |
| // TODO: check that any declared content-length matches, like |
| // server.go's (*stream).endStream method. |
| if !cs.readClosed { |
| cs.readClosed = true |
| // Close cs.bufPipe and cs.peerClosed with cc.mu held to avoid a |
| // race condition: The caller can read io.EOF from Response.Body |
| // and close the body before we close cs.peerClosed, causing |
| // cleanupWriteRequest to send a RST_STREAM. |
| rl.cc.mu.Lock() |
| defer rl.cc.mu.Unlock() |
| cs.bufPipe.closeWithErrorAndCode(io.EOF, cs.copyTrailers) |
| close(cs.peerClosed) |
| } |
| } |
| |
| func (rl *clientConnReadLoop) endStreamError(cs *clientStream, err error) { |
| cs.readAborted = true |
| cs.abortStream(err) |
| } |
| |
| func (rl *clientConnReadLoop) streamByID(id uint32) *clientStream { |
| rl.cc.mu.Lock() |
| defer rl.cc.mu.Unlock() |
| cs := rl.cc.streams[id] |
| if cs != nil && !cs.readAborted { |
| return cs |
| } |
| return nil |
| } |
| |
| func (cs *clientStream) copyTrailers() { |
| for k, vv := range cs.trailer { |
| t := cs.resTrailer |
| if *t == nil { |
| *t = make(http.Header) |
| } |
| (*t)[k] = vv |
| } |
| } |
| |
| func (rl *clientConnReadLoop) processGoAway(f *GoAwayFrame) error { |
| cc := rl.cc |
| cc.t.connPool().MarkDead(cc) |
| if f.ErrCode != 0 { |
| // TODO: deal with GOAWAY more. particularly the error code |
| cc.vlogf("transport got GOAWAY with error code = %v", f.ErrCode) |
| if fn := cc.t.CountError; fn != nil { |
| fn("recv_goaway_" + f.ErrCode.stringToken()) |
| } |
| } |
| cc.setGoAway(f) |
| return nil |
| } |
| |
| func (rl *clientConnReadLoop) processSettings(f *SettingsFrame) error { |
| cc := rl.cc |
| // Locking both mu and wmu here allows frame encoding to read settings with only wmu held. |
| // Acquiring wmu when f.IsAck() is unnecessary, but convenient and mostly harmless. |
| cc.wmu.Lock() |
| defer cc.wmu.Unlock() |
| |
| if err := rl.processSettingsNoWrite(f); err != nil { |
| return err |
| } |
| if !f.IsAck() { |
| cc.fr.WriteSettingsAck() |
| cc.bw.Flush() |
| } |
| return nil |
| } |
| |
| func (rl *clientConnReadLoop) processSettingsNoWrite(f *SettingsFrame) error { |
| cc := rl.cc |
| cc.mu.Lock() |
| defer cc.mu.Unlock() |
| |
| if f.IsAck() { |
| if cc.wantSettingsAck { |
| cc.wantSettingsAck = false |
| return nil |
| } |
| return ConnectionError(ErrCodeProtocol) |
| } |
| |
| var seenMaxConcurrentStreams bool |
| err := f.ForeachSetting(func(s Setting) error { |
| switch s.ID { |
| case SettingMaxFrameSize: |
| cc.maxFrameSize = s.Val |
| case SettingMaxConcurrentStreams: |
| cc.maxConcurrentStreams = s.Val |
| seenMaxConcurrentStreams = true |
| case SettingMaxHeaderListSize: |
| cc.peerMaxHeaderListSize = uint64(s.Val) |
| case SettingInitialWindowSize: |
| // Values above the maximum flow-control |
| // window size of 2^31-1 MUST be treated as a |
| // connection error (Section 5.4.1) of type |
| // FLOW_CONTROL_ERROR. |
| if s.Val > math.MaxInt32 { |
| return ConnectionError(ErrCodeFlowControl) |
| } |
| |
| // Adjust flow control of currently-open |
| // frames by the difference of the old initial |
| // window size and this one. |
| delta := int32(s.Val) - int32(cc.initialWindowSize) |
| for _, cs := range cc.streams { |
| cs.flow.add(delta) |
| } |
| cc.cond.Broadcast() |
| |
| cc.initialWindowSize = s.Val |
| case SettingHeaderTableSize: |
| cc.henc.SetMaxDynamicTableSize(s.Val) |
| cc.peerMaxHeaderTableSize = s.Val |
| default: |
| cc.vlogf("Unhandled Setting: %v", s) |
| } |
| return nil |
| }) |
| if err != nil { |
| return err |
| } |
| |
| if !cc.seenSettings { |
| if !seenMaxConcurrentStreams { |
| // This was the servers initial SETTINGS frame and it |
| // didn't contain a MAX_CONCURRENT_STREAMS field so |
| // increase the number of concurrent streams this |
| // connection can establish to our default. |
| cc.maxConcurrentStreams = defaultMaxConcurrentStreams |
| } |
| cc.seenSettings = true |
| } |
| |
| return nil |
| } |
| |
| func (rl *clientConnReadLoop) processWindowUpdate(f *WindowUpdateFrame) error { |
| cc := rl.cc |
| cs := rl.streamByID(f.StreamID) |
| if f.StreamID != 0 && cs == nil { |
| return nil |
| } |
| |
| cc.mu.Lock() |
| defer cc.mu.Unlock() |
| |
| fl := &cc.flow |
| if cs != nil { |
| fl = &cs.flow |
| } |
| if !fl.add(int32(f.Increment)) { |
| // For stream, the sender sends RST_STREAM with an error code of FLOW_CONTROL_ERROR |
| if cs != nil { |
| rl.endStreamError(cs, StreamError{ |
| StreamID: f.StreamID, |
| Code: ErrCodeFlowControl, |
| }) |
| return nil |
| } |
| |
| return ConnectionError(ErrCodeFlowControl) |
| } |
| cc.cond.Broadcast() |
| return nil |
| } |
| |
| func (rl *clientConnReadLoop) processResetStream(f *RSTStreamFrame) error { |
| cs := rl.streamByID(f.StreamID) |
| if cs == nil { |
| // TODO: return error if server tries to RST_STREAM an idle stream |
| return nil |
| } |
| serr := streamError(cs.ID, f.ErrCode) |
| serr.Cause = errFromPeer |
| if f.ErrCode == ErrCodeProtocol { |
| rl.cc.SetDoNotReuse() |
| } |
| if fn := cs.cc.t.CountError; fn != nil { |
| fn("recv_rststream_" + f.ErrCode.stringToken()) |
| } |
| cs.abortStream(serr) |
| |
| cs.bufPipe.CloseWithError(serr) |
| return nil |
| } |
| |
| // Ping sends a PING frame to the server and waits for the ack. |
| func (cc *ClientConn) Ping(ctx context.Context) error { |
| c := make(chan struct{}) |
| // Generate a random payload |
| var p [8]byte |
| for { |
| if _, err := rand.Read(p[:]); err != nil { |
| return err |
| } |
| cc.mu.Lock() |
| // check for dup before insert |
| if _, found := cc.pings[p]; !found { |
| cc.pings[p] = c |
| cc.mu.Unlock() |
| break |
| } |
| cc.mu.Unlock() |
| } |
| var pingError error |
| errc := make(chan struct{}) |
| go func() { |
| cc.t.markNewGoroutine() |
| cc.wmu.Lock() |
| defer cc.wmu.Unlock() |
| if pingError = cc.fr.WritePing(false, p); pingError != nil { |
| close(errc) |
| return |
| } |
| if pingError = cc.bw.Flush(); pingError != nil { |
| close(errc) |
| return |
| } |
| }() |
| select { |
| case <-c: |
| return nil |
| case <-errc: |
| return pingError |
| case <-ctx.Done(): |
| return ctx.Err() |
| case <-cc.readerDone: |
| // connection closed |
| return cc.readerErr |
| } |
| } |
| |
| func (rl *clientConnReadLoop) processPing(f *PingFrame) error { |
| if f.IsAck() { |
| cc := rl.cc |
| cc.mu.Lock() |
| defer cc.mu.Unlock() |
| // If ack, notify listener if any |
| if c, ok := cc.pings[f.Data]; ok { |
| close(c) |
| delete(cc.pings, f.Data) |
| } |
| if cc.pendingResets > 0 { |
| // See clientStream.cleanupWriteRequest. |
| cc.pendingResets = 0 |
| cc.cond.Broadcast() |
| } |
| return nil |
| } |
| cc := rl.cc |
| cc.wmu.Lock() |
| defer cc.wmu.Unlock() |
| if err := cc.fr.WritePing(true, f.Data); err != nil { |
| return err |
| } |
| return cc.bw.Flush() |
| } |
| |
| func (rl *clientConnReadLoop) processPushPromise(f *PushPromiseFrame) error { |
| // We told the peer we don't want them. |
| // Spec says: |
| // "PUSH_PROMISE MUST NOT be sent if the SETTINGS_ENABLE_PUSH |
| // setting of the peer endpoint is set to 0. An endpoint that |
| // has set this setting and has received acknowledgement MUST |
| // treat the receipt of a PUSH_PROMISE frame as a connection |
| // error (Section 5.4.1) of type PROTOCOL_ERROR." |
| return ConnectionError(ErrCodeProtocol) |
| } |
| |
| // writeStreamReset sends a RST_STREAM frame. |
| // When ping is true, it also sends a PING frame with a random payload. |
| func (cc *ClientConn) writeStreamReset(streamID uint32, code ErrCode, ping bool, err error) { |
| // TODO: map err to more interesting error codes, once the |
| // HTTP community comes up with some. But currently for |
| // RST_STREAM there's no equivalent to GOAWAY frame's debug |
| // data, and the error codes are all pretty vague ("cancel"). |
| cc.wmu.Lock() |
| cc.fr.WriteRSTStream(streamID, code) |
| if ping { |
| var payload [8]byte |
| rand.Read(payload[:]) |
| cc.fr.WritePing(false, payload) |
| } |
| cc.bw.Flush() |
| cc.wmu.Unlock() |
| } |
| |
| var ( |
| errResponseHeaderListSize = errors.New("http2: response header list larger than advertised limit") |
| errRequestHeaderListSize = errors.New("http2: request header list larger than peer's advertised limit") |
| ) |
| |
| func (cc *ClientConn) logf(format string, args ...interface{}) { |
| cc.t.logf(format, args...) |
| } |
| |
| func (cc *ClientConn) vlogf(format string, args ...interface{}) { |
| cc.t.vlogf(format, args...) |
| } |
| |
| func (t *Transport) vlogf(format string, args ...interface{}) { |
| if VerboseLogs { |
| t.logf(format, args...) |
| } |
| } |
| |
| func (t *Transport) logf(format string, args ...interface{}) { |
| log.Printf(format, args...) |
| } |
| |
| var noBody io.ReadCloser = noBodyReader{} |
| |
| type noBodyReader struct{} |
| |
| func (noBodyReader) Close() error { return nil } |
| func (noBodyReader) Read([]byte) (int, error) { return 0, io.EOF } |
| |
| type missingBody struct{} |
| |
| func (missingBody) Close() error { return nil } |
| func (missingBody) Read([]byte) (int, error) { return 0, io.ErrUnexpectedEOF } |
| |
| func strSliceContains(ss []string, s string) bool { |
| for _, v := range ss { |
| if v == s { |
| return true |
| } |
| } |
| return false |
| } |
| |
| type erringRoundTripper struct{ err error } |
| |
| func (rt erringRoundTripper) RoundTripErr() error { return rt.err } |
| func (rt erringRoundTripper) RoundTrip(*http.Request) (*http.Response, error) { return nil, rt.err } |
| |
| // gzipReader wraps a response body so it can lazily |
| // call gzip.NewReader on the first call to Read |
| type gzipReader struct { |
| _ incomparable |
| body io.ReadCloser // underlying Response.Body |
| zr *gzip.Reader // lazily-initialized gzip reader |
| zerr error // sticky error |
| } |
| |
| func (gz *gzipReader) Read(p []byte) (n int, err error) { |
| if gz.zerr != nil { |
| return 0, gz.zerr |
| } |
| if gz.zr == nil { |
| gz.zr, err = gzip.NewReader(gz.body) |
| if err != nil { |
| gz.zerr = err |
| return 0, err |
| } |
| } |
| return gz.zr.Read(p) |
| } |
| |
| func (gz *gzipReader) Close() error { |
| if err := gz.body.Close(); err != nil { |
| return err |
| } |
| gz.zerr = fs.ErrClosed |
| return nil |
| } |
| |
| type errorReader struct{ err error } |
| |
| func (r errorReader) Read(p []byte) (int, error) { return 0, r.err } |
| |
| // isConnectionCloseRequest reports whether req should use its own |
| // connection for a single request and then close the connection. |
| func isConnectionCloseRequest(req *http.Request) bool { |
| return req.Close || httpguts.HeaderValuesContainsToken(req.Header["Connection"], "close") |
| } |
| |
| // registerHTTPSProtocol calls Transport.RegisterProtocol but |
| // converting panics into errors. |
| func registerHTTPSProtocol(t *http.Transport, rt noDialH2RoundTripper) (err error) { |
| defer func() { |
| if e := recover(); e != nil { |
| err = fmt.Errorf("%v", e) |
| } |
| }() |
| t.RegisterProtocol("https", rt) |
| return nil |
| } |
| |
| // noDialH2RoundTripper is a RoundTripper which only tries to complete the request |
| // if there's already has a cached connection to the host. |
| // (The field is exported so it can be accessed via reflect from net/http; tested |
| // by TestNoDialH2RoundTripperType) |
| type noDialH2RoundTripper struct{ *Transport } |
| |
| func (rt noDialH2RoundTripper) RoundTrip(req *http.Request) (*http.Response, error) { |
| res, err := rt.Transport.RoundTrip(req) |
| if isNoCachedConnError(err) { |
| return nil, http.ErrSkipAltProtocol |
| } |
| return res, err |
| } |
| |
| func (t *Transport) idleConnTimeout() time.Duration { |
| // to keep things backwards compatible, we use non-zero values of |
| // IdleConnTimeout, followed by using the IdleConnTimeout on the underlying |
| // http1 transport, followed by 0 |
| if t.IdleConnTimeout != 0 { |
| return t.IdleConnTimeout |
| } |
| |
| if t.t1 != nil { |
| return t.t1.IdleConnTimeout |
| } |
| |
| return 0 |
| } |
| |
| func traceGetConn(req *http.Request, hostPort string) { |
| trace := httptrace.ContextClientTrace(req.Context()) |
| if trace == nil || trace.GetConn == nil { |
| return |
| } |
| trace.GetConn(hostPort) |
| } |
| |
| func traceGotConn(req *http.Request, cc *ClientConn, reused bool) { |
| trace := httptrace.ContextClientTrace(req.Context()) |
| if trace == nil || trace.GotConn == nil { |
| return |
| } |
| ci := httptrace.GotConnInfo{Conn: cc.tconn} |
| ci.Reused = reused |
| cc.mu.Lock() |
| ci.WasIdle = len(cc.streams) == 0 && reused |
| if ci.WasIdle && !cc.lastActive.IsZero() { |
| ci.IdleTime = time.Since(cc.lastActive) |
| } |
| cc.mu.Unlock() |
| |
| trace.GotConn(ci) |
| } |
| |
| func traceWroteHeaders(trace *httptrace.ClientTrace) { |
| if trace != nil && trace.WroteHeaders != nil { |
| trace.WroteHeaders() |
| } |
| } |
| |
| func traceGot100Continue(trace *httptrace.ClientTrace) { |
| if trace != nil && trace.Got100Continue != nil { |
| trace.Got100Continue() |
| } |
| } |
| |
| func traceWait100Continue(trace *httptrace.ClientTrace) { |
| if trace != nil && trace.Wait100Continue != nil { |
| trace.Wait100Continue() |
| } |
| } |
| |
| func traceWroteRequest(trace *httptrace.ClientTrace, err error) { |
| if trace != nil && trace.WroteRequest != nil { |
| trace.WroteRequest(httptrace.WroteRequestInfo{Err: err}) |
| } |
| } |
| |
| func traceFirstResponseByte(trace *httptrace.ClientTrace) { |
| if trace != nil && trace.GotFirstResponseByte != nil { |
| trace.GotFirstResponseByte() |
| } |
| } |
| |
| func traceHasWroteHeaderField(trace *httptrace.ClientTrace) bool { |
| return trace != nil && trace.WroteHeaderField != nil |
| } |
| |
| func traceWroteHeaderField(trace *httptrace.ClientTrace, k, v string) { |
| if trace != nil && trace.WroteHeaderField != nil { |
| trace.WroteHeaderField(k, []string{v}) |
| } |
| } |
| |
| func traceGot1xxResponseFunc(trace *httptrace.ClientTrace) func(int, textproto.MIMEHeader) error { |
| if trace != nil { |
| return trace.Got1xxResponse |
| } |
| return nil |
| } |
| |
| // dialTLSWithContext uses tls.Dialer, added in Go 1.15, to open a TLS |
| // connection. |
| func (t *Transport) dialTLSWithContext(ctx context.Context, network, addr string, cfg *tls.Config) (*tls.Conn, error) { |
| dialer := &tls.Dialer{ |
| Config: cfg, |
| } |
| cn, err := dialer.DialContext(ctx, network, addr) |
| if err != nil { |
| return nil, err |
| } |
| tlsCn := cn.(*tls.Conn) // DialContext comment promises this will always succeed |
| return tlsCn, nil |
| } |