| // Copyright 2015 The Go Authors. All rights reserved. |
| // Use of this source code is governed by a BSD-style |
| // license that can be found in the LICENSE file. |
| |
| // Transport code. |
| |
| package http2 |
| |
| import ( |
| "bufio" |
| "bytes" |
| "crypto/tls" |
| "errors" |
| "fmt" |
| "io" |
| "io/ioutil" |
| "log" |
| "net" |
| "net/http" |
| "strconv" |
| "strings" |
| "sync" |
| |
| "golang.org/x/net/http2/hpack" |
| ) |
| |
| const ( |
| // transportDefaultConnFlow is how many connection-level flow control |
| // tokens we give the server at start-up, past the default 64k. |
| transportDefaultConnFlow = 1 << 30 |
| |
| // transportDefaultStreamFlow is how many stream-level flow |
| // control tokens we announce to the peer, and how many bytes |
| // we buffer per stream. |
| transportDefaultStreamFlow = 4 << 20 |
| |
| // transportDefaultStreamMinRefresh is the minimum number of bytes we'll send |
| // a stream-level WINDOW_UPDATE for at a time. |
| transportDefaultStreamMinRefresh = 4 << 10 |
| ) |
| |
| // Transport is an HTTP/2 Transport. |
| // |
| // A Transport internally caches connections to servers. It is safe |
| // for concurrent use by multiple goroutines. |
| type Transport struct { |
| // DialTLS specifies an optional dial function for creating |
| // TLS connections for requests. |
| // |
| // If DialTLS is nil, tls.Dial is used. |
| // |
| // If the returned net.Conn has a ConnectionState method like tls.Conn, |
| // it will be used to set http.Response.TLS. |
| DialTLS func(network, addr string, cfg *tls.Config) (net.Conn, error) |
| |
| // TLSClientConfig specifies the TLS configuration to use with |
| // tls.Client. If nil, the default configuration is used. |
| TLSClientConfig *tls.Config |
| |
| // TODO: switch to RWMutex |
| // TODO: add support for sharing conns based on cert names |
| // (e.g. share conn for googleapis.com and appspot.com) |
| connMu sync.Mutex |
| conns map[string][]*clientConn // key is host:port |
| } |
| |
| // clientConn is the state of a single HTTP/2 client connection to an |
| // HTTP/2 server. |
| type clientConn struct { |
| t *Transport |
| tconn net.Conn |
| tlsState *tls.ConnectionState |
| connKey []string // key(s) this connection is cached in, in t.conns |
| |
| // readLoop goroutine fields: |
| readerDone chan struct{} // closed on error |
| readerErr error // set before readerDone is closed |
| |
| mu sync.Mutex // guards following |
| cond *sync.Cond // hold mu; broadcast on flow/closed changes |
| flow flow // our conn-level flow control quota (cs.flow is per stream) |
| inflow flow // peer's conn-level flow control |
| closed bool |
| goAway *GoAwayFrame // if non-nil, the GoAwayFrame we received |
| streams map[uint32]*clientStream // client-initiated |
| nextStreamID uint32 |
| bw *bufio.Writer |
| br *bufio.Reader |
| fr *Framer |
| // Settings from peer: |
| maxFrameSize uint32 |
| maxConcurrentStreams uint32 |
| initialWindowSize uint32 |
| hbuf bytes.Buffer // HPACK encoder writes into this |
| henc *hpack.Encoder |
| freeBuf [][]byte |
| |
| wmu sync.Mutex // held while writing; acquire AFTER wmu if holding both |
| werr error // first write error that has occurred |
| } |
| |
| // clientStream is the state for a single HTTP/2 stream. One of these |
| // is created for each Transport.RoundTrip call. |
| type clientStream struct { |
| cc *clientConn |
| ID uint32 |
| resc chan resAndError |
| bufPipe pipe // buffered pipe with the flow-controlled response payload |
| |
| flow flow // guarded by cc.mu |
| inflow flow // guarded by cc.mu |
| |
| peerReset chan struct{} // closed on peer reset |
| resetErr error // populated before peerReset is closed |
| } |
| |
| // checkReset reports any error sent in a RST_STREAM frame by the |
| // server. |
| func (cs *clientStream) checkReset() error { |
| select { |
| case <-cs.peerReset: |
| return cs.resetErr |
| default: |
| return nil |
| } |
| } |
| |
| type stickyErrWriter struct { |
| w io.Writer |
| err *error |
| } |
| |
| func (sew stickyErrWriter) Write(p []byte) (n int, err error) { |
| if *sew.err != nil { |
| return 0, *sew.err |
| } |
| n, err = sew.w.Write(p) |
| *sew.err = err |
| return |
| } |
| |
| func (t *Transport) RoundTrip(req *http.Request) (*http.Response, error) { |
| if req.URL.Scheme != "https" { |
| return nil, errors.New("http2: unsupported scheme") |
| } |
| |
| host, port, err := net.SplitHostPort(req.URL.Host) |
| if err != nil { |
| host = req.URL.Host |
| port = "443" |
| } |
| |
| for { |
| cc, err := t.getClientConn(host, port) |
| if err != nil { |
| return nil, err |
| } |
| res, err := cc.roundTrip(req) |
| if shouldRetryRequest(err) { // TODO: or clientconn is overloaded (too many outstanding requests)? |
| continue |
| } |
| if err != nil { |
| return nil, err |
| } |
| return res, nil |
| } |
| } |
| |
| // CloseIdleConnections closes any connections which were previously |
| // connected from previous requests but are now sitting idle. |
| // It does not interrupt any connections currently in use. |
| func (t *Transport) CloseIdleConnections() { |
| t.connMu.Lock() |
| defer t.connMu.Unlock() |
| for _, vv := range t.conns { |
| for _, cc := range vv { |
| cc.closeIfIdle() |
| } |
| } |
| } |
| |
| var errClientConnClosed = errors.New("http2: client conn is closed") |
| |
| func shouldRetryRequest(err error) bool { |
| // TODO: or GOAWAY graceful shutdown stuff |
| return err == errClientConnClosed |
| } |
| |
| func (t *Transport) removeClientConn(cc *clientConn) { |
| t.connMu.Lock() |
| defer t.connMu.Unlock() |
| for _, key := range cc.connKey { |
| vv, ok := t.conns[key] |
| if !ok { |
| continue |
| } |
| newList := filterOutClientConn(vv, cc) |
| if len(newList) > 0 { |
| t.conns[key] = newList |
| } else { |
| delete(t.conns, key) |
| } |
| } |
| } |
| |
| func filterOutClientConn(in []*clientConn, exclude *clientConn) []*clientConn { |
| out := in[:0] |
| for _, v := range in { |
| if v != exclude { |
| out = append(out, v) |
| } |
| } |
| // If we filtered it out, zero out the last item to prevent |
| // the GC from seeing it. |
| if len(in) != len(out) { |
| in[len(in)-1] = nil |
| } |
| return out |
| } |
| |
| // AddIdleConn adds c as an idle conn for Transport. |
| // It assumes that c has not yet exchanged SETTINGS frames. |
| // The addr maybe be either "host" or "host:port". |
| func (t *Transport) AddIdleConn(addr string, c *tls.Conn) error { |
| var key string |
| _, _, err := net.SplitHostPort(addr) |
| if err == nil { |
| key = addr |
| } else { |
| key = addr + ":443" |
| } |
| cc, err := t.newClientConn(key, c) |
| if err != nil { |
| return err |
| } |
| |
| t.addConn(key, cc) |
| return nil |
| } |
| |
| func (t *Transport) addConn(key string, cc *clientConn) { |
| t.connMu.Lock() |
| defer t.connMu.Unlock() |
| if t.conns == nil { |
| t.conns = make(map[string][]*clientConn) |
| } |
| t.conns[key] = append(t.conns[key], cc) |
| } |
| |
| func (t *Transport) getClientConn(host, port string) (*clientConn, error) { |
| key := net.JoinHostPort(host, port) |
| |
| t.connMu.Lock() |
| for _, cc := range t.conns[key] { |
| if cc.canTakeNewRequest() { |
| t.connMu.Unlock() |
| return cc, nil |
| } |
| } |
| t.connMu.Unlock() |
| |
| // TODO(bradfitz): use a singleflight.Group to only lock once per 'key'. |
| // Probably need to vendor it in as github.com/golang/sync/singleflight |
| // though, since the net package already uses it? Also lines up with |
| // sameer, bcmills, et al wanting to open source some sync stuff. |
| cc, err := t.dialClientConn(host, port, key) |
| if err != nil { |
| return nil, err |
| } |
| t.addConn(key, cc) |
| return cc, nil |
| } |
| |
| func (t *Transport) dialClientConn(host, port, key string) (*clientConn, error) { |
| tconn, err := t.dialTLS()("tcp", net.JoinHostPort(host, port), t.newTLSConfig(host)) |
| if err != nil { |
| return nil, err |
| } |
| return t.newClientConn(key, tconn) |
| } |
| |
| func (t *Transport) newTLSConfig(host string) *tls.Config { |
| cfg := new(tls.Config) |
| if t.TLSClientConfig != nil { |
| *cfg = *t.TLSClientConfig |
| } |
| cfg.NextProtos = []string{NextProtoTLS} // TODO: don't override if already in list |
| cfg.ServerName = host |
| return cfg |
| } |
| |
| func (t *Transport) dialTLS() func(string, string, *tls.Config) (net.Conn, error) { |
| if t.DialTLS != nil { |
| return t.DialTLS |
| } |
| return t.dialTLSDefault |
| } |
| |
| func (t *Transport) dialTLSDefault(network, addr string, cfg *tls.Config) (net.Conn, error) { |
| cn, err := tls.Dial(network, addr, cfg) |
| if err != nil { |
| return nil, err |
| } |
| if err := cn.Handshake(); err != nil { |
| return nil, err |
| } |
| if !cfg.InsecureSkipVerify { |
| if err := cn.VerifyHostname(cfg.ServerName); err != nil { |
| return nil, err |
| } |
| } |
| state := cn.ConnectionState() |
| if p := state.NegotiatedProtocol; p != NextProtoTLS { |
| return nil, fmt.Errorf("http2: unexpected ALPN protocol %q; want %q", p, NextProtoTLS) |
| } |
| if !state.NegotiatedProtocolIsMutual { |
| return nil, errors.New("http2: could not negotiate protocol mutually") |
| } |
| return cn, nil |
| } |
| |
| func (t *Transport) newClientConn(key string, tconn net.Conn) (*clientConn, error) { |
| if _, err := tconn.Write(clientPreface); err != nil { |
| return nil, err |
| } |
| |
| cc := &clientConn{ |
| t: t, |
| tconn: tconn, |
| connKey: []string{key}, // TODO: cert's validated hostnames too |
| readerDone: make(chan struct{}), |
| nextStreamID: 1, |
| maxFrameSize: 16 << 10, // spec default |
| initialWindowSize: 65535, // spec default |
| maxConcurrentStreams: 1000, // "infinite", per spec. 1000 seems good enough. |
| streams: make(map[uint32]*clientStream), |
| } |
| cc.cond = sync.NewCond(&cc.mu) |
| cc.flow.add(int32(initialWindowSize)) |
| |
| // TODO: adjust this writer size to account for frame size + |
| // MTU + crypto/tls record padding. |
| cc.bw = bufio.NewWriter(stickyErrWriter{tconn, &cc.werr}) |
| cc.br = bufio.NewReader(tconn) |
| cc.fr = NewFramer(cc.bw, cc.br) |
| cc.henc = hpack.NewEncoder(&cc.hbuf) |
| |
| type connectionStater interface { |
| ConnectionState() tls.ConnectionState |
| } |
| if cs, ok := tconn.(connectionStater); ok { |
| state := cs.ConnectionState() |
| cc.tlsState = &state |
| } |
| |
| cc.fr.WriteSettings( |
| Setting{ID: SettingEnablePush, Val: 0}, |
| Setting{ID: SettingInitialWindowSize, Val: transportDefaultStreamFlow}, |
| ) |
| cc.fr.WriteWindowUpdate(0, transportDefaultConnFlow) |
| cc.inflow.add(transportDefaultConnFlow + initialWindowSize) |
| cc.bw.Flush() |
| if cc.werr != nil { |
| return nil, cc.werr |
| } |
| |
| // Read the obligatory SETTINGS frame |
| f, err := cc.fr.ReadFrame() |
| if err != nil { |
| return nil, err |
| } |
| sf, ok := f.(*SettingsFrame) |
| if !ok { |
| return nil, fmt.Errorf("expected settings frame, got: %T", f) |
| } |
| cc.fr.WriteSettingsAck() |
| cc.bw.Flush() |
| |
| sf.ForeachSetting(func(s Setting) error { |
| switch s.ID { |
| case SettingMaxFrameSize: |
| cc.maxFrameSize = s.Val |
| case SettingMaxConcurrentStreams: |
| cc.maxConcurrentStreams = s.Val |
| case SettingInitialWindowSize: |
| cc.initialWindowSize = s.Val |
| default: |
| // TODO(bradfitz): handle more |
| t.vlogf("Unhandled Setting: %v", s) |
| } |
| return nil |
| }) |
| |
| go cc.readLoop() |
| return cc, nil |
| } |
| |
| func (cc *clientConn) setGoAway(f *GoAwayFrame) { |
| cc.mu.Lock() |
| defer cc.mu.Unlock() |
| cc.goAway = f |
| } |
| |
| func (cc *clientConn) canTakeNewRequest() bool { |
| cc.mu.Lock() |
| defer cc.mu.Unlock() |
| return cc.goAway == nil && |
| int64(len(cc.streams)+1) < int64(cc.maxConcurrentStreams) && |
| cc.nextStreamID < 2147483647 |
| } |
| |
| func (cc *clientConn) closeIfIdle() { |
| cc.mu.Lock() |
| if len(cc.streams) > 0 { |
| cc.mu.Unlock() |
| return |
| } |
| cc.closed = true |
| // TODO: do clients send GOAWAY too? maybe? Just Close: |
| cc.mu.Unlock() |
| |
| cc.tconn.Close() |
| } |
| |
| const maxAllocFrameSize = 512 << 10 |
| |
| // frameBuffer returns a scratch buffer suitable for writing DATA frames. |
| // They're capped at the min of the peer's max frame size or 512KB |
| // (kinda arbitrarily), but definitely capped so we don't allocate 4GB |
| // bufers. |
| func (cc *clientConn) frameScratchBuffer() []byte { |
| cc.mu.Lock() |
| size := cc.maxFrameSize |
| if size > maxAllocFrameSize { |
| size = maxAllocFrameSize |
| } |
| for i, buf := range cc.freeBuf { |
| if len(buf) >= int(size) { |
| cc.freeBuf[i] = nil |
| cc.mu.Unlock() |
| return buf[:size] |
| } |
| } |
| cc.mu.Unlock() |
| return make([]byte, size) |
| } |
| |
| func (cc *clientConn) putFrameScratchBuffer(buf []byte) { |
| cc.mu.Lock() |
| defer cc.mu.Unlock() |
| const maxBufs = 4 // arbitrary; 4 concurrent requests per conn? investigate. |
| if len(cc.freeBuf) < maxBufs { |
| cc.freeBuf = append(cc.freeBuf, buf) |
| return |
| } |
| for i, old := range cc.freeBuf { |
| if old == nil { |
| cc.freeBuf[i] = buf |
| return |
| } |
| } |
| // forget about it. |
| } |
| |
| func (cc *clientConn) roundTrip(req *http.Request) (*http.Response, error) { |
| cc.mu.Lock() |
| |
| if cc.closed { |
| cc.mu.Unlock() |
| return nil, errClientConnClosed |
| } |
| |
| cs := cc.newStream() |
| hasBody := req.Body != nil |
| |
| // we send: HEADERS[+CONTINUATION] + (DATA?) |
| hdrs := cc.encodeHeaders(req) |
| first := true |
| |
| cc.wmu.Lock() |
| frameSize := int(cc.maxFrameSize) |
| for len(hdrs) > 0 && cc.werr == nil { |
| chunk := hdrs |
| if len(chunk) > frameSize { |
| chunk = chunk[:frameSize] |
| } |
| hdrs = hdrs[len(chunk):] |
| endHeaders := len(hdrs) == 0 |
| if first { |
| cc.fr.WriteHeaders(HeadersFrameParam{ |
| StreamID: cs.ID, |
| BlockFragment: chunk, |
| EndStream: !hasBody, |
| EndHeaders: endHeaders, |
| }) |
| first = false |
| } else { |
| cc.fr.WriteContinuation(cs.ID, endHeaders, chunk) |
| } |
| } |
| cc.bw.Flush() |
| werr := cc.werr |
| cc.wmu.Unlock() |
| cc.mu.Unlock() |
| |
| if werr != nil { |
| return nil, werr |
| } |
| |
| var bodyCopyErrc chan error |
| var gotResHeaders chan struct{} // closed on resheaders |
| if hasBody { |
| bodyCopyErrc = make(chan error, 1) |
| gotResHeaders = make(chan struct{}) |
| go func() { |
| bodyCopyErrc <- cs.writeRequestBody(req.Body, gotResHeaders) |
| }() |
| } |
| |
| for { |
| select { |
| case re := <-cs.resc: |
| if gotResHeaders != nil { |
| close(gotResHeaders) |
| } |
| if re.err != nil { |
| return nil, re.err |
| } |
| res := re.res |
| res.Request = req |
| res.TLS = cc.tlsState |
| return res, nil |
| case err := <-bodyCopyErrc: |
| if err != nil { |
| return nil, err |
| } |
| } |
| } |
| } |
| |
| var errServerResponseBeforeRequestBody = errors.New("http2: server sent response while still writing request body") |
| |
| func (cs *clientStream) writeRequestBody(body io.Reader, gotResHeaders <-chan struct{}) error { |
| cc := cs.cc |
| done := false |
| buf := cc.frameScratchBuffer() |
| defer cc.putFrameScratchBuffer(buf) |
| |
| for !done { |
| n, err := io.ReadFull(body, buf) |
| if err == io.ErrUnexpectedEOF { |
| done = true |
| err = nil |
| } else if err == io.EOF { |
| break |
| } else if err != nil { |
| return err |
| } |
| |
| toWrite := buf[:n] |
| for len(toWrite) > 0 && err == nil { |
| var allowed int32 |
| allowed, err = cs.awaitFlowControl(int32(len(toWrite))) |
| if err != nil { |
| return err |
| } |
| |
| cc.wmu.Lock() |
| select { |
| case <-gotResHeaders: |
| err = errServerResponseBeforeRequestBody |
| case <-cs.peerReset: |
| err = cs.resetErr |
| default: |
| err = cc.fr.WriteData(cs.ID, done, toWrite[:allowed]) |
| toWrite = toWrite[allowed:] |
| } |
| cc.wmu.Unlock() |
| } |
| if err != nil { |
| return err |
| } |
| } |
| |
| var err error |
| |
| cc.wmu.Lock() |
| if !done { |
| err = cc.fr.WriteData(cs.ID, true, nil) |
| } |
| if ferr := cc.bw.Flush(); ferr != nil && err == nil { |
| err = ferr |
| } |
| cc.wmu.Unlock() |
| |
| return err |
| } |
| |
| // awaitFlowControl waits for [1, min(maxBytes, cc.cs.maxFrameSize)] flow |
| // control tokens from the server. |
| // It returns either the non-zero number of tokens taken or an error |
| // if the stream is dead. |
| func (cs *clientStream) awaitFlowControl(maxBytes int32) (taken int32, err error) { |
| cc := cs.cc |
| cc.mu.Lock() |
| defer cc.mu.Unlock() |
| for { |
| if cc.closed { |
| return 0, errClientConnClosed |
| } |
| if err := cs.checkReset(); err != nil { |
| return 0, err |
| } |
| if a := cs.flow.available(); a > 0 { |
| take := a |
| if take > maxBytes { |
| take = maxBytes |
| } |
| if take > int32(cc.maxFrameSize) { |
| take = int32(cc.maxFrameSize) |
| } |
| cs.flow.take(take) |
| return take, nil |
| } |
| cc.cond.Wait() |
| } |
| } |
| |
| // requires cc.mu be held. |
| func (cc *clientConn) encodeHeaders(req *http.Request) []byte { |
| cc.hbuf.Reset() |
| |
| // TODO(bradfitz): figure out :authority-vs-Host stuff between http2 and Go |
| host := req.Host |
| if host == "" { |
| host = req.URL.Host |
| } |
| |
| // 8.1.2.3 Request Pseudo-Header Fields |
| // The :path pseudo-header field includes the path and query parts of the |
| // target URI (the path-absolute production and optionally a '?' character |
| // followed by the query production (see Sections 3.3 and 3.4 of |
| // [RFC3986]). |
| cc.writeHeader(":authority", host) // probably not right for all sites |
| cc.writeHeader(":method", req.Method) |
| cc.writeHeader(":path", req.URL.RequestURI()) |
| cc.writeHeader(":scheme", "https") |
| |
| for k, vv := range req.Header { |
| lowKey := strings.ToLower(k) |
| if lowKey == "host" { |
| continue |
| } |
| for _, v := range vv { |
| cc.writeHeader(lowKey, v) |
| } |
| } |
| return cc.hbuf.Bytes() |
| } |
| |
| func (cc *clientConn) writeHeader(name, value string) { |
| cc.henc.WriteField(hpack.HeaderField{Name: name, Value: value}) |
| } |
| |
| type resAndError struct { |
| res *http.Response |
| err error |
| } |
| |
| // requires cc.mu be held. |
| func (cc *clientConn) newStream() *clientStream { |
| cs := &clientStream{ |
| cc: cc, |
| ID: cc.nextStreamID, |
| resc: make(chan resAndError, 1), |
| peerReset: make(chan struct{}), |
| } |
| cs.flow.add(int32(cc.initialWindowSize)) |
| cs.flow.setConnFlow(&cc.flow) |
| cs.inflow.add(transportDefaultStreamFlow) |
| cs.inflow.setConnFlow(&cc.inflow) |
| cc.nextStreamID += 2 |
| cc.streams[cs.ID] = cs |
| return cs |
| } |
| |
| func (cc *clientConn) streamByID(id uint32, andRemove bool) *clientStream { |
| cc.mu.Lock() |
| defer cc.mu.Unlock() |
| cs := cc.streams[id] |
| if andRemove { |
| delete(cc.streams, id) |
| } |
| return cs |
| } |
| |
| // clientConnReadLoop is the state owned by the clientConn's frame-reading readLoop. |
| type clientConnReadLoop struct { |
| cc *clientConn |
| activeRes map[uint32]*clientStream // keyed by streamID |
| |
| // continueStreamID is the stream ID we're waiting for |
| // continuation frames for. |
| continueStreamID uint32 |
| |
| hdec *hpack.Decoder |
| |
| // Fields reset on each HEADERS: |
| nextRes *http.Response |
| sawRegHeader bool // saw non-pseudo header |
| reqMalformed error // non-nil once known to be malformed |
| } |
| |
| // readLoop runs in its own goroutine and reads and dispatches frames. |
| func (cc *clientConn) readLoop() { |
| rl := &clientConnReadLoop{ |
| cc: cc, |
| activeRes: make(map[uint32]*clientStream), |
| } |
| // TODO: figure out henc size |
| rl.hdec = hpack.NewDecoder(initialHeaderTableSize, rl.onNewHeaderField) |
| |
| defer rl.cleanup() |
| cc.readerErr = rl.run() |
| if ce, ok := cc.readerErr.(ConnectionError); ok { |
| cc.wmu.Lock() |
| cc.fr.WriteGoAway(0, ErrCode(ce), nil) |
| cc.wmu.Unlock() |
| } |
| } |
| |
| func (rl *clientConnReadLoop) cleanup() { |
| cc := rl.cc |
| defer cc.tconn.Close() |
| defer cc.t.removeClientConn(cc) |
| defer close(cc.readerDone) |
| |
| // Close any response bodies if the server closes prematurely. |
| // TODO: also do this if we've written the headers but not |
| // gotten a response yet. |
| err := cc.readerErr |
| if err == io.EOF { |
| err = io.ErrUnexpectedEOF |
| } |
| cc.mu.Lock() |
| for _, cs := range rl.activeRes { |
| cs.bufPipe.CloseWithError(err) |
| } |
| for _, cs := range cc.streams { |
| select { |
| case cs.resc <- resAndError{err: err}: |
| default: |
| } |
| } |
| cc.closed = true |
| cc.cond.Broadcast() |
| cc.mu.Unlock() |
| } |
| |
| func (rl *clientConnReadLoop) run() error { |
| cc := rl.cc |
| for { |
| f, err := cc.fr.ReadFrame() |
| if se, ok := err.(StreamError); ok { |
| // TODO: deal with stream errors from the framer. |
| return se |
| } else if err != nil { |
| return err |
| } |
| cc.vlogf("Transport received %v: %#v", f.Header(), f) |
| |
| streamID := f.Header().StreamID |
| |
| _, isContinue := f.(*ContinuationFrame) |
| if isContinue { |
| if streamID != rl.continueStreamID { |
| cc.logf("Protocol violation: got CONTINUATION with id %d; want %d", streamID, rl.continueStreamID) |
| return ConnectionError(ErrCodeProtocol) |
| } |
| } else if rl.continueStreamID != 0 { |
| // Continue frames need to be adjacent in the stream |
| // and we were in the middle of headers. |
| cc.logf("Protocol violation: got %T for stream %d, want CONTINUATION for %d", f, streamID, rl.continueStreamID) |
| return ConnectionError(ErrCodeProtocol) |
| } |
| |
| switch f := f.(type) { |
| case *HeadersFrame: |
| err = rl.processHeaders(f) |
| case *ContinuationFrame: |
| err = rl.processContinuation(f) |
| case *DataFrame: |
| err = rl.processData(f) |
| case *GoAwayFrame: |
| err = rl.processGoAway(f) |
| case *RSTStreamFrame: |
| err = rl.processResetStream(f) |
| case *SettingsFrame: |
| err = rl.processSettings(f) |
| case *PushPromiseFrame: |
| err = rl.processPushPromise(f) |
| case *WindowUpdateFrame: |
| err = rl.processWindowUpdate(f) |
| default: |
| cc.logf("Transport: unhandled response frame type %T", f) |
| } |
| if err != nil { |
| return err |
| } |
| } |
| } |
| |
| func (rl *clientConnReadLoop) processHeaders(f *HeadersFrame) error { |
| rl.sawRegHeader = false |
| rl.reqMalformed = nil |
| rl.nextRes = &http.Response{ |
| Proto: "HTTP/2.0", |
| ProtoMajor: 2, |
| Header: make(http.Header), |
| } |
| return rl.processHeaderBlockFragment(f.HeaderBlockFragment(), f.StreamID, f.HeadersEnded(), f.StreamEnded()) |
| } |
| |
| func (rl *clientConnReadLoop) processContinuation(f *ContinuationFrame) error { |
| return rl.processHeaderBlockFragment(f.HeaderBlockFragment(), f.StreamID, f.HeadersEnded(), f.StreamEnded()) |
| } |
| |
| func (rl *clientConnReadLoop) processHeaderBlockFragment(frag []byte, streamID uint32, headersEnded, streamEnded bool) error { |
| cc := rl.cc |
| cs := cc.streamByID(streamID, streamEnded) |
| if cs == nil { |
| // We could return a ConnectionError(ErrCodeProtocol) |
| // here, except that in the case of us canceling |
| // client requests, we may also delete from the |
| // streams map, in which case we forgot that we sent |
| // this request. So, just ignore any responses for |
| // now. They might've been in-flight before the |
| // server got our RST_STREAM. |
| return nil |
| } |
| _, err := rl.hdec.Write(frag) |
| if err != nil { |
| return err |
| } |
| if !headersEnded { |
| rl.continueStreamID = cs.ID |
| return nil |
| } |
| |
| // HEADERS (or CONTINUATION) are now over. |
| rl.continueStreamID = 0 |
| |
| if rl.reqMalformed != nil { |
| cs.resc <- resAndError{err: rl.reqMalformed} |
| rl.cc.writeStreamReset(cs.ID, ErrCodeProtocol, rl.reqMalformed) |
| return nil |
| } |
| |
| res := rl.nextRes |
| if streamEnded { |
| res.Body = noBody |
| } else { |
| buf := new(bytes.Buffer) // TODO(bradfitz): recycle this garbage |
| cs.bufPipe = pipe{b: buf} |
| res.Body = transportResponseBody{cs} |
| } |
| rl.activeRes[cs.ID] = cs |
| cs.resc <- resAndError{res: res} |
| rl.nextRes = nil // unused now; will be reset next HEADERS frame |
| return nil |
| } |
| |
| // transportResponseBody is the concrete type of Transport.RoundTrip's |
| // Response.Body. It is an io.ReadCloser. On Read, it reads from cs.body. |
| // On Close it sends RST_STREAM if EOF wasn't already seen. |
| type transportResponseBody struct { |
| cs *clientStream |
| } |
| |
| func (b transportResponseBody) Read(p []byte) (n int, err error) { |
| n, err = b.cs.bufPipe.Read(p) |
| if n == 0 { |
| return |
| } |
| |
| cs := b.cs |
| cc := cs.cc |
| cc.mu.Lock() |
| defer cc.mu.Unlock() |
| |
| var connAdd, streamAdd int32 |
| // Check the conn-level first, before the stream-level. |
| if v := cc.inflow.available(); v < transportDefaultConnFlow/2 { |
| connAdd = transportDefaultConnFlow - v |
| cc.inflow.add(connAdd) |
| } |
| if err == nil { // No need to refresh if the stream is over or failed. |
| if v := cs.inflow.available(); v < transportDefaultStreamFlow-transportDefaultStreamMinRefresh { |
| streamAdd = transportDefaultStreamFlow - v |
| cs.inflow.add(streamAdd) |
| } |
| } |
| if connAdd != 0 || streamAdd != 0 { |
| cc.wmu.Lock() |
| defer cc.wmu.Unlock() |
| if connAdd != 0 { |
| cc.fr.WriteWindowUpdate(0, mustUint31(connAdd)) |
| } |
| if streamAdd != 0 { |
| cc.fr.WriteWindowUpdate(cs.ID, mustUint31(streamAdd)) |
| } |
| cc.bw.Flush() |
| } |
| return |
| } |
| |
| func (b transportResponseBody) Close() error { |
| if b.cs.bufPipe.Err() != io.EOF { |
| // TODO: write test for this |
| b.cs.cc.writeStreamReset(b.cs.ID, ErrCodeCancel, nil) |
| } |
| return nil |
| } |
| |
| func (rl *clientConnReadLoop) processData(f *DataFrame) error { |
| cc := rl.cc |
| cs := cc.streamByID(f.StreamID, f.StreamEnded()) |
| if cs == nil { |
| return nil |
| } |
| data := f.Data() |
| if VerboseLogs { |
| rl.cc.logf("DATA: %q", data) |
| } |
| |
| // Check connection-level flow control. |
| cc.mu.Lock() |
| if cs.inflow.available() >= int32(len(data)) { |
| cs.inflow.take(int32(len(data))) |
| } else { |
| cc.mu.Unlock() |
| return ConnectionError(ErrCodeFlowControl) |
| } |
| cc.mu.Unlock() |
| |
| if _, err := cs.bufPipe.Write(data); err != nil { |
| return err |
| } |
| |
| if f.StreamEnded() { |
| cs.bufPipe.CloseWithError(io.EOF) |
| delete(rl.activeRes, cs.ID) |
| } |
| return nil |
| } |
| |
| func (rl *clientConnReadLoop) processGoAway(f *GoAwayFrame) error { |
| cc := rl.cc |
| cc.t.removeClientConn(cc) |
| if f.ErrCode != 0 { |
| // TODO: deal with GOAWAY more. particularly the error code |
| cc.vlogf("transport got GOAWAY with error code = %v", f.ErrCode) |
| } |
| cc.setGoAway(f) |
| return nil |
| } |
| |
| func (rl *clientConnReadLoop) processSettings(f *SettingsFrame) error { |
| cc := rl.cc |
| cc.mu.Lock() |
| defer cc.mu.Unlock() |
| return f.ForeachSetting(func(s Setting) error { |
| switch s.ID { |
| case SettingMaxFrameSize: |
| cc.maxFrameSize = s.Val |
| case SettingMaxConcurrentStreams: |
| cc.maxConcurrentStreams = s.Val |
| case SettingInitialWindowSize: |
| // TODO: error if this is too large. |
| |
| // TODO: adjust flow control of still-open |
| // frames by the difference of the old initial |
| // window size and this one. |
| cc.initialWindowSize = s.Val |
| default: |
| // TODO(bradfitz): handle more settings? |
| cc.vlogf("Unhandled Setting: %v", s) |
| } |
| return nil |
| }) |
| } |
| |
| func (rl *clientConnReadLoop) processWindowUpdate(f *WindowUpdateFrame) error { |
| cc := rl.cc |
| cs := cc.streamByID(f.StreamID, false) |
| if f.StreamID != 0 && cs == nil { |
| return nil |
| } |
| |
| cc.mu.Lock() |
| defer cc.mu.Unlock() |
| |
| fl := &cc.flow |
| if cs != nil { |
| fl = &cs.flow |
| } |
| if !fl.add(int32(f.Increment)) { |
| return ConnectionError(ErrCodeFlowControl) |
| } |
| cc.cond.Broadcast() |
| return nil |
| } |
| |
| func (rl *clientConnReadLoop) processResetStream(f *RSTStreamFrame) error { |
| cs := rl.cc.streamByID(f.StreamID, true) |
| if cs == nil { |
| // TODO: return error if server tries to RST_STEAM an idle stream |
| return nil |
| } |
| select { |
| case <-cs.peerReset: |
| // Already reset. |
| // This is the only goroutine |
| // which closes this, so there |
| // isn't a race. |
| default: |
| err := StreamError{cs.ID, f.ErrCode} |
| cs.resetErr = err |
| close(cs.peerReset) |
| cs.bufPipe.CloseWithError(err) |
| } |
| delete(rl.activeRes, cs.ID) |
| return nil |
| } |
| |
| func (rl *clientConnReadLoop) processPushPromise(f *PushPromiseFrame) error { |
| // We told the peer we don't want them. |
| // Spec says: |
| // "PUSH_PROMISE MUST NOT be sent if the SETTINGS_ENABLE_PUSH |
| // setting of the peer endpoint is set to 0. An endpoint that |
| // has set this setting and has received acknowledgement MUST |
| // treat the receipt of a PUSH_PROMISE frame as a connection |
| // error (Section 5.4.1) of type PROTOCOL_ERROR." |
| return ConnectionError(ErrCodeProtocol) |
| } |
| |
| func (cc *clientConn) writeStreamReset(streamID uint32, code ErrCode, err error) { |
| // TODO: do something with err? send it as a debug frame to the peer? |
| // But that's only in GOAWAY. Invent a new frame type? Is there one already? |
| cc.wmu.Lock() |
| cc.fr.WriteRSTStream(streamID, code) |
| cc.wmu.Unlock() |
| } |
| |
| // onNewHeaderField runs on the readLoop goroutine whenever a new |
| // hpack header field is decoded. |
| func (rl *clientConnReadLoop) onNewHeaderField(f hpack.HeaderField) { |
| cc := rl.cc |
| if VerboseLogs { |
| cc.logf("Header field: %+v", f) |
| } |
| isPseudo := strings.HasPrefix(f.Name, ":") |
| if isPseudo { |
| if rl.sawRegHeader { |
| rl.reqMalformed = errors.New("http2: invalid pseudo header after regular header") |
| return |
| } |
| switch f.Name { |
| case ":status": |
| code, err := strconv.Atoi(f.Value) |
| if err != nil { |
| rl.reqMalformed = errors.New("http2: invalid :status") |
| return |
| } |
| rl.nextRes.Status = f.Value + " " + http.StatusText(code) |
| rl.nextRes.StatusCode = code |
| default: |
| // "Endpoints MUST NOT generate pseudo-header |
| // fields other than those defined in this |
| // document." |
| rl.reqMalformed = fmt.Errorf("http2: unknown response pseudo header %q", f.Name) |
| } |
| } else { |
| rl.sawRegHeader = true |
| rl.nextRes.Header.Add(http.CanonicalHeaderKey(f.Name), f.Value) |
| } |
| } |
| |
| func (cc *clientConn) logf(format string, args ...interface{}) { |
| cc.t.logf(format, args...) |
| } |
| |
| func (cc *clientConn) vlogf(format string, args ...interface{}) { |
| cc.t.vlogf(format, args...) |
| } |
| |
| func (t *Transport) vlogf(format string, args ...interface{}) { |
| if VerboseLogs { |
| t.logf(format, args...) |
| } |
| } |
| |
| func (t *Transport) logf(format string, args ...interface{}) { |
| log.Printf(format, args...) |
| } |
| |
| var noBody io.ReadCloser = ioutil.NopCloser(bytes.NewReader(nil)) |