| // Copyright 2015 The Go Authors. All rights reserved. |
| // Use of this source code is governed by a BSD-style |
| // license that can be found in the LICENSE file. |
| |
| // Transport code. |
| // |
| // TODO: send flow control WINDOW_UPDATE frames as DATA is received, |
| // for both stream and conn levels. Currently this code will hang |
| // after 1GB is downloaded total. |
| // TODO: buffer each stream's data (~few MB), flow controlled separately, |
| // so one ignored Response.Body doesn't block all goroutines. |
| |
| package http2 |
| |
| import ( |
| "bufio" |
| "bytes" |
| "crypto/tls" |
| "errors" |
| "fmt" |
| "io" |
| "log" |
| "net" |
| "net/http" |
| "strconv" |
| "strings" |
| "sync" |
| |
| "golang.org/x/net/http2/hpack" |
| ) |
| |
| const ( |
| // transportDefaultConnFlow is how many connection-level flow control |
| // tokens we give the server at start-up, past the default 64k. |
| transportDefaultConnFlow = 1 << 30 |
| |
| // transportDefaultStreamFlow is how many stream-level flow |
| // control tokens we announce to the peer, and how many bytes |
| // we buffer per stream. |
| transportDefaultStreamFlow = 4 << 20 |
| ) |
| |
| // Transport is an HTTP/2 Transport. |
| // |
| // A Transport internally caches connections to servers. It is safe |
| // for concurrent use by multiple goroutines. |
| type Transport struct { |
| // TODO: remove this and make more general with a TLS dial hook, like http |
| InsecureTLSDial bool |
| |
| // TODO: switch to RWMutex |
| // TODO: add support for sharing conns based on cert names |
| // (e.g. share conn for googleapis.com and appspot.com) |
| connMu sync.Mutex |
| conns map[string][]*clientConn // key is host:port |
| } |
| |
| // clientConn is the state of a single HTTP/2 client connection to an |
| // HTTP/2 server. |
| type clientConn struct { |
| t *Transport |
| tconn *tls.Conn |
| tlsState *tls.ConnectionState |
| connKey []string // key(s) this connection is cached in, in t.conns |
| |
| // readLoop goroutine fields: |
| readerDone chan struct{} // closed on error |
| readerErr error // set before readerDone is closed |
| |
| mu sync.Mutex // guards following |
| cond *sync.Cond // hold mu; broadcast on flow/closed changes |
| flow flow // our conn-level flow control quota (cs.flow is per stream) |
| inflow flow // peer's conn-level flow control |
| closed bool |
| goAway *GoAwayFrame // if non-nil, the GoAwayFrame we received |
| streams map[uint32]*clientStream // client-initiated |
| nextStreamID uint32 |
| bw *bufio.Writer |
| br *bufio.Reader |
| fr *Framer |
| // Settings from peer: |
| maxFrameSize uint32 |
| maxConcurrentStreams uint32 |
| initialWindowSize uint32 |
| hbuf bytes.Buffer // HPACK encoder writes into this |
| henc *hpack.Encoder |
| freeBuf [][]byte |
| |
| wmu sync.Mutex // held while writing; acquire AFTER wmu if holding both |
| werr error // first write error that has occurred |
| } |
| |
| type clientStream struct { |
| cc *clientConn |
| ID uint32 |
| resc chan resAndError |
| pw *io.PipeWriter |
| pr *io.PipeReader |
| |
| flow flow // guarded by cc.mu |
| inflow flow // guarded by cc.mu |
| |
| peerReset chan struct{} // closed on peer reset |
| resetErr error // populated before peerReset is closed |
| } |
| |
| func (cs *clientStream) checkReset() error { |
| select { |
| case <-cs.peerReset: |
| return cs.resetErr |
| default: |
| return nil |
| } |
| } |
| |
| type stickyErrWriter struct { |
| w io.Writer |
| err *error |
| } |
| |
| func (sew stickyErrWriter) Write(p []byte) (n int, err error) { |
| if *sew.err != nil { |
| return 0, *sew.err |
| } |
| n, err = sew.w.Write(p) |
| *sew.err = err |
| return |
| } |
| |
| func (t *Transport) RoundTrip(req *http.Request) (*http.Response, error) { |
| if req.URL.Scheme != "https" { |
| return nil, errors.New("http2: unsupported scheme") |
| } |
| |
| host, port, err := net.SplitHostPort(req.URL.Host) |
| if err != nil { |
| host = req.URL.Host |
| port = "443" |
| } |
| |
| for { |
| cc, err := t.getClientConn(host, port) |
| if err != nil { |
| return nil, err |
| } |
| res, err := cc.roundTrip(req) |
| if shouldRetryRequest(err) { // TODO: or clientconn is overloaded (too many outstanding requests)? |
| continue |
| } |
| if err != nil { |
| return nil, err |
| } |
| return res, nil |
| } |
| } |
| |
| // CloseIdleConnections closes any connections which were previously |
| // connected from previous requests but are now sitting idle. |
| // It does not interrupt any connections currently in use. |
| func (t *Transport) CloseIdleConnections() { |
| t.connMu.Lock() |
| defer t.connMu.Unlock() |
| for _, vv := range t.conns { |
| for _, cc := range vv { |
| cc.closeIfIdle() |
| } |
| } |
| } |
| |
| var errClientConnClosed = errors.New("http2: client conn is closed") |
| |
| func shouldRetryRequest(err error) bool { |
| // TODO: or GOAWAY graceful shutdown stuff |
| return err == errClientConnClosed |
| } |
| |
| func (t *Transport) removeClientConn(cc *clientConn) { |
| t.connMu.Lock() |
| defer t.connMu.Unlock() |
| for _, key := range cc.connKey { |
| vv, ok := t.conns[key] |
| if !ok { |
| continue |
| } |
| newList := filterOutClientConn(vv, cc) |
| if len(newList) > 0 { |
| t.conns[key] = newList |
| } else { |
| delete(t.conns, key) |
| } |
| } |
| } |
| |
| func filterOutClientConn(in []*clientConn, exclude *clientConn) []*clientConn { |
| out := in[:0] |
| for _, v := range in { |
| if v != exclude { |
| out = append(out, v) |
| } |
| } |
| // If we filtered it out, zero out the last item to prevent |
| // the GC from seeing it. |
| if len(in) != len(out) { |
| in[len(in)-1] = nil |
| } |
| return out |
| } |
| |
| // AddIdleConn adds c as an idle conn for Transport. |
| // It assumes that c has not yet exchanged SETTINGS frames. |
| // The addr maybe be either "host" or "host:port". |
| func (t *Transport) AddIdleConn(addr string, c *tls.Conn) error { |
| var key string |
| host, _, err := net.SplitHostPort(addr) |
| if err == nil { |
| key = addr |
| } else { |
| host = addr |
| key = addr + ":443" |
| } |
| cc, err := t.newClientConn(host, key, c) |
| if err != nil { |
| return err |
| } |
| |
| t.addConn(key, cc) |
| return nil |
| } |
| |
| func (t *Transport) addConn(key string, cc *clientConn) { |
| t.connMu.Lock() |
| defer t.connMu.Unlock() |
| if t.conns == nil { |
| t.conns = make(map[string][]*clientConn) |
| } |
| t.conns[key] = append(t.conns[key], cc) |
| } |
| |
| func (t *Transport) getClientConn(host, port string) (*clientConn, error) { |
| key := net.JoinHostPort(host, port) |
| |
| t.connMu.Lock() |
| for _, cc := range t.conns[key] { |
| if cc.canTakeNewRequest() { |
| t.connMu.Unlock() |
| return cc, nil |
| } |
| } |
| t.connMu.Unlock() |
| |
| // TODO(bradfitz): use a singleflight.Group to only lock once per 'key'. |
| // Probably need to vendor it in as github.com/golang/sync/singleflight |
| // though, since the net package already uses it? Also lines up with |
| // sameer, bcmills, et al wanting to open source some sync stuff. |
| cc, err := t.dialClientConn(host, port, key) |
| if err != nil { |
| return nil, err |
| } |
| t.addConn(key, cc) |
| return cc, nil |
| } |
| |
| func (t *Transport) dialClientConn(host, port, key string) (*clientConn, error) { |
| cfg := &tls.Config{ |
| ServerName: host, |
| NextProtos: []string{NextProtoTLS}, |
| InsecureSkipVerify: t.InsecureTLSDial, |
| } |
| tconn, err := tls.Dial("tcp", net.JoinHostPort(host, port), cfg) |
| if err != nil { |
| return nil, err |
| } |
| return t.newClientConn(host, key, tconn) |
| } |
| |
| func (t *Transport) newClientConn(host, key string, tconn *tls.Conn) (*clientConn, error) { |
| if err := tconn.Handshake(); err != nil { |
| return nil, err |
| } |
| if !t.InsecureTLSDial { |
| if err := tconn.VerifyHostname(host); err != nil { |
| return nil, err |
| } |
| } |
| state := tconn.ConnectionState() |
| if p := state.NegotiatedProtocol; p != NextProtoTLS { |
| return nil, fmt.Errorf("http2: unexpected ALPN protocol %q; want %q", p, NextProtoTLS) |
| } |
| if !state.NegotiatedProtocolIsMutual { |
| return nil, errors.New("http2: could not negotiate protocol mutually") |
| } |
| if _, err := tconn.Write(clientPreface); err != nil { |
| return nil, err |
| } |
| |
| cc := &clientConn{ |
| t: t, |
| tconn: tconn, |
| connKey: []string{key}, // TODO: cert's validated hostnames too |
| tlsState: &state, |
| readerDone: make(chan struct{}), |
| nextStreamID: 1, |
| maxFrameSize: 16 << 10, // spec default |
| initialWindowSize: 65535, // spec default |
| maxConcurrentStreams: 1000, // "infinite", per spec. 1000 seems good enough. |
| streams: make(map[uint32]*clientStream), |
| } |
| cc.cond = sync.NewCond(&cc.mu) |
| cc.flow.add(int32(initialWindowSize)) |
| |
| // TODO: adjust this writer size to account for frame size + |
| // MTU + crypto/tls record padding. |
| cc.bw = bufio.NewWriter(stickyErrWriter{tconn, &cc.werr}) |
| cc.br = bufio.NewReader(tconn) |
| cc.fr = NewFramer(cc.bw, cc.br) |
| cc.henc = hpack.NewEncoder(&cc.hbuf) |
| cc.fr.WriteSettings( |
| Setting{ID: SettingEnablePush, Val: 0}, |
| Setting{ID: SettingInitialWindowSize, Val: transportDefaultStreamFlow}, |
| ) |
| cc.fr.WriteWindowUpdate(0, transportDefaultConnFlow) |
| cc.inflow.add(transportDefaultConnFlow + initialWindowSize) |
| cc.bw.Flush() |
| if cc.werr != nil { |
| return nil, cc.werr |
| } |
| |
| // Read the obligatory SETTINGS frame |
| f, err := cc.fr.ReadFrame() |
| if err != nil { |
| return nil, err |
| } |
| sf, ok := f.(*SettingsFrame) |
| if !ok { |
| return nil, fmt.Errorf("expected settings frame, got: %T", f) |
| } |
| cc.fr.WriteSettingsAck() |
| cc.bw.Flush() |
| |
| sf.ForeachSetting(func(s Setting) error { |
| switch s.ID { |
| case SettingMaxFrameSize: |
| cc.maxFrameSize = s.Val |
| case SettingMaxConcurrentStreams: |
| cc.maxConcurrentStreams = s.Val |
| case SettingInitialWindowSize: |
| cc.initialWindowSize = s.Val |
| default: |
| // TODO(bradfitz): handle more |
| t.vlogf("Unhandled Setting: %v", s) |
| } |
| return nil |
| }) |
| |
| go cc.readLoop() |
| return cc, nil |
| } |
| |
| func (cc *clientConn) setGoAway(f *GoAwayFrame) { |
| cc.mu.Lock() |
| defer cc.mu.Unlock() |
| cc.goAway = f |
| } |
| |
| func (cc *clientConn) canTakeNewRequest() bool { |
| cc.mu.Lock() |
| defer cc.mu.Unlock() |
| return cc.goAway == nil && |
| int64(len(cc.streams)+1) < int64(cc.maxConcurrentStreams) && |
| cc.nextStreamID < 2147483647 |
| } |
| |
| func (cc *clientConn) closeIfIdle() { |
| cc.mu.Lock() |
| if len(cc.streams) > 0 { |
| cc.mu.Unlock() |
| return |
| } |
| cc.closed = true |
| // TODO: do clients send GOAWAY too? maybe? Just Close: |
| cc.mu.Unlock() |
| |
| cc.tconn.Close() |
| } |
| |
| const maxAllocFrameSize = 512 << 10 |
| |
| // frameBuffer returns a scratch buffer suitable for writing DATA frames. |
| // They're capped at the min of the peer's max frame size or 512KB |
| // (kinda arbitrarily), but definitely capped so we don't allocate 4GB |
| // bufers. |
| func (cc *clientConn) frameScratchBuffer() []byte { |
| cc.mu.Lock() |
| size := cc.maxFrameSize |
| if size > maxAllocFrameSize { |
| size = maxAllocFrameSize |
| } |
| for i, buf := range cc.freeBuf { |
| if len(buf) >= int(size) { |
| cc.freeBuf[i] = nil |
| cc.mu.Unlock() |
| return buf[:size] |
| } |
| } |
| cc.mu.Unlock() |
| return make([]byte, size) |
| } |
| |
| func (cc *clientConn) putFrameScratchBuffer(buf []byte) { |
| cc.mu.Lock() |
| defer cc.mu.Unlock() |
| const maxBufs = 4 // arbitrary; 4 concurrent requests per conn? investigate. |
| if len(cc.freeBuf) < maxBufs { |
| cc.freeBuf = append(cc.freeBuf, buf) |
| return |
| } |
| for i, old := range cc.freeBuf { |
| if old == nil { |
| cc.freeBuf[i] = buf |
| return |
| } |
| } |
| // forget about it. |
| } |
| |
| func (cc *clientConn) roundTrip(req *http.Request) (*http.Response, error) { |
| cc.mu.Lock() |
| |
| if cc.closed { |
| cc.mu.Unlock() |
| return nil, errClientConnClosed |
| } |
| |
| cs := cc.newStream() |
| hasBody := req.Body != nil |
| |
| // we send: HEADERS[+CONTINUATION] + (DATA?) |
| hdrs := cc.encodeHeaders(req) |
| first := true |
| |
| cc.wmu.Lock() |
| frameSize := int(cc.maxFrameSize) |
| for len(hdrs) > 0 && cc.werr == nil { |
| chunk := hdrs |
| if len(chunk) > frameSize { |
| chunk = chunk[:frameSize] |
| } |
| hdrs = hdrs[len(chunk):] |
| endHeaders := len(hdrs) == 0 |
| if first { |
| cc.fr.WriteHeaders(HeadersFrameParam{ |
| StreamID: cs.ID, |
| BlockFragment: chunk, |
| EndStream: !hasBody, |
| EndHeaders: endHeaders, |
| }) |
| first = false |
| } else { |
| cc.fr.WriteContinuation(cs.ID, endHeaders, chunk) |
| } |
| } |
| cc.bw.Flush() |
| werr := cc.werr |
| cc.wmu.Unlock() |
| cc.mu.Unlock() |
| |
| if werr != nil { |
| return nil, werr |
| } |
| |
| var bodyCopyErrc chan error |
| var gotResHeaders chan struct{} // closed on resheaders |
| if hasBody { |
| bodyCopyErrc = make(chan error, 1) |
| gotResHeaders = make(chan struct{}) |
| go func() { |
| bodyCopyErrc <- cs.writeRequestBody(req.Body, gotResHeaders) |
| }() |
| } |
| |
| for { |
| select { |
| case re := <-cs.resc: |
| if gotResHeaders != nil { |
| close(gotResHeaders) |
| } |
| if re.err != nil { |
| return nil, re.err |
| } |
| res := re.res |
| res.Request = req |
| res.TLS = cc.tlsState |
| return res, nil |
| case err := <-bodyCopyErrc: |
| if err != nil { |
| return nil, err |
| } |
| } |
| } |
| } |
| |
| var errServerResponseBeforeRequestBody = errors.New("http2: server sent response while still writing request body") |
| |
| func (cs *clientStream) writeRequestBody(body io.Reader, gotResHeaders <-chan struct{}) error { |
| cc := cs.cc |
| done := false |
| for !done { |
| buf := cc.frameScratchBuffer() |
| n, err := io.ReadFull(body, buf) |
| if err == io.ErrUnexpectedEOF { |
| done = true |
| } else if err != nil { |
| return err |
| } |
| |
| // Await for n flow control tokens. |
| if err := cs.awaitFlowControl(int32(n)); err != nil { |
| return err |
| } |
| |
| cc.wmu.Lock() |
| select { |
| case <-gotResHeaders: |
| err = errServerResponseBeforeRequestBody |
| case <-cs.peerReset: |
| err = cs.resetErr |
| default: |
| err = cc.fr.WriteData(cs.ID, done, buf[:n]) |
| } |
| cc.wmu.Unlock() |
| |
| cc.putFrameScratchBuffer(buf) |
| if err != nil { |
| return err |
| } |
| } |
| |
| var err error |
| |
| cc.wmu.Lock() |
| if !done { |
| err = cc.fr.WriteData(cs.ID, true, nil) |
| } |
| if ferr := cc.bw.Flush(); ferr != nil && err == nil { |
| err = ferr |
| } |
| cc.wmu.Unlock() |
| |
| return err |
| } |
| |
| func (cs *clientStream) awaitFlowControl(n int32) error { |
| cc := cs.cc |
| cc.mu.Lock() |
| defer cc.mu.Unlock() |
| for { |
| if cc.closed { |
| return errClientConnClosed |
| } |
| if err := cs.checkReset(); err != nil { |
| return err |
| } |
| if cs.flow.available() >= n { |
| cs.flow.take(n) |
| return nil |
| } |
| cc.cond.Wait() |
| } |
| } |
| |
| // requires cc.mu be held. |
| func (cc *clientConn) encodeHeaders(req *http.Request) []byte { |
| cc.hbuf.Reset() |
| |
| // TODO(bradfitz): figure out :authority-vs-Host stuff between http2 and Go |
| host := req.Host |
| if host == "" { |
| host = req.URL.Host |
| } |
| |
| path := req.URL.Path |
| if path == "" { |
| path = "/" |
| } |
| |
| cc.writeHeader(":authority", host) // probably not right for all sites |
| cc.writeHeader(":method", req.Method) |
| cc.writeHeader(":path", path) |
| cc.writeHeader(":scheme", "https") |
| |
| for k, vv := range req.Header { |
| lowKey := strings.ToLower(k) |
| if lowKey == "host" { |
| continue |
| } |
| for _, v := range vv { |
| cc.writeHeader(lowKey, v) |
| } |
| } |
| return cc.hbuf.Bytes() |
| } |
| |
| func (cc *clientConn) writeHeader(name, value string) { |
| cc.henc.WriteField(hpack.HeaderField{Name: name, Value: value}) |
| } |
| |
| type resAndError struct { |
| res *http.Response |
| err error |
| } |
| |
| // requires cc.mu be held. |
| func (cc *clientConn) newStream() *clientStream { |
| cs := &clientStream{ |
| cc: cc, |
| ID: cc.nextStreamID, |
| resc: make(chan resAndError, 1), |
| peerReset: make(chan struct{}), |
| } |
| cs.flow.add(int32(cc.initialWindowSize)) |
| cs.flow.setConnFlow(&cc.flow) |
| cs.inflow.add(transportDefaultStreamFlow) |
| cs.inflow.setConnFlow(&cc.inflow) |
| cc.nextStreamID += 2 |
| cc.streams[cs.ID] = cs |
| return cs |
| } |
| |
| func (cc *clientConn) streamByID(id uint32, andRemove bool) *clientStream { |
| cc.mu.Lock() |
| defer cc.mu.Unlock() |
| cs := cc.streams[id] |
| if andRemove { |
| delete(cc.streams, id) |
| } |
| return cs |
| } |
| |
| // clientConnReadLoop is the state owned by the clientConn's frame-reading readLoop. |
| type clientConnReadLoop struct { |
| cc *clientConn |
| activeRes map[uint32]*clientStream // keyed by streamID |
| |
| // continueStreamID is the stream ID we're waiting for |
| // continuation frames for. |
| continueStreamID uint32 |
| |
| hdec *hpack.Decoder |
| |
| // Fields reset on each HEADERS: |
| nextRes *http.Response |
| sawRegHeader bool // saw non-pseudo header |
| reqMalformed error // non-nil once known to be malformed |
| } |
| |
| // readLoop runs in its own goroutine and reads and dispatches frames. |
| func (cc *clientConn) readLoop() { |
| rl := &clientConnReadLoop{ |
| cc: cc, |
| activeRes: make(map[uint32]*clientStream), |
| } |
| // TODO: figure out henc size |
| rl.hdec = hpack.NewDecoder(initialHeaderTableSize, rl.onNewHeaderField) |
| |
| defer rl.cleanup() |
| cc.readerErr = rl.run() |
| if ce, ok := cc.readerErr.(ConnectionError); ok { |
| cc.wmu.Lock() |
| cc.fr.WriteGoAway(0, ErrCode(ce), nil) |
| cc.wmu.Unlock() |
| } |
| cc.tconn.Close() |
| } |
| |
| func (rl *clientConnReadLoop) cleanup() { |
| cc := rl.cc |
| defer cc.t.removeClientConn(cc) |
| defer close(cc.readerDone) |
| |
| // Close any response bodies if the server closes prematurely. |
| // TODO: also do this if we've written the headers but not |
| // gotten a response yet. |
| err := cc.readerErr |
| if err == io.EOF { |
| err = io.ErrUnexpectedEOF |
| } |
| for _, cs := range rl.activeRes { |
| cs.pw.CloseWithError(err) |
| } |
| |
| cc.mu.Lock() |
| cc.closed = true |
| cc.cond.Broadcast() |
| cc.mu.Unlock() |
| } |
| |
| func (rl *clientConnReadLoop) run() error { |
| cc := rl.cc |
| for { |
| f, err := cc.fr.ReadFrame() |
| if se, ok := err.(StreamError); ok { |
| // TODO: deal with stream errors from the framer. |
| return se |
| } else if err != nil { |
| return err |
| } |
| cc.vlogf("Transport received %v: %#v", f.Header(), f) |
| |
| streamID := f.Header().StreamID |
| |
| _, isContinue := f.(*ContinuationFrame) |
| if isContinue { |
| if streamID != rl.continueStreamID { |
| cc.logf("Protocol violation: got CONTINUATION with id %d; want %d", streamID, rl.continueStreamID) |
| return ConnectionError(ErrCodeProtocol) |
| } |
| } else if rl.continueStreamID != 0 { |
| // Continue frames need to be adjacent in the stream |
| // and we were in the middle of headers. |
| cc.logf("Protocol violation: got %T for stream %d, want CONTINUATION for %d", f, streamID, rl.continueStreamID) |
| return ConnectionError(ErrCodeProtocol) |
| } |
| |
| if streamID%2 == 0 { |
| // Ignore streams pushed from the server for now. |
| // These always have an even stream id. |
| continue |
| } |
| |
| // TODO(bradfitz): push all this (streamEnded + lookup |
| // of clientStream) into per-frame methods, so the |
| // cc.mu Lock/Unlock is only done once. |
| streamEnded := false |
| if ff, ok := f.(streamEnder); ok { |
| streamEnded = ff.StreamEnded() |
| } |
| var cs *clientStream |
| if streamID != 0 { |
| cs = cc.streamByID(streamID, streamEnded) |
| if cs == nil { |
| rl.cc.logf("Received frame for untracked stream ID %d", streamID) |
| } |
| } |
| |
| switch f := f.(type) { |
| case *HeadersFrame: |
| err = rl.processHeaders(f, cs) |
| case *ContinuationFrame: |
| err = rl.processContinuation(f, cs) |
| case *DataFrame: |
| err = rl.processData(f, cs) |
| case *GoAwayFrame: |
| err = rl.processGoAway(f) |
| case *RSTStreamFrame: |
| err = rl.processResetStream(f, cs) |
| case *SettingsFrame: |
| err = rl.processSettings(f) |
| case *PushPromiseFrame: |
| // Skip it. We told the peer we don't want |
| // them anyway. And we already skipped even |
| // stream IDs above. So really shouldn't be |
| // here. |
| case *WindowUpdateFrame: |
| err = rl.processWindowUpdate(f, cs) |
| default: |
| cc.logf("Transport: unhandled response frame type %T", f) |
| } |
| if err != nil { |
| return err |
| } |
| } |
| } |
| |
| func (rl *clientConnReadLoop) processHeaders(f *HeadersFrame, cs *clientStream) error { |
| rl.sawRegHeader = false |
| rl.reqMalformed = nil |
| rl.nextRes = &http.Response{ |
| Proto: "HTTP/2.0", |
| ProtoMajor: 2, |
| Header: make(http.Header), |
| } |
| cs.pr, cs.pw = io.Pipe() |
| return rl.processHeaderBlockFragment(cs, f.HeaderBlockFragment(), f.HeadersEnded(), f.StreamEnded()) |
| } |
| |
| func (rl *clientConnReadLoop) processContinuation(f *ContinuationFrame, cs *clientStream) error { |
| return rl.processHeaderBlockFragment(cs, f.HeaderBlockFragment(), f.HeadersEnded(), f.StreamEnded()) |
| } |
| |
| func (rl *clientConnReadLoop) processHeaderBlockFragment(cs *clientStream, frag []byte, headersEnded, streamEnded bool) error { |
| if cs == nil { |
| return nil |
| } |
| _, err := rl.hdec.Write(frag) |
| if err != nil { |
| return err |
| } |
| if !headersEnded { |
| rl.continueStreamID = cs.ID |
| return nil |
| } |
| |
| // HEADERS (or CONTINUATION) are now over. |
| rl.continueStreamID = 0 |
| |
| if rl.reqMalformed != nil { |
| cs.resc <- resAndError{err: rl.reqMalformed} |
| rl.cc.writeStreamReset(cs.ID, ErrCodeProtocol, rl.reqMalformed) |
| return nil |
| } |
| |
| // TODO: set the Body to one which notes the |
| // Close and also sends the server a |
| // RST_STREAM |
| rl.nextRes.Body = cs.pr |
| res := rl.nextRes |
| rl.activeRes[cs.ID] = cs |
| cs.resc <- resAndError{res: res} |
| rl.nextRes = nil // unused now; will be reset next HEADERS frame |
| return nil |
| } |
| |
| func (rl *clientConnReadLoop) processData(f *DataFrame, cs *clientStream) error { |
| if cs == nil { |
| return nil |
| } |
| data := f.Data() |
| // TODO: decrement cs.inflow and cc.inflow, sending errors as appropriate. |
| if VerboseLogs { |
| rl.cc.logf("DATA: %q", data) |
| } |
| cc := rl.cc |
| |
| // Check connection-level flow control. |
| cc.mu.Lock() |
| if cs.inflow.available() >= int32(len(data)) { |
| cs.inflow.take(int32(len(data))) |
| } else { |
| cc.mu.Unlock() |
| return ConnectionError(ErrCodeFlowControl) |
| } |
| cc.mu.Unlock() |
| |
| if _, err := cs.pw.Write(data); err != nil { |
| return err |
| } |
| // send WINDOW_UPDATE frames occasionally as per-stream flow control depletes |
| |
| if f.StreamEnded() { |
| cs.pw.Close() |
| delete(rl.activeRes, cs.ID) |
| } |
| return nil |
| } |
| |
| func (rl *clientConnReadLoop) processGoAway(f *GoAwayFrame) error { |
| cc := rl.cc |
| cc.t.removeClientConn(cc) |
| if f.ErrCode != 0 { |
| // TODO: deal with GOAWAY more. particularly the error code |
| cc.vlogf("transport got GOAWAY with error code = %v", f.ErrCode) |
| } |
| cc.setGoAway(f) |
| return nil |
| } |
| |
| func (rl *clientConnReadLoop) processSettings(f *SettingsFrame) error { |
| cc := rl.cc |
| cc.mu.Lock() |
| defer cc.mu.Unlock() |
| return f.ForeachSetting(func(s Setting) error { |
| switch s.ID { |
| case SettingMaxFrameSize: |
| cc.maxFrameSize = s.Val |
| case SettingMaxConcurrentStreams: |
| cc.maxConcurrentStreams = s.Val |
| case SettingInitialWindowSize: |
| // TODO: error if this is too large. |
| |
| // TODO: adjust flow control of still-open |
| // frames by the difference of the old initial |
| // window size and this one. |
| cc.initialWindowSize = s.Val |
| default: |
| // TODO(bradfitz): handle more settings? |
| cc.vlogf("Unhandled Setting: %v", s) |
| } |
| return nil |
| }) |
| } |
| |
| func (rl *clientConnReadLoop) processWindowUpdate(f *WindowUpdateFrame, cs *clientStream) error { |
| if f.StreamID != 0 && cs == nil { |
| return nil |
| } |
| cc := rl.cc |
| |
| cc.mu.Lock() |
| defer cc.mu.Unlock() |
| |
| fl := &cc.flow |
| if cs != nil { |
| fl = &cs.flow |
| } |
| if !fl.add(int32(f.Increment)) { |
| return ConnectionError(ErrCodeFlowControl) |
| } |
| cc.cond.Broadcast() |
| return nil |
| } |
| |
| func (rl *clientConnReadLoop) processResetStream(f *RSTStreamFrame, cs *clientStream) error { |
| if cs == nil { |
| // TODO: return error if rst of idle stream? |
| return nil |
| } |
| select { |
| case <-cs.peerReset: |
| // Already reset. |
| // This is the only goroutine |
| // which closes this, so there |
| // isn't a race. |
| default: |
| err := StreamError{cs.ID, f.ErrCode} |
| cs.resetErr = err |
| close(cs.peerReset) |
| cs.pw.CloseWithError(err) |
| } |
| delete(rl.activeRes, cs.ID) |
| return nil |
| } |
| |
| func (cc *clientConn) writeStreamReset(streamID uint32, code ErrCode, err error) { |
| // TODO: do something with err? send it as a debug frame to the peer? |
| cc.wmu.Lock() |
| cc.fr.WriteRSTStream(streamID, code) |
| cc.wmu.Unlock() |
| } |
| |
| // onNewHeaderField runs on the readLoop goroutine whenever a new |
| // hpack header field is decoded. |
| func (rl *clientConnReadLoop) onNewHeaderField(f hpack.HeaderField) { |
| cc := rl.cc |
| if VerboseLogs { |
| cc.logf("Header field: %+v", f) |
| } |
| isPseudo := strings.HasPrefix(f.Name, ":") |
| if isPseudo { |
| if rl.sawRegHeader { |
| rl.reqMalformed = errors.New("http2: invalid pseudo header after regular header") |
| return |
| } |
| switch f.Name { |
| case ":status": |
| code, err := strconv.Atoi(f.Value) |
| if err != nil { |
| rl.reqMalformed = errors.New("http2: invalid :status") |
| return |
| } |
| rl.nextRes.Status = f.Value + " " + http.StatusText(code) |
| rl.nextRes.StatusCode = code |
| default: |
| // "Endpoints MUST NOT generate pseudo-header |
| // fields other than those defined in this |
| // document." |
| rl.reqMalformed = fmt.Errorf("http2: unknown response pseudo header %q", f.Name) |
| } |
| } else { |
| rl.sawRegHeader = true |
| rl.nextRes.Header.Add(http.CanonicalHeaderKey(f.Name), f.Value) |
| } |
| } |
| |
| func (cc *clientConn) logf(format string, args ...interface{}) { |
| cc.t.logf(format, args...) |
| } |
| |
| func (cc *clientConn) vlogf(format string, args ...interface{}) { |
| cc.t.vlogf(format, args...) |
| } |
| |
| func (t *Transport) vlogf(format string, args ...interface{}) { |
| if VerboseLogs { |
| t.logf(format, args...) |
| } |
| } |
| |
| func (t *Transport) logf(format string, args ...interface{}) { |
| log.Printf(format, args...) |
| } |