| // Copyright 2009 The Go Authors. All rights reserved. |
| // Use of this source code is governed by a BSD-style |
| // license that can be found in the LICENSE file. |
| |
| // The wire protocol for HTTP's "chunked" Transfer-Encoding. |
| |
| // Package internal contains HTTP internals shared by net/http and |
| // net/http/httputil. |
| package internal |
| |
| import ( |
| "bufio" |
| "bytes" |
| "errors" |
| "fmt" |
| "io" |
| ) |
| |
| const maxLineLength = 4096 // assumed <= bufio.defaultBufSize |
| |
| var ErrLineTooLong = errors.New("header line too long") |
| |
| // NewChunkedReader returns a new chunkedReader that translates the data read from r |
| // out of HTTP "chunked" format before returning it. |
| // The chunkedReader returns io.EOF when the final 0-length chunk is read. |
| // |
| // NewChunkedReader is not needed by normal applications. The http package |
| // automatically decodes chunking when reading response bodies. |
| func NewChunkedReader(r io.Reader) io.Reader { |
| br, ok := r.(*bufio.Reader) |
| if !ok { |
| br = bufio.NewReader(r) |
| } |
| return &chunkedReader{r: br} |
| } |
| |
| type chunkedReader struct { |
| r *bufio.Reader |
| n uint64 // unread bytes in chunk |
| err error |
| buf [2]byte |
| checkEnd bool // whether need to check for \r\n chunk footer |
| } |
| |
| func (cr *chunkedReader) beginChunk() { |
| // chunk-size CRLF |
| var line []byte |
| line, cr.err = readChunkLine(cr.r) |
| if cr.err != nil { |
| return |
| } |
| cr.n, cr.err = parseHexUint(line) |
| if cr.err != nil { |
| return |
| } |
| if cr.n == 0 { |
| cr.err = io.EOF |
| } |
| } |
| |
| func (cr *chunkedReader) chunkHeaderAvailable() bool { |
| n := cr.r.Buffered() |
| if n > 0 { |
| peek, _ := cr.r.Peek(n) |
| return bytes.IndexByte(peek, '\n') >= 0 |
| } |
| return false |
| } |
| |
| func (cr *chunkedReader) Read(b []uint8) (n int, err error) { |
| for cr.err == nil { |
| if cr.checkEnd { |
| if n > 0 && cr.r.Buffered() < 2 { |
| // We have some data. Return early (per the io.Reader |
| // contract) instead of potentially blocking while |
| // reading more. |
| break |
| } |
| if _, cr.err = io.ReadFull(cr.r, cr.buf[:2]); cr.err == nil { |
| if string(cr.buf[:]) != "\r\n" { |
| cr.err = errors.New("malformed chunked encoding") |
| break |
| } |
| } |
| cr.checkEnd = false |
| } |
| if cr.n == 0 { |
| if n > 0 && !cr.chunkHeaderAvailable() { |
| // We've read enough. Don't potentially block |
| // reading a new chunk header. |
| break |
| } |
| cr.beginChunk() |
| continue |
| } |
| if len(b) == 0 { |
| break |
| } |
| rbuf := b |
| if uint64(len(rbuf)) > cr.n { |
| rbuf = rbuf[:cr.n] |
| } |
| var n0 int |
| n0, cr.err = cr.r.Read(rbuf) |
| n += n0 |
| b = b[n0:] |
| cr.n -= uint64(n0) |
| // If we're at the end of a chunk, read the next two |
| // bytes to verify they are "\r\n". |
| if cr.n == 0 && cr.err == nil { |
| cr.checkEnd = true |
| } |
| } |
| return n, cr.err |
| } |
| |
| // Read a line of bytes (up to \n) from b. |
| // Give up if the line exceeds maxLineLength. |
| // The returned bytes are owned by the bufio.Reader |
| // so they are only valid until the next bufio read. |
| func readChunkLine(b *bufio.Reader) ([]byte, error) { |
| p, err := b.ReadSlice('\n') |
| if err != nil { |
| // We always know when EOF is coming. |
| // If the caller asked for a line, there should be a line. |
| if err == io.EOF { |
| err = io.ErrUnexpectedEOF |
| } else if err == bufio.ErrBufferFull { |
| err = ErrLineTooLong |
| } |
| return nil, err |
| } |
| if len(p) >= maxLineLength { |
| return nil, ErrLineTooLong |
| } |
| p = trimTrailingWhitespace(p) |
| p, err = removeChunkExtension(p) |
| if err != nil { |
| return nil, err |
| } |
| return p, nil |
| } |
| |
| func trimTrailingWhitespace(b []byte) []byte { |
| for len(b) > 0 && isASCIISpace(b[len(b)-1]) { |
| b = b[:len(b)-1] |
| } |
| return b |
| } |
| |
| func isASCIISpace(b byte) bool { |
| return b == ' ' || b == '\t' || b == '\n' || b == '\r' |
| } |
| |
| // removeChunkExtension removes any chunk-extension from p. |
| // For example, |
| // "0" => "0" |
| // "0;token" => "0" |
| // "0;token=val" => "0" |
| // `0;token="quoted string"` => "0" |
| func removeChunkExtension(p []byte) ([]byte, error) { |
| semi := bytes.IndexByte(p, ';') |
| if semi == -1 { |
| return p, nil |
| } |
| // TODO: care about exact syntax of chunk extensions? We're |
| // ignoring and stripping them anyway. For now just never |
| // return an error. |
| return p[:semi], nil |
| } |
| |
| // NewChunkedWriter returns a new chunkedWriter that translates writes into HTTP |
| // "chunked" format before writing them to w. Closing the returned chunkedWriter |
| // sends the final 0-length chunk that marks the end of the stream but does |
| // not send the final CRLF that appears after trailers; trailers and the last |
| // CRLF must be written separately. |
| // |
| // NewChunkedWriter is not needed by normal applications. The http |
| // package adds chunking automatically if handlers don't set a |
| // Content-Length header. Using newChunkedWriter inside a handler |
| // would result in double chunking or chunking with a Content-Length |
| // length, both of which are wrong. |
| func NewChunkedWriter(w io.Writer) io.WriteCloser { |
| return &chunkedWriter{w} |
| } |
| |
| // Writing to chunkedWriter translates to writing in HTTP chunked Transfer |
| // Encoding wire format to the underlying Wire chunkedWriter. |
| type chunkedWriter struct { |
| Wire io.Writer |
| } |
| |
| // Write the contents of data as one chunk to Wire. |
| // NOTE: Note that the corresponding chunk-writing procedure in Conn.Write has |
| // a bug since it does not check for success of io.WriteString |
| func (cw *chunkedWriter) Write(data []byte) (n int, err error) { |
| |
| // Don't send 0-length data. It looks like EOF for chunked encoding. |
| if len(data) == 0 { |
| return 0, nil |
| } |
| |
| if _, err = fmt.Fprintf(cw.Wire, "%x\r\n", len(data)); err != nil { |
| return 0, err |
| } |
| if n, err = cw.Wire.Write(data); err != nil { |
| return |
| } |
| if n != len(data) { |
| err = io.ErrShortWrite |
| return |
| } |
| if _, err = io.WriteString(cw.Wire, "\r\n"); err != nil { |
| return |
| } |
| if bw, ok := cw.Wire.(*FlushAfterChunkWriter); ok { |
| err = bw.Flush() |
| } |
| return |
| } |
| |
| func (cw *chunkedWriter) Close() error { |
| _, err := io.WriteString(cw.Wire, "0\r\n") |
| return err |
| } |
| |
| // FlushAfterChunkWriter signals from the caller of NewChunkedWriter |
| // that each chunk should be followed by a flush. It is used by the |
| // http.Transport code to keep the buffering behavior for headers and |
| // trailers, but flush out chunks aggressively in the middle for |
| // request bodies which may be generated slowly. See Issue 6574. |
| type FlushAfterChunkWriter struct { |
| *bufio.Writer |
| } |
| |
| func parseHexUint(v []byte) (n uint64, err error) { |
| for i, b := range v { |
| switch { |
| case '0' <= b && b <= '9': |
| b = b - '0' |
| case 'a' <= b && b <= 'f': |
| b = b - 'a' + 10 |
| case 'A' <= b && b <= 'F': |
| b = b - 'A' + 10 |
| default: |
| return 0, errors.New("invalid byte in chunk length") |
| } |
| if i == 16 { |
| return 0, errors.New("http chunk length too large") |
| } |
| n <<= 4 |
| n |= uint64(b) |
| } |
| return |
| } |