|  | // Copyright 2009 The Go Authors. All rights reserved. | 
|  | // Use of this source code is governed by a BSD-style | 
|  | // license that can be found in the LICENSE file. | 
|  |  | 
|  | // The wire protocol for HTTP's "chunked" Transfer-Encoding. | 
|  |  | 
|  | // Package internal contains HTTP internals shared by net/http and | 
|  | // net/http/httputil. | 
|  | package internal | 
|  |  | 
|  | import ( | 
|  | "bufio" | 
|  | "bytes" | 
|  | "errors" | 
|  | "fmt" | 
|  | "io" | 
|  | ) | 
|  |  | 
|  | const maxLineLength = 4096 // assumed <= bufio.defaultBufSize | 
|  |  | 
|  | var ErrLineTooLong = errors.New("header line too long") | 
|  |  | 
|  | // NewChunkedReader returns a new chunkedReader that translates the data read from r | 
|  | // out of HTTP "chunked" format before returning it. | 
|  | // The chunkedReader returns io.EOF when the final 0-length chunk is read. | 
|  | // | 
|  | // NewChunkedReader is not needed by normal applications. The http package | 
|  | // automatically decodes chunking when reading response bodies. | 
|  | func NewChunkedReader(r io.Reader) io.Reader { | 
|  | br, ok := r.(*bufio.Reader) | 
|  | if !ok { | 
|  | br = bufio.NewReader(r) | 
|  | } | 
|  | return &chunkedReader{r: br} | 
|  | } | 
|  |  | 
|  | type chunkedReader struct { | 
|  | r        *bufio.Reader | 
|  | n        uint64 // unread bytes in chunk | 
|  | err      error | 
|  | buf      [2]byte | 
|  | checkEnd bool // whether need to check for \r\n chunk footer | 
|  | } | 
|  |  | 
|  | func (cr *chunkedReader) beginChunk() { | 
|  | // chunk-size CRLF | 
|  | var line []byte | 
|  | line, cr.err = readChunkLine(cr.r) | 
|  | if cr.err != nil { | 
|  | return | 
|  | } | 
|  | cr.n, cr.err = parseHexUint(line) | 
|  | if cr.err != nil { | 
|  | return | 
|  | } | 
|  | if cr.n == 0 { | 
|  | cr.err = io.EOF | 
|  | } | 
|  | } | 
|  |  | 
|  | func (cr *chunkedReader) chunkHeaderAvailable() bool { | 
|  | n := cr.r.Buffered() | 
|  | if n > 0 { | 
|  | peek, _ := cr.r.Peek(n) | 
|  | return bytes.IndexByte(peek, '\n') >= 0 | 
|  | } | 
|  | return false | 
|  | } | 
|  |  | 
|  | func (cr *chunkedReader) Read(b []uint8) (n int, err error) { | 
|  | for cr.err == nil { | 
|  | if cr.checkEnd { | 
|  | if n > 0 && cr.r.Buffered() < 2 { | 
|  | // We have some data. Return early (per the io.Reader | 
|  | // contract) instead of potentially blocking while | 
|  | // reading more. | 
|  | break | 
|  | } | 
|  | if _, cr.err = io.ReadFull(cr.r, cr.buf[:2]); cr.err == nil { | 
|  | if string(cr.buf[:]) != "\r\n" { | 
|  | cr.err = errors.New("malformed chunked encoding") | 
|  | break | 
|  | } | 
|  | } | 
|  | cr.checkEnd = false | 
|  | } | 
|  | if cr.n == 0 { | 
|  | if n > 0 && !cr.chunkHeaderAvailable() { | 
|  | // We've read enough. Don't potentially block | 
|  | // reading a new chunk header. | 
|  | break | 
|  | } | 
|  | cr.beginChunk() | 
|  | continue | 
|  | } | 
|  | if len(b) == 0 { | 
|  | break | 
|  | } | 
|  | rbuf := b | 
|  | if uint64(len(rbuf)) > cr.n { | 
|  | rbuf = rbuf[:cr.n] | 
|  | } | 
|  | var n0 int | 
|  | n0, cr.err = cr.r.Read(rbuf) | 
|  | n += n0 | 
|  | b = b[n0:] | 
|  | cr.n -= uint64(n0) | 
|  | // If we're at the end of a chunk, read the next two | 
|  | // bytes to verify they are "\r\n". | 
|  | if cr.n == 0 && cr.err == nil { | 
|  | cr.checkEnd = true | 
|  | } | 
|  | } | 
|  | return n, cr.err | 
|  | } | 
|  |  | 
|  | // Read a line of bytes (up to \n) from b. | 
|  | // Give up if the line exceeds maxLineLength. | 
|  | // The returned bytes are owned by the bufio.Reader | 
|  | // so they are only valid until the next bufio read. | 
|  | func readChunkLine(b *bufio.Reader) ([]byte, error) { | 
|  | p, err := b.ReadSlice('\n') | 
|  | if err != nil { | 
|  | // We always know when EOF is coming. | 
|  | // If the caller asked for a line, there should be a line. | 
|  | if err == io.EOF { | 
|  | err = io.ErrUnexpectedEOF | 
|  | } else if err == bufio.ErrBufferFull { | 
|  | err = ErrLineTooLong | 
|  | } | 
|  | return nil, err | 
|  | } | 
|  | if len(p) >= maxLineLength { | 
|  | return nil, ErrLineTooLong | 
|  | } | 
|  | p = trimTrailingWhitespace(p) | 
|  | p, err = removeChunkExtension(p) | 
|  | if err != nil { | 
|  | return nil, err | 
|  | } | 
|  | return p, nil | 
|  | } | 
|  |  | 
|  | func trimTrailingWhitespace(b []byte) []byte { | 
|  | for len(b) > 0 && isASCIISpace(b[len(b)-1]) { | 
|  | b = b[:len(b)-1] | 
|  | } | 
|  | return b | 
|  | } | 
|  |  | 
|  | func isASCIISpace(b byte) bool { | 
|  | return b == ' ' || b == '\t' || b == '\n' || b == '\r' | 
|  | } | 
|  |  | 
|  | // removeChunkExtension removes any chunk-extension from p. | 
|  | // For example, | 
|  | //     "0" => "0" | 
|  | //     "0;token" => "0" | 
|  | //     "0;token=val" => "0" | 
|  | //     `0;token="quoted string"` => "0" | 
|  | func removeChunkExtension(p []byte) ([]byte, error) { | 
|  | semi := bytes.IndexByte(p, ';') | 
|  | if semi == -1 { | 
|  | return p, nil | 
|  | } | 
|  | // TODO: care about exact syntax of chunk extensions? We're | 
|  | // ignoring and stripping them anyway. For now just never | 
|  | // return an error. | 
|  | return p[:semi], nil | 
|  | } | 
|  |  | 
|  | // NewChunkedWriter returns a new chunkedWriter that translates writes into HTTP | 
|  | // "chunked" format before writing them to w. Closing the returned chunkedWriter | 
|  | // sends the final 0-length chunk that marks the end of the stream but does | 
|  | // not send the final CRLF that appears after trailers; trailers and the last | 
|  | // CRLF must be written separately. | 
|  | // | 
|  | // NewChunkedWriter is not needed by normal applications. The http | 
|  | // package adds chunking automatically if handlers don't set a | 
|  | // Content-Length header. Using newChunkedWriter inside a handler | 
|  | // would result in double chunking or chunking with a Content-Length | 
|  | // length, both of which are wrong. | 
|  | func NewChunkedWriter(w io.Writer) io.WriteCloser { | 
|  | return &chunkedWriter{w} | 
|  | } | 
|  |  | 
|  | // Writing to chunkedWriter translates to writing in HTTP chunked Transfer | 
|  | // Encoding wire format to the underlying Wire chunkedWriter. | 
|  | type chunkedWriter struct { | 
|  | Wire io.Writer | 
|  | } | 
|  |  | 
|  | // Write the contents of data as one chunk to Wire. | 
|  | // NOTE: Note that the corresponding chunk-writing procedure in Conn.Write has | 
|  | // a bug since it does not check for success of io.WriteString | 
|  | func (cw *chunkedWriter) Write(data []byte) (n int, err error) { | 
|  |  | 
|  | // Don't send 0-length data. It looks like EOF for chunked encoding. | 
|  | if len(data) == 0 { | 
|  | return 0, nil | 
|  | } | 
|  |  | 
|  | if _, err = fmt.Fprintf(cw.Wire, "%x\r\n", len(data)); err != nil { | 
|  | return 0, err | 
|  | } | 
|  | if n, err = cw.Wire.Write(data); err != nil { | 
|  | return | 
|  | } | 
|  | if n != len(data) { | 
|  | err = io.ErrShortWrite | 
|  | return | 
|  | } | 
|  | if _, err = io.WriteString(cw.Wire, "\r\n"); err != nil { | 
|  | return | 
|  | } | 
|  | if bw, ok := cw.Wire.(*FlushAfterChunkWriter); ok { | 
|  | err = bw.Flush() | 
|  | } | 
|  | return | 
|  | } | 
|  |  | 
|  | func (cw *chunkedWriter) Close() error { | 
|  | _, err := io.WriteString(cw.Wire, "0\r\n") | 
|  | return err | 
|  | } | 
|  |  | 
|  | // FlushAfterChunkWriter signals from the caller of NewChunkedWriter | 
|  | // that each chunk should be followed by a flush. It is used by the | 
|  | // http.Transport code to keep the buffering behavior for headers and | 
|  | // trailers, but flush out chunks aggressively in the middle for | 
|  | // request bodies which may be generated slowly. See Issue 6574. | 
|  | type FlushAfterChunkWriter struct { | 
|  | *bufio.Writer | 
|  | } | 
|  |  | 
|  | func parseHexUint(v []byte) (n uint64, err error) { | 
|  | for i, b := range v { | 
|  | switch { | 
|  | case '0' <= b && b <= '9': | 
|  | b = b - '0' | 
|  | case 'a' <= b && b <= 'f': | 
|  | b = b - 'a' + 10 | 
|  | case 'A' <= b && b <= 'F': | 
|  | b = b - 'A' + 10 | 
|  | default: | 
|  | return 0, errors.New("invalid byte in chunk length") | 
|  | } | 
|  | if i == 16 { | 
|  | return 0, errors.New("http chunk length too large") | 
|  | } | 
|  | n <<= 4 | 
|  | n |= uint64(b) | 
|  | } | 
|  | return | 
|  | } |