mirror of
				https://codeberg.org/forgejo/forgejo.git
				synced 2025-11-04 08:21:11 +00:00 
			
		
		
		
	Use [chi](https://github.com/go-chi/chi) instead of the forked [macaron](https://gitea.com/macaron/macaron). Since macaron and chi have conflicts with session share, this big PR becomes a have-to thing. According my previous idea, we can replace macaron step by step but I'm wrong. :( Below is a list of big changes on this PR. - [x] Define `context.ResponseWriter` interface with an implementation `context.Response`. - [x] Use chi instead of macaron, and also a customize `Route` to wrap chi so that the router usage is similar as before. - [x] Create different routers for `web`, `api`, `internal` and `install` so that the codes will be more clear and no magic . - [x] Use https://github.com/unrolled/render instead of macaron's internal render - [x] Use https://github.com/NYTimes/gziphandler instead of https://gitea.com/macaron/gzip - [x] Use https://gitea.com/go-chi/session which is a modified version of https://gitea.com/macaron/session and removed `nodb` support since it will not be maintained. **BREAK** - [x] Use https://gitea.com/go-chi/captcha which is a modified version of https://gitea.com/macaron/captcha - [x] Use https://gitea.com/go-chi/cache which is a modified version of https://gitea.com/macaron/cache - [x] Use https://gitea.com/go-chi/binding which is a modified version of https://gitea.com/macaron/binding - [x] Use https://github.com/go-chi/cors instead of https://gitea.com/macaron/cors - [x] Dropped https://gitea.com/macaron/i18n and make a new one in `code.gitea.io/gitea/modules/translation` - [x] Move validation form structs from `code.gitea.io/gitea/modules/auth` to `code.gitea.io/gitea/modules/forms` to avoid dependency cycle. - [x] Removed macaron log service because it's not need any more. **BREAK** - [x] All form structs have to be get by `web.GetForm(ctx)` in the route function but not as a function parameter on routes definition. - [x] Move Git HTTP protocol implementation to use routers directly. - [x] Fix the problem that chi routes don't support trailing slash but macaron did. - [x] `/api/v1/swagger` now will be redirect to `/api/swagger` but not render directly so that `APIContext` will not create a html render. Notices: - Chi router don't support request with trailing slash - Integration test `TestUserHeatmap` maybe mysql version related. It's failed on my macOS(mysql 5.7.29 installed via brew) but succeed on CI. Co-authored-by: 6543 <6543@obermui.de>
		
			
				
	
	
		
			532 lines
		
	
	
	
		
			16 KiB
		
	
	
	
		
			Go
		
	
	
	
		
			Vendored
		
	
	
	
			
		
		
	
	
			532 lines
		
	
	
	
		
			16 KiB
		
	
	
	
		
			Go
		
	
	
	
		
			Vendored
		
	
	
	
package gziphandler // import "github.com/NYTimes/gziphandler"
 | 
						||
 | 
						||
import (
 | 
						||
	"bufio"
 | 
						||
	"compress/gzip"
 | 
						||
	"fmt"
 | 
						||
	"io"
 | 
						||
	"mime"
 | 
						||
	"net"
 | 
						||
	"net/http"
 | 
						||
	"strconv"
 | 
						||
	"strings"
 | 
						||
	"sync"
 | 
						||
)
 | 
						||
 | 
						||
const (
 | 
						||
	vary            = "Vary"
 | 
						||
	acceptEncoding  = "Accept-Encoding"
 | 
						||
	contentEncoding = "Content-Encoding"
 | 
						||
	contentType     = "Content-Type"
 | 
						||
	contentLength   = "Content-Length"
 | 
						||
)
 | 
						||
 | 
						||
type codings map[string]float64
 | 
						||
 | 
						||
const (
 | 
						||
	// DefaultQValue is the default qvalue to assign to an encoding if no explicit qvalue is set.
 | 
						||
	// This is actually kind of ambiguous in RFC 2616, so hopefully it's correct.
 | 
						||
	// The examples seem to indicate that it is.
 | 
						||
	DefaultQValue = 1.0
 | 
						||
 | 
						||
	// DefaultMinSize is the default minimum size until we enable gzip compression.
 | 
						||
	// 1500 bytes is the MTU size for the internet since that is the largest size allowed at the network layer.
 | 
						||
	// If you take a file that is 1300 bytes and compress it to 800 bytes, it’s still transmitted in that same 1500 byte packet regardless, so you’ve gained nothing.
 | 
						||
	// That being the case, you should restrict the gzip compression to files with a size greater than a single packet, 1400 bytes (1.4KB) is a safe value.
 | 
						||
	DefaultMinSize = 1400
 | 
						||
)
 | 
						||
 | 
						||
// gzipWriterPools stores a sync.Pool for each compression level for reuse of
 | 
						||
// gzip.Writers. Use poolIndex to covert a compression level to an index into
 | 
						||
// gzipWriterPools.
 | 
						||
var gzipWriterPools [gzip.BestCompression - gzip.BestSpeed + 2]*sync.Pool
 | 
						||
 | 
						||
func init() {
 | 
						||
	for i := gzip.BestSpeed; i <= gzip.BestCompression; i++ {
 | 
						||
		addLevelPool(i)
 | 
						||
	}
 | 
						||
	addLevelPool(gzip.DefaultCompression)
 | 
						||
}
 | 
						||
 | 
						||
// poolIndex maps a compression level to its index into gzipWriterPools. It
 | 
						||
// assumes that level is a valid gzip compression level.
 | 
						||
func poolIndex(level int) int {
 | 
						||
	// gzip.DefaultCompression == -1, so we need to treat it special.
 | 
						||
	if level == gzip.DefaultCompression {
 | 
						||
		return gzip.BestCompression - gzip.BestSpeed + 1
 | 
						||
	}
 | 
						||
	return level - gzip.BestSpeed
 | 
						||
}
 | 
						||
 | 
						||
func addLevelPool(level int) {
 | 
						||
	gzipWriterPools[poolIndex(level)] = &sync.Pool{
 | 
						||
		New: func() interface{} {
 | 
						||
			// NewWriterLevel only returns error on a bad level, we are guaranteeing
 | 
						||
			// that this will be a valid level so it is okay to ignore the returned
 | 
						||
			// error.
 | 
						||
			w, _ := gzip.NewWriterLevel(nil, level)
 | 
						||
			return w
 | 
						||
		},
 | 
						||
	}
 | 
						||
}
 | 
						||
 | 
						||
// GzipResponseWriter provides an http.ResponseWriter interface, which gzips
 | 
						||
// bytes before writing them to the underlying response. This doesn't close the
 | 
						||
// writers, so don't forget to do that.
 | 
						||
// It can be configured to skip response smaller than minSize.
 | 
						||
type GzipResponseWriter struct {
 | 
						||
	http.ResponseWriter
 | 
						||
	index int // Index for gzipWriterPools.
 | 
						||
	gw    *gzip.Writer
 | 
						||
 | 
						||
	code int // Saves the WriteHeader value.
 | 
						||
 | 
						||
	minSize int    // Specifed the minimum response size to gzip. If the response length is bigger than this value, it is compressed.
 | 
						||
	buf     []byte // Holds the first part of the write before reaching the minSize or the end of the write.
 | 
						||
	ignore  bool   // If true, then we immediately passthru writes to the underlying ResponseWriter.
 | 
						||
 | 
						||
	contentTypes []parsedContentType // Only compress if the response is one of these content-types. All are accepted if empty.
 | 
						||
}
 | 
						||
 | 
						||
type GzipResponseWriterWithCloseNotify struct {
 | 
						||
	*GzipResponseWriter
 | 
						||
}
 | 
						||
 | 
						||
func (w GzipResponseWriterWithCloseNotify) CloseNotify() <-chan bool {
 | 
						||
	return w.ResponseWriter.(http.CloseNotifier).CloseNotify()
 | 
						||
}
 | 
						||
 | 
						||
// Write appends data to the gzip writer.
 | 
						||
func (w *GzipResponseWriter) Write(b []byte) (int, error) {
 | 
						||
	// GZIP responseWriter is initialized. Use the GZIP responseWriter.
 | 
						||
	if w.gw != nil {
 | 
						||
		return w.gw.Write(b)
 | 
						||
	}
 | 
						||
 | 
						||
	// If we have already decided not to use GZIP, immediately passthrough.
 | 
						||
	if w.ignore {
 | 
						||
		return w.ResponseWriter.Write(b)
 | 
						||
	}
 | 
						||
 | 
						||
	// Save the write into a buffer for later use in GZIP responseWriter (if content is long enough) or at close with regular responseWriter.
 | 
						||
	// On the first write, w.buf changes from nil to a valid slice
 | 
						||
	w.buf = append(w.buf, b...)
 | 
						||
 | 
						||
	var (
 | 
						||
		cl, _ = strconv.Atoi(w.Header().Get(contentLength))
 | 
						||
		ct    = w.Header().Get(contentType)
 | 
						||
		ce    = w.Header().Get(contentEncoding)
 | 
						||
	)
 | 
						||
	// Only continue if they didn't already choose an encoding or a known unhandled content length or type.
 | 
						||
	if ce == "" && (cl == 0 || cl >= w.minSize) && (ct == "" || handleContentType(w.contentTypes, ct)) {
 | 
						||
		// If the current buffer is less than minSize and a Content-Length isn't set, then wait until we have more data.
 | 
						||
		if len(w.buf) < w.minSize && cl == 0 {
 | 
						||
			return len(b), nil
 | 
						||
		}
 | 
						||
		// If the Content-Length is larger than minSize or the current buffer is larger than minSize, then continue.
 | 
						||
		if cl >= w.minSize || len(w.buf) >= w.minSize {
 | 
						||
			// If a Content-Type wasn't specified, infer it from the current buffer.
 | 
						||
			if ct == "" {
 | 
						||
				ct = http.DetectContentType(w.buf)
 | 
						||
				w.Header().Set(contentType, ct)
 | 
						||
			}
 | 
						||
			// If the Content-Type is acceptable to GZIP, initialize the GZIP writer.
 | 
						||
			if handleContentType(w.contentTypes, ct) {
 | 
						||
				if err := w.startGzip(); err != nil {
 | 
						||
					return 0, err
 | 
						||
				}
 | 
						||
				return len(b), nil
 | 
						||
			}
 | 
						||
		}
 | 
						||
	}
 | 
						||
	// If we got here, we should not GZIP this response.
 | 
						||
	if err := w.startPlain(); err != nil {
 | 
						||
		return 0, err
 | 
						||
	}
 | 
						||
	return len(b), nil
 | 
						||
}
 | 
						||
 | 
						||
// startGzip initializes a GZIP writer and writes the buffer.
 | 
						||
func (w *GzipResponseWriter) startGzip() error {
 | 
						||
	// Set the GZIP header.
 | 
						||
	w.Header().Set(contentEncoding, "gzip")
 | 
						||
 | 
						||
	// if the Content-Length is already set, then calls to Write on gzip
 | 
						||
	// will fail to set the Content-Length header since its already set
 | 
						||
	// See: https://github.com/golang/go/issues/14975.
 | 
						||
	w.Header().Del(contentLength)
 | 
						||
 | 
						||
	// Write the header to gzip response.
 | 
						||
	if w.code != 0 {
 | 
						||
		w.ResponseWriter.WriteHeader(w.code)
 | 
						||
		// Ensure that no other WriteHeader's happen
 | 
						||
		w.code = 0
 | 
						||
	}
 | 
						||
 | 
						||
	// Initialize and flush the buffer into the gzip response if there are any bytes.
 | 
						||
	// If there aren't any, we shouldn't initialize it yet because on Close it will
 | 
						||
	// write the gzip header even if nothing was ever written.
 | 
						||
	if len(w.buf) > 0 {
 | 
						||
		// Initialize the GZIP response.
 | 
						||
		w.init()
 | 
						||
		n, err := w.gw.Write(w.buf)
 | 
						||
 | 
						||
		// This should never happen (per io.Writer docs), but if the write didn't
 | 
						||
		// accept the entire buffer but returned no specific error, we have no clue
 | 
						||
		// what's going on, so abort just to be safe.
 | 
						||
		if err == nil && n < len(w.buf) {
 | 
						||
			err = io.ErrShortWrite
 | 
						||
		}
 | 
						||
		return err
 | 
						||
	}
 | 
						||
	return nil
 | 
						||
}
 | 
						||
 | 
						||
// startPlain writes to sent bytes and buffer the underlying ResponseWriter without gzip.
 | 
						||
func (w *GzipResponseWriter) startPlain() error {
 | 
						||
	if w.code != 0 {
 | 
						||
		w.ResponseWriter.WriteHeader(w.code)
 | 
						||
		// Ensure that no other WriteHeader's happen
 | 
						||
		w.code = 0
 | 
						||
	}
 | 
						||
	w.ignore = true
 | 
						||
	// If Write was never called then don't call Write on the underlying ResponseWriter.
 | 
						||
	if w.buf == nil {
 | 
						||
		return nil
 | 
						||
	}
 | 
						||
	n, err := w.ResponseWriter.Write(w.buf)
 | 
						||
	w.buf = nil
 | 
						||
	// This should never happen (per io.Writer docs), but if the write didn't
 | 
						||
	// accept the entire buffer but returned no specific error, we have no clue
 | 
						||
	// what's going on, so abort just to be safe.
 | 
						||
	if err == nil && n < len(w.buf) {
 | 
						||
		err = io.ErrShortWrite
 | 
						||
	}
 | 
						||
	return err
 | 
						||
}
 | 
						||
 | 
						||
// WriteHeader just saves the response code until close or GZIP effective writes.
 | 
						||
func (w *GzipResponseWriter) WriteHeader(code int) {
 | 
						||
	if w.code == 0 {
 | 
						||
		w.code = code
 | 
						||
	}
 | 
						||
}
 | 
						||
 | 
						||
// init graps a new gzip writer from the gzipWriterPool and writes the correct
 | 
						||
// content encoding header.
 | 
						||
func (w *GzipResponseWriter) init() {
 | 
						||
	// Bytes written during ServeHTTP are redirected to this gzip writer
 | 
						||
	// before being written to the underlying response.
 | 
						||
	gzw := gzipWriterPools[w.index].Get().(*gzip.Writer)
 | 
						||
	gzw.Reset(w.ResponseWriter)
 | 
						||
	w.gw = gzw
 | 
						||
}
 | 
						||
 | 
						||
// Close will close the gzip.Writer and will put it back in the gzipWriterPool.
 | 
						||
func (w *GzipResponseWriter) Close() error {
 | 
						||
	if w.ignore {
 | 
						||
		return nil
 | 
						||
	}
 | 
						||
 | 
						||
	if w.gw == nil {
 | 
						||
		// GZIP not triggered yet, write out regular response.
 | 
						||
		err := w.startPlain()
 | 
						||
		// Returns the error if any at write.
 | 
						||
		if err != nil {
 | 
						||
			err = fmt.Errorf("gziphandler: write to regular responseWriter at close gets error: %q", err.Error())
 | 
						||
		}
 | 
						||
		return err
 | 
						||
	}
 | 
						||
 | 
						||
	err := w.gw.Close()
 | 
						||
	gzipWriterPools[w.index].Put(w.gw)
 | 
						||
	w.gw = nil
 | 
						||
	return err
 | 
						||
}
 | 
						||
 | 
						||
// Flush flushes the underlying *gzip.Writer and then the underlying
 | 
						||
// http.ResponseWriter if it is an http.Flusher. This makes GzipResponseWriter
 | 
						||
// an http.Flusher.
 | 
						||
func (w *GzipResponseWriter) Flush() {
 | 
						||
	if w.gw == nil && !w.ignore {
 | 
						||
		// Only flush once startGzip or startPlain has been called.
 | 
						||
		//
 | 
						||
		// Flush is thus a no-op until we're certain whether a plain
 | 
						||
		// or gzipped response will be served.
 | 
						||
		return
 | 
						||
	}
 | 
						||
 | 
						||
	if w.gw != nil {
 | 
						||
		w.gw.Flush()
 | 
						||
	}
 | 
						||
 | 
						||
	if fw, ok := w.ResponseWriter.(http.Flusher); ok {
 | 
						||
		fw.Flush()
 | 
						||
	}
 | 
						||
}
 | 
						||
 | 
						||
// Hijack implements http.Hijacker. If the underlying ResponseWriter is a
 | 
						||
// Hijacker, its Hijack method is returned. Otherwise an error is returned.
 | 
						||
func (w *GzipResponseWriter) Hijack() (net.Conn, *bufio.ReadWriter, error) {
 | 
						||
	if hj, ok := w.ResponseWriter.(http.Hijacker); ok {
 | 
						||
		return hj.Hijack()
 | 
						||
	}
 | 
						||
	return nil, nil, fmt.Errorf("http.Hijacker interface is not supported")
 | 
						||
}
 | 
						||
 | 
						||
// verify Hijacker interface implementation
 | 
						||
var _ http.Hijacker = &GzipResponseWriter{}
 | 
						||
 | 
						||
// MustNewGzipLevelHandler behaves just like NewGzipLevelHandler except that in
 | 
						||
// an error case it panics rather than returning an error.
 | 
						||
func MustNewGzipLevelHandler(level int) func(http.Handler) http.Handler {
 | 
						||
	wrap, err := NewGzipLevelHandler(level)
 | 
						||
	if err != nil {
 | 
						||
		panic(err)
 | 
						||
	}
 | 
						||
	return wrap
 | 
						||
}
 | 
						||
 | 
						||
// NewGzipLevelHandler returns a wrapper function (often known as middleware)
 | 
						||
// which can be used to wrap an HTTP handler to transparently gzip the response
 | 
						||
// body if the client supports it (via the Accept-Encoding header). Responses will
 | 
						||
// be encoded at the given gzip compression level. An error will be returned only
 | 
						||
// if an invalid gzip compression level is given, so if one can ensure the level
 | 
						||
// is valid, the returned error can be safely ignored.
 | 
						||
func NewGzipLevelHandler(level int) (func(http.Handler) http.Handler, error) {
 | 
						||
	return NewGzipLevelAndMinSize(level, DefaultMinSize)
 | 
						||
}
 | 
						||
 | 
						||
// NewGzipLevelAndMinSize behave as NewGzipLevelHandler except it let the caller
 | 
						||
// specify the minimum size before compression.
 | 
						||
func NewGzipLevelAndMinSize(level, minSize int) (func(http.Handler) http.Handler, error) {
 | 
						||
	return GzipHandlerWithOpts(CompressionLevel(level), MinSize(minSize))
 | 
						||
}
 | 
						||
 | 
						||
func GzipHandlerWithOpts(opts ...option) (func(http.Handler) http.Handler, error) {
 | 
						||
	c := &config{
 | 
						||
		level:   gzip.DefaultCompression,
 | 
						||
		minSize: DefaultMinSize,
 | 
						||
	}
 | 
						||
 | 
						||
	for _, o := range opts {
 | 
						||
		o(c)
 | 
						||
	}
 | 
						||
 | 
						||
	if err := c.validate(); err != nil {
 | 
						||
		return nil, err
 | 
						||
	}
 | 
						||
 | 
						||
	return func(h http.Handler) http.Handler {
 | 
						||
		index := poolIndex(c.level)
 | 
						||
 | 
						||
		return http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {
 | 
						||
			w.Header().Add(vary, acceptEncoding)
 | 
						||
			if acceptsGzip(r) {
 | 
						||
				gw := &GzipResponseWriter{
 | 
						||
					ResponseWriter: w,
 | 
						||
					index:          index,
 | 
						||
					minSize:        c.minSize,
 | 
						||
					contentTypes:   c.contentTypes,
 | 
						||
				}
 | 
						||
				defer gw.Close()
 | 
						||
 | 
						||
				if _, ok := w.(http.CloseNotifier); ok {
 | 
						||
					gwcn := GzipResponseWriterWithCloseNotify{gw}
 | 
						||
					h.ServeHTTP(gwcn, r)
 | 
						||
				} else {
 | 
						||
					h.ServeHTTP(gw, r)
 | 
						||
				}
 | 
						||
 | 
						||
			} else {
 | 
						||
				h.ServeHTTP(w, r)
 | 
						||
			}
 | 
						||
		})
 | 
						||
	}, nil
 | 
						||
}
 | 
						||
 | 
						||
// Parsed representation of one of the inputs to ContentTypes.
 | 
						||
// See https://golang.org/pkg/mime/#ParseMediaType
 | 
						||
type parsedContentType struct {
 | 
						||
	mediaType string
 | 
						||
	params    map[string]string
 | 
						||
}
 | 
						||
 | 
						||
// equals returns whether this content type matches another content type.
 | 
						||
func (pct parsedContentType) equals(mediaType string, params map[string]string) bool {
 | 
						||
	if pct.mediaType != mediaType {
 | 
						||
		return false
 | 
						||
	}
 | 
						||
	// if pct has no params, don't care about other's params
 | 
						||
	if len(pct.params) == 0 {
 | 
						||
		return true
 | 
						||
	}
 | 
						||
 | 
						||
	// if pct has any params, they must be identical to other's.
 | 
						||
	if len(pct.params) != len(params) {
 | 
						||
		return false
 | 
						||
	}
 | 
						||
	for k, v := range pct.params {
 | 
						||
		if w, ok := params[k]; !ok || v != w {
 | 
						||
			return false
 | 
						||
		}
 | 
						||
	}
 | 
						||
	return true
 | 
						||
}
 | 
						||
 | 
						||
// Used for functional configuration.
 | 
						||
type config struct {
 | 
						||
	minSize      int
 | 
						||
	level        int
 | 
						||
	contentTypes []parsedContentType
 | 
						||
}
 | 
						||
 | 
						||
func (c *config) validate() error {
 | 
						||
	if c.level != gzip.DefaultCompression && (c.level < gzip.BestSpeed || c.level > gzip.BestCompression) {
 | 
						||
		return fmt.Errorf("invalid compression level requested: %d", c.level)
 | 
						||
	}
 | 
						||
 | 
						||
	if c.minSize < 0 {
 | 
						||
		return fmt.Errorf("minimum size must be more than zero")
 | 
						||
	}
 | 
						||
 | 
						||
	return nil
 | 
						||
}
 | 
						||
 | 
						||
type option func(c *config)
 | 
						||
 | 
						||
func MinSize(size int) option {
 | 
						||
	return func(c *config) {
 | 
						||
		c.minSize = size
 | 
						||
	}
 | 
						||
}
 | 
						||
 | 
						||
func CompressionLevel(level int) option {
 | 
						||
	return func(c *config) {
 | 
						||
		c.level = level
 | 
						||
	}
 | 
						||
}
 | 
						||
 | 
						||
// ContentTypes specifies a list of content types to compare
 | 
						||
// the Content-Type header to before compressing. If none
 | 
						||
// match, the response will be returned as-is.
 | 
						||
//
 | 
						||
// Content types are compared in a case-insensitive, whitespace-ignored
 | 
						||
// manner.
 | 
						||
//
 | 
						||
// A MIME type without any other directive will match a content type
 | 
						||
// that has the same MIME type, regardless of that content type's other
 | 
						||
// directives. I.e., "text/html" will match both "text/html" and
 | 
						||
// "text/html; charset=utf-8".
 | 
						||
//
 | 
						||
// A MIME type with any other directive will only match a content type
 | 
						||
// that has the same MIME type and other directives. I.e.,
 | 
						||
// "text/html; charset=utf-8" will only match "text/html; charset=utf-8".
 | 
						||
//
 | 
						||
// By default, responses are gzipped regardless of
 | 
						||
// Content-Type.
 | 
						||
func ContentTypes(types []string) option {
 | 
						||
	return func(c *config) {
 | 
						||
		c.contentTypes = []parsedContentType{}
 | 
						||
		for _, v := range types {
 | 
						||
			mediaType, params, err := mime.ParseMediaType(v)
 | 
						||
			if err == nil {
 | 
						||
				c.contentTypes = append(c.contentTypes, parsedContentType{mediaType, params})
 | 
						||
			}
 | 
						||
		}
 | 
						||
	}
 | 
						||
}
 | 
						||
 | 
						||
// GzipHandler wraps an HTTP handler, to transparently gzip the response body if
 | 
						||
// the client supports it (via the Accept-Encoding header). This will compress at
 | 
						||
// the default compression level.
 | 
						||
func GzipHandler(h http.Handler) http.Handler {
 | 
						||
	wrapper, _ := NewGzipLevelHandler(gzip.DefaultCompression)
 | 
						||
	return wrapper(h)
 | 
						||
}
 | 
						||
 | 
						||
// acceptsGzip returns true if the given HTTP request indicates that it will
 | 
						||
// accept a gzipped response.
 | 
						||
func acceptsGzip(r *http.Request) bool {
 | 
						||
	acceptedEncodings, _ := parseEncodings(r.Header.Get(acceptEncoding))
 | 
						||
	return acceptedEncodings["gzip"] > 0.0
 | 
						||
}
 | 
						||
 | 
						||
// returns true if we've been configured to compress the specific content type.
 | 
						||
func handleContentType(contentTypes []parsedContentType, ct string) bool {
 | 
						||
	// If contentTypes is empty we handle all content types.
 | 
						||
	if len(contentTypes) == 0 {
 | 
						||
		return true
 | 
						||
	}
 | 
						||
 | 
						||
	mediaType, params, err := mime.ParseMediaType(ct)
 | 
						||
	if err != nil {
 | 
						||
		return false
 | 
						||
	}
 | 
						||
 | 
						||
	for _, c := range contentTypes {
 | 
						||
		if c.equals(mediaType, params) {
 | 
						||
			return true
 | 
						||
		}
 | 
						||
	}
 | 
						||
 | 
						||
	return false
 | 
						||
}
 | 
						||
 | 
						||
// parseEncodings attempts to parse a list of codings, per RFC 2616, as might
 | 
						||
// appear in an Accept-Encoding header. It returns a map of content-codings to
 | 
						||
// quality values, and an error containing the errors encountered. It's probably
 | 
						||
// safe to ignore those, because silently ignoring errors is how the internet
 | 
						||
// works.
 | 
						||
//
 | 
						||
// See: http://tools.ietf.org/html/rfc2616#section-14.3.
 | 
						||
func parseEncodings(s string) (codings, error) {
 | 
						||
	c := make(codings)
 | 
						||
	var e []string
 | 
						||
 | 
						||
	for _, ss := range strings.Split(s, ",") {
 | 
						||
		coding, qvalue, err := parseCoding(ss)
 | 
						||
 | 
						||
		if err != nil {
 | 
						||
			e = append(e, err.Error())
 | 
						||
		} else {
 | 
						||
			c[coding] = qvalue
 | 
						||
		}
 | 
						||
	}
 | 
						||
 | 
						||
	// TODO (adammck): Use a proper multi-error struct, so the individual errors
 | 
						||
	//                 can be extracted if anyone cares.
 | 
						||
	if len(e) > 0 {
 | 
						||
		return c, fmt.Errorf("errors while parsing encodings: %s", strings.Join(e, ", "))
 | 
						||
	}
 | 
						||
 | 
						||
	return c, nil
 | 
						||
}
 | 
						||
 | 
						||
// parseCoding parses a single conding (content-coding with an optional qvalue),
 | 
						||
// as might appear in an Accept-Encoding header. It attempts to forgive minor
 | 
						||
// formatting errors.
 | 
						||
func parseCoding(s string) (coding string, qvalue float64, err error) {
 | 
						||
	for n, part := range strings.Split(s, ";") {
 | 
						||
		part = strings.TrimSpace(part)
 | 
						||
		qvalue = DefaultQValue
 | 
						||
 | 
						||
		if n == 0 {
 | 
						||
			coding = strings.ToLower(part)
 | 
						||
		} else if strings.HasPrefix(part, "q=") {
 | 
						||
			qvalue, err = strconv.ParseFloat(strings.TrimPrefix(part, "q="), 64)
 | 
						||
 | 
						||
			if qvalue < 0.0 {
 | 
						||
				qvalue = 0.0
 | 
						||
			} else if qvalue > 1.0 {
 | 
						||
				qvalue = 1.0
 | 
						||
			}
 | 
						||
		}
 | 
						||
	}
 | 
						||
 | 
						||
	if coding == "" {
 | 
						||
		err = fmt.Errorf("empty content-coding")
 | 
						||
	}
 | 
						||
 | 
						||
	return
 | 
						||
}
 |